code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class __lowerCAmelCase : """simple docstring""" def __init__( self : Any , _lowerCAmelCase : int , _lowerCAmelCase : Any=1_3 , _lowerCAmelCase : Dict=7 , _lowerCAmelCase : Optional[int]=True , _lowerCAmelCase : Any=True , _lowerCAmelCase : Union[str, Any]=True , _lowerCAmelCase : str=9_9 , _lowerCAmelCase : List[str]=3_2 , _lowerCAmelCase : Optional[int]=5 , _lowerCAmelCase : Dict=4 , _lowerCAmelCase : str=3_7 , _lowerCAmelCase : Union[str, Any]="gelu" , _lowerCAmelCase : Any=0.1 , _lowerCAmelCase : str=0.1 , _lowerCAmelCase : Optional[int]=5_1_2 , _lowerCAmelCase : Optional[Any]=1_6 , _lowerCAmelCase : int=2 , _lowerCAmelCase : Tuple=0.02 , _lowerCAmelCase : str=3 , _lowerCAmelCase : Union[str, Any]=4 , _lowerCAmelCase : List[str]=None , ) -> Union[str, Any]: """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_labels snake_case_ = num_choices snake_case_ = scope snake_case_ = self.vocab_size - 1 def lowerCAmelCase__ ( self : Dict ) -> Tuple: """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_token_type_ids: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) snake_case_ = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def lowerCAmelCase__ ( self : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : Any , _lowerCAmelCase : Tuple , _lowerCAmelCase : int , *_lowerCAmelCase : str ) -> List[str]: """simple docstring""" snake_case_ = OpenAIGPTModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() snake_case_ = model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase , head_mask=_lowerCAmelCase ) snake_case_ = model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase ) snake_case_ = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCAmelCase__ ( self : str , _lowerCAmelCase : int , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] , *_lowerCAmelCase : Dict ) -> List[str]: """simple docstring""" snake_case_ = OpenAIGPTLMHeadModel(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() snake_case_ = model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase__ ( self : List[str] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , *_lowerCAmelCase : List[Any] ) -> List[Any]: """simple docstring""" snake_case_ = OpenAIGPTDoubleHeadsModel(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() snake_case_ = model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCAmelCase__ ( self : List[Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , *_lowerCAmelCase : Optional[Any] ) -> List[Any]: """simple docstring""" snake_case_ = self.num_labels snake_case_ = OpenAIGPTForSequenceClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = model(_lowerCAmelCase , token_type_ids=_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase__ ( self : Optional[Any] ) -> Tuple: """simple docstring""" snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = config_and_inputs snake_case_ = { "input_ids": input_ids, "token_type_ids": token_type_ids, "head_mask": head_mask, } return config, inputs_dict @require_torch class __lowerCAmelCase ( a , a , a , unittest.TestCase ): """simple docstring""" _SCREAMING_SNAKE_CASE = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) _SCREAMING_SNAKE_CASE = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly _SCREAMING_SNAKE_CASE = ( { 'feature-extraction': OpenAIGPTModel, 'text-classification': OpenAIGPTForSequenceClassification, 'text-generation': OpenAIGPTLMHeadModel, 'zero-shot': OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def lowerCAmelCase__ ( self : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[Any] ) -> Dict: """simple docstring""" if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def lowerCAmelCase__ ( self : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any]=False ) -> int: """simple docstring""" snake_case_ = super()._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": snake_case_ = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=_lowerCAmelCase , ) snake_case_ = inputs_dict["labels"] snake_case_ = inputs_dict["labels"] snake_case_ = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=_lowerCAmelCase , ) snake_case_ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_lowerCAmelCase ) return inputs_dict def lowerCAmelCase__ ( self : str ) -> Union[str, Any]: """simple docstring""" snake_case_ = OpenAIGPTModelTester(self ) snake_case_ = ConfigTester(self , config_class=_lowerCAmelCase , n_embd=3_7 ) def lowerCAmelCase__ ( self : List[str] ) -> List[str]: """simple docstring""" self.config_tester.run_common_tests() def lowerCAmelCase__ ( self : List[Any] ) -> Optional[int]: """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*_lowerCAmelCase ) def lowerCAmelCase__ ( self : Dict ) -> Optional[Any]: """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*_lowerCAmelCase ) def lowerCAmelCase__ ( self : Any ) -> List[Any]: """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*_lowerCAmelCase ) def lowerCAmelCase__ ( self : Any ) -> int: """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*_lowerCAmelCase ) @slow def lowerCAmelCase__ ( self : int ) -> Union[str, Any]: """simple docstring""" for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = OpenAIGPTModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) @require_torch class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def lowerCAmelCase__ ( self : str ) -> Union[str, Any]: """simple docstring""" snake_case_ = OpenAIGPTLMHeadModel.from_pretrained("openai-gpt" ) model.to(_lowerCAmelCase ) snake_case_ = torch.tensor([[4_8_1, 4_7_3_5, 5_4_4]] , dtype=torch.long , device=_lowerCAmelCase ) # the president is snake_case_ = [ 4_8_1, 4_7_3_5, 5_4_4, 2_4_6, 9_6_3, 8_7_0, 7_6_2, 2_3_9, 2_4_4, 4_0_4_7_7, 2_4_4, 2_4_9, 7_1_9, 8_8_1, 4_8_7, 5_4_4, 2_4_0, 2_4_4, 6_0_3, 4_8_1, ] # the president is a very good man. " \n " i\'m sure he is, " said the snake_case_ = model.generate(_lowerCAmelCase , do_sample=_lowerCAmelCase ) self.assertListEqual(output_ids[0].tolist() , _lowerCAmelCase )
283
import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC SCREAMING_SNAKE_CASE :Union[str, Any] = parse(importlib.metadata.version('''torch''')) def _lowerCAmelCase ( lowerCAmelCase_ :Union[str, Version] , lowerCAmelCase_ :str , lowerCAmelCase_ :str )->Any: '''simple docstring''' if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(F'''`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}''' ) snake_case_ = STR_OPERATION_TO_FUNC[operation] if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): snake_case_ = parse(importlib.metadata.version(lowerCAmelCase_ ) ) return operation(lowerCAmelCase_ , parse(lowerCAmelCase_ ) ) def _lowerCAmelCase ( lowerCAmelCase_ :str , lowerCAmelCase_ :str )->Any: '''simple docstring''' return compare_versions(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ )
283
1
"""simple docstring""" import copy import fnmatch import json import os import pickle as pkl import shutil import sys import tarfile import tempfile from collections import OrderedDict from contextlib import contextmanager from functools import partial from hashlib import shaaaa from io import BytesIO from pathlib import Path from urllib.parse import urlparse from zipfile import ZipFile, is_zipfile import cva import numpy as np import requests import wget from filelock import FileLock from PIL import Image from tqdm.auto import tqdm from yaml import Loader, dump, load try: import torch lowerCamelCase__ = True except ImportError: lowerCamelCase__ = False try: from torch.hub import _get_torch_home lowerCamelCase__ = _get_torch_home() except ImportError: lowerCamelCase__ = os.path.expanduser( os.getenv("""TORCH_HOME""", os.path.join(os.getenv("""XDG_CACHE_HOME""", """~/.cache"""), """torch""")) ) lowerCamelCase__ = os.path.join(torch_cache_home, """transformers""") lowerCamelCase__ = """https://cdn.huggingface.co""" lowerCamelCase__ = """https://s3.amazonaws.com/models.huggingface.co/bert""" lowerCamelCase__ = """/""".join(str(Path(__file__).resolve()).split("""/""")[:-1]) lowerCamelCase__ = os.path.join(PATH, """config.yaml""") lowerCamelCase__ = os.path.join(PATH, """attributes.txt""") lowerCamelCase__ = os.path.join(PATH, """objects.txt""") lowerCamelCase__ = os.getenv("""PYTORCH_PRETRAINED_BERT_CACHE""", default_cache_path) lowerCamelCase__ = os.getenv("""PYTORCH_TRANSFORMERS_CACHE""", PYTORCH_PRETRAINED_BERT_CACHE) lowerCamelCase__ = os.getenv("""TRANSFORMERS_CACHE""", PYTORCH_TRANSFORMERS_CACHE) lowerCamelCase__ = """pytorch_model.bin""" lowerCamelCase__ = """config.yaml""" def __lowerCAmelCase (_UpperCamelCase=OBJECTS , _UpperCamelCase=ATTRIBUTES ): __lowerCAmelCase : Dict = [] with open(_UpperCamelCase ) as f: for object in f.readlines(): vg_classes.append(object.split(',' )[0].lower().strip() ) __lowerCAmelCase : str = [] with open(_UpperCamelCase ) as f: for object in f.readlines(): vg_attrs.append(object.split(',' )[0].lower().strip() ) return vg_classes, vg_attrs def __lowerCAmelCase (_UpperCamelCase ): __lowerCAmelCase : List[str] = OrderedDict() with open(_UpperCamelCase , 'rb' ) as f: __lowerCAmelCase : Union[str, Any] = pkl.load(_UpperCamelCase )['model'] for k in copy.deepcopy(list(ckp.keys() ) ): __lowerCAmelCase : str = ckp.pop(_UpperCamelCase ) if isinstance(_UpperCamelCase , np.ndarray ): __lowerCAmelCase : Dict = torch.tensor(_UpperCamelCase ) else: assert isinstance(_UpperCamelCase , torch.tensor ), type(_UpperCamelCase ) __lowerCAmelCase : Tuple = v return r class A__ : A_ : str = {} def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = "root" , _SCREAMING_SNAKE_CASE=0 ): __lowerCAmelCase : Any = name __lowerCAmelCase : Any = level __lowerCAmelCase : Optional[Any] = {} for k, v in dictionary.items(): if v is None: raise ValueError() __lowerCAmelCase : Optional[int] = copy.deepcopy(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[str] = copy.deepcopy(_SCREAMING_SNAKE_CASE ) if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): __lowerCAmelCase : Dict = Config(_SCREAMING_SNAKE_CASE , name=_SCREAMING_SNAKE_CASE , level=level + 1 ) __lowerCAmelCase : Union[str, Any] = v setattr(self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : int = d def __repr__( self ): return str(list((self._pointer.keys()) ) ) def __setattr__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): __lowerCAmelCase : List[str] = val __lowerCAmelCase : Optional[Any] = val __lowerCAmelCase : Optional[Any] = key.split('.' ) __lowerCAmelCase : Tuple = len(_SCREAMING_SNAKE_CASE ) - 1 __lowerCAmelCase : Tuple = self._pointer if len(_SCREAMING_SNAKE_CASE ) > 1: for i, l in enumerate(_SCREAMING_SNAKE_CASE ): if hasattr(self , _SCREAMING_SNAKE_CASE ) and isinstance(getattr(self , _SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ): setattr(getattr(self , _SCREAMING_SNAKE_CASE ) , '.'.join(levels[i:] ) , _SCREAMING_SNAKE_CASE ) if l == last_level: __lowerCAmelCase : Tuple = val else: __lowerCAmelCase : str = pointer[l] def __lowerCamelCase ( self ): return self._pointer def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): with open(f"{file_name}" , 'w' ) as stream: dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): with open(f"{file_name}" , 'w' ) as stream: json.dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @staticmethod def __lowerCamelCase ( _SCREAMING_SNAKE_CASE ): with open(_SCREAMING_SNAKE_CASE ) as stream: __lowerCAmelCase : Dict = load(_SCREAMING_SNAKE_CASE , Loader=_SCREAMING_SNAKE_CASE ) return data def __str__( self ): __lowerCAmelCase : Union[str, Any] = ' ' if self._name != "root": __lowerCAmelCase : List[Any] = f"{t * (self._level-1)}{self._name}:\n" else: __lowerCAmelCase : Union[str, Any] = '' __lowerCAmelCase : List[Any] = self._level for i, (k, v) in enumerate(self._pointer.items() ): if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): r += f"{t * (self._level)}{v}\n" self._level += 1 else: r += f"{t * (self._level)}{k}: {v} ({type(_SCREAMING_SNAKE_CASE ).__name__})\n" __lowerCAmelCase : Optional[Any] = level return r[:-1] @classmethod def __lowerCamelCase ( cls , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): __lowerCAmelCase : Dict = cls.get_config_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) return cls(_SCREAMING_SNAKE_CASE ) @classmethod def __lowerCamelCase ( cls , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): __lowerCAmelCase : Optional[Any] = kwargs.pop('cache_dir' , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[str] = kwargs.pop('force_download' , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Tuple = kwargs.pop('resume_download' , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[Any] = kwargs.pop('proxies' , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[Any] = kwargs.pop('local_files_only' , _SCREAMING_SNAKE_CASE ) if os.path.isdir(_SCREAMING_SNAKE_CASE ): __lowerCAmelCase : List[str] = os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) elif os.path.isfile(_SCREAMING_SNAKE_CASE ) or is_remote_url(_SCREAMING_SNAKE_CASE ): __lowerCAmelCase : Any = pretrained_model_name_or_path else: __lowerCAmelCase : Union[str, Any] = hf_bucket_url(_SCREAMING_SNAKE_CASE , filename=_SCREAMING_SNAKE_CASE , use_cdn=_SCREAMING_SNAKE_CASE ) try: # Load from URL or cache if already cached __lowerCAmelCase : Optional[int] = cached_path( _SCREAMING_SNAKE_CASE , cache_dir=_SCREAMING_SNAKE_CASE , force_download=_SCREAMING_SNAKE_CASE , proxies=_SCREAMING_SNAKE_CASE , resume_download=_SCREAMING_SNAKE_CASE , local_files_only=_SCREAMING_SNAKE_CASE , ) # Load config dict if resolved_config_file is None: raise EnvironmentError __lowerCAmelCase : Tuple = Config.load_yaml(_SCREAMING_SNAKE_CASE ) except EnvironmentError: __lowerCAmelCase : str = 'Can\'t load config for' raise EnvironmentError(_SCREAMING_SNAKE_CASE ) if resolved_config_file == config_file: print('loading configuration file from path' ) else: print('loading configuration file cache' ) return Config.load_yaml(_SCREAMING_SNAKE_CASE ), kwargs def __lowerCAmelCase (_UpperCamelCase ): __lowerCAmelCase : int = torch.load('dump.pt' , map_location=in_tensor.device ) __lowerCAmelCase : Tuple = in_tensor.numpy() __lowerCAmelCase : Dict = out_tensor.numpy()[0] print(na.shape , na[0, 0, :5] ) print(na.shape , na[0, 0, :5] ) assert np.allclose(_UpperCamelCase , _UpperCamelCase , rtol=0.01 , atol=0.1 ), ( F"{sum([1 for x in np.isclose(_UpperCamelCase , _UpperCamelCase , rtol=0.01 , atol=0.1 ).flatten() if x is False] )/len(na.flatten() )*100:.4f} %" " element-wise mismatch" ) raise Exception('tensors are all good' ) # Hugging face functions below def __lowerCAmelCase (_UpperCamelCase ): __lowerCAmelCase : Dict = urlparse(_UpperCamelCase ) return parsed.scheme in ("http", "https") def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase=True ): __lowerCAmelCase : List[str] = CLOUDFRONT_DISTRIB_PREFIX if use_cdn else S3_BUCKET_PREFIX __lowerCAmelCase : Optional[int] = '/' not in model_id if legacy_format: return F"{endpoint}/{model_id}-{filename}" else: return F"{endpoint}/{model_id}/{filename}" def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase , _UpperCamelCase=None , _UpperCamelCase=0 , _UpperCamelCase=None , ): __lowerCAmelCase : Union[str, Any] = 'python/{}'.format(sys.version.split()[0] ) if _torch_available: ua += "; torch/{}".format(torch.__version__ ) if isinstance(_UpperCamelCase , _UpperCamelCase ): ua += "; " + "; ".join('{}/{}'.format(_UpperCamelCase , _UpperCamelCase ) for k, v in user_agent.items() ) elif isinstance(_UpperCamelCase , _UpperCamelCase ): ua += "; " + user_agent __lowerCAmelCase : Optional[Any] = {'user-agent': ua} if resume_size > 0: __lowerCAmelCase : Union[str, Any] = 'bytes=%d-' % (resume_size,) __lowerCAmelCase : int = requests.get(_UpperCamelCase , stream=_UpperCamelCase , proxies=_UpperCamelCase , headers=_UpperCamelCase ) if response.status_code == 416: # Range not satisfiable return __lowerCAmelCase : Dict = response.headers.get('Content-Length' ) __lowerCAmelCase : int = resume_size + int(_UpperCamelCase ) if content_length is not None else None __lowerCAmelCase : Union[str, Any] = tqdm( unit='B' , unit_scale=_UpperCamelCase , total=_UpperCamelCase , initial=_UpperCamelCase , desc='Downloading' , ) for chunk in response.iter_content(chunk_size=1024 ): if chunk: # filter out keep-alive new chunks progress.update(len(_UpperCamelCase ) ) temp_file.write(_UpperCamelCase ) progress.close() def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase=None , _UpperCamelCase=False , _UpperCamelCase=None , _UpperCamelCase=10 , _UpperCamelCase=False , _UpperCamelCase=None , _UpperCamelCase=False , ): if cache_dir is None: __lowerCAmelCase : Any = TRANSFORMERS_CACHE if isinstance(_UpperCamelCase , _UpperCamelCase ): __lowerCAmelCase : Any = str(_UpperCamelCase ) os.makedirs(_UpperCamelCase , exist_ok=_UpperCamelCase ) __lowerCAmelCase : Tuple = None if not local_files_only: try: __lowerCAmelCase : Optional[Any] = requests.head(_UpperCamelCase , allow_redirects=_UpperCamelCase , proxies=_UpperCamelCase , timeout=_UpperCamelCase ) if response.status_code == 200: __lowerCAmelCase : Tuple = response.headers.get('ETag' ) except (EnvironmentError, requests.exceptions.Timeout): # etag is already None pass __lowerCAmelCase : Optional[int] = url_to_filename(_UpperCamelCase , _UpperCamelCase ) # get cache path to put the file __lowerCAmelCase : List[Any] = os.path.join(_UpperCamelCase , _UpperCamelCase ) # etag is None = we don't have a connection, or url doesn't exist, or is otherwise inaccessible. # try to get the last downloaded one if etag is None: if os.path.exists(_UpperCamelCase ): return cache_path else: __lowerCAmelCase : List[Any] = [ file for file in fnmatch.filter(os.listdir(_UpperCamelCase ) , filename + '.*' ) if not file.endswith('.json' ) and not file.endswith('.lock' ) ] if len(_UpperCamelCase ) > 0: return os.path.join(_UpperCamelCase , matching_files[-1] ) else: # If files cannot be found and local_files_only=True, # the models might've been found if local_files_only=False # Notify the user about that if local_files_only: raise ValueError( 'Cannot find the requested files in the cached path and outgoing traffic has been' ' disabled. To enable model look-ups and downloads online, set \'local_files_only\'' ' to False.' ) return None # From now on, etag is not None. if os.path.exists(_UpperCamelCase ) and not force_download: return cache_path # Prevent parallel downloads of the same file with a lock. __lowerCAmelCase : List[Any] = cache_path + '.lock' with FileLock(_UpperCamelCase ): # If the download just completed while the lock was activated. if os.path.exists(_UpperCamelCase ) and not force_download: # Even if returning early like here, the lock will be released. return cache_path if resume_download: __lowerCAmelCase : List[str] = cache_path + '.incomplete' @contextmanager def _resumable_file_manager(): with open(_UpperCamelCase , 'a+b' ) as f: yield f __lowerCAmelCase : str = _resumable_file_manager if os.path.exists(_UpperCamelCase ): __lowerCAmelCase : Optional[Any] = os.stat(_UpperCamelCase ).st_size else: __lowerCAmelCase : str = 0 else: __lowerCAmelCase : str = partial(tempfile.NamedTemporaryFile , dir=_UpperCamelCase , delete=_UpperCamelCase ) __lowerCAmelCase : Tuple = 0 # Download to temporary file, then copy to cache dir once finished. # Otherwise you get corrupt cache entries if the download gets interrupted. with temp_file_manager() as temp_file: print( '%s not found in cache or force_download set to True, downloading to %s' , _UpperCamelCase , temp_file.name , ) http_get( _UpperCamelCase , _UpperCamelCase , proxies=_UpperCamelCase , resume_size=_UpperCamelCase , user_agent=_UpperCamelCase , ) os.replace(temp_file.name , _UpperCamelCase ) __lowerCAmelCase : List[Any] = {'url': url, 'etag': etag} __lowerCAmelCase : Any = cache_path + '.json' with open(_UpperCamelCase , 'w' ) as meta_file: json.dump(_UpperCamelCase , _UpperCamelCase ) return cache_path def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase=None ): __lowerCAmelCase : int = url.encode('utf-8' ) __lowerCAmelCase : str = shaaaa(_UpperCamelCase ) __lowerCAmelCase : Dict = url_hash.hexdigest() if etag: __lowerCAmelCase : Optional[Any] = etag.encode('utf-8' ) __lowerCAmelCase : int = shaaaa(_UpperCamelCase ) filename += "." + etag_hash.hexdigest() if url.endswith('.h5' ): filename += ".h5" return filename def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase=None , _UpperCamelCase=False , _UpperCamelCase=None , _UpperCamelCase=False , _UpperCamelCase=None , _UpperCamelCase=False , _UpperCamelCase=False , _UpperCamelCase=False , ): if cache_dir is None: __lowerCAmelCase : List[Any] = TRANSFORMERS_CACHE if isinstance(_UpperCamelCase , _UpperCamelCase ): __lowerCAmelCase : int = str(_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ): __lowerCAmelCase : Any = str(_UpperCamelCase ) if is_remote_url(_UpperCamelCase ): # URL, so get it from the cache (downloading if necessary) __lowerCAmelCase : Optional[Any] = get_from_cache( _UpperCamelCase , cache_dir=_UpperCamelCase , force_download=_UpperCamelCase , proxies=_UpperCamelCase , resume_download=_UpperCamelCase , user_agent=_UpperCamelCase , local_files_only=_UpperCamelCase , ) elif os.path.exists(_UpperCamelCase ): # File, and it exists. __lowerCAmelCase : Optional[int] = url_or_filename elif urlparse(_UpperCamelCase ).scheme == "": # File, but it doesn't exist. raise EnvironmentError('file {} not found'.format(_UpperCamelCase ) ) else: # Something unknown raise ValueError('unable to parse {} as a URL or as a local path'.format(_UpperCamelCase ) ) if extract_compressed_file: if not is_zipfile(_UpperCamelCase ) and not tarfile.is_tarfile(_UpperCamelCase ): return output_path # Path where we extract compressed archives # We avoid '.' in dir name and add "-extracted" at the end: "./model.zip" => "./model-zip-extracted/" __lowerCAmelCase : Tuple = os.path.split(_UpperCamelCase ) __lowerCAmelCase : Optional[int] = output_file.replace('.' , '-' ) + '-extracted' __lowerCAmelCase : Optional[Any] = os.path.join(_UpperCamelCase , _UpperCamelCase ) if os.path.isdir(_UpperCamelCase ) and os.listdir(_UpperCamelCase ) and not force_extract: return output_path_extracted # Prevent parallel extractions __lowerCAmelCase : Union[str, Any] = output_path + '.lock' with FileLock(_UpperCamelCase ): shutil.rmtree(_UpperCamelCase , ignore_errors=_UpperCamelCase ) os.makedirs(_UpperCamelCase ) if is_zipfile(_UpperCamelCase ): with ZipFile(_UpperCamelCase , 'r' ) as zip_file: zip_file.extractall(_UpperCamelCase ) zip_file.close() elif tarfile.is_tarfile(_UpperCamelCase ): __lowerCAmelCase : int = tarfile.open(_UpperCamelCase ) tar_file.extractall(_UpperCamelCase ) tar_file.close() else: raise EnvironmentError('Archive format of {} could not be identified'.format(_UpperCamelCase ) ) return output_path_extracted return output_path def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase="," ): assert isinstance(_UpperCamelCase , _UpperCamelCase ) if os.path.isfile(_UpperCamelCase ): with open(_UpperCamelCase ) as f: __lowerCAmelCase : Optional[Any] = eval(f.read() ) else: __lowerCAmelCase : Optional[Any] = requests.get(_UpperCamelCase ) try: __lowerCAmelCase : Union[str, Any] = requests.json() except Exception: __lowerCAmelCase : List[Any] = req.content.decode() assert data is not None, "could not connect" try: __lowerCAmelCase : str = eval(_UpperCamelCase ) except Exception: __lowerCAmelCase : int = data.split('\n' ) req.close() return data def __lowerCAmelCase (_UpperCamelCase ): __lowerCAmelCase : Tuple = requests.get(_UpperCamelCase ) __lowerCAmelCase : List[Any] = np.array(Image.open(BytesIO(response.content ) ) ) return img def __lowerCAmelCase (_UpperCamelCase ): __lowerCAmelCase : List[str] = url.split('/' )[-1] if fn not in os.listdir(os.getcwd() ): wget.download(_UpperCamelCase ) with open(_UpperCamelCase , 'rb' ) as stream: __lowerCAmelCase : int = pkl.load(_UpperCamelCase ) __lowerCAmelCase : Dict = weights.pop('model' ) __lowerCAmelCase : Dict = {} for k, v in model.items(): __lowerCAmelCase : List[str] = torch.from_numpy(_UpperCamelCase ) if "running_var" in k: __lowerCAmelCase : List[str] = torch.tensor([0] ) __lowerCAmelCase : Optional[int] = k.replace('running_var' , 'num_batches_tracked' ) __lowerCAmelCase : Dict = zero return new def __lowerCAmelCase (): print(F"{os.path.abspath(os.path.join(_UpperCamelCase , os.pardir ) )}/demo.ipynb" ) def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase="RGB" ): assert isinstance(_UpperCamelCase , _UpperCamelCase ) if os.path.isfile(_UpperCamelCase ): __lowerCAmelCase : Union[str, Any] = cva.imread(_UpperCamelCase ) else: __lowerCAmelCase : Any = get_image_from_url(_UpperCamelCase ) assert img is not None, F"could not connect to: {im}" __lowerCAmelCase : Optional[int] = cva.cvtColor(_UpperCamelCase , cva.COLOR_BGR2RGB ) if input_format == "RGB": __lowerCAmelCase : Optional[int] = img[:, :, ::-1] return img def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase=1 ): return (images[i : i + batch] for i in range(0 , len(_UpperCamelCase ) , _UpperCamelCase ))
701
"""simple docstring""" import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class A__ ( unittest.TestCase): def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=13 , _SCREAMING_SNAKE_CASE=7 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=99 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=5 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE=37 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=5_12 , _SCREAMING_SNAKE_CASE=16 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=4 , ): __lowerCAmelCase : List[Any] = parent __lowerCAmelCase : int = batch_size __lowerCAmelCase : List[str] = seq_length __lowerCAmelCase : str = is_training __lowerCAmelCase : Optional[Any] = use_attention_mask __lowerCAmelCase : Union[str, Any] = use_token_type_ids __lowerCAmelCase : Dict = use_labels __lowerCAmelCase : Union[str, Any] = vocab_size __lowerCAmelCase : List[str] = hidden_size __lowerCAmelCase : List[str] = num_hidden_layers __lowerCAmelCase : Optional[int] = num_attention_heads __lowerCAmelCase : Dict = intermediate_size __lowerCAmelCase : Dict = hidden_act __lowerCAmelCase : Optional[int] = hidden_dropout_prob __lowerCAmelCase : Dict = attention_probs_dropout_prob __lowerCAmelCase : Union[str, Any] = max_position_embeddings __lowerCAmelCase : Any = type_vocab_size __lowerCAmelCase : Tuple = type_sequence_label_size __lowerCAmelCase : Union[str, Any] = initializer_range __lowerCAmelCase : int = num_choices def __lowerCamelCase ( self ): __lowerCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase : List[str] = None if self.use_attention_mask: __lowerCAmelCase : str = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCAmelCase : Dict = None if self.use_token_type_ids: __lowerCAmelCase : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCAmelCase : str = RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def __lowerCamelCase ( self ): __lowerCAmelCase : Dict = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : List[Any] = config_and_inputs __lowerCAmelCase : List[Any] = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict def __lowerCamelCase ( self ): __lowerCAmelCase : Optional[int] = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Any = config_and_inputs __lowerCAmelCase : int = True __lowerCAmelCase : Any = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __lowerCAmelCase : int = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class A__ ( _lowerCamelCase , unittest.TestCase): A_ : List[Any] = True A_ : Optional[Any] = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def __lowerCamelCase ( self ): __lowerCAmelCase : Tuple = FlaxRobertaModelTester(self ) @slow def __lowerCamelCase ( self ): for model_class_name in self.all_model_classes: __lowerCAmelCase : str = model_class_name.from_pretrained('roberta-base' , from_pt=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE )
549
0
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("4.31.0") lowercase_ = logging.getLogger(__name__) @dataclass class __A : '''simple docstring''' __lowerCamelCase : str = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) __lowerCamelCase : Optional[str] = field( default=A , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) __lowerCamelCase : Optional[str] = field( default=A , metadata={'help': 'Pretrained tokenizer name or path if not the same as model_name'} ) __lowerCamelCase : Optional[str] = field( default=A , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) __lowerCamelCase : bool = field( default=A , metadata={'help': 'Whether to use one of the fast tokenizer (backed by the tokenizers library) or not.'} , ) __lowerCamelCase : str = field( default='main' , metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'} , ) __lowerCamelCase : bool = field( default=A , metadata={ 'help': ( 'Will use the token generated when running `huggingface-cli login` (necessary to use this script ' 'with private models).' ) } , ) @dataclass class __A : '''simple docstring''' __lowerCamelCase : Optional[str] = field(default=A , metadata={'help': 'The input training data file (a text file).'} ) __lowerCamelCase : Optional[str] = field( default=A , metadata={'help': 'An optional input evaluation data file to evaluate the perplexity on (a text file).'} , ) __lowerCamelCase : bool = field( default=A , metadata={'help': 'Overwrite the cached training and evaluation sets'} ) __lowerCamelCase : Optional[int] = field( default=A , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) __lowerCamelCase : Optional[int] = field( default=A , metadata={ 'help': ( 'The maximum total input sequence length after tokenization. If passed, sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) } , ) __lowerCamelCase : bool = field( default=A , metadata={ 'help': ( 'Whether to pad all samples to the maximum sentence length. ' 'If False, will pad the samples dynamically when batching to the maximum length in the batch. More ' 'efficient on GPU but very bad for TPU.' ) } , ) __lowerCamelCase : Optional[int] = field( default=A , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) __lowerCamelCase : Optional[int] = field( default=A , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of evaluation examples to this ' 'value if set.' ) } , ) def a__ (self ) -> str: """simple docstring""" if self.train_file is not None: _a = self.train_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: _a = self.validation_file.split('''.''' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class __A : '''simple docstring''' __lowerCamelCase : PreTrainedTokenizerBase __lowerCamelCase : Union[bool, str, PaddingStrategy] = True __lowerCamelCase : Optional[int] = None __lowerCamelCase : Optional[int] = None def __call__(self , A ) -> int: """simple docstring""" _a = '''label''' if '''label''' in features[0].keys() else '''labels''' _a = [feature.pop(A ) for feature in features] _a = len(A ) _a = len(features[0]['''input_ids'''] ) _a = [ [{k: v[i] for k, v in feature.items()} for i in range(A )] for feature in features ] _a = list(chain(*A ) ) _a = self.tokenizer.pad( A , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' , ) # Un-flatten _a = {k: v.view(A , A , -1 ) for k, v in batch.items()} # Add back labels _a = torch.tensor(A , dtype=torch.intaa ) return batch def lowerCAmelCase (): """simple docstring""" _a = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments)) if len(sys.argv) == 2 and sys.argv[1].endswith('''.json'''): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _a , _a , _a = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1])) else: _a , _a , _a = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('''run_swag''' , __A , __A) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout)] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _a = training_args.get_process_log_level() logger.setLevel(__A) datasets.utils.logging.set_verbosity(__A) transformers.utils.logging.set_verbosity(__A) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1)}, 16-bits training: {training_args.fpaa}''') logger.info(F'''Training/evaluation parameters {training_args}''') # Detecting last checkpoint. _a = None if os.path.isdir(training_args.output_dir) and training_args.do_train and not training_args.overwrite_output_dir: _a = get_last_checkpoint(training_args.output_dir) if last_checkpoint is None and len(os.listdir(training_args.output_dir)) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' '''Use --overwrite_output_dir to overcome.''') elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''') # Set seed before initializing model. set_seed(training_args.seed) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: _a = {} if data_args.train_file is not None: _a = data_args.train_file if data_args.validation_file is not None: _a = data_args.validation_file _a = data_args.train_file.split('''.''')[-1] _a = load_dataset( __A , data_files=__A , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. _a = load_dataset( '''swag''' , '''regular''' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _a = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _a = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) _a = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('''.ckpt''' in model_args.model_name_or_path) , config=__A , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. _a = [F'''ending{i}''' for i in range(4)] _a = '''sent1''' _a = '''sent2''' if data_args.max_seq_length is None: _a = tokenizer.model_max_length if max_seq_length > 1_024: logger.warning( '''The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value''' ''' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can''' ''' override this default with `--block_size xxx`.''') _a = 1_024 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F'''The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the''' F'''model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.''') _a = min(data_args.max_seq_length , tokenizer.model_max_length) # Preprocessing the datasets. def preprocess_function(__A): _a = [[context] * 4 for context in examples[context_name]] _a = examples[question_header_name] _a = [ [F'''{header} {examples[end][i]}''' for end in ending_names] for i, header in enumerate(__A) ] # Flatten out _a = list(chain(*__A)) _a = list(chain(*__A)) # Tokenize _a = tokenizer( __A , __A , truncation=__A , max_length=__A , padding='''max_length''' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(__A) , 4)] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('''--do_train requires a train dataset''') _a = raw_datasets['''train'''] if data_args.max_train_samples is not None: _a = min(len(__A) , data_args.max_train_samples) _a = train_dataset.select(range(__A)) with training_args.main_process_first(desc='''train dataset map pre-processing'''): _a = train_dataset.map( __A , batched=__A , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('''--do_eval requires a validation dataset''') _a = raw_datasets['''validation'''] if data_args.max_eval_samples is not None: _a = min(len(__A) , data_args.max_eval_samples) _a = eval_dataset.select(range(__A)) with training_args.main_process_first(desc='''validation dataset map pre-processing'''): _a = eval_dataset.map( __A , batched=__A , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator _a = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=__A , pad_to_multiple_of=8 if training_args.fpaa else None) ) # Metric def compute_metrics(__A): _a , _a = eval_predictions _a = np.argmax(__A , axis=1) return {"accuracy": (preds == label_ids).astype(np.floataa).mean().item()} # Initialize our Trainer _a = Trainer( model=__A , args=__A , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=__A , data_collator=__A , compute_metrics=__A , ) # Training if training_args.do_train: _a = None if training_args.resume_from_checkpoint is not None: _a = training_args.resume_from_checkpoint elif last_checkpoint is not None: _a = last_checkpoint _a = trainer.train(resume_from_checkpoint=__A) trainer.save_model() # Saves the tokenizer too for easy upload _a = train_result.metrics _a = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(__A) ) _a = min(__A , len(__A)) trainer.log_metrics('''train''' , __A) trainer.save_metrics('''train''' , __A) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('''*** Evaluate ***''') _a = trainer.evaluate() _a = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(__A) _a = min(__A , len(__A)) trainer.log_metrics('''eval''' , __A) trainer.save_metrics('''eval''' , __A) _a = { '''finetuned_from''': model_args.model_name_or_path, '''tasks''': '''multiple-choice''', '''dataset_tags''': '''swag''', '''dataset_args''': '''regular''', '''dataset''': '''SWAG''', '''language''': '''en''', } if training_args.push_to_hub: trainer.push_to_hub(**__A) else: trainer.create_model_card(**__A) def lowerCAmelCase (__A): """simple docstring""" main() if __name__ == "__main__": main()
11
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __SCREAMING_SNAKE_CASE =logging.get_logger(__name__) __SCREAMING_SNAKE_CASE ={"""vocab_file""": """spiece.model"""} __SCREAMING_SNAKE_CASE ={ """vocab_file""": { """AI-Sweden/gpt-sw3-126m""": """https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-350m""": """https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-1.6b""": """https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-6.7b""": """https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model""", """AI-Sweden/gpt-sw3-20b""": """https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model""", } } __SCREAMING_SNAKE_CASE ={ """AI-Sweden/gpt-sw3-126m""": 2_048, """AI-Sweden/gpt-sw3-350m""": 2_048, """AI-Sweden/gpt-sw3-1.6b""": 2_048, """AI-Sweden/gpt-sw3-6.7b""": 2_048, """AI-Sweden/gpt-sw3-20b""": 2_048, } class __magic_name__ ( __UpperCAmelCase): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : List[Any] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : Optional[Any] = ["input_ids", "attention_mask"] def __init__( self: Optional[Any] , _lowerCamelCase: Optional[int] , _lowerCamelCase: Any=False , _lowerCamelCase: Union[str, Any]=False , _lowerCamelCase: str=False , _lowerCamelCase: Optional[Any]=None , _lowerCamelCase: Any=None , _lowerCamelCase: int=None , _lowerCamelCase: List[Any]=None , _lowerCamelCase: Optional[Dict[str, Any]] = None , **_lowerCamelCase: List[str] , ): SCREAMING_SNAKE_CASE_ = {} if sp_model_kwargs is None else sp_model_kwargs SCREAMING_SNAKE_CASE_ = kwargs.get('''name_or_path''' ) if name_or_path is None: logger.warning( '''name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,''' ''' you are testing the model, this can safely be ignored''' ) SCREAMING_SNAKE_CASE_ = '''None''' # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing SCREAMING_SNAKE_CASE_ = '''<|endoftext|>''' if eos_token is None else eos_token SCREAMING_SNAKE_CASE_ = '''<unk>''' if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: SCREAMING_SNAKE_CASE_ = unk_token if pad_token is None else pad_token SCREAMING_SNAKE_CASE_ = eos_token if bos_token is None else bos_token else: SCREAMING_SNAKE_CASE_ = '''<pad>''' if pad_token is None else pad_token SCREAMING_SNAKE_CASE_ = '''<s>''' if bos_token is None else bos_token super().__init__( do_lower_case=_lowerCamelCase , remove_space=_lowerCamelCase , keep_accents=_lowerCamelCase , bos_token=_lowerCamelCase , eos_token=_lowerCamelCase , unk_token=_lowerCamelCase , pad_token=_lowerCamelCase , sp_model_kwargs=self.sp_model_kwargs , **_lowerCamelCase , ) SCREAMING_SNAKE_CASE_ = do_lower_case SCREAMING_SNAKE_CASE_ = remove_space SCREAMING_SNAKE_CASE_ = keep_accents SCREAMING_SNAKE_CASE_ = vocab_file SCREAMING_SNAKE_CASE_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_lowerCamelCase ) # Used for whitespace normalization in input texts # fmt : off SCREAMING_SNAKE_CASE_ = {''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', ''' ''', '''''', '''„'''} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing SCREAMING_SNAKE_CASE_ = re.compile( f"[{''.join(map(_lowerCamelCase , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(1_27 , 1_60 ) ) + [1_60, 1_73, 82_03] ) )}]" ) def __getstate__( self: int ): SCREAMING_SNAKE_CASE_ = self.__dict__.copy() SCREAMING_SNAKE_CASE_ = None return state def __setstate__( self: List[Any] , _lowerCamelCase: Optional[Any] ): SCREAMING_SNAKE_CASE_ = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): SCREAMING_SNAKE_CASE_ = {} SCREAMING_SNAKE_CASE_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def _A ( self: int ): return len(self.sp_model ) def _A ( self: int , _lowerCamelCase: str ): SCREAMING_SNAKE_CASE_ = self.non_printing_characters_re.sub('''''' , _lowerCamelCase ) # Normalize whitespaces SCREAMING_SNAKE_CASE_ = ''''''.join([char if char not in self.whitespaces else ''' ''' for char in text] ) # NFC Unicode normalization SCREAMING_SNAKE_CASE_ = unicodedata.normalize('''NFC''' , _lowerCamelCase ) return text def _A ( self: Tuple , _lowerCamelCase: str , **_lowerCamelCase: Optional[Any] ): SCREAMING_SNAKE_CASE_ = self.preprocess_text(_lowerCamelCase ) return self.sp_model.encode(_lowerCamelCase , out_type=_lowerCamelCase ) def _A ( self: Union[str, Any] , _lowerCamelCase: str ): return self.sp_model.PieceToId(_lowerCamelCase ) def _A ( self: Optional[int] , _lowerCamelCase: int ): return self.sp_model.IdToPiece(_lowerCamelCase ) @staticmethod def _A ( _lowerCamelCase: str ): return out_string def _A ( self: Tuple , _lowerCamelCase: List[str] ): SCREAMING_SNAKE_CASE_ = [] SCREAMING_SNAKE_CASE_ = '''''' SCREAMING_SNAKE_CASE_ = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(_lowerCamelCase ) + token SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = [] else: current_sub_tokens.append(_lowerCamelCase ) SCREAMING_SNAKE_CASE_ = False out_string += self.sp_model.decode(_lowerCamelCase ) return out_string def _A ( self: Any ): SCREAMING_SNAKE_CASE_ = {self.convert_ids_to_tokens(_lowerCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _A ( self: Any , _lowerCamelCase: str , _lowerCamelCase: Optional[str] = None ): if not os.path.isdir(_lowerCamelCase ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return SCREAMING_SNAKE_CASE_ = os.path.join( _lowerCamelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_lowerCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _lowerCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(_lowerCamelCase , '''wb''' ) as fi: SCREAMING_SNAKE_CASE_ = self.sp_model.serialized_model_proto() fi.write(_lowerCamelCase ) return (out_vocab_file,) def _A ( self: Dict , _lowerCamelCase: Union[str, List[str]] , _lowerCamelCase: Union[str, bool] = False ): if isinstance(_lowerCamelCase , _lowerCamelCase ): SCREAMING_SNAKE_CASE_ = self.preprocess_text(_lowerCamelCase ) SCREAMING_SNAKE_CASE_ = self.sp_model.encode(_lowerCamelCase ) else: SCREAMING_SNAKE_CASE_ = [self.preprocess_text(_lowerCamelCase ) for t in text] SCREAMING_SNAKE_CASE_ = self.sp_model.encode(_lowerCamelCase ) if return_tensors is True or return_tensors == "pt": SCREAMING_SNAKE_CASE_ = torch.tensor(_lowerCamelCase ) return token_ids def _A ( self: List[Any] , _lowerCamelCase: Union[int, List[int]] ): return self.sp_model.decode(_lowerCamelCase ) def _A ( self: Union[str, Any] , _lowerCamelCase: "Conversation" ): SCREAMING_SNAKE_CASE_ = [f"User: {text}" if is_user else f"Bot: {text}" for is_user, text in conversation.iter_texts()] SCREAMING_SNAKE_CASE_ = ( f"{self.eos_token}{self.bos_token}" + f"{self.bos_token}".join(_lowerCamelCase ) + f"{self.bos_token}Bot:" ) return self.encode(text=_lowerCamelCase )
234
0
"""simple docstring""" from PIL import Image def lowercase (_snake_case ,_snake_case ) -> Image: '''simple docstring''' def brightness(_snake_case ) -> float: return 128 + level + (c - 128) if not -2_5_5.0 <= level <= 2_5_5.0: raise ValueError("level must be between -255.0 (black) and 255.0 (white)" ) return img.point(_snake_case ) if __name__ == "__main__": # Load image with Image.open("image_data/lena.jpg") as img: # Change brightness to 100 _A = change_brightness(img, 100) brigt_img.save("image_data/lena_brightness.png", format="png")
228
"""simple docstring""" from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar _A = TypeVar("T") class __UpperCAmelCase ( Generic[T] ): """simple docstring""" _snake_case : deque[T] # Cache store of keys _snake_case : set[T] # References of the keys in cache _snake_case : int = 1_0 # Maximum capacity of cache def __init__( self : List[Any] , A_ : int )-> None: __UpperCamelCase = deque() __UpperCamelCase = set() if not n: __UpperCamelCase = sys.maxsize elif n < 0: raise ValueError("n should be an integer greater than 0." ) else: __UpperCamelCase = n def A ( self : Dict , A_ : T )-> None: if x not in self.key_reference: if len(self.dq_store ) == LRUCache._MAX_CAPACITY: __UpperCamelCase = self.dq_store.pop() self.key_reference.remove(A_ ) else: self.dq_store.remove(A_ ) self.dq_store.appendleft(A_ ) self.key_reference.add(A_ ) def A ( self : int )-> None: for k in self.dq_store: print(A_ ) def __repr__( self : int )-> str: return f"""LRUCache({self._MAX_CAPACITY}) => {list(self.dq_store )}""" if __name__ == "__main__": import doctest doctest.testmod() _A = LRUCache(4) lru_cache.refer("A") lru_cache.refer(2) lru_cache.refer(3) lru_cache.refer("A") lru_cache.refer(4) lru_cache.refer(5) lru_cache.display() print(lru_cache) assert str(lru_cache) == "LRUCache(4) => [5, 4, 'A', 3]"
228
1
'''simple docstring''' def A_ ( _lowerCAmelCase : list[list] ): """simple docstring""" _lowerCamelCase : List[str] = current_set.copy() for row_index, row in enumerate(_lowerCAmelCase ): _lowerCamelCase : Optional[int] = row[0] for column_index, column in enumerate(_lowerCAmelCase ): if magnitude == 0: _lowerCamelCase : Optional[Any] = column continue _lowerCamelCase : List[str] = column / magnitude # Subtract to cancel term _lowerCamelCase : List[str] = current_set[0] _lowerCamelCase : Dict = [first_row] _lowerCamelCase : Union[str, Any] = current_set[1::] for row in current_set: _lowerCamelCase : int = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(_lowerCAmelCase ) continue for column_index in range(len(_lowerCAmelCase ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(_lowerCAmelCase ) # Create next recursion iteration set if len(final_set[0] ) != 3: _lowerCamelCase : Optional[Any] = final_set[0] _lowerCamelCase : List[str] = [] _lowerCamelCase : Optional[Any] = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) _lowerCamelCase : Optional[int] = simplify(_lowerCAmelCase ) for i in range(len(_lowerCAmelCase ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , _lowerCAmelCase ) _lowerCamelCase : Any = resultant return final_set def A_ ( _lowerCAmelCase : list[list] ): """simple docstring""" if len(_lowerCAmelCase ) == 0: raise IndexError("solve_simultaneous() requires n lists of length n+1" ) _lowerCamelCase : str = len(_lowerCAmelCase ) + 1 if any(len(_lowerCAmelCase ) != _length for item in equations ): raise IndexError("solve_simultaneous() requires n lists of length n+1" ) for row in equations: if any(not isinstance(_lowerCAmelCase , (int, float) ) for column in row ): raise ValueError("solve_simultaneous() requires lists of integers" ) if len(_lowerCAmelCase ) == 1: return [equations[0][-1] / equations[0][0]] _lowerCamelCase : int = equations.copy() if any(0 in row for row in data_set ): _lowerCamelCase : Union[str, Any] = data_set.copy() _lowerCamelCase : Optional[Any] = [] for row_index, row in enumerate(_lowerCAmelCase ): if 0 not in row: _lowerCamelCase : Dict = data_set.pop(_lowerCAmelCase ) break if not full_row: raise ValueError("solve_simultaneous() requires at least 1 full equation" ) data_set.insert(0 , _lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = data_set.copy() _lowerCamelCase : Union[str, Any] = simplify(_lowerCAmelCase ) _lowerCamelCase : int = simplified[::-1] _lowerCamelCase : list = [] for row in simplified: _lowerCamelCase : List[Any] = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue _lowerCamelCase : Optional[int] = row.copy()[: len(_lowerCAmelCase ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(_lowerCAmelCase ) == 0: solutions.append(0 ) continue _lowerCamelCase : List[str] = temp_row[1::] _lowerCamelCase : Union[str, Any] = temp_row[::-1] for column_index, column in enumerate(_lowerCAmelCase ): current_solution -= column * solutions[column_index] solutions.append(_lowerCAmelCase ) _lowerCamelCase : Tuple = [] for item in solutions: final.append(float(round(_lowerCAmelCase , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase_ : Union[str, Any] = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
44
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from typing import Optional import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor from torchvision.transforms.functional import InterpolationMode import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, ViTImageProcessor, ViTMAEConfig, ViTMAEForPreTraining, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version UpperCAmelCase = logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("""4.31.0""") require_version("""datasets>=1.8.0""", """To fix: pip install -r examples/pytorch/image-pretraining/requirements.txt""") @dataclass class lowercase__ : __UpperCAmelCase = field( default='''cifar10''' ,metadata={'''help''': '''Name of a dataset from the datasets package'''} ) __UpperCAmelCase = field( default=A_ ,metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) __UpperCAmelCase = field( default=A_ ,metadata={'''help''': '''The column name of the images in the files.'''} ) __UpperCAmelCase = field(default=A_ ,metadata={'''help''': '''A folder containing the training data.'''} ) __UpperCAmelCase = field(default=A_ ,metadata={'''help''': '''A folder containing the validation data.'''} ) __UpperCAmelCase = field( default=0.1_5 ,metadata={'''help''': '''Percent to split off of train for validation.'''} ) __UpperCAmelCase = field( default=A_ ,metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of training examples to this ''' '''value if set.''' ) } ,) __UpperCAmelCase = field( default=A_ ,metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of evaluation examples to this ''' '''value if set.''' ) } ,) def UpperCamelCase_ ( self) -> Any: _lowerCamelCase : Any = {} if self.train_dir is not None: _lowerCamelCase : int = self.train_dir if self.validation_dir is not None: _lowerCamelCase : Tuple = self.validation_dir _lowerCamelCase : Optional[int] = data_files if data_files else None @dataclass class lowercase__ : __UpperCAmelCase = field( default=A_ ,metadata={ '''help''': ( '''The model checkpoint for weights initialization.Don\'t set if you want to train a model from scratch.''' ) } ,) __UpperCAmelCase = field( default=A_ ,metadata={'''help''': '''Pretrained config name or path if not the same as model_name_or_path'''} ) __UpperCAmelCase = field( default=A_ ,metadata={ '''help''': ( '''Override some existing default config settings when a model is trained from scratch. Example: ''' '''n_embd=10,resid_pdrop=0.2,scale_attn_weights=false,summary_type=cls_index''' ) } ,) __UpperCAmelCase = field( default=A_ ,metadata={'''help''': '''Where do you want to store the pretrained models downloaded from s3'''} ) __UpperCAmelCase = field( default='''main''' ,metadata={'''help''': '''The specific model version to use (can be a branch name, tag name or commit id).'''} ,) __UpperCAmelCase = field(default=A_ ,metadata={'''help''': '''Name or path of preprocessor config.'''} ) __UpperCAmelCase = field( default=A_ ,metadata={ '''help''': ( '''Will use the token generated when running `huggingface-cli login` (necessary to use this script ''' '''with private models).''' ) } ,) __UpperCAmelCase = field( default=0.7_5 ,metadata={'''help''': '''The ratio of the number of masked tokens in the input sequence.'''} ) __UpperCAmelCase = field( default=A_ ,metadata={'''help''': '''Whether or not to train with normalized pixel values as target.'''} ) @dataclass class lowercase__ ( A_ ): __UpperCAmelCase = field( default=1e-3 ,metadata={'''help''': '''Base learning rate: absolute_lr = base_lr * total_batch_size / 256.'''} ) def _snake_case ( __snake_case : Optional[Any] ): """simple docstring""" _lowerCamelCase : int = torch.stack([example["""pixel_values"""] for example in examples] ) return {"pixel_values": pixel_values} def _snake_case ( ): """simple docstring""" _lowerCamelCase : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, CustomTrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _lowerCamelCase , _lowerCamelCase , _lowerCamelCase : Union[str, Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase : Dict = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry("""run_mae""" , __snake_case , __snake_case ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() _lowerCamelCase : Union[str, Any] = training_args.get_process_log_level() logger.setLevel(__snake_case ) transformers.utils.logging.set_verbosity(__snake_case ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}' + F'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) logger.info(F'Training/evaluation parameters {training_args}' ) # Detecting last checkpoint. _lowerCamelCase : List[Any] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _lowerCamelCase : Optional[int] = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'Output directory ({training_args.output_dir}) already exists and is not empty. ' """Use --overwrite_output_dir to overcome.""" ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' """the `--output_dir` or add `--overwrite_output_dir` to train from scratch.""" ) # Initialize our dataset. _lowerCamelCase : Optional[Any] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , data_files=data_args.data_files , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # If we don't have a validation split, split off a percentage of train as validation. _lowerCamelCase : Tuple = None if """validation""" in ds.keys() else data_args.train_val_split if isinstance(data_args.train_val_split , __snake_case ) and data_args.train_val_split > 0.0: _lowerCamelCase : List[str] = ds["""train"""].train_test_split(data_args.train_val_split ) _lowerCamelCase : Union[str, Any] = split["""train"""] _lowerCamelCase : Optional[int] = split["""test"""] # Load pretrained model and image processor # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _lowerCamelCase : str = { """cache_dir""": model_args.cache_dir, """revision""": model_args.model_revision, """use_auth_token""": True if model_args.use_auth_token else None, } if model_args.config_name: _lowerCamelCase : Dict = ViTMAEConfig.from_pretrained(model_args.config_name , **__snake_case ) elif model_args.model_name_or_path: _lowerCamelCase : Union[str, Any] = ViTMAEConfig.from_pretrained(model_args.model_name_or_path , **__snake_case ) else: _lowerCamelCase : Optional[Any] = ViTMAEConfig() logger.warning("""You are instantiating a new config instance from scratch.""" ) if model_args.config_overrides is not None: logger.info(F'Overriding config: {model_args.config_overrides}' ) config.update_from_string(model_args.config_overrides ) logger.info(F'New config: {config}' ) # adapt config config.update( { """mask_ratio""": model_args.mask_ratio, """norm_pix_loss""": model_args.norm_pix_loss, } ) # create image processor if model_args.image_processor_name: _lowerCamelCase : str = ViTImageProcessor.from_pretrained(model_args.image_processor_name , **__snake_case ) elif model_args.model_name_or_path: _lowerCamelCase : Dict = ViTImageProcessor.from_pretrained(model_args.model_name_or_path , **__snake_case ) else: _lowerCamelCase : Union[str, Any] = ViTImageProcessor() # create model if model_args.model_name_or_path: _lowerCamelCase : List[Any] = ViTMAEForPreTraining.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=__snake_case , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) else: logger.info("""Training new model from scratch""" ) _lowerCamelCase : Union[str, Any] = ViTMAEForPreTraining(__snake_case ) if training_args.do_train: _lowerCamelCase : List[Any] = ds["""train"""].column_names else: _lowerCamelCase : Union[str, Any] = ds["""validation"""].column_names if data_args.image_column_name is not None: _lowerCamelCase : str = data_args.image_column_name elif "image" in column_names: _lowerCamelCase : Optional[Any] = """image""" elif "img" in column_names: _lowerCamelCase : List[Any] = """img""" else: _lowerCamelCase : str = column_names[0] # transformations as done in original MAE paper # source: https://github.com/facebookresearch/mae/blob/main/main_pretrain.py if "shortest_edge" in image_processor.size: _lowerCamelCase : Dict = image_processor.size["""shortest_edge"""] else: _lowerCamelCase : List[Any] = (image_processor.size["""height"""], image_processor.size["""width"""]) _lowerCamelCase : Tuple = Compose( [ Lambda(lambda __snake_case : img.convert("""RGB""" ) if img.mode != "RGB" else img ), RandomResizedCrop(__snake_case , scale=(0.2, 1.0) , interpolation=InterpolationMode.BICUBIC ), RandomHorizontalFlip(), ToTensor(), Normalize(mean=image_processor.image_mean , std=image_processor.image_std ), ] ) def preprocess_images(__snake_case : Optional[Any] ): _lowerCamelCase : Dict = [transforms(__snake_case ) for image in examples[image_column_name]] return examples if training_args.do_train: if "train" not in ds: raise ValueError("""--do_train requires a train dataset""" ) if data_args.max_train_samples is not None: _lowerCamelCase : int = ds["""train"""].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) # Set the training transforms ds["train"].set_transform(__snake_case ) if training_args.do_eval: if "validation" not in ds: raise ValueError("""--do_eval requires a validation dataset""" ) if data_args.max_eval_samples is not None: _lowerCamelCase : Union[str, Any] = ( ds["""validation"""].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms ds["validation"].set_transform(__snake_case ) # Compute absolute learning rate _lowerCamelCase : Optional[Any] = ( training_args.train_batch_size * training_args.gradient_accumulation_steps * training_args.world_size ) if training_args.base_learning_rate is not None: _lowerCamelCase : Tuple = training_args.base_learning_rate * total_train_batch_size / 256 # Initialize our trainer _lowerCamelCase : Optional[Any] = Trainer( model=__snake_case , args=__snake_case , train_dataset=ds["""train"""] if training_args.do_train else None , eval_dataset=ds["""validation"""] if training_args.do_eval else None , tokenizer=__snake_case , data_collator=__snake_case , ) # Training if training_args.do_train: _lowerCamelCase : Any = None if training_args.resume_from_checkpoint is not None: _lowerCamelCase : List[Any] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _lowerCamelCase : Union[str, Any] = last_checkpoint _lowerCamelCase : Optional[Any] = trainer.train(resume_from_checkpoint=__snake_case ) trainer.save_model() trainer.log_metrics("""train""" , train_result.metrics ) trainer.save_metrics("""train""" , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: _lowerCamelCase : int = trainer.evaluate() trainer.log_metrics("""eval""" , __snake_case ) trainer.save_metrics("""eval""" , __snake_case ) # Write model card and (optionally) push to hub _lowerCamelCase : Optional[Any] = { """tasks""": """masked-auto-encoding""", """dataset""": data_args.dataset_name, """tags""": ["""masked-auto-encoding"""], } if training_args.push_to_hub: trainer.push_to_hub(**__snake_case ) else: trainer.create_model_card(**__snake_case ) def _snake_case ( __snake_case : Dict ): """simple docstring""" main() if __name__ == "__main__": main()
88
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _A = { "configuration_roberta_prelayernorm": [ "ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP", "RobertaPreLayerNormConfig", "RobertaPreLayerNormOnnxConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ "ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST", "RobertaPreLayerNormForCausalLM", "RobertaPreLayerNormForMaskedLM", "RobertaPreLayerNormForMultipleChoice", "RobertaPreLayerNormForQuestionAnswering", "RobertaPreLayerNormForSequenceClassification", "RobertaPreLayerNormForTokenClassification", "RobertaPreLayerNormModel", "RobertaPreLayerNormPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ "TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFRobertaPreLayerNormForCausalLM", "TFRobertaPreLayerNormForMaskedLM", "TFRobertaPreLayerNormForMultipleChoice", "TFRobertaPreLayerNormForQuestionAnswering", "TFRobertaPreLayerNormForSequenceClassification", "TFRobertaPreLayerNormForTokenClassification", "TFRobertaPreLayerNormMainLayer", "TFRobertaPreLayerNormModel", "TFRobertaPreLayerNormPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ "FlaxRobertaPreLayerNormForCausalLM", "FlaxRobertaPreLayerNormForMaskedLM", "FlaxRobertaPreLayerNormForMultipleChoice", "FlaxRobertaPreLayerNormForQuestionAnswering", "FlaxRobertaPreLayerNormForSequenceClassification", "FlaxRobertaPreLayerNormForTokenClassification", "FlaxRobertaPreLayerNormModel", "FlaxRobertaPreLayerNormPreTrainedModel", ] if TYPE_CHECKING: from .configuration_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaPreLayerNormConfig, RobertaPreLayerNormOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta_prelayernorm import ( ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaPreLayerNormForCausalLM, RobertaPreLayerNormForMaskedLM, RobertaPreLayerNormForMultipleChoice, RobertaPreLayerNormForQuestionAnswering, RobertaPreLayerNormForSequenceClassification, RobertaPreLayerNormForTokenClassification, RobertaPreLayerNormModel, RobertaPreLayerNormPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta_prelayernorm import ( TF_ROBERTA_PRELAYERNORM_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaPreLayerNormForCausalLM, TFRobertaPreLayerNormForMaskedLM, TFRobertaPreLayerNormForMultipleChoice, TFRobertaPreLayerNormForQuestionAnswering, TFRobertaPreLayerNormForSequenceClassification, TFRobertaPreLayerNormForTokenClassification, TFRobertaPreLayerNormMainLayer, TFRobertaPreLayerNormModel, TFRobertaPreLayerNormPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormPreTrainedModel, ) else: import sys _A = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
709
"""simple docstring""" import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _A = logging.get_logger(__name__) _A = { "microsoft/conditional-detr-resnet-50": ( "https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json" ), } class __UpperCAmelCase ( snake_case__ ): """simple docstring""" _snake_case : Union[str, Any] = 'conditional_detr' _snake_case : Any = ['past_key_values'] _snake_case : Dict = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self : List[Any] , A_ : int=True , A_ : List[Any]=None , A_ : int=3 , A_ : Union[str, Any]=3_00 , A_ : int=6 , A_ : List[Any]=20_48 , A_ : str=8 , A_ : Dict=6 , A_ : str=20_48 , A_ : str=8 , A_ : str=0.0 , A_ : List[Any]=0.0 , A_ : Union[str, Any]=True , A_ : List[str]="relu" , A_ : Optional[Any]=2_56 , A_ : Optional[int]=0.1 , A_ : Tuple=0.0 , A_ : List[str]=0.0 , A_ : Any=0.02 , A_ : int=1.0 , A_ : Any=False , A_ : Tuple="sine" , A_ : int="resnet50" , A_ : Dict=True , A_ : List[str]=False , A_ : Optional[Any]=2 , A_ : List[Any]=5 , A_ : List[str]=2 , A_ : Union[str, Any]=1 , A_ : Dict=1 , A_ : str=2 , A_ : Any=5 , A_ : Optional[int]=2 , A_ : List[str]=0.25 , **A_ : Union[str, Any] , )-> Optional[Any]: if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`." ) if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) __UpperCamelCase = CONFIG_MAPPING["resnet"](out_features=["stage4"] ) elif isinstance(A_ , A_ ): __UpperCamelCase = backbone_config.get("model_type" ) __UpperCamelCase = CONFIG_MAPPING[backbone_model_type] __UpperCamelCase = config_class.from_dict(A_ ) __UpperCamelCase = use_timm_backbone __UpperCamelCase = backbone_config __UpperCamelCase = num_channels __UpperCamelCase = num_queries __UpperCamelCase = d_model __UpperCamelCase = encoder_ffn_dim __UpperCamelCase = encoder_layers __UpperCamelCase = encoder_attention_heads __UpperCamelCase = decoder_ffn_dim __UpperCamelCase = decoder_layers __UpperCamelCase = decoder_attention_heads __UpperCamelCase = dropout __UpperCamelCase = attention_dropout __UpperCamelCase = activation_dropout __UpperCamelCase = activation_function __UpperCamelCase = init_std __UpperCamelCase = init_xavier_std __UpperCamelCase = encoder_layerdrop __UpperCamelCase = decoder_layerdrop __UpperCamelCase = encoder_layers __UpperCamelCase = auxiliary_loss __UpperCamelCase = position_embedding_type __UpperCamelCase = backbone __UpperCamelCase = use_pretrained_backbone __UpperCamelCase = dilation # Hungarian matcher __UpperCamelCase = class_cost __UpperCamelCase = bbox_cost __UpperCamelCase = giou_cost # Loss coefficients __UpperCamelCase = mask_loss_coefficient __UpperCamelCase = dice_loss_coefficient __UpperCamelCase = cls_loss_coefficient __UpperCamelCase = bbox_loss_coefficient __UpperCamelCase = giou_loss_coefficient __UpperCamelCase = focal_alpha super().__init__(is_encoder_decoder=A_ , **A_ ) @property def A ( self : int )-> int: return self.encoder_attention_heads @property def A ( self : List[Any] )-> int: return self.d_model def A ( self : List[Any] )-> Tuple: __UpperCamelCase = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: __UpperCamelCase = self.backbone_config.to_dict() __UpperCamelCase = self.__class__.model_type return output class __UpperCAmelCase ( snake_case__ ): """simple docstring""" _snake_case : Union[str, Any] = version.parse('1.11' ) @property def A ( self : str )-> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("pixel_mask", {0: "batch"}), ] ) @property def A ( self : Optional[Any] )-> float: return 1e-5 @property def A ( self : List[Any] )-> int: return 12
228
0
import os import posixpath import uuid from dataclasses import dataclass from typing import TYPE_CHECKING, Iterable, List, Optional, Tuple, Union import numpy as np import pyarrow as pa import datasets from datasets.arrow_writer import ArrowWriter, ParquetWriter from datasets.config import MAX_SHARD_SIZE from datasets.filesystems import ( is_remote_filesystem, rename, ) from datasets.iterable_dataset import _BaseExamplesIterable from datasets.utils.py_utils import convert_file_size_to_int __lowerCamelCase : Dict = datasets.utils.logging.get_logger(__name__) if TYPE_CHECKING: import pyspark @dataclass class a__ ( datasets.BuilderConfig ): A = None def _snake_case ( lowerCAmelCase : "pyspark.sql.DataFrame" , lowerCAmelCase : List[int] , ): """simple docstring""" import pyspark def generate_fn(): SCREAMING_SNAKE_CASE_ : Optional[Any] = df.select("*" , pyspark.sql.functions.spark_partition_id().alias("part_id" ) ) for partition_id in partition_order: SCREAMING_SNAKE_CASE_ : Optional[Any] = df_with_partition_id.select("*" ).where(f'part_id = {partition_id}' ).drop("part_id" ) SCREAMING_SNAKE_CASE_ : List[str] = partition_df.collect() SCREAMING_SNAKE_CASE_ : Tuple = 0 for row in rows: yield f'{partition_id}_{row_id}', row.asDict() row_id += 1 return generate_fn class a__ ( _BaseExamplesIterable ): def __init__( self : Union[str, Any],_A : "pyspark.sql.DataFrame",_A : Any=None,): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = df SCREAMING_SNAKE_CASE_ : Tuple = partition_order or range(self.df.rdd.getNumPartitions() ) SCREAMING_SNAKE_CASE_ : Optional[Any] = _generate_iterable_examples(self.df,self.partition_order ) def __iter__( self : Union[str, Any] ): """simple docstring""" yield from self.generate_examples_fn() def __UpperCamelCase ( self : int,_A : np.random.Generator ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = list(range(self.df.rdd.getNumPartitions() ) ) generator.shuffle(_A ) return SparkExamplesIterable(self.df,partition_order=_A ) def __UpperCamelCase ( self : str,_A : int,_A : int ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.split_shard_indices_by_worker(_A,_A ) return SparkExamplesIterable(self.df,partition_order=_A ) @property def __UpperCamelCase ( self : str ): """simple docstring""" return len(self.partition_order ) class a__ ( datasets.DatasetBuilder ): A = SparkConfig def __init__( self : List[str],_A : "pyspark.sql.DataFrame",_A : str = None,_A : str = None,**_A : Optional[Any],): """simple docstring""" import pyspark SCREAMING_SNAKE_CASE_ : Union[str, Any] = pyspark.sql.SparkSession.builder.getOrCreate() SCREAMING_SNAKE_CASE_ : Optional[int] = df SCREAMING_SNAKE_CASE_ : Dict = working_dir super().__init__( cache_dir=_A,config_name=str(self.df.semanticHash() ),**_A,) def __UpperCamelCase ( self : int ): """simple docstring""" def create_cache_and_write_probe(_A : int ): # makedirs with exist_ok will recursively create the directory. It will not throw an error if directories # already exist. os.makedirs(self._cache_dir,exist_ok=_A ) SCREAMING_SNAKE_CASE_ : int = os.path.join(self._cache_dir,"fs_test" + uuid.uuida().hex ) # Opening the file in append mode will create a new file unless it already exists, in which case it will not # change the file contents. open(_A,"a" ) return [probe_file] if self._spark.conf.get("spark.master","" ).startswith("local" ): return # If the cluster is multi-node, make sure that the user provided a cache_dir and that it is on an NFS # accessible to the driver. # TODO: Stream batches to the driver using ArrowCollectSerializer instead of throwing an error. if self._cache_dir: SCREAMING_SNAKE_CASE_ : Optional[Any] = ( self._spark.sparkContext.parallelize(range(1 ),1 ).mapPartitions(_A ).collect() ) if os.path.isfile(probe[0] ): return raise ValueError( "When using Dataset.from_spark on a multi-node cluster, the driver and all workers should be able to access cache_dir" ) def __UpperCamelCase ( self : List[str] ): """simple docstring""" return datasets.DatasetInfo(features=self.config.features ) def __UpperCamelCase ( self : Tuple,_A : datasets.download.download_manager.DownloadManager ): """simple docstring""" return [datasets.SplitGenerator(name=datasets.Split.TRAIN )] def __UpperCamelCase ( self : Union[str, Any],_A : List[Any] ): """simple docstring""" import pyspark def get_arrow_batch_size(_A : Optional[Any] ): for batch in it: yield pa.RecordBatch.from_pydict({"batch_bytes": [batch.nbytes]} ) SCREAMING_SNAKE_CASE_ : Optional[int] = self.df.count() SCREAMING_SNAKE_CASE_ : str = df_num_rows if df_num_rows <= 100 else 100 # Approximate the size of each row (in Arrow format) by averaging over a max-100-row sample. SCREAMING_SNAKE_CASE_ : List[str] = ( self.df.limit(_A ) .repartition(1 ) .mapInArrow(_A,"batch_bytes: long" ) .agg(pyspark.sql.functions.sum("batch_bytes" ).alias("sample_bytes" ) ) .collect()[0] .sample_bytes / sample_num_rows ) SCREAMING_SNAKE_CASE_ : str = approx_bytes_per_row * df_num_rows if approx_total_size > max_shard_size: # Make sure there is at least one row per partition. SCREAMING_SNAKE_CASE_ : int = min(_A,int(approx_total_size / max_shard_size ) ) SCREAMING_SNAKE_CASE_ : List[Any] = self.df.repartition(_A ) def __UpperCamelCase ( self : Any,_A : str,_A : str,_A : int,): """simple docstring""" import pyspark SCREAMING_SNAKE_CASE_ : Union[str, Any] = ParquetWriter if file_format == "parquet" else ArrowWriter SCREAMING_SNAKE_CASE_ : Tuple = os.path.join(self._working_dir,os.path.basename(_A ) ) if self._working_dir else fpath SCREAMING_SNAKE_CASE_ : Tuple = file_format == "parquet" # Define these so that we don't reference self in write_arrow, which will result in a pickling error due to # pickling the SparkContext. SCREAMING_SNAKE_CASE_ : Dict = self.config.features SCREAMING_SNAKE_CASE_ : Optional[int] = self._writer_batch_size SCREAMING_SNAKE_CASE_ : Tuple = self._fs.storage_options def write_arrow(_A : Optional[int] ): # Within the same SparkContext, no two task attempts will share the same attempt ID. SCREAMING_SNAKE_CASE_ : Any = pyspark.TaskContext().taskAttemptId() SCREAMING_SNAKE_CASE_ : List[Any] = next(_A,_A ) if first_batch is None: # Some partitions might not receive any data. return pa.RecordBatch.from_arrays( [[task_id], [0], [0]],names=["task_id", "num_examples", "num_bytes"],) SCREAMING_SNAKE_CASE_ : Optional[int] = 0 SCREAMING_SNAKE_CASE_ : Tuple = writer_class( features=_A,path=working_fpath.replace("SSSSS",F'{shard_id:05d}' ).replace("TTTTT",F'{task_id:05d}' ),writer_batch_size=_A,storage_options=_A,embed_local_files=_A,) SCREAMING_SNAKE_CASE_ : Dict = pa.Table.from_batches([first_batch] ) writer.write_table(_A ) for batch in it: if max_shard_size is not None and writer._num_bytes >= max_shard_size: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Union[str, Any] = writer.finalize() writer.close() yield pa.RecordBatch.from_arrays( [[task_id], [num_examples], [num_bytes]],names=["task_id", "num_examples", "num_bytes"],) shard_id += 1 SCREAMING_SNAKE_CASE_ : List[str] = writer_class( features=writer._features,path=working_fpath.replace("SSSSS",F'{shard_id:05d}' ).replace("TTTTT",F'{task_id:05d}' ),writer_batch_size=_A,storage_options=_A,embed_local_files=_A,) SCREAMING_SNAKE_CASE_ : List[Any] = pa.Table.from_batches([batch] ) writer.write_table(_A ) if writer._num_bytes > 0: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : str = writer.finalize() writer.close() yield pa.RecordBatch.from_arrays( [[task_id], [num_examples], [num_bytes]],names=["task_id", "num_examples", "num_bytes"],) if working_fpath != fpath: for file in os.listdir(os.path.dirname(_A ) ): SCREAMING_SNAKE_CASE_ : List[str] = os.path.join(os.path.dirname(_A ),os.path.basename(_A ) ) shutil.move(_A,_A ) SCREAMING_SNAKE_CASE_ : Any = ( self.df.mapInArrow(_A,"task_id: long, num_examples: long, num_bytes: long" ) .groupBy("task_id" ) .agg( pyspark.sql.functions.sum("num_examples" ).alias("total_num_examples" ),pyspark.sql.functions.sum("num_bytes" ).alias("total_num_bytes" ),pyspark.sql.functions.count("num_bytes" ).alias("num_shards" ),pyspark.sql.functions.collect_list("num_examples" ).alias("shard_lengths" ),) .collect() ) for row in stats: yield row.task_id, (row.total_num_examples, row.total_num_bytes, row.num_shards, row.shard_lengths) def __UpperCamelCase ( self : Optional[int],_A : "datasets.SplitGenerator",_A : str = "arrow",_A : Optional[Union[str, int]] = None,_A : Optional[int] = None,**_A : Union[str, Any],): """simple docstring""" self._validate_cache_dir() SCREAMING_SNAKE_CASE_ : int = convert_file_size_to_int(max_shard_size or MAX_SHARD_SIZE ) self._repartition_df_if_needed(_A ) SCREAMING_SNAKE_CASE_ : List[Any] = not is_remote_filesystem(self._fs ) SCREAMING_SNAKE_CASE_ : int = os.path.join if is_local else posixpath.join SCREAMING_SNAKE_CASE_ : str = "-TTTTT-SSSSS-of-NNNNN" SCREAMING_SNAKE_CASE_ : Any = F'{self.name}-{split_generator.name}{SUFFIX}.{file_format}' SCREAMING_SNAKE_CASE_ : Dict = path_join(self._output_dir,_A ) SCREAMING_SNAKE_CASE_ : Optional[Any] = 0 SCREAMING_SNAKE_CASE_ : Optional[int] = 0 SCREAMING_SNAKE_CASE_ : List[str] = 0 SCREAMING_SNAKE_CASE_ : List[Any] = [] SCREAMING_SNAKE_CASE_ : Any = [] for task_id, content in self._prepare_split_single(_A,_A,_A ): ( ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ) : Union[str, Any] = content if num_bytes > 0: total_num_examples += num_examples total_num_bytes += num_bytes total_shards += num_shards task_id_and_num_shards.append((task_id, num_shards) ) all_shard_lengths.extend(_A ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = total_num_examples SCREAMING_SNAKE_CASE_ : Optional[int] = total_num_bytes # should rename everything at the end logger.debug(F'Renaming {total_shards} shards.' ) if total_shards > 1: SCREAMING_SNAKE_CASE_ : int = all_shard_lengths # Define fs outside of _rename_shard so that we don't reference self in the function, which will result in a # pickling error due to pickling the SparkContext. SCREAMING_SNAKE_CASE_ : List[Any] = self._fs # use the -SSSSS-of-NNNNN pattern def _rename_shard( _A : int,_A : int,_A : int,): rename( _A,fpath.replace("SSSSS",F'{shard_id:05d}' ).replace("TTTTT",F'{task_id:05d}' ),fpath.replace("TTTTT-SSSSS",F'{global_shard_id:05d}' ).replace("NNNNN",F'{total_shards:05d}' ),) SCREAMING_SNAKE_CASE_ : List[Any] = [] SCREAMING_SNAKE_CASE_ : List[str] = 0 for i in range(len(_A ) ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Any = task_id_and_num_shards[i] for shard_id in range(_A ): args.append([task_id, shard_id, global_shard_id] ) global_shard_id += 1 self._spark.sparkContext.parallelize(_A,len(_A ) ).map(lambda _A : _rename_shard(*_A ) ).collect() else: # don't use any pattern SCREAMING_SNAKE_CASE_ : Any = 0 SCREAMING_SNAKE_CASE_ : Optional[int] = task_id_and_num_shards[0][0] self._rename( fpath.replace("SSSSS",F'{shard_id:05d}' ).replace("TTTTT",F'{task_id:05d}' ),fpath.replace(_A,"" ),) def __UpperCamelCase ( self : List[str],_A : "datasets.SplitGenerator",): """simple docstring""" return SparkExamplesIterable(self.df )
216
import os import posixpath import uuid from dataclasses import dataclass from typing import TYPE_CHECKING, Iterable, List, Optional, Tuple, Union import numpy as np import pyarrow as pa import datasets from datasets.arrow_writer import ArrowWriter, ParquetWriter from datasets.config import MAX_SHARD_SIZE from datasets.filesystems import ( is_remote_filesystem, rename, ) from datasets.iterable_dataset import _BaseExamplesIterable from datasets.utils.py_utils import convert_file_size_to_int __lowerCamelCase : Dict = datasets.utils.logging.get_logger(__name__) if TYPE_CHECKING: import pyspark @dataclass class a__ ( datasets.BuilderConfig ): A = None def _snake_case ( lowerCAmelCase : "pyspark.sql.DataFrame" , lowerCAmelCase : List[int] , ): """simple docstring""" import pyspark def generate_fn(): SCREAMING_SNAKE_CASE_ : Optional[Any] = df.select("*" , pyspark.sql.functions.spark_partition_id().alias("part_id" ) ) for partition_id in partition_order: SCREAMING_SNAKE_CASE_ : Optional[Any] = df_with_partition_id.select("*" ).where(f'part_id = {partition_id}' ).drop("part_id" ) SCREAMING_SNAKE_CASE_ : List[str] = partition_df.collect() SCREAMING_SNAKE_CASE_ : Tuple = 0 for row in rows: yield f'{partition_id}_{row_id}', row.asDict() row_id += 1 return generate_fn class a__ ( _BaseExamplesIterable ): def __init__( self : Union[str, Any],_A : "pyspark.sql.DataFrame",_A : Any=None,): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = df SCREAMING_SNAKE_CASE_ : Tuple = partition_order or range(self.df.rdd.getNumPartitions() ) SCREAMING_SNAKE_CASE_ : Optional[Any] = _generate_iterable_examples(self.df,self.partition_order ) def __iter__( self : Union[str, Any] ): """simple docstring""" yield from self.generate_examples_fn() def __UpperCamelCase ( self : int,_A : np.random.Generator ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = list(range(self.df.rdd.getNumPartitions() ) ) generator.shuffle(_A ) return SparkExamplesIterable(self.df,partition_order=_A ) def __UpperCamelCase ( self : str,_A : int,_A : int ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.split_shard_indices_by_worker(_A,_A ) return SparkExamplesIterable(self.df,partition_order=_A ) @property def __UpperCamelCase ( self : str ): """simple docstring""" return len(self.partition_order ) class a__ ( datasets.DatasetBuilder ): A = SparkConfig def __init__( self : List[str],_A : "pyspark.sql.DataFrame",_A : str = None,_A : str = None,**_A : Optional[Any],): """simple docstring""" import pyspark SCREAMING_SNAKE_CASE_ : Union[str, Any] = pyspark.sql.SparkSession.builder.getOrCreate() SCREAMING_SNAKE_CASE_ : Optional[int] = df SCREAMING_SNAKE_CASE_ : Dict = working_dir super().__init__( cache_dir=_A,config_name=str(self.df.semanticHash() ),**_A,) def __UpperCamelCase ( self : int ): """simple docstring""" def create_cache_and_write_probe(_A : int ): # makedirs with exist_ok will recursively create the directory. It will not throw an error if directories # already exist. os.makedirs(self._cache_dir,exist_ok=_A ) SCREAMING_SNAKE_CASE_ : int = os.path.join(self._cache_dir,"fs_test" + uuid.uuida().hex ) # Opening the file in append mode will create a new file unless it already exists, in which case it will not # change the file contents. open(_A,"a" ) return [probe_file] if self._spark.conf.get("spark.master","" ).startswith("local" ): return # If the cluster is multi-node, make sure that the user provided a cache_dir and that it is on an NFS # accessible to the driver. # TODO: Stream batches to the driver using ArrowCollectSerializer instead of throwing an error. if self._cache_dir: SCREAMING_SNAKE_CASE_ : Optional[Any] = ( self._spark.sparkContext.parallelize(range(1 ),1 ).mapPartitions(_A ).collect() ) if os.path.isfile(probe[0] ): return raise ValueError( "When using Dataset.from_spark on a multi-node cluster, the driver and all workers should be able to access cache_dir" ) def __UpperCamelCase ( self : List[str] ): """simple docstring""" return datasets.DatasetInfo(features=self.config.features ) def __UpperCamelCase ( self : Tuple,_A : datasets.download.download_manager.DownloadManager ): """simple docstring""" return [datasets.SplitGenerator(name=datasets.Split.TRAIN )] def __UpperCamelCase ( self : Union[str, Any],_A : List[Any] ): """simple docstring""" import pyspark def get_arrow_batch_size(_A : Optional[Any] ): for batch in it: yield pa.RecordBatch.from_pydict({"batch_bytes": [batch.nbytes]} ) SCREAMING_SNAKE_CASE_ : Optional[int] = self.df.count() SCREAMING_SNAKE_CASE_ : str = df_num_rows if df_num_rows <= 100 else 100 # Approximate the size of each row (in Arrow format) by averaging over a max-100-row sample. SCREAMING_SNAKE_CASE_ : List[str] = ( self.df.limit(_A ) .repartition(1 ) .mapInArrow(_A,"batch_bytes: long" ) .agg(pyspark.sql.functions.sum("batch_bytes" ).alias("sample_bytes" ) ) .collect()[0] .sample_bytes / sample_num_rows ) SCREAMING_SNAKE_CASE_ : str = approx_bytes_per_row * df_num_rows if approx_total_size > max_shard_size: # Make sure there is at least one row per partition. SCREAMING_SNAKE_CASE_ : int = min(_A,int(approx_total_size / max_shard_size ) ) SCREAMING_SNAKE_CASE_ : List[Any] = self.df.repartition(_A ) def __UpperCamelCase ( self : Any,_A : str,_A : str,_A : int,): """simple docstring""" import pyspark SCREAMING_SNAKE_CASE_ : Union[str, Any] = ParquetWriter if file_format == "parquet" else ArrowWriter SCREAMING_SNAKE_CASE_ : Tuple = os.path.join(self._working_dir,os.path.basename(_A ) ) if self._working_dir else fpath SCREAMING_SNAKE_CASE_ : Tuple = file_format == "parquet" # Define these so that we don't reference self in write_arrow, which will result in a pickling error due to # pickling the SparkContext. SCREAMING_SNAKE_CASE_ : Dict = self.config.features SCREAMING_SNAKE_CASE_ : Optional[int] = self._writer_batch_size SCREAMING_SNAKE_CASE_ : Tuple = self._fs.storage_options def write_arrow(_A : Optional[int] ): # Within the same SparkContext, no two task attempts will share the same attempt ID. SCREAMING_SNAKE_CASE_ : Any = pyspark.TaskContext().taskAttemptId() SCREAMING_SNAKE_CASE_ : List[Any] = next(_A,_A ) if first_batch is None: # Some partitions might not receive any data. return pa.RecordBatch.from_arrays( [[task_id], [0], [0]],names=["task_id", "num_examples", "num_bytes"],) SCREAMING_SNAKE_CASE_ : Optional[int] = 0 SCREAMING_SNAKE_CASE_ : Tuple = writer_class( features=_A,path=working_fpath.replace("SSSSS",F'{shard_id:05d}' ).replace("TTTTT",F'{task_id:05d}' ),writer_batch_size=_A,storage_options=_A,embed_local_files=_A,) SCREAMING_SNAKE_CASE_ : Dict = pa.Table.from_batches([first_batch] ) writer.write_table(_A ) for batch in it: if max_shard_size is not None and writer._num_bytes >= max_shard_size: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Union[str, Any] = writer.finalize() writer.close() yield pa.RecordBatch.from_arrays( [[task_id], [num_examples], [num_bytes]],names=["task_id", "num_examples", "num_bytes"],) shard_id += 1 SCREAMING_SNAKE_CASE_ : List[str] = writer_class( features=writer._features,path=working_fpath.replace("SSSSS",F'{shard_id:05d}' ).replace("TTTTT",F'{task_id:05d}' ),writer_batch_size=_A,storage_options=_A,embed_local_files=_A,) SCREAMING_SNAKE_CASE_ : List[Any] = pa.Table.from_batches([batch] ) writer.write_table(_A ) if writer._num_bytes > 0: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : str = writer.finalize() writer.close() yield pa.RecordBatch.from_arrays( [[task_id], [num_examples], [num_bytes]],names=["task_id", "num_examples", "num_bytes"],) if working_fpath != fpath: for file in os.listdir(os.path.dirname(_A ) ): SCREAMING_SNAKE_CASE_ : List[str] = os.path.join(os.path.dirname(_A ),os.path.basename(_A ) ) shutil.move(_A,_A ) SCREAMING_SNAKE_CASE_ : Any = ( self.df.mapInArrow(_A,"task_id: long, num_examples: long, num_bytes: long" ) .groupBy("task_id" ) .agg( pyspark.sql.functions.sum("num_examples" ).alias("total_num_examples" ),pyspark.sql.functions.sum("num_bytes" ).alias("total_num_bytes" ),pyspark.sql.functions.count("num_bytes" ).alias("num_shards" ),pyspark.sql.functions.collect_list("num_examples" ).alias("shard_lengths" ),) .collect() ) for row in stats: yield row.task_id, (row.total_num_examples, row.total_num_bytes, row.num_shards, row.shard_lengths) def __UpperCamelCase ( self : Optional[int],_A : "datasets.SplitGenerator",_A : str = "arrow",_A : Optional[Union[str, int]] = None,_A : Optional[int] = None,**_A : Union[str, Any],): """simple docstring""" self._validate_cache_dir() SCREAMING_SNAKE_CASE_ : int = convert_file_size_to_int(max_shard_size or MAX_SHARD_SIZE ) self._repartition_df_if_needed(_A ) SCREAMING_SNAKE_CASE_ : List[Any] = not is_remote_filesystem(self._fs ) SCREAMING_SNAKE_CASE_ : int = os.path.join if is_local else posixpath.join SCREAMING_SNAKE_CASE_ : str = "-TTTTT-SSSSS-of-NNNNN" SCREAMING_SNAKE_CASE_ : Any = F'{self.name}-{split_generator.name}{SUFFIX}.{file_format}' SCREAMING_SNAKE_CASE_ : Dict = path_join(self._output_dir,_A ) SCREAMING_SNAKE_CASE_ : Optional[Any] = 0 SCREAMING_SNAKE_CASE_ : Optional[int] = 0 SCREAMING_SNAKE_CASE_ : List[str] = 0 SCREAMING_SNAKE_CASE_ : List[Any] = [] SCREAMING_SNAKE_CASE_ : Any = [] for task_id, content in self._prepare_split_single(_A,_A,_A ): ( ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ) : Union[str, Any] = content if num_bytes > 0: total_num_examples += num_examples total_num_bytes += num_bytes total_shards += num_shards task_id_and_num_shards.append((task_id, num_shards) ) all_shard_lengths.extend(_A ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = total_num_examples SCREAMING_SNAKE_CASE_ : Optional[int] = total_num_bytes # should rename everything at the end logger.debug(F'Renaming {total_shards} shards.' ) if total_shards > 1: SCREAMING_SNAKE_CASE_ : int = all_shard_lengths # Define fs outside of _rename_shard so that we don't reference self in the function, which will result in a # pickling error due to pickling the SparkContext. SCREAMING_SNAKE_CASE_ : List[Any] = self._fs # use the -SSSSS-of-NNNNN pattern def _rename_shard( _A : int,_A : int,_A : int,): rename( _A,fpath.replace("SSSSS",F'{shard_id:05d}' ).replace("TTTTT",F'{task_id:05d}' ),fpath.replace("TTTTT-SSSSS",F'{global_shard_id:05d}' ).replace("NNNNN",F'{total_shards:05d}' ),) SCREAMING_SNAKE_CASE_ : List[Any] = [] SCREAMING_SNAKE_CASE_ : List[str] = 0 for i in range(len(_A ) ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Any = task_id_and_num_shards[i] for shard_id in range(_A ): args.append([task_id, shard_id, global_shard_id] ) global_shard_id += 1 self._spark.sparkContext.parallelize(_A,len(_A ) ).map(lambda _A : _rename_shard(*_A ) ).collect() else: # don't use any pattern SCREAMING_SNAKE_CASE_ : Any = 0 SCREAMING_SNAKE_CASE_ : Optional[int] = task_id_and_num_shards[0][0] self._rename( fpath.replace("SSSSS",F'{shard_id:05d}' ).replace("TTTTT",F'{task_id:05d}' ),fpath.replace(_A,"" ),) def __UpperCamelCase ( self : List[str],_A : "datasets.SplitGenerator",): """simple docstring""" return SparkExamplesIterable(self.df )
216
1
from typing import Dict, Optional import numpy as np import datasets _lowerCamelCase = '\nIoU is the area of overlap between the predicted segmentation and the ground truth divided by the area of union\nbetween the predicted segmentation and the ground truth. For binary (two classes) or multi-class segmentation,\nthe mean IoU of the image is calculated by taking the IoU of each class and averaging them.\n' _lowerCamelCase = '\nArgs:\n predictions (`List[ndarray]`):\n List of predicted segmentation maps, each of shape (height, width). Each segmentation map can be of a different size.\n references (`List[ndarray]`):\n List of ground truth segmentation maps, each of shape (height, width). Each segmentation map can be of a different size.\n num_labels (`int`):\n Number of classes (categories).\n ignore_index (`int`):\n Index that will be ignored during evaluation.\n nan_to_num (`int`, *optional*):\n If specified, NaN values will be replaced by the number defined by the user.\n label_map (`dict`, *optional*):\n If specified, dictionary mapping old label indices to new label indices.\n reduce_labels (`bool`, *optional*, defaults to `False`):\n Whether or not to reduce all label values of segmentation maps by 1. Usually used for datasets where 0 is used for background,\n and background itself is not included in all classes of a dataset (e.g. ADE20k). The background label will be replaced by 255.\n\nReturns:\n `Dict[str, float | ndarray]` comprising various elements:\n - *mean_iou* (`float`):\n Mean Intersection-over-Union (IoU averaged over all categories).\n - *mean_accuracy* (`float`):\n Mean accuracy (averaged over all categories).\n - *overall_accuracy* (`float`):\n Overall accuracy on all images.\n - *per_category_accuracy* (`ndarray` of shape `(num_labels,)`):\n Per category accuracy.\n - *per_category_iou* (`ndarray` of shape `(num_labels,)`):\n Per category IoU.\n\nExamples:\n\n >>> import numpy as np\n\n >>> mean_iou = datasets.load_metric("mean_iou")\n\n >>> # suppose one has 3 different segmentation maps predicted\n >>> predicted_1 = np.array([[1, 2], [3, 4], [5, 255]])\n >>> actual_1 = np.array([[0, 3], [5, 4], [6, 255]])\n\n >>> predicted_2 = np.array([[2, 7], [9, 2], [3, 6]])\n >>> actual_2 = np.array([[1, 7], [9, 2], [3, 6]])\n\n >>> predicted_3 = np.array([[2, 2, 3], [8, 2, 4], [3, 255, 2]])\n >>> actual_3 = np.array([[1, 2, 2], [8, 2, 1], [3, 255, 1]])\n\n >>> predicted = [predicted_1, predicted_2, predicted_3]\n >>> ground_truth = [actual_1, actual_2, actual_3]\n\n >>> results = mean_iou.compute(predictions=predicted, references=ground_truth, num_labels=10, ignore_index=255, reduce_labels=False)\n >>> print(results) # doctest: +NORMALIZE_WHITESPACE\n {\'mean_iou\': 0.47750000000000004, \'mean_accuracy\': 0.5916666666666666, \'overall_accuracy\': 0.5263157894736842, \'per_category_iou\': array([0. , 0. , 0.375, 0.4 , 0.5 , 0. , 0.5 , 1. , 1. , 1. ]), \'per_category_accuracy\': array([0. , 0. , 0.75 , 0.66666667, 1. , 0. , 0.5 , 1. , 1. , 1. ])}\n' _lowerCamelCase = '\\n@software{MMSegmentation_Contributors_OpenMMLab_Semantic_Segmentation_2020,\nauthor = {{MMSegmentation Contributors}},\nlicense = {Apache-2.0},\nmonth = {7},\ntitle = {{OpenMMLab Semantic Segmentation Toolbox and Benchmark}},\nurl = {https://github.com/open-mmlab/mmsegmentation},\nyear = {2020}\n}' def __UpperCAmelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ = None , lowercase_ = False , ): if label_map is not None: for old_id, new_id in label_map.items(): _lowerCamelCase : Optional[int] = new_id # turn into Numpy arrays _lowerCamelCase : int = np.array(lowercase_ ) _lowerCamelCase : List[str] = np.array(lowercase_ ) if reduce_labels: _lowerCamelCase : Any = 2_55 _lowerCamelCase : List[str] = label - 1 _lowerCamelCase : Optional[Any] = 2_55 _lowerCamelCase : Optional[Any] = label != ignore_index _lowerCamelCase : Dict = np.not_equal(lowercase_ , lowercase_ ) _lowerCamelCase : Any = pred_label[mask] _lowerCamelCase : List[str] = np.array(lowercase_ )[mask] _lowerCamelCase : int = pred_label[pred_label == label] _lowerCamelCase : Optional[int] = np.histogram(lowercase_ , bins=lowercase_ , range=(0, num_labels - 1) )[0] _lowerCamelCase : str = np.histogram(lowercase_ , bins=lowercase_ , range=(0, num_labels - 1) )[0] _lowerCamelCase : Optional[Any] = np.histogram(lowercase_ , bins=lowercase_ , range=(0, num_labels - 1) )[0] _lowerCamelCase : Any = area_pred_label + area_label - area_intersect return area_intersect, area_union, area_pred_label, area_label def __UpperCAmelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ = None , lowercase_ = False , ): _lowerCamelCase : List[str] = np.zeros((num_labels,) , dtype=np.floataa ) _lowerCamelCase : Optional[int] = np.zeros((num_labels,) , dtype=np.floataa ) _lowerCamelCase : Optional[int] = np.zeros((num_labels,) , dtype=np.floataa ) _lowerCamelCase : Tuple = np.zeros((num_labels,) , dtype=np.floataa ) for result, gt_seg_map in zip(lowercase_ , lowercase_ ): _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Union[str, Any] = intersect_and_union( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) total_area_intersect += area_intersect total_area_union += area_union total_area_pred_label += area_pred_label total_area_label += area_label return total_area_intersect, total_area_union, total_area_pred_label, total_area_label def __UpperCAmelCase( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = False , ): _lowerCamelCase, _lowerCamelCase, _lowerCamelCase, _lowerCamelCase : Optional[Any] = total_intersect_and_union( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) # compute metrics _lowerCamelCase : Optional[int] = {} _lowerCamelCase : str = total_area_intersect.sum() / total_area_label.sum() _lowerCamelCase : Tuple = total_area_intersect / total_area_union _lowerCamelCase : int = total_area_intersect / total_area_label _lowerCamelCase : Dict = np.nanmean(lowercase_ ) _lowerCamelCase : Optional[Any] = np.nanmean(lowercase_ ) _lowerCamelCase : Any = all_acc _lowerCamelCase : Dict = iou _lowerCamelCase : str = acc if nan_to_num is not None: _lowerCamelCase : List[Any] = {metric: np.nan_to_num(lowercase_ , nan=lowercase_ ) for metric, metric_value in metrics.items()} return metrics @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class __A ( datasets.Metric ): """simple docstring""" def __snake_case ( self): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( # 1st Seq - height dim, 2nd - width dim { '''predictions''': datasets.Sequence(datasets.Sequence(datasets.Value('''uint16'''))), '''references''': datasets.Sequence(datasets.Sequence(datasets.Value('''uint16'''))), }) , reference_urls=[ '''https://github.com/open-mmlab/mmsegmentation/blob/71c201b1813267d78764f306a297ca717827c4bf/mmseg/core/evaluation/metrics.py''' ] , ) def __snake_case ( self , a__ , a__ , a__ , a__ , a__ = None , a__ = None , a__ = False , ): """simple docstring""" _lowerCamelCase : Dict = mean_iou( results=a__ , gt_seg_maps=a__ , num_labels=a__ , ignore_index=a__ , nan_to_num=a__ , label_map=a__ , reduce_labels=a__ , ) return iou_result
613
import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, temp_seed, temporary_assignment, zip_dict, ) from .utils import require_tf, require_torch def __UpperCAmelCase( lowercase_ ): # picklable for multiprocessing return x.sum() def __UpperCAmelCase( lowercase_ ): # picklable for multiprocessing return i + 1 @dataclass class __A : """simple docstring""" UpperCAmelCase__ = 42 UpperCAmelCase__ = 42 class __A ( lowerCamelCase__ ): """simple docstring""" def __snake_case ( self): """simple docstring""" _lowerCamelCase : Tuple = {} _lowerCamelCase : Dict = [] _lowerCamelCase : Optional[Any] = 1 _lowerCamelCase : Optional[int] = [1, 2] _lowerCamelCase : str = {'''a''': 1, '''b''': 2} _lowerCamelCase : Dict = {'''a''': [1, 2], '''b''': [3, 4]} _lowerCamelCase : Any = {'''a''': {'''1''': 1}, '''b''': 2} _lowerCamelCase : Optional[Any] = {'''a''': 1, '''b''': 2, '''c''': 3, '''d''': 4} _lowerCamelCase : str = {} _lowerCamelCase : int = [] _lowerCamelCase : str = 2 _lowerCamelCase : int = [2, 3] _lowerCamelCase : str = {'''a''': 2, '''b''': 3} _lowerCamelCase : Tuple = {'''a''': [2, 3], '''b''': [4, 5]} _lowerCamelCase : List[str] = {'''a''': {'''1''': 2}, '''b''': 3} _lowerCamelCase : str = {'''a''': 2, '''b''': 3, '''c''': 4, '''d''': 5} self.assertEqual(map_nested(a__ , a__) , a__) self.assertEqual(map_nested(a__ , a__) , a__) self.assertEqual(map_nested(a__ , a__) , a__) self.assertEqual(map_nested(a__ , a__) , a__) self.assertEqual(map_nested(a__ , a__) , a__) self.assertEqual(map_nested(a__ , a__) , a__) self.assertEqual(map_nested(a__ , a__) , a__) self.assertEqual(map_nested(a__ , a__) , a__) _lowerCamelCase : Dict = 2 self.assertEqual(map_nested(a__ , a__ , num_proc=a__) , a__) self.assertEqual(map_nested(a__ , a__ , num_proc=a__) , a__) self.assertEqual(map_nested(a__ , a__ , num_proc=a__) , a__) self.assertEqual(map_nested(a__ , a__ , num_proc=a__) , a__) self.assertEqual(map_nested(a__ , a__ , num_proc=a__) , a__) self.assertEqual(map_nested(a__ , a__ , num_proc=a__) , a__) self.assertEqual(map_nested(a__ , a__ , num_proc=a__) , a__) self.assertEqual(map_nested(a__ , a__ , num_proc=a__) , a__) _lowerCamelCase : Any = {'''a''': np.eye(2), '''b''': np.zeros(3), '''c''': np.ones(2)} _lowerCamelCase : Optional[int] = {'''a''': 2, '''b''': 0, '''c''': 2} _lowerCamelCase : Optional[int] = { '''a''': np.eye(2).astype(a__), '''b''': np.zeros(3).astype(a__), '''c''': np.ones(2).astype(a__), } self.assertEqual(map_nested(a__ , a__ , map_numpy=a__) , a__) self.assertEqual( {k: v.tolist() for k, v in map_nested(a__ , a__ , map_numpy=a__).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) self.assertEqual(map_nested(a__ , a__ , map_numpy=a__ , num_proc=a__) , a__) self.assertEqual( {k: v.tolist() for k, v in map_nested(a__ , a__ , map_numpy=a__ , num_proc=a__).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) with self.assertRaises(a__): # can't pickle a local lambda map_nested(lambda a__: x + 1 , a__ , num_proc=a__) def __snake_case ( self): """simple docstring""" _lowerCamelCase : Dict = {'''a''': 1, '''b''': 2} _lowerCamelCase : Optional[int] = {'''a''': 3, '''b''': 4} _lowerCamelCase : int = {'''a''': 5, '''b''': 6} _lowerCamelCase : Optional[int] = sorted([('''a''', (1, 3, 5)), ('''b''', (2, 4, 6))]) self.assertEqual(sorted(zip_dict(a__ , a__ , a__)) , a__) def __snake_case ( self): """simple docstring""" class __A : """simple docstring""" UpperCAmelCase__ = """bar""" _lowerCamelCase : Any = Foo() self.assertEqual(foo.my_attr , '''bar''') with temporary_assignment(a__ , '''my_attr''' , '''BAR'''): self.assertEqual(foo.my_attr , '''BAR''') self.assertEqual(foo.my_attr , '''bar''') @pytest.mark.parametrize( '''iterable_length, num_proc, expected_num_proc''' , [ (1, None, 1), (1, 1, 1), (2, None, 1), (2, 1, 1), (2, 2, 1), (2, 3, 1), (3, 2, 1), (16, 16, 16), (16, 17, 16), (17, 16, 16), ] , ) def __UpperCAmelCase( lowercase_ , lowercase_ , lowercase_ ): with patch('''datasets.utils.py_utils._single_map_nested''' ) as mock_single_map_nested, patch( '''datasets.parallel.parallel.Pool''' ) as mock_multiprocessing_pool: _lowerCamelCase : Union[str, Any] = {F"""{i}""": i for i in range(lowercase_ )} _lowerCamelCase : List[str] = map_nested(lambda lowercase_ : x + 10 , lowercase_ , num_proc=lowercase_ , parallel_min_length=16 ) if expected_num_proc == 1: assert mock_single_map_nested.called assert not mock_multiprocessing_pool.called else: assert not mock_single_map_nested.called assert mock_multiprocessing_pool.called assert mock_multiprocessing_pool.call_args[0][0] == expected_num_proc class __A ( lowerCamelCase__ ): """simple docstring""" @require_tf def __snake_case ( self): """simple docstring""" import tensorflow as tf from tensorflow.keras import layers _lowerCamelCase : int = layers.Dense(2) def gen_random_output(): _lowerCamelCase : Union[str, Any] = tf.random.uniform((1, 3)) return model(a__).numpy() with temp_seed(42 , set_tensorflow=a__): _lowerCamelCase : List[str] = gen_random_output() with temp_seed(42 , set_tensorflow=a__): _lowerCamelCase : Any = gen_random_output() _lowerCamelCase : str = gen_random_output() np.testing.assert_equal(a__ , a__) self.assertGreater(np.abs(outa - outa).sum() , 0) @require_torch def __snake_case ( self): """simple docstring""" import torch def gen_random_output(): _lowerCamelCase : Union[str, Any] = torch.nn.Linear(3 , 2) _lowerCamelCase : Dict = torch.rand(1 , 3) return model(a__).detach().numpy() with temp_seed(42 , set_pytorch=a__): _lowerCamelCase : Any = gen_random_output() with temp_seed(42 , set_pytorch=a__): _lowerCamelCase : Optional[int] = gen_random_output() _lowerCamelCase : Union[str, Any] = gen_random_output() np.testing.assert_equal(a__ , a__) self.assertGreater(np.abs(outa - outa).sum() , 0) def __snake_case ( self): """simple docstring""" def gen_random_output(): return np.random.rand(1 , 3) with temp_seed(42): _lowerCamelCase : Union[str, Any] = gen_random_output() with temp_seed(42): _lowerCamelCase : List[str] = gen_random_output() _lowerCamelCase : str = gen_random_output() np.testing.assert_equal(a__ , a__) self.assertGreater(np.abs(outa - outa).sum() , 0) @pytest.mark.parametrize('''input_data''' , [{}] ) def __UpperCAmelCase( lowercase_ ): _lowerCamelCase : List[Any] = NestedDataStructure(lowercase_ ).data assert output_data == input_data @pytest.mark.parametrize( '''data, expected_output''' , [ ({}, []), ([], []), ('''foo''', ['''foo''']), (['''foo''', '''bar'''], ['''foo''', '''bar''']), ([['''foo''', '''bar''']], ['''foo''', '''bar''']), ([[['''foo'''], ['''bar''']]], ['''foo''', '''bar''']), ([[['''foo'''], '''bar''']], ['''foo''', '''bar''']), ({'''a''': 1, '''b''': 2}, [1, 2]), ({'''a''': [1, 2], '''b''': [3, 4]}, [1, 2, 3, 4]), ({'''a''': [[1, 2]], '''b''': [[3, 4]]}, [1, 2, 3, 4]), ({'''a''': [[1, 2]], '''b''': [3, 4]}, [1, 2, 3, 4]), ({'''a''': [[[1], [2]]], '''b''': [[[3], [4]]]}, [1, 2, 3, 4]), ({'''a''': [[[1], [2]]], '''b''': [[3, 4]]}, [1, 2, 3, 4]), ({'''a''': [[[1], [2]]], '''b''': [3, 4]}, [1, 2, 3, 4]), ({'''a''': [[[1], [2]]], '''b''': [3, [4]]}, [1, 2, 3, 4]), ({'''a''': {'''1''': 1}, '''b''': 2}, [1, 2]), ({'''a''': {'''1''': [1]}, '''b''': 2}, [1, 2]), ({'''a''': {'''1''': [1]}, '''b''': [2]}, [1, 2]), ] , ) def __UpperCAmelCase( lowercase_ , lowercase_ ): _lowerCamelCase : int = NestedDataStructure(lowercase_ ).flatten() assert output == expected_output def __UpperCAmelCase( ): _lowerCamelCase : Any = A(x=1 , y='''foobar''' ) _lowerCamelCase : Union[str, Any] = {'''x''': 1, '''y''': '''foobar'''} assert asdict(lowercase_ ) == expected_output _lowerCamelCase : Optional[int] = {'''a''': {'''b''': A(x=10 , y='''foo''' )}, '''c''': [A(x=20 , y='''bar''' )]} _lowerCamelCase : Union[str, Any] = {'''a''': {'''b''': {'''x''': 10, '''y''': '''foo'''}}, '''c''': [{'''x''': 20, '''y''': '''bar'''}]} assert asdict(lowercase_ ) == expected_output with pytest.raises(lowercase_ ): asdict([1, A(x=10 , y='''foo''' )] ) def __UpperCAmelCase( lowercase_ ): return text.split() def __UpperCAmelCase( lowercase_ ): yield (time.time(), content) time.sleep(2 ) yield (time.time(), content) def __UpperCAmelCase( ): with Pool(2 ) as pool: _lowerCamelCase : Tuple = list(iflatmap_unordered(lowercase_ , _split_text , kwargs_iterable=[{'''text''': '''hello there'''}] * 10 ) ) assert out.count('''hello''' ) == 10 assert out.count('''there''' ) == 10 assert len(lowercase_ ) == 20 # check multiprocess from pathos (uses dill for pickling) with multiprocess.Pool(2 ) as pool: _lowerCamelCase : Dict = list(iflatmap_unordered(lowercase_ , _split_text , kwargs_iterable=[{'''text''': '''hello there'''}] * 10 ) ) assert out.count('''hello''' ) == 10 assert out.count('''there''' ) == 10 assert len(lowercase_ ) == 20 # check that we get items as fast as possible with Pool(2 ) as pool: _lowerCamelCase : str = [] for yield_time, content in iflatmap_unordered( lowercase_ , _aseconds_generator_of_aitems_with_timing , kwargs_iterable=[{'''content''': '''a'''}, {'''content''': '''b'''}] ): assert yield_time < time.time() + 0.1, "we should each item directly after it was yielded" out.append(lowercase_ ) assert out.count('''a''' ) == 2 assert out.count('''b''' ) == 2 assert len(lowercase_ ) == 4
613
1
"""simple docstring""" import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def __A ( a_ :BertModel , a_ :str , a_ :str) -> str: __a : List[str] = ('''dense.weight''', '''attention.self.query''', '''attention.self.key''', '''attention.self.value''') __a : Any = ( ('''layer.''', '''layer_'''), ('''word_embeddings.weight''', '''word_embeddings'''), ('''position_embeddings.weight''', '''position_embeddings'''), ('''token_type_embeddings.weight''', '''token_type_embeddings'''), ('''.''', '''/'''), ('''LayerNorm/weight''', '''LayerNorm/gamma'''), ('''LayerNorm/bias''', '''LayerNorm/beta'''), ('''weight''', '''kernel'''), ) if not os.path.isdir(a_): os.makedirs(a_) __a : List[Any] = model.state_dict() def to_tf_var_name(a_ :str): for patt, repl in iter(a_): __a : int = name.replace(a_ , a_) return F"""bert/{name}""" def create_tf_var(a_ :np.ndarray , a_ :str , a_ :tf.Session): __a : int = tf.dtypes.as_dtype(tensor.dtype) __a : Optional[int] = tf.get_variable(dtype=a_ , shape=tensor.shape , name=a_ , initializer=tf.zeros_initializer()) session.run(tf.variables_initializer([tf_var])) session.run(a_) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: __a : Any = to_tf_var_name(a_) __a : Tuple = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose): __a : List[Any] = torch_tensor.T __a : Optional[Any] = create_tf_var(tensor=a_ , name=a_ , session=a_) tf.keras.backend.set_value(a_ , a_) __a : int = session.run(a_) print(F"""Successfully created {tf_name}: {np.allclose(a_ , a_)}""") __a : Tuple = tf.train.Saver(tf.trainable_variables()) saver.save(a_ , os.path.join(a_ , model_name.replace('''-''' , '''_''') + '''.ckpt''')) def __A ( a_ :int=None) -> str: __a : str = argparse.ArgumentParser() parser.add_argument('''--model_name''' , type=a_ , required=a_ , help='''model name e.g. bert-base-uncased''') parser.add_argument( '''--cache_dir''' , type=a_ , default=a_ , required=a_ , help='''Directory containing pytorch model''') parser.add_argument('''--pytorch_model_path''' , type=a_ , required=a_ , help='''/path/to/<pytorch-model-name>.bin''') parser.add_argument('''--tf_cache_dir''' , type=a_ , required=a_ , help='''Directory in which to save tensorflow model''') __a : Optional[Any] = parser.parse_args(a_) __a : Optional[Any] = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=a_ , ckpt_dir=args.tf_cache_dir , model_name=args.model_name) if __name__ == "__main__": main()
52
import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def _A ( __snake_case :Dict ) -> Optional[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = VideoMAEConfig() set_architecture_configs(__snake_case , __snake_case ) if "finetuned" not in model_name: __SCREAMING_SNAKE_CASE = False if "finetuned" in model_name: __SCREAMING_SNAKE_CASE = "huggingface/label-files" if "kinetics" in model_name: __SCREAMING_SNAKE_CASE = 400 __SCREAMING_SNAKE_CASE = "kinetics400-id2label.json" elif "ssv2" in model_name: __SCREAMING_SNAKE_CASE = 174 __SCREAMING_SNAKE_CASE = "something-something-v2-id2label.json" else: raise ValueError("Model name should either contain 'kinetics' or 'ssv2' in case it's fine-tuned." ) __SCREAMING_SNAKE_CASE = json.load(open(hf_hub_download(__snake_case , __snake_case , repo_type="dataset" ) , "r" ) ) __SCREAMING_SNAKE_CASE = {int(__snake_case ): v for k, v in idalabel.items()} __SCREAMING_SNAKE_CASE = idalabel __SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} return config def _A ( __snake_case :Dict , __snake_case :Optional[Any] ) -> List[Any]: """simple docstring""" if "small" in model_name: __SCREAMING_SNAKE_CASE = 384 __SCREAMING_SNAKE_CASE = 1536 __SCREAMING_SNAKE_CASE = 12 __SCREAMING_SNAKE_CASE = 16 __SCREAMING_SNAKE_CASE = 12 __SCREAMING_SNAKE_CASE = 3 __SCREAMING_SNAKE_CASE = 192 __SCREAMING_SNAKE_CASE = 768 elif "large" in model_name: __SCREAMING_SNAKE_CASE = 1024 __SCREAMING_SNAKE_CASE = 4096 __SCREAMING_SNAKE_CASE = 24 __SCREAMING_SNAKE_CASE = 16 __SCREAMING_SNAKE_CASE = 12 __SCREAMING_SNAKE_CASE = 8 __SCREAMING_SNAKE_CASE = 512 __SCREAMING_SNAKE_CASE = 2048 elif "huge" in model_name: __SCREAMING_SNAKE_CASE = 1280 __SCREAMING_SNAKE_CASE = 5120 __SCREAMING_SNAKE_CASE = 32 __SCREAMING_SNAKE_CASE = 16 __SCREAMING_SNAKE_CASE = 12 __SCREAMING_SNAKE_CASE = 8 __SCREAMING_SNAKE_CASE = 640 __SCREAMING_SNAKE_CASE = 2560 elif "base" not in model_name: raise ValueError("Model name should include either \"small\", \"base\", \"large\", or \"huge\"" ) def _A ( __snake_case :List[Any] ) -> Optional[int]: """simple docstring""" if "encoder." in name: __SCREAMING_SNAKE_CASE = name.replace("encoder." , "" ) if "cls_token" in name: __SCREAMING_SNAKE_CASE = name.replace("cls_token" , "videomae.embeddings.cls_token" ) if "decoder_pos_embed" in name: __SCREAMING_SNAKE_CASE = name.replace("decoder_pos_embed" , "decoder.decoder_pos_embed" ) if "pos_embed" in name and "decoder" not in name: __SCREAMING_SNAKE_CASE = name.replace("pos_embed" , "videomae.embeddings.position_embeddings" ) if "patch_embed.proj" in name: __SCREAMING_SNAKE_CASE = name.replace("patch_embed.proj" , "videomae.embeddings.patch_embeddings.projection" ) if "patch_embed.norm" in name: __SCREAMING_SNAKE_CASE = name.replace("patch_embed.norm" , "videomae.embeddings.norm" ) if "decoder.blocks" in name: __SCREAMING_SNAKE_CASE = name.replace("decoder.blocks" , "decoder.decoder_layers" ) if "blocks" in name: __SCREAMING_SNAKE_CASE = name.replace("blocks" , "videomae.encoder.layer" ) if "attn.proj" in name: __SCREAMING_SNAKE_CASE = name.replace("attn.proj" , "attention.output.dense" ) if "attn" in name and "bias" not in name: __SCREAMING_SNAKE_CASE = name.replace("attn" , "attention.self" ) if "attn" in name: __SCREAMING_SNAKE_CASE = name.replace("attn" , "attention.attention" ) if "norm1" in name: __SCREAMING_SNAKE_CASE = name.replace("norm1" , "layernorm_before" ) if "norm2" in name: __SCREAMING_SNAKE_CASE = name.replace("norm2" , "layernorm_after" ) if "mlp.fc1" in name: __SCREAMING_SNAKE_CASE = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: __SCREAMING_SNAKE_CASE = name.replace("mlp.fc2" , "output.dense" ) if "decoder_embed" in name: __SCREAMING_SNAKE_CASE = name.replace("decoder_embed" , "decoder.decoder_embed" ) if "decoder_norm" in name: __SCREAMING_SNAKE_CASE = name.replace("decoder_norm" , "decoder.decoder_norm" ) if "decoder_pred" in name: __SCREAMING_SNAKE_CASE = name.replace("decoder_pred" , "decoder.decoder_pred" ) if "norm.weight" in name and "decoder" not in name and "fc" not in name: __SCREAMING_SNAKE_CASE = name.replace("norm.weight" , "videomae.layernorm.weight" ) if "norm.bias" in name and "decoder" not in name and "fc" not in name: __SCREAMING_SNAKE_CASE = name.replace("norm.bias" , "videomae.layernorm.bias" ) if "head" in name and "decoder" not in name: __SCREAMING_SNAKE_CASE = name.replace("head" , "classifier" ) return name def _A ( __snake_case :Union[str, Any] , __snake_case :Optional[int] ) -> Optional[Any]: """simple docstring""" for key in orig_state_dict.copy().keys(): __SCREAMING_SNAKE_CASE = orig_state_dict.pop(__snake_case ) if key.startswith("encoder." ): __SCREAMING_SNAKE_CASE = key.replace("encoder." , "" ) if "qkv" in key: __SCREAMING_SNAKE_CASE = key.split("." ) if key.startswith("decoder.blocks" ): __SCREAMING_SNAKE_CASE = config.decoder_hidden_size __SCREAMING_SNAKE_CASE = int(key_split[2] ) __SCREAMING_SNAKE_CASE = "decoder.decoder_layers." if "weight" in key: __SCREAMING_SNAKE_CASE = val[:dim, :] __SCREAMING_SNAKE_CASE = val[dim : dim * 2, :] __SCREAMING_SNAKE_CASE = val[-dim:, :] else: __SCREAMING_SNAKE_CASE = config.hidden_size __SCREAMING_SNAKE_CASE = int(key_split[1] ) __SCREAMING_SNAKE_CASE = "videomae.encoder.layer." if "weight" in key: __SCREAMING_SNAKE_CASE = val[:dim, :] __SCREAMING_SNAKE_CASE = val[dim : dim * 2, :] __SCREAMING_SNAKE_CASE = val[-dim:, :] else: __SCREAMING_SNAKE_CASE = val return orig_state_dict def _A ( ) -> Dict: """simple docstring""" __SCREAMING_SNAKE_CASE = hf_hub_download( repo_id="hf-internal-testing/spaghetti-video" , filename="eating_spaghetti.npy" , repo_type="dataset" ) __SCREAMING_SNAKE_CASE = np.load(__snake_case ) return list(__snake_case ) def _A ( __snake_case :Optional[int] , __snake_case :List[str] , __snake_case :Union[str, Any] , __snake_case :Optional[Any] ) -> Optional[int]: """simple docstring""" __SCREAMING_SNAKE_CASE = get_videomae_config(__snake_case ) if "finetuned" in model_name: __SCREAMING_SNAKE_CASE = VideoMAEForVideoClassification(__snake_case ) else: __SCREAMING_SNAKE_CASE = VideoMAEForPreTraining(__snake_case ) # download original checkpoint, hosted on Google Drive __SCREAMING_SNAKE_CASE = "pytorch_model.bin" gdown.cached_download(__snake_case , __snake_case , quiet=__snake_case ) __SCREAMING_SNAKE_CASE = torch.load(__snake_case , map_location="cpu" ) if "model" in files: __SCREAMING_SNAKE_CASE = files["model"] else: __SCREAMING_SNAKE_CASE = files["module"] __SCREAMING_SNAKE_CASE = convert_state_dict(__snake_case , __snake_case ) model.load_state_dict(__snake_case ) model.eval() # verify model on basic input __SCREAMING_SNAKE_CASE = VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) __SCREAMING_SNAKE_CASE = prepare_video() __SCREAMING_SNAKE_CASE = image_processor(__snake_case , return_tensors="pt" ) if "finetuned" not in model_name: __SCREAMING_SNAKE_CASE = hf_hub_download(repo_id="hf-internal-testing/bool-masked-pos" , filename="bool_masked_pos.pt" ) __SCREAMING_SNAKE_CASE = torch.load(__snake_case ) __SCREAMING_SNAKE_CASE = model(**__snake_case ) __SCREAMING_SNAKE_CASE = outputs.logits __SCREAMING_SNAKE_CASE = [ "videomae-small-finetuned-kinetics", "videomae-small-finetuned-ssv2", # Kinetics-400 checkpoints (short = pretrained only for 800 epochs instead of 1600) "videomae-base-short", "videomae-base-short-finetuned-kinetics", "videomae-base", "videomae-base-finetuned-kinetics", "videomae-large", "videomae-large-finetuned-kinetics", "videomae-huge-finetuned-kinetics", # Something-Something-v2 checkpoints (short = pretrained only for 800 epochs instead of 2400) "videomae-base-short-ssv2", "videomae-base-short-finetuned-ssv2", "videomae-base-ssv2", "videomae-base-finetuned-ssv2", ] # NOTE: logits were tested with image_mean and image_std equal to [0.5, 0.5, 0.5] and [0.5, 0.5, 0.5] if model_name == "videomae-small-finetuned-kinetics": __SCREAMING_SNAKE_CASE = torch.Size([1, 400] ) __SCREAMING_SNAKE_CASE = torch.tensor([-0.9_2_9_1, -0.4_0_6_1, -0.9_3_0_7] ) elif model_name == "videomae-small-finetuned-ssv2": __SCREAMING_SNAKE_CASE = torch.Size([1, 174] ) __SCREAMING_SNAKE_CASE = torch.tensor([0.2_6_7_1, -0.4_6_8_9, -0.8_2_3_5] ) elif model_name == "videomae-base": __SCREAMING_SNAKE_CASE = torch.Size([1, 1408, 1536] ) __SCREAMING_SNAKE_CASE = torch.tensor([[0.7_7_3_9, 0.7_9_6_8, 0.7_0_8_9], [0.6_7_0_1, 0.7_4_8_7, 0.6_2_0_9], [0.4_2_8_7, 0.5_1_5_8, 0.4_7_7_3]] ) elif model_name == "videomae-base-short": __SCREAMING_SNAKE_CASE = torch.Size([1, 1408, 1536] ) __SCREAMING_SNAKE_CASE = torch.tensor([[0.7_9_9_4, 0.9_6_1_2, 0.8_5_0_8], [0.7_4_0_1, 0.8_9_5_8, 0.8_3_0_2], [0.5_8_6_2, 0.7_4_6_8, 0.7_3_2_5]] ) # we verified the loss both for normalized and unnormalized targets for this one __SCREAMING_SNAKE_CASE = torch.tensor([0.5_1_4_2] ) if config.norm_pix_loss else torch.tensor([0.6_4_6_9] ) elif model_name == "videomae-large": __SCREAMING_SNAKE_CASE = torch.Size([1, 1408, 1536] ) __SCREAMING_SNAKE_CASE = torch.tensor([[0.7_1_4_9, 0.7_9_9_7, 0.6_9_6_6], [0.6_7_6_8, 0.7_8_6_9, 0.6_9_4_8], [0.5_1_3_9, 0.6_2_2_1, 0.5_6_0_5]] ) elif model_name == "videomae-large-finetuned-kinetics": __SCREAMING_SNAKE_CASE = torch.Size([1, 400] ) __SCREAMING_SNAKE_CASE = torch.tensor([0.0_7_7_1, 0.0_0_1_1, -0.3_6_2_5] ) elif model_name == "videomae-huge-finetuned-kinetics": __SCREAMING_SNAKE_CASE = torch.Size([1, 400] ) __SCREAMING_SNAKE_CASE = torch.tensor([0.2_4_3_3, 0.1_6_3_2, -0.4_8_9_4] ) elif model_name == "videomae-base-short-finetuned-kinetics": __SCREAMING_SNAKE_CASE = torch.Size([1, 400] ) __SCREAMING_SNAKE_CASE = torch.tensor([0.6_5_8_8, 0.0_9_9_0, -0.2_4_9_3] ) elif model_name == "videomae-base-finetuned-kinetics": __SCREAMING_SNAKE_CASE = torch.Size([1, 400] ) __SCREAMING_SNAKE_CASE = torch.tensor([0.3_6_6_9, -0.0_6_8_8, -0.2_4_2_1] ) elif model_name == "videomae-base-short-ssv2": __SCREAMING_SNAKE_CASE = torch.Size([1, 1408, 1536] ) __SCREAMING_SNAKE_CASE = torch.tensor([[0.4_7_1_2, 0.5_2_9_6, 0.5_7_8_6], [0.2_2_7_8, 0.2_7_2_9, 0.4_0_2_6], [0.0_3_5_2, 0.0_7_3_0, 0.2_5_0_6]] ) elif model_name == "videomae-base-short-finetuned-ssv2": __SCREAMING_SNAKE_CASE = torch.Size([1, 174] ) __SCREAMING_SNAKE_CASE = torch.tensor([-0.0_5_3_7, -0.1_5_3_9, -0.3_2_6_6] ) elif model_name == "videomae-base-ssv2": __SCREAMING_SNAKE_CASE = torch.Size([1, 1408, 1536] ) __SCREAMING_SNAKE_CASE = torch.tensor([[0.8_1_3_1, 0.8_7_2_7, 0.8_5_4_6], [0.7_3_6_6, 0.9_3_7_7, 0.8_8_7_0], [0.5_9_3_5, 0.8_8_7_4, 0.8_5_6_4]] ) elif model_name == "videomae-base-finetuned-ssv2": __SCREAMING_SNAKE_CASE = torch.Size([1, 174] ) __SCREAMING_SNAKE_CASE = torch.tensor([0.1_9_6_1, -0.8_3_3_7, -0.6_3_8_9] ) else: raise ValueError(f'''Model name not supported. Should be one of {model_names}''' ) # verify logits assert logits.shape == expected_shape if "finetuned" in model_name: assert torch.allclose(logits[0, :3] , __snake_case , atol=1e-4 ) else: print("Logits:" , logits[0, :3, :3] ) assert torch.allclose(logits[0, :3, :3] , __snake_case , atol=1e-4 ) print("Logits ok!" ) # verify loss, if applicable if model_name == "videomae-base-short": __SCREAMING_SNAKE_CASE = outputs.loss assert torch.allclose(__snake_case , __snake_case , atol=1e-4 ) print("Loss ok!" ) if pytorch_dump_folder_path is not None: print(f'''Saving model and image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__snake_case ) model.save_pretrained(__snake_case ) if push_to_hub: print("Pushing to the hub..." ) model.push_to_hub(__snake_case , organization="nielsr" ) if __name__ == "__main__": _snake_case : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://drive.google.com/u/1/uc?id=1tEhLyskjb755TJ65ptsrafUG2llSwQE1&amp;export=download&amp;confirm=t&amp;uuid=aa3276eb-fb7e-482a-adec-dc7171df14c4', type=str, help=( 'URL of the original PyTorch checkpoint (on Google Drive) you\'d like to convert. Should be a direct' ' download link.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default='/Users/nielsrogge/Documents/VideoMAE/Test', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--model_name', default='videomae-base', type=str, help='Name of the model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) _snake_case : Optional[int] = parser.parse_args() convert_videomae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
693
0
'''simple docstring''' from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxCrossAttnUpBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, FlaxUpBlockaD, ) @flax.struct.dataclass class UpperCAmelCase_ ( __snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE = 42 @flax_register_to_config class UpperCAmelCase_ ( nn.Module , __snake_case , __snake_case ): """simple docstring""" __SCREAMING_SNAKE_CASE = 32 __SCREAMING_SNAKE_CASE = 4 __SCREAMING_SNAKE_CASE = 4 __SCREAMING_SNAKE_CASE = ( '''CrossAttnDownBlock2D''', '''CrossAttnDownBlock2D''', '''CrossAttnDownBlock2D''', '''DownBlock2D''', ) __SCREAMING_SNAKE_CASE = ('''UpBlock2D''', '''CrossAttnUpBlock2D''', '''CrossAttnUpBlock2D''', '''CrossAttnUpBlock2D''') __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = (3_20, 6_40, 12_80, 12_80) __SCREAMING_SNAKE_CASE = 2 __SCREAMING_SNAKE_CASE = 8 __SCREAMING_SNAKE_CASE = None __SCREAMING_SNAKE_CASE = 12_80 __SCREAMING_SNAKE_CASE = 0.0 __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = jnp.floataa __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = 0 __SCREAMING_SNAKE_CASE = False def SCREAMING_SNAKE_CASE__ ( self , lowerCamelCase ) -> Optional[int]: '''simple docstring''' UpperCamelCase : Dict = (1, self.in_channels, self.sample_size, self.sample_size) UpperCamelCase : Optional[int] = jnp.zeros(_lowercase , dtype=jnp.floataa ) UpperCamelCase : List[str] = jnp.ones((1,) , dtype=jnp.intaa ) UpperCamelCase : str = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) UpperCamelCase : List[str] = jax.random.split(_lowercase ) UpperCamelCase : Tuple = {"""params""": params_rng, """dropout""": dropout_rng} return self.init(_lowercase , _lowercase , _lowercase , _lowercase )["params"] def SCREAMING_SNAKE_CASE__ ( self ) -> Tuple: '''simple docstring''' UpperCamelCase : List[Any] = self.block_out_channels UpperCamelCase : List[Any] = block_out_channels[0] * 4 if self.num_attention_heads is not None: raise ValueError( "At the moment it is not possible to define the number of attention heads via `num_attention_heads` because of a naming issue as described in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131. Passing `num_attention_heads` will only be supported in diffusers v0.19." ) # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. UpperCamelCase : Union[str, Any] = self.num_attention_heads or self.attention_head_dim # input UpperCamelCase : List[Any] = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time UpperCamelCase : Optional[Any] = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) UpperCamelCase : Optional[Any] = FlaxTimestepEmbedding(_lowercase , dtype=self.dtype ) UpperCamelCase : str = self.only_cross_attention if isinstance(_lowercase , _lowercase ): UpperCamelCase : Dict = (only_cross_attention,) * len(self.down_block_types ) if isinstance(_lowercase , _lowercase ): UpperCamelCase : int = (num_attention_heads,) * len(self.down_block_types ) # down UpperCamelCase : Union[str, Any] = [] UpperCamelCase : Optional[Any] = block_out_channels[0] for i, down_block_type in enumerate(self.down_block_types ): UpperCamelCase : List[Any] = output_channel UpperCamelCase : Dict = block_out_channels[i] UpperCamelCase : Dict = i == len(_lowercase ) - 1 if down_block_type == "CrossAttnDownBlock2D": UpperCamelCase : Optional[Any] = FlaxCrossAttnDownBlockaD( in_channels=_lowercase , out_channels=_lowercase , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: UpperCamelCase : Optional[Any] = FlaxDownBlockaD( in_channels=_lowercase , out_channels=_lowercase , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(_lowercase ) UpperCamelCase : Optional[int] = down_blocks # mid UpperCamelCase : Tuple = FlaxUNetMidBlockaDCrossAttn( in_channels=block_out_channels[-1] , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) # up UpperCamelCase : Optional[int] = [] UpperCamelCase : Union[str, Any] = list(reversed(_lowercase ) ) UpperCamelCase : List[str] = list(reversed(_lowercase ) ) UpperCamelCase : Tuple = list(reversed(_lowercase ) ) UpperCamelCase : Dict = reversed_block_out_channels[0] for i, up_block_type in enumerate(self.up_block_types ): UpperCamelCase : Union[str, Any] = output_channel UpperCamelCase : Tuple = reversed_block_out_channels[i] UpperCamelCase : Optional[int] = reversed_block_out_channels[min(i + 1 , len(_lowercase ) - 1 )] UpperCamelCase : List[str] = i == len(_lowercase ) - 1 if up_block_type == "CrossAttnUpBlock2D": UpperCamelCase : List[Any] = FlaxCrossAttnUpBlockaD( in_channels=_lowercase , out_channels=_lowercase , prev_output_channel=_lowercase , num_layers=self.layers_per_block + 1 , num_attention_heads=reversed_num_attention_heads[i] , add_upsample=not is_final_block , dropout=self.dropout , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: UpperCamelCase : Any = FlaxUpBlockaD( in_channels=_lowercase , out_channels=_lowercase , prev_output_channel=_lowercase , num_layers=self.layers_per_block + 1 , add_upsample=not is_final_block , dropout=self.dropout , dtype=self.dtype , ) up_blocks.append(_lowercase ) UpperCamelCase : Union[str, Any] = output_channel UpperCamelCase : int = up_blocks # out UpperCamelCase : Optional[Any] = nn.GroupNorm(num_groups=32 , epsilon=1e-5 ) UpperCamelCase : Optional[Any] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=None , lowerCamelCase=None , lowerCamelCase = True , lowerCamelCase = False , ) -> str: '''simple docstring''' if not isinstance(_lowercase , jnp.ndarray ): UpperCamelCase : Union[str, Any] = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(_lowercase , jnp.ndarray ) and len(timesteps.shape ) == 0: UpperCamelCase : Union[str, Any] = timesteps.astype(dtype=jnp.floataa ) UpperCamelCase : List[str] = jnp.expand_dims(_lowercase , 0 ) UpperCamelCase : Any = self.time_proj(_lowercase ) UpperCamelCase : Tuple = self.time_embedding(_lowercase ) # 2. pre-process UpperCamelCase : Optional[Any] = jnp.transpose(_lowercase , (0, 2, 3, 1) ) UpperCamelCase : List[str] = self.conv_in(_lowercase ) # 3. down UpperCamelCase : Dict = (sample,) for down_block in self.down_blocks: if isinstance(_lowercase , _lowercase ): UpperCamelCase : Tuple = down_block(_lowercase , _lowercase , _lowercase , deterministic=not train ) else: UpperCamelCase : Tuple = down_block(_lowercase , _lowercase , deterministic=not train ) down_block_res_samples += res_samples if down_block_additional_residuals is not None: UpperCamelCase : Dict = () for down_block_res_sample, down_block_additional_residual in zip( _lowercase , _lowercase ): down_block_res_sample += down_block_additional_residual new_down_block_res_samples += (down_block_res_sample,) UpperCamelCase : Dict = new_down_block_res_samples # 4. mid UpperCamelCase : Dict = self.mid_block(_lowercase , _lowercase , _lowercase , deterministic=not train ) if mid_block_additional_residual is not None: sample += mid_block_additional_residual # 5. up for up_block in self.up_blocks: UpperCamelCase : List[str] = down_block_res_samples[-(self.layers_per_block + 1) :] UpperCamelCase : Optional[Any] = down_block_res_samples[: -(self.layers_per_block + 1)] if isinstance(_lowercase , _lowercase ): UpperCamelCase : List[str] = up_block( _lowercase , temb=_lowercase , encoder_hidden_states=_lowercase , res_hidden_states_tuple=_lowercase , deterministic=not train , ) else: UpperCamelCase : Tuple = up_block(_lowercase , temb=_lowercase , res_hidden_states_tuple=_lowercase , deterministic=not train ) # 6. post-process UpperCamelCase : Optional[Any] = self.conv_norm_out(_lowercase ) UpperCamelCase : Optional[Any] = nn.silu(_lowercase ) UpperCamelCase : Any = self.conv_out(_lowercase ) UpperCamelCase : int = jnp.transpose(_lowercase , (0, 3, 1, 2) ) if not return_dict: return (sample,) return FlaxUNetaDConditionOutput(sample=_lowercase )
719
'''simple docstring''' import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class UpperCAmelCase_ ( lowerCamelCase_ , unittest.TestCase ): """simple docstring""" __SCREAMING_SNAKE_CASE = BioGptTokenizer __SCREAMING_SNAKE_CASE = False def SCREAMING_SNAKE_CASE__ ( self ) -> Optional[int]: '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCamelCase : Optional[Any] = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "w</w>", "r</w>", "t</w>", "lo", "low", "er</w>", "low</w>", "lowest</w>", "newer</w>", "wider</w>", "<unk>", ] UpperCamelCase : Optional[int] = dict(zip(lowerCamelCase , range(len(lowerCamelCase ) ) ) ) UpperCamelCase : Union[str, Any] = ["l o 123", "lo w 1456", "e r</w> 1789", ""] UpperCamelCase : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) UpperCamelCase : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" ) as fp: fp.write(json.dumps(lowerCamelCase ) ) with open(self.merges_file , "w" ) as fp: fp.write("\n".join(lowerCamelCase ) ) def SCREAMING_SNAKE_CASE__ ( self , lowerCamelCase ) -> int: '''simple docstring''' UpperCamelCase : Optional[int] = "lower newer" UpperCamelCase : Optional[int] = "lower newer" return input_text, output_text def SCREAMING_SNAKE_CASE__ ( self ) -> str: '''simple docstring''' UpperCamelCase : Optional[Any] = BioGptTokenizer(self.vocab_file , self.merges_file ) UpperCamelCase : int = "lower" UpperCamelCase : List[Any] = ["low", "er</w>"] UpperCamelCase : Dict = tokenizer.tokenize(lowerCamelCase ) self.assertListEqual(lowerCamelCase , lowerCamelCase ) UpperCamelCase : List[Any] = tokens + ["<unk>"] UpperCamelCase : Tuple = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase ) , lowerCamelCase ) @slow def SCREAMING_SNAKE_CASE__ ( self ) -> Optional[Any]: '''simple docstring''' UpperCamelCase : List[Any] = BioGptTokenizer.from_pretrained("microsoft/biogpt" ) UpperCamelCase : Dict = tokenizer.encode("sequence builders" , add_special_tokens=lowerCamelCase ) UpperCamelCase : Dict = tokenizer.encode("multi-sequence build" , add_special_tokens=lowerCamelCase ) UpperCamelCase : Tuple = tokenizer.build_inputs_with_special_tokens(lowerCamelCase ) UpperCamelCase : List[Any] = tokenizer.build_inputs_with_special_tokens(lowerCamelCase , lowerCamelCase ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
435
0
from scipy.stats import pearsonr import datasets a_ = """ Pearson correlation coefficient and p-value for testing non-correlation. The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. """ a_ = """ Args: predictions (`list` of `int`): Predicted class labels, as returned by a model. references (`list` of `int`): Ground truth labels. return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`. Returns: pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation. p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities. Examples: Example 1-A simple example using only predictions and references. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5]) >>> print(round(results['pearsonr'], 2)) -0.74 Example 2-The same as Example 1, but that also returns the `p-value`. >>> pearsonr_metric = datasets.load_metric(\"pearsonr\") >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True) >>> print(sorted(list(results.keys()))) ['p-value', 'pearsonr'] >>> print(round(results['pearsonr'], 2)) -0.74 >>> print(round(results['p-value'], 2)) 0.15 """ a_ = """ @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, Ilhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Antonio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase__ ( datasets.Metric ): """simple docstring""" def _UpperCAmelCase ( self: int ) -> str: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } ) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html"] , ) def _UpperCAmelCase ( self: Tuple , __lowerCAmelCase: int , __lowerCAmelCase: Union[str, Any] , __lowerCAmelCase: List[Any]=False ) -> int: '''simple docstring''' if return_pvalue: __UpperCAmelCase = pearsonr(__lowerCAmelCase , __lowerCAmelCase ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(__lowerCAmelCase , __lowerCAmelCase )[0] )}
221
import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class UpperCAmelCase__ : """simple docstring""" def __init__( self: str , __lowerCAmelCase: Any , __lowerCAmelCase: Union[str, Any]=13 , __lowerCAmelCase: Dict=30 , __lowerCAmelCase: List[Any]=2 , __lowerCAmelCase: Tuple=3 , __lowerCAmelCase: Tuple=True , __lowerCAmelCase: Dict=True , __lowerCAmelCase: List[str]=32 , __lowerCAmelCase: Tuple=5 , __lowerCAmelCase: List[str]=4 , __lowerCAmelCase: Any=37 , __lowerCAmelCase: int="gelu" , __lowerCAmelCase: str=0.1 , __lowerCAmelCase: List[Any]=0.1 , __lowerCAmelCase: List[Any]=10 , __lowerCAmelCase: Optional[int]=0.02 , __lowerCAmelCase: int=None , __lowerCAmelCase: str=2 , ) -> Optional[int]: '''simple docstring''' __UpperCAmelCase = parent __UpperCAmelCase = batch_size __UpperCAmelCase = image_size __UpperCAmelCase = patch_size __UpperCAmelCase = num_channels __UpperCAmelCase = is_training __UpperCAmelCase = use_labels __UpperCAmelCase = hidden_size __UpperCAmelCase = num_hidden_layers __UpperCAmelCase = num_attention_heads __UpperCAmelCase = intermediate_size __UpperCAmelCase = hidden_act __UpperCAmelCase = hidden_dropout_prob __UpperCAmelCase = attention_probs_dropout_prob __UpperCAmelCase = type_sequence_label_size __UpperCAmelCase = initializer_range __UpperCAmelCase = scope __UpperCAmelCase = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) __UpperCAmelCase = (image_size // patch_size) ** 2 __UpperCAmelCase = num_patches + 1 def _UpperCAmelCase ( self: Optional[int] ) -> List[Any]: '''simple docstring''' __UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __UpperCAmelCase = None if self.use_labels: __UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCAmelCase = self.get_config() return config, pixel_values, labels def _UpperCAmelCase ( self: Tuple ) -> Optional[int]: '''simple docstring''' return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCAmelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _UpperCAmelCase ( self: int , __lowerCAmelCase: List[str] , __lowerCAmelCase: int , __lowerCAmelCase: int ) -> Any: '''simple docstring''' __UpperCAmelCase = ViTModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() __UpperCAmelCase = model(__lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCAmelCase ( self: Union[str, Any] , __lowerCAmelCase: Dict , __lowerCAmelCase: Any , __lowerCAmelCase: Tuple ) -> List[Any]: '''simple docstring''' __UpperCAmelCase = ViTForMaskedImageModeling(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() __UpperCAmelCase = model(__lowerCAmelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __UpperCAmelCase = 1 __UpperCAmelCase = ViTForMaskedImageModeling(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() __UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __UpperCAmelCase = model(__lowerCAmelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _UpperCAmelCase ( self: str , __lowerCAmelCase: Any , __lowerCAmelCase: str , __lowerCAmelCase: Union[str, Any] ) -> Tuple: '''simple docstring''' __UpperCAmelCase = self.type_sequence_label_size __UpperCAmelCase = ViTForImageClassification(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() __UpperCAmelCase = model(__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images __UpperCAmelCase = 1 __UpperCAmelCase = ViTForImageClassification(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() __UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __UpperCAmelCase = model(__lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _UpperCAmelCase ( self: Tuple ) -> Dict: '''simple docstring''' __UpperCAmelCase = self.prepare_config_and_inputs() ( ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ( __UpperCAmelCase ) , ) = config_and_inputs __UpperCAmelCase = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase__ ( snake_case , snake_case , unittest.TestCase ): """simple docstring""" lowerCAmelCase__ : Optional[Any] = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) lowerCAmelCase__ : Optional[int] = ( {'feature-extraction': ViTModel, 'image-classification': ViTForImageClassification} if is_torch_available() else {} ) lowerCAmelCase__ : List[str] = True lowerCAmelCase__ : Dict = False lowerCAmelCase__ : Any = False lowerCAmelCase__ : Union[str, Any] = False def _UpperCAmelCase ( self: Tuple ) -> Optional[int]: '''simple docstring''' __UpperCAmelCase = ViTModelTester(self ) __UpperCAmelCase = ConfigTester(self , config_class=__lowerCAmelCase , has_text_modality=__lowerCAmelCase , hidden_size=37 ) def _UpperCAmelCase ( self: str ) -> Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="ViT does not use inputs_embeds" ) def _UpperCAmelCase ( self: Dict ) -> int: '''simple docstring''' pass def _UpperCAmelCase ( self: Optional[Any] ) -> Any: '''simple docstring''' __UpperCAmelCase , __UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase = model_class(__lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __UpperCAmelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCAmelCase , nn.Linear ) ) def _UpperCAmelCase ( self: str ) -> Union[str, Any]: '''simple docstring''' __UpperCAmelCase , __UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase = model_class(__lowerCAmelCase ) __UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase = [*signature.parameters.keys()] __UpperCAmelCase = ["pixel_values"] self.assertListEqual(arg_names[:1] , __lowerCAmelCase ) def _UpperCAmelCase ( self: Optional[int] ) -> List[str]: '''simple docstring''' __UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCAmelCase ) def _UpperCAmelCase ( self: Optional[int] ) -> Tuple: '''simple docstring''' __UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__lowerCAmelCase ) def _UpperCAmelCase ( self: Optional[int] ) -> List[str]: '''simple docstring''' __UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCAmelCase ) @slow def _UpperCAmelCase ( self: Union[str, Any] ) -> List[str]: '''simple docstring''' for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCAmelCase = ViTModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) def __lowerCAmelCase ( ) -> List[str]: __UpperCAmelCase = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def _UpperCAmelCase ( self: Dict ) -> Optional[int]: '''simple docstring''' return ViTImageProcessor.from_pretrained("google/vit-base-patch16-224" ) if is_vision_available() else None @slow def _UpperCAmelCase ( self: List[Any] ) -> Optional[Any]: '''simple docstring''' __UpperCAmelCase = ViTForImageClassification.from_pretrained("google/vit-base-patch16-224" ).to(__lowerCAmelCase ) __UpperCAmelCase = self.default_image_processor __UpperCAmelCase = prepare_img() __UpperCAmelCase = image_processor(images=__lowerCAmelCase , return_tensors="pt" ).to(__lowerCAmelCase ) # forward pass with torch.no_grad(): __UpperCAmelCase = model(**__lowerCAmelCase ) # verify the logits __UpperCAmelCase = torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , __lowerCAmelCase ) __UpperCAmelCase = torch.tensor([-0.2744, 0.8215, -0.0836] ).to(__lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCAmelCase , atol=1E-4 ) ) @slow def _UpperCAmelCase ( self: Tuple ) -> Tuple: '''simple docstring''' __UpperCAmelCase = ViTModel.from_pretrained("facebook/dino-vits8" ).to(__lowerCAmelCase ) __UpperCAmelCase = ViTImageProcessor.from_pretrained("facebook/dino-vits8" , size=480 ) __UpperCAmelCase = prepare_img() __UpperCAmelCase = image_processor(images=__lowerCAmelCase , return_tensors="pt" ) __UpperCAmelCase = inputs.pixel_values.to(__lowerCAmelCase ) # forward pass with torch.no_grad(): __UpperCAmelCase = model(__lowerCAmelCase , interpolate_pos_encoding=__lowerCAmelCase ) # verify the logits __UpperCAmelCase = torch.Size((1, 3_601, 384) ) self.assertEqual(outputs.last_hidden_state.shape , __lowerCAmelCase ) __UpperCAmelCase = torch.tensor( [[4.2340, 4.3906, -6.6692], [4.5463, 1.8928, -6.7257], [4.4429, 0.8496, -5.8585]] ).to(__lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , __lowerCAmelCase , atol=1E-4 ) ) @slow @require_accelerate @require_torch_gpu def _UpperCAmelCase ( self: Tuple ) -> str: '''simple docstring''' __UpperCAmelCase = ViTModel.from_pretrained("facebook/dino-vits8" , torch_dtype=torch.floataa , device_map="auto" ) __UpperCAmelCase = self.default_image_processor __UpperCAmelCase = prepare_img() __UpperCAmelCase = image_processor(images=__lowerCAmelCase , return_tensors="pt" ) __UpperCAmelCase = inputs.pixel_values.to(__lowerCAmelCase ) # forward pass to make sure inference works in fp16 with torch.no_grad(): __UpperCAmelCase = model(__lowerCAmelCase )
221
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase = { "configuration_roformer": ["ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "RoFormerConfig", "RoFormerOnnxConfig"], "tokenization_roformer": ["RoFormerTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = ["RoFormerTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ "ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "RoFormerForCausalLM", "RoFormerForMaskedLM", "RoFormerForMultipleChoice", "RoFormerForQuestionAnswering", "RoFormerForSequenceClassification", "RoFormerForTokenClassification", "RoFormerLayer", "RoFormerModel", "RoFormerPreTrainedModel", "load_tf_weights_in_roformer", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ "TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TFRoFormerForCausalLM", "TFRoFormerForMaskedLM", "TFRoFormerForMultipleChoice", "TFRoFormerForQuestionAnswering", "TFRoFormerForSequenceClassification", "TFRoFormerForTokenClassification", "TFRoFormerLayer", "TFRoFormerModel", "TFRoFormerPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase = [ "FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "FlaxRoFormerForMaskedLM", "FlaxRoFormerForMultipleChoice", "FlaxRoFormerForQuestionAnswering", "FlaxRoFormerForSequenceClassification", "FlaxRoFormerForTokenClassification", "FlaxRoFormerModel", "FlaxRoFormerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys lowerCamelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
707
"""simple docstring""" import argparse import json import pickle from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import MaskFormerConfig, MaskFormerForInstanceSegmentation, MaskFormerImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase = logging.get_logger(__name__) def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = SwinConfig.from_pretrained( "microsoft/swin-tiny-patch4-window7-224" , out_features=["stage1", "stage2", "stage3", "stage4"] ) UpperCAmelCase_ = MaskFormerConfig(backbone_config=lowerCAmelCase__ ) UpperCAmelCase_ = "huggingface/label-files" if "ade20k-full" in model_name: # this should be ok UpperCAmelCase_ = 847 UpperCAmelCase_ = "maskformer-ade20k-full-id2label.json" elif "ade" in model_name: # this should be ok UpperCAmelCase_ = 150 UpperCAmelCase_ = "ade20k-id2label.json" elif "coco-stuff" in model_name: # this should be ok UpperCAmelCase_ = 171 UpperCAmelCase_ = "maskformer-coco-stuff-id2label.json" elif "coco" in model_name: # TODO UpperCAmelCase_ = 133 UpperCAmelCase_ = "coco-panoptic-id2label.json" elif "cityscapes" in model_name: # this should be ok UpperCAmelCase_ = 19 UpperCAmelCase_ = "cityscapes-id2label.json" elif "vistas" in model_name: # this should be ok UpperCAmelCase_ = 65 UpperCAmelCase_ = "mapillary-vistas-id2label.json" UpperCAmelCase_ = json.load(open(hf_hub_download(lowerCAmelCase__ , lowerCAmelCase__ , repo_type="dataset" ) , "r" ) ) UpperCAmelCase_ = {int(lowerCAmelCase__ ): v for k, v in idalabel.items()} return config def a__ ( lowerCAmelCase__ ): UpperCAmelCase_ = [] # stem # fmt: off rename_keys.append(("backbone.patch_embed.proj.weight", "model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.weight") ) rename_keys.append(("backbone.patch_embed.proj.bias", "model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.bias") ) rename_keys.append(("backbone.patch_embed.norm.weight", "model.pixel_level_module.encoder.model.embeddings.norm.weight") ) rename_keys.append(("backbone.patch_embed.norm.bias", "model.pixel_level_module.encoder.model.embeddings.norm.bias") ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.norm1.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.norm1.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.attn.relative_position_bias_table""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.attn.relative_position_index""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.attn.proj.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.attn.proj.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.norm2.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.norm2.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.mlp.fc1.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.mlp.fc1.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.mlp.fc2.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.weight""") ) rename_keys.append((f"""backbone.layers.{i}.blocks.{j}.mlp.fc2.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.bias""") ) if i < 3: rename_keys.append((f"""backbone.layers.{i}.downsample.reduction.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.reduction.weight""") ) rename_keys.append((f"""backbone.layers.{i}.downsample.norm.weight""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.weight""") ) rename_keys.append((f"""backbone.layers.{i}.downsample.norm.bias""", f"""model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.bias""") ) rename_keys.append((f"""backbone.norm{i}.weight""", f"""model.pixel_level_module.encoder.hidden_states_norms.{i}.weight""") ) rename_keys.append((f"""backbone.norm{i}.bias""", f"""model.pixel_level_module.encoder.hidden_states_norms.{i}.bias""") ) # FPN rename_keys.append(("sem_seg_head.layer_4.weight", "model.pixel_level_module.decoder.fpn.stem.0.weight") ) rename_keys.append(("sem_seg_head.layer_4.norm.weight", "model.pixel_level_module.decoder.fpn.stem.1.weight") ) rename_keys.append(("sem_seg_head.layer_4.norm.bias", "model.pixel_level_module.decoder.fpn.stem.1.bias") ) for source_index, target_index in zip(range(3 , 0 , -1 ) , range(0 , 3 ) ): rename_keys.append((f"""sem_seg_head.adapter_{source_index}.weight""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.0.weight""") ) rename_keys.append((f"""sem_seg_head.adapter_{source_index}.norm.weight""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.weight""") ) rename_keys.append((f"""sem_seg_head.adapter_{source_index}.norm.bias""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.bias""") ) rename_keys.append((f"""sem_seg_head.layer_{source_index}.weight""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.block.0.weight""") ) rename_keys.append((f"""sem_seg_head.layer_{source_index}.norm.weight""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.weight""") ) rename_keys.append((f"""sem_seg_head.layer_{source_index}.norm.bias""", f"""model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.bias""") ) rename_keys.append(("sem_seg_head.mask_features.weight", "model.pixel_level_module.decoder.mask_projection.weight") ) rename_keys.append(("sem_seg_head.mask_features.bias", "model.pixel_level_module.decoder.mask_projection.bias") ) # Transformer decoder for idx in range(config.decoder_config.decoder_layers ): # self-attention out projection rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.weight""", f"""model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.bias""", f"""model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.bias""") ) # cross-attention out projection rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.weight""", f"""model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.bias""", f"""model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.bias""") ) # MLP 1 rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.weight""", f"""model.transformer_module.decoder.layers.{idx}.fc1.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.bias""", f"""model.transformer_module.decoder.layers.{idx}.fc1.bias""") ) # MLP 2 rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.weight""", f"""model.transformer_module.decoder.layers.{idx}.fc2.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.bias""", f"""model.transformer_module.decoder.layers.{idx}.fc2.bias""") ) # layernorm 1 (self-attention layernorm) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.weight""", f"""model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.bias""", f"""model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.bias""") ) # layernorm 2 (cross-attention layernorm) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.weight""", f"""model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.bias""", f"""model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.bias""") ) # layernorm 3 (final layernorm) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.weight""", f"""model.transformer_module.decoder.layers.{idx}.final_layer_norm.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.bias""", f"""model.transformer_module.decoder.layers.{idx}.final_layer_norm.bias""") ) rename_keys.append(("sem_seg_head.predictor.transformer.decoder.norm.weight", "model.transformer_module.decoder.layernorm.weight") ) rename_keys.append(("sem_seg_head.predictor.transformer.decoder.norm.bias", "model.transformer_module.decoder.layernorm.bias") ) # heads on top rename_keys.append(("sem_seg_head.predictor.query_embed.weight", "model.transformer_module.queries_embedder.weight") ) rename_keys.append(("sem_seg_head.predictor.input_proj.weight", "model.transformer_module.input_projection.weight") ) rename_keys.append(("sem_seg_head.predictor.input_proj.bias", "model.transformer_module.input_projection.bias") ) rename_keys.append(("sem_seg_head.predictor.class_embed.weight", "class_predictor.weight") ) rename_keys.append(("sem_seg_head.predictor.class_embed.bias", "class_predictor.bias") ) for i in range(3 ): rename_keys.append((f"""sem_seg_head.predictor.mask_embed.layers.{i}.weight""", f"""mask_embedder.{i}.0.weight""") ) rename_keys.append((f"""sem_seg_head.predictor.mask_embed.layers.{i}.bias""", f"""mask_embedder.{i}.0.bias""") ) # fmt: on return rename_keys def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = dct.pop(lowerCAmelCase__ ) UpperCAmelCase_ = val def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): UpperCAmelCase_ = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) UpperCAmelCase_ = state_dict.pop(f"""backbone.layers.{i}.blocks.{j}.attn.qkv.weight""" ) UpperCAmelCase_ = state_dict.pop(f"""backbone.layers.{i}.blocks.{j}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ = in_proj_weight[:dim, :] UpperCAmelCase_ = in_proj_bias[: dim] UpperCAmelCase_ = in_proj_weight[ dim : dim * 2, : ] UpperCAmelCase_ = in_proj_bias[ dim : dim * 2 ] UpperCAmelCase_ = in_proj_weight[ -dim :, : ] UpperCAmelCase_ = in_proj_bias[-dim :] # fmt: on def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): # fmt: off UpperCAmelCase_ = config.decoder_config.hidden_size for idx in range(config.decoder_config.decoder_layers ): # read in weights + bias of self-attention input projection layer (in the original implementation, this is a single matrix + bias) UpperCAmelCase_ = state_dict.pop(f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_weight""" ) UpperCAmelCase_ = state_dict.pop(f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ = in_proj_weight[: hidden_size, :] UpperCAmelCase_ = in_proj_bias[:config.hidden_size] UpperCAmelCase_ = in_proj_weight[hidden_size : hidden_size * 2, :] UpperCAmelCase_ = in_proj_bias[hidden_size : hidden_size * 2] UpperCAmelCase_ = in_proj_weight[-hidden_size :, :] UpperCAmelCase_ = in_proj_bias[-hidden_size :] # read in weights + bias of cross-attention input projection layer (in the original implementation, this is a single matrix + bias) UpperCAmelCase_ = state_dict.pop(f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_weight""" ) UpperCAmelCase_ = state_dict.pop(f"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ = in_proj_weight[: hidden_size, :] UpperCAmelCase_ = in_proj_bias[:config.hidden_size] UpperCAmelCase_ = in_proj_weight[hidden_size : hidden_size * 2, :] UpperCAmelCase_ = in_proj_bias[hidden_size : hidden_size * 2] UpperCAmelCase_ = in_proj_weight[-hidden_size :, :] UpperCAmelCase_ = in_proj_bias[-hidden_size :] # fmt: on def a__ ( ): UpperCAmelCase_ = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ = Image.open(requests.get(lowerCAmelCase__ , stream=lowerCAmelCase__ ).raw ) return im @torch.no_grad() def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = False ): UpperCAmelCase_ = get_maskformer_config(lowerCAmelCase__ ) # load original state_dict with open(lowerCAmelCase__ , "rb" ) as f: UpperCAmelCase_ = pickle.load(lowerCAmelCase__ ) UpperCAmelCase_ = data["model"] # for name, param in state_dict.items(): # print(name, param.shape) # rename keys UpperCAmelCase_ = create_rename_keys(lowerCAmelCase__ ) for src, dest in rename_keys: rename_key(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) read_in_swin_q_k_v(lowerCAmelCase__ , config.backbone_config ) read_in_decoder_q_k_v(lowerCAmelCase__ , lowerCAmelCase__ ) # update to torch tensors for key, value in state_dict.items(): UpperCAmelCase_ = torch.from_numpy(lowerCAmelCase__ ) # load 🤗 model UpperCAmelCase_ = MaskFormerForInstanceSegmentation(lowerCAmelCase__ ) model.eval() for name, param in model.named_parameters(): print(lowerCAmelCase__ , param.shape ) UpperCAmelCase_ , UpperCAmelCase_ = model.load_state_dict(lowerCAmelCase__ , strict=lowerCAmelCase__ ) assert missing_keys == [ "model.pixel_level_module.encoder.model.layernorm.weight", "model.pixel_level_module.encoder.model.layernorm.bias", ] assert len(lowerCAmelCase__ ) == 0, f"""Unexpected keys: {unexpected_keys}""" # verify results UpperCAmelCase_ = prepare_img() if "vistas" in model_name: UpperCAmelCase_ = 65 elif "cityscapes" in model_name: UpperCAmelCase_ = 65535 else: UpperCAmelCase_ = 255 UpperCAmelCase_ = True if "ade" in model_name else False UpperCAmelCase_ = MaskFormerImageProcessor(ignore_index=lowerCAmelCase__ , reduce_labels=lowerCAmelCase__ ) UpperCAmelCase_ = image_processor(lowerCAmelCase__ , return_tensors="pt" ) UpperCAmelCase_ = model(**lowerCAmelCase__ ) print("Logits:" , outputs.class_queries_logits[0, :3, :3] ) if model_name == "maskformer-swin-tiny-ade": UpperCAmelCase_ = torch.tensor( [[3.6353, -4.4770, -2.6065], [0.5081, -4.2394, -3.5343], [2.1909, -5.0353, -1.9323]] ) assert torch.allclose(outputs.class_queries_logits[0, :3, :3] , lowerCAmelCase__ , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: print(f"""Saving model and image processor to {pytorch_dump_folder_path}""" ) Path(lowerCAmelCase__ ).mkdir(exist_ok=lowerCAmelCase__ ) model.save_pretrained(lowerCAmelCase__ ) image_processor.save_pretrained(lowerCAmelCase__ ) if push_to_hub: print("Pushing model and image processor to the hub..." ) model.push_to_hub(f"""nielsr/{model_name}""" ) image_processor.push_to_hub(f"""nielsr/{model_name}""" ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""maskformer-swin-tiny-ade""", type=str, help=("""Name of the MaskFormer model you'd like to convert""",), ) parser.add_argument( """--checkpoint_path""", default="""/Users/nielsrogge/Documents/MaskFormer_checkpoints/MaskFormer-Swin-tiny-ADE20k/model.pkl""", type=str, help="""Path to the original state dict (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) lowerCamelCase = parser.parse_args() convert_maskformer_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
14
0
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class lowerCamelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' def lowerCAmelCase_ ( self : Dict ): SCREAMING_SNAKE_CASE_ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(_lowerCAmelCase , 'tf_padding' ) ) self.parent.assertTrue(hasattr(_lowerCAmelCase , 'depth_multiplier' ) ) class lowerCamelCase_ : '''simple docstring''' def __init__( self : Dict , _lowerCAmelCase : int , _lowerCAmelCase : List[Any]=13 , _lowerCAmelCase : Optional[Any]=3 , _lowerCAmelCase : Optional[int]=32 , _lowerCAmelCase : Dict=0.25 , _lowerCAmelCase : str=8 , _lowerCAmelCase : Dict=8 , _lowerCAmelCase : Union[str, Any]=6 , _lowerCAmelCase : str=32 , _lowerCAmelCase : int=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Optional[int]=True , _lowerCAmelCase : Tuple="relu6" , _lowerCAmelCase : Optional[int]=1_280 , _lowerCAmelCase : Union[str, Any]=0.1 , _lowerCAmelCase : Any=0.02 , _lowerCAmelCase : int=True , _lowerCAmelCase : Dict=True , _lowerCAmelCase : Union[str, Any]=10 , _lowerCAmelCase : Optional[Any]=None , ): SCREAMING_SNAKE_CASE_ = parent SCREAMING_SNAKE_CASE_ = batch_size SCREAMING_SNAKE_CASE_ = num_channels SCREAMING_SNAKE_CASE_ = image_size SCREAMING_SNAKE_CASE_ = depth_multiplier SCREAMING_SNAKE_CASE_ = depth_divisible_by SCREAMING_SNAKE_CASE_ = min_depth SCREAMING_SNAKE_CASE_ = expand_ratio SCREAMING_SNAKE_CASE_ = tf_padding SCREAMING_SNAKE_CASE_ = output_stride SCREAMING_SNAKE_CASE_ = first_layer_is_expansion SCREAMING_SNAKE_CASE_ = finegrained_output SCREAMING_SNAKE_CASE_ = hidden_act SCREAMING_SNAKE_CASE_ = last_hidden_size if finegrained_output else int(last_hidden_size * depth_multiplier ) SCREAMING_SNAKE_CASE_ = classifier_dropout_prob SCREAMING_SNAKE_CASE_ = use_labels SCREAMING_SNAKE_CASE_ = is_training SCREAMING_SNAKE_CASE_ = num_labels SCREAMING_SNAKE_CASE_ = initializer_range SCREAMING_SNAKE_CASE_ = scope def lowerCAmelCase_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = None if self.use_labels: SCREAMING_SNAKE_CASE_ = ids_tensor([self.batch_size] , self.num_labels ) SCREAMING_SNAKE_CASE_ = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) SCREAMING_SNAKE_CASE_ = self.get_config() return config, pixel_values, labels, pixel_labels def lowerCAmelCase_ ( self : str ): return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , depth_divisible_by=self.depth_divisible_by , min_depth=self.min_depth , expand_ratio=self.expand_ratio , output_stride=self.output_stride , first_layer_is_expansion=self.first_layer_is_expansion , finegrained_output=self.finegrained_output , hidden_act=self.hidden_act , tf_padding=self.tf_padding , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def lowerCAmelCase_ ( self : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str ): SCREAMING_SNAKE_CASE_ = MobileNetVaModel(config=_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) self.parent.assertEqual( result.pooler_output.shape , (self.batch_size, self.last_hidden_size) , ) def lowerCAmelCase_ ( self : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] ): SCREAMING_SNAKE_CASE_ = self.num_labels SCREAMING_SNAKE_CASE_ = MobileNetVaForImageClassification(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCAmelCase_ ( self : List[Any] , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : Dict , _lowerCAmelCase : Dict ): SCREAMING_SNAKE_CASE_ = self.num_labels SCREAMING_SNAKE_CASE_ = MobileNetVaForSemanticSegmentation(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) SCREAMING_SNAKE_CASE_ = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowerCAmelCase_ ( self : List[str] ): SCREAMING_SNAKE_CASE_ = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = config_and_inputs SCREAMING_SNAKE_CASE_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class lowerCamelCase_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' lowercase_ = ( (MobileNetVaModel, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation) if is_torch_available() else () ) lowercase_ = ( { "feature-extraction": MobileNetVaModel, "image-classification": MobileNetVaForImageClassification, "image-segmentation": MobileNetVaForSemanticSegmentation, } if is_torch_available() else {} ) lowercase_ = False lowercase_ = False lowercase_ = False lowercase_ = False def lowerCAmelCase_ ( self : List[str] ): SCREAMING_SNAKE_CASE_ = MobileNetVaModelTester(self ) SCREAMING_SNAKE_CASE_ = MobileNetVaConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase ) def lowerCAmelCase_ ( self : List[Any] ): self.config_tester.run_common_tests() @unittest.skip(reason='MobileNetV2 does not use inputs_embeds' ) def lowerCAmelCase_ ( self : Any ): pass @unittest.skip(reason='MobileNetV2 does not support input and output embeddings' ) def lowerCAmelCase_ ( self : Optional[int] ): pass @unittest.skip(reason='MobileNetV2 does not output attentions' ) def lowerCAmelCase_ ( self : Optional[int] ): pass def lowerCAmelCase_ ( self : str ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = model_class(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE_ = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def lowerCAmelCase_ ( self : List[Any] ): SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def lowerCAmelCase_ ( self : int ): def check_hidden_states_output(_lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple ): SCREAMING_SNAKE_CASE_ = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(**self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) SCREAMING_SNAKE_CASE_ = outputs.hidden_states SCREAMING_SNAKE_CASE_ = 16 self.assertEqual(len(_lowerCAmelCase ) , _lowerCAmelCase ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] SCREAMING_SNAKE_CASE_ = True check_hidden_states_output(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) def lowerCAmelCase_ ( self : int ): SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) def lowerCAmelCase_ ( self : List[Any] ): SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*_lowerCAmelCase ) @slow def lowerCAmelCase_ ( self : Any ): for model_name in MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE_ = MobileNetVaModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def UpperCAmelCase_ ( ) -> Dict: SCREAMING_SNAKE_CASE_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def lowerCAmelCase_ ( self : Optional[Any] ): return ( MobileNetVaImageProcessor.from_pretrained('google/mobilenet_v2_1.0_224' ) if is_vision_available() else None ) @slow def lowerCAmelCase_ ( self : List[str] ): SCREAMING_SNAKE_CASE_ = MobileNetVaForImageClassification.from_pretrained('google/mobilenet_v2_1.0_224' ).to(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = self.default_image_processor SCREAMING_SNAKE_CASE_ = prepare_img() SCREAMING_SNAKE_CASE_ = image_processor(images=_lowerCAmelCase , return_tensors='pt' ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(**_lowerCAmelCase ) # verify the logits SCREAMING_SNAKE_CASE_ = torch.Size((1, 1_001) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = torch.tensor([0.2445, -1.1993, 0.1905] ).to(_lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1E-4 ) ) @slow def lowerCAmelCase_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE_ = MobileNetVaForSemanticSegmentation.from_pretrained('google/deeplabv3_mobilenet_v2_1.0_513' ) SCREAMING_SNAKE_CASE_ = model.to(_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = MobileNetVaImageProcessor.from_pretrained('google/deeplabv3_mobilenet_v2_1.0_513' ) SCREAMING_SNAKE_CASE_ = prepare_img() SCREAMING_SNAKE_CASE_ = image_processor(images=_lowerCAmelCase , return_tensors='pt' ).to(_lowerCAmelCase ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(**_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = outputs.logits # verify the logits SCREAMING_SNAKE_CASE_ = torch.Size((1, 21, 65, 65) ) self.assertEqual(logits.shape , _lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = torch.tensor( [ [[17.5790, 17.7581, 18.3355], [18.3257, 18.4230, 18.8973], [18.6169, 18.8650, 19.2187]], [[-2.1595, -2.0977, -2.3741], [-2.4226, -2.3028, -2.6835], [-2.7819, -2.5991, -2.7706]], [[4.2058, 4.8317, 4.7638], [4.4136, 5.0361, 4.9383], [4.5028, 4.9644, 4.8734]], ] , device=_lowerCAmelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , _lowerCAmelCase , atol=1E-4 ) )
31
# Copyright (c) 2021-, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #################################################################################################### # # Note: If when running this conversion script you're getting an exception: # ModuleNotFoundError: No module named 'megatron.model.enums' # you need to tell python where to find the clone of Megatron-LM, e.g.: # # cd /tmp # git clone https://github.com/NVIDIA/Megatron-LM # PYTHONPATH=/tmp/Megatron-LM python src/transformers/models/megatron_gpt2/convert_megatron_gpt2_checkpoint.py ... # # if you already have it cloned elsewhere, simply adjust the path to the existing path # # If the training was done using a Megatron-LM fork, e.g., # https://github.com/microsoft/Megatron-DeepSpeed/ then chances are that you need to have that one # in your path, i.e., /path/to/Megatron-DeepSpeed/ # import argparse import os import re import zipfile import torch from transformers import AutoTokenizer, GPTaConfig def _lowercase ( lowercase__ , lowercase__ , lowercase__=0 ): # Format the message. if name is None: __lowerCAmelCase : Any = None else: __lowerCAmelCase : Union[str, Any] = '''.''' * max(0 , spaces - 2 ) + '''# {:''' + str(5_0 - spaces ) + '''s}''' __lowerCAmelCase : Dict = fmt.format(lowercase__ ) # Print and recurse (if needed). if isinstance(lowercase__ , lowercase__ ): if msg is not None: print(lowercase__ ) for k in val.keys(): recursive_print(lowercase__ , val[k] , spaces + 2 ) elif isinstance(lowercase__ , torch.Tensor ): print(lowercase__ , ''':''' , val.size() ) else: print(lowercase__ , ''':''' , lowercase__ ) def _lowercase ( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): # Permutes layout of param tensor to [num_splits * num_heads * hidden_size, :] # for compatibility with later versions of NVIDIA Megatron-LM. # The inverse operation is performed inside Megatron-LM to read checkpoints: # https://github.com/NVIDIA/Megatron-LM/blob/v2.4/megatron/checkpointing.py#L209 # If param is the weight tensor of the self-attention block, the returned tensor # will have to be transposed one more time to be read by HuggingFace GPT2. __lowerCAmelCase : str = param.size() if checkpoint_version == 1.0: # version 1.0 stores [num_heads * hidden_size * num_splits, :] __lowerCAmelCase : Tuple = (num_heads, hidden_size, num_splits) + input_shape[1:] __lowerCAmelCase : int = param.view(*lowercase__ ) __lowerCAmelCase : int = param.transpose(0 , 2 ) __lowerCAmelCase : Any = param.transpose(1 , 2 ).contiguous() elif checkpoint_version >= 2.0: # other versions store [num_heads * num_splits * hidden_size, :] __lowerCAmelCase : Union[str, Any] = (num_heads, num_splits, hidden_size) + input_shape[1:] __lowerCAmelCase : str = param.view(*lowercase__ ) __lowerCAmelCase : List[str] = param.transpose(0 , 1 ).contiguous() __lowerCAmelCase : List[str] = param.view(*lowercase__ ) return param def _lowercase ( lowercase__ , lowercase__ , lowercase__ ): # The converted output model. __lowerCAmelCase : Any = {} # old versions did not store training args __lowerCAmelCase : Optional[int] = input_state_dict.get('''args''' , lowercase__ ) if ds_args is not None: # do not make the user write a config file when the exact dimensions/sizes are already in the checkpoint # from pprint import pprint # pprint(vars(ds_args)) __lowerCAmelCase : List[str] = ds_args.padded_vocab_size __lowerCAmelCase : List[str] = ds_args.max_position_embeddings __lowerCAmelCase : Optional[int] = ds_args.hidden_size __lowerCAmelCase : Tuple = ds_args.num_layers __lowerCAmelCase : int = ds_args.num_attention_heads __lowerCAmelCase : Union[str, Any] = ds_args.ffn_hidden_size # pprint(config) # The number of heads. __lowerCAmelCase : Union[str, Any] = config.n_head # The hidden_size per head. __lowerCAmelCase : List[Any] = config.n_embd // config.n_head # Megatron-LM checkpoint version if "checkpoint_version" in input_state_dict.keys(): __lowerCAmelCase : Union[str, Any] = input_state_dict['''checkpoint_version'''] else: __lowerCAmelCase : Optional[Any] = 0.0 # The model. __lowerCAmelCase : int = input_state_dict['''model'''] # The language model. __lowerCAmelCase : str = model['''language_model'''] # The embeddings. __lowerCAmelCase : Optional[Any] = lm['''embedding'''] # The word embeddings. __lowerCAmelCase : Tuple = embeddings['''word_embeddings''']['''weight'''] # Truncate the embedding table to vocab_size rows. __lowerCAmelCase : Dict = word_embeddings[: config.vocab_size, :] __lowerCAmelCase : int = word_embeddings # The position embeddings. __lowerCAmelCase : Optional[Any] = embeddings['''position_embeddings''']['''weight'''] # Read the causal mask dimension (seqlen). [max_sequence_length, hidden_size] __lowerCAmelCase : Dict = pos_embeddings.size(0 ) if n_positions != config.n_positions: raise ValueError( f"""pos_embeddings.max_sequence_length={n_positions} and config.n_positions={config.n_positions} don't match""" ) # Store the position embeddings. __lowerCAmelCase : Any = pos_embeddings # The transformer. __lowerCAmelCase : Dict = lm['''transformer'''] if '''transformer''' in lm.keys() else lm['''encoder'''] # The regex to extract layer names. __lowerCAmelCase : Any = re.compile(r'''layers\.(\d+)\.([a-z0-9_.]+)\.([a-z]+)''' ) # The simple map of names for "automated" rules. __lowerCAmelCase : Optional[Any] = { '''attention.dense''': '''.attn.c_proj.''', '''self_attention.dense''': '''.attn.c_proj.''', '''mlp.dense_h_to_4h''': '''.mlp.c_fc.''', '''mlp.dense_4h_to_h''': '''.mlp.c_proj.''', } # Extract the layers. for key, val in transformer.items(): # Match the name. __lowerCAmelCase : str = layer_re.match(lowercase__ ) # Stop if that's not a layer if m is None: break # The index of the layer. __lowerCAmelCase : Tuple = int(m.group(1 ) ) # The name of the operation. __lowerCAmelCase : Dict = m.group(2 ) # Is it a weight or a bias? __lowerCAmelCase : Optional[Any] = m.group(3 ) # The name of the layer. __lowerCAmelCase : Dict = f"""transformer.h.{layer_idx}""" # For layernorm(s), simply store the layer norm. if op_name.endswith('''layernorm''' ): __lowerCAmelCase : List[Any] = '''ln_1''' if op_name.startswith('''input''' ) else '''ln_2''' __lowerCAmelCase : Optional[int] = val # Transpose the QKV matrix. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "weight": # Insert a tensor of 1x1xDxD bias. __lowerCAmelCase : List[str] = torch.tril(torch.ones((n_positions, n_positions) , dtype=torch.floataa ) ).view( 1 , 1 , lowercase__ , lowercase__ ) __lowerCAmelCase : List[Any] = causal_mask # Insert a "dummy" tensor for masked_bias. __lowerCAmelCase : str = torch.tensor(-1E4 , dtype=torch.floataa ) __lowerCAmelCase : str = masked_bias __lowerCAmelCase : Tuple = fix_query_key_value_ordering(lowercase__ , lowercase__ , 3 , lowercase__ , lowercase__ ) # Megatron stores (3*D) x D but transformers-GPT2 expects D x 3*D. __lowerCAmelCase : int = out_val.transpose(0 , 1 ).contiguous() # Store. __lowerCAmelCase : Dict = out_val # Transpose the bias. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "bias": __lowerCAmelCase : List[str] = fix_query_key_value_ordering(lowercase__ , lowercase__ , 3 , lowercase__ , lowercase__ ) # Store. No change of shape. __lowerCAmelCase : int = out_val # Transpose the weights. elif weight_or_bias == "weight": __lowerCAmelCase : List[Any] = megatron_to_transformers[op_name] __lowerCAmelCase : Optional[int] = val.transpose(0 , 1 ) # Copy the bias. elif weight_or_bias == "bias": __lowerCAmelCase : Optional[int] = megatron_to_transformers[op_name] __lowerCAmelCase : Any = val # DEBUG. assert config.n_layer == layer_idx + 1 # The final layernorm. __lowerCAmelCase : str = transformer['''final_layernorm.weight'''] __lowerCAmelCase : List[Any] = transformer['''final_layernorm.bias'''] # For LM head, transformers' wants the matrix to weight embeddings. __lowerCAmelCase : List[Any] = word_embeddings # It should be done! return output_state_dict def _lowercase ( ): # Create the argument parser. __lowerCAmelCase : Union[str, Any] = argparse.ArgumentParser() parser.add_argument('''--print-checkpoint-structure''' , action='''store_true''' ) parser.add_argument( '''path_to_checkpoint''' , type=lowercase__ , help='''Path to the checkpoint file (.zip archive or direct .pt file)''' , ) parser.add_argument( '''--config_file''' , default='''''' , type=lowercase__ , help='''An optional config json file describing the pre-trained model.''' , ) __lowerCAmelCase : Optional[int] = parser.parse_args() # Extract the basename. __lowerCAmelCase : Union[str, Any] = os.path.dirname(args.path_to_checkpoint ) # Load the model. # the .zip is very optional, let's keep it for backward compatibility print(f"""Extracting PyTorch state dictionary from {args.path_to_checkpoint}""" ) if args.path_to_checkpoint.endswith('''.zip''' ): with zipfile.ZipFile(args.path_to_checkpoint , '''r''' ) as checkpoint: with checkpoint.open('''release/mp_rank_00/model_optim_rng.pt''' ) as pytorch_dict: __lowerCAmelCase : Union[str, Any] = torch.load(lowercase__ , map_location='''cpu''' ) else: __lowerCAmelCase : Optional[int] = torch.load(args.path_to_checkpoint , map_location='''cpu''' ) __lowerCAmelCase : int = input_state_dict.get('''args''' , lowercase__ ) # Read the config, or default to the model released by NVIDIA. if args.config_file == "": if ds_args is not None: if ds_args.bias_gelu_fusion: __lowerCAmelCase : List[Any] = '''gelu_fast''' elif ds_args.openai_gelu: __lowerCAmelCase : Union[str, Any] = '''gelu_new''' else: __lowerCAmelCase : str = '''gelu''' else: # in the very early days this used to be "gelu_new" __lowerCAmelCase : int = '''gelu_new''' # Spell out all parameters in case the defaults change. __lowerCAmelCase : str = GPTaConfig( vocab_size=5_0_2_5_7 , n_positions=1_0_2_4 , n_embd=1_0_2_4 , n_layer=2_4 , n_head=1_6 , n_inner=4_0_9_6 , activation_function=lowercase__ , resid_pdrop=0.1 , embd_pdrop=0.1 , attn_pdrop=0.1 , layer_norm_epsilon=1E-5 , initializer_range=0.0_2 , summary_type='''cls_index''' , summary_use_proj=lowercase__ , summary_activation=lowercase__ , summary_proj_to_labels=lowercase__ , summary_first_dropout=0.1 , scale_attn_weights=lowercase__ , use_cache=lowercase__ , bos_token_id=5_0_2_5_6 , eos_token_id=5_0_2_5_6 , ) else: __lowerCAmelCase : List[str] = GPTaConfig.from_json_file(args.config_file ) __lowerCAmelCase : List[str] = ['''GPT2LMHeadModel'''] # Convert. print('''Converting''' ) __lowerCAmelCase : List[Any] = convert_megatron_checkpoint(lowercase__ , lowercase__ , lowercase__ ) # Print the structure of converted state dict. if args.print_checkpoint_structure: recursive_print(lowercase__ , lowercase__ ) # Add tokenizer class info to config # see https://github.com/huggingface/transformers/issues/13906) if ds_args is not None: __lowerCAmelCase : Any = ds_args.tokenizer_type if tokenizer_type == "GPT2BPETokenizer": __lowerCAmelCase : Any = '''gpt2''' elif tokenizer_type == "PretrainedFromHF": __lowerCAmelCase : Optional[int] = ds_args.tokenizer_name_or_path else: raise ValueError(f"""Unrecognized tokenizer_type {tokenizer_type}""" ) else: __lowerCAmelCase : str = '''gpt2''' __lowerCAmelCase : int = AutoTokenizer.from_pretrained(lowercase__ ) __lowerCAmelCase : Tuple = type(lowercase__ ).__name__ __lowerCAmelCase : Dict = tokenizer_class # Store the config to file. print('''Saving config''' ) config.save_pretrained(lowercase__ ) # Save tokenizer based on args print(f"""Adding {tokenizer_class} tokenizer files""" ) tokenizer.save_pretrained(lowercase__ ) # Store the state_dict to file. __lowerCAmelCase : List[str] = os.path.join(lowercase__ , '''pytorch_model.bin''' ) print(f"""Saving checkpoint to \"{output_checkpoint_file}\"""" ) torch.save(lowercase__ , lowercase__ ) #################################################################################################### if __name__ == "__main__": main() ####################################################################################################
492
0
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # and perform gradient accumulation # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## SCREAMING_SNAKE_CASE__ : Dict = 16 SCREAMING_SNAKE_CASE__ : Optional[Any] = 32 def __lowercase ( snake_case, snake_case = 1_6 ): """simple docstring""" __magic_name__ :Optional[Any] = AutoTokenizer.from_pretrained('''bert-base-cased''' ) __magic_name__ :List[str] = load_dataset('''glue''', '''mrpc''' ) def tokenize_function(snake_case ): # max_length=None => use the model max length (it's actually the default) __magic_name__ :Dict = tokenizer(examples['''sentence1'''], examples['''sentence2'''], truncation=snake_case, max_length=snake_case ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): __magic_name__ :Any = datasets.map( snake_case, batched=snake_case, remove_columns=['''idx''', '''sentence1''', '''sentence2'''], ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __magic_name__ :Optional[Any] = tokenized_datasets.rename_column('''label''', '''labels''' ) def collate_fn(snake_case ): # On TPU it's best to pad everything to the same length or training will be very slow. __magic_name__ :str = 1_2_8 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": __magic_name__ :Optional[Any] = 1_6 elif accelerator.mixed_precision != "no": __magic_name__ :List[Any] = 8 else: __magic_name__ :Optional[int] = None return tokenizer.pad( snake_case, padding='''longest''', max_length=snake_case, pad_to_multiple_of=snake_case, return_tensors='''pt''', ) # Instantiate dataloaders. __magic_name__ :Optional[int] = DataLoader( tokenized_datasets['''train'''], shuffle=snake_case, collate_fn=snake_case, batch_size=snake_case ) __magic_name__ :int = DataLoader( tokenized_datasets['''validation'''], shuffle=snake_case, collate_fn=snake_case, batch_size=snake_case ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", None) == "1": from accelerate.test_utils.training import mocked_dataloaders SCREAMING_SNAKE_CASE__ : Optional[int] = mocked_dataloaders # noqa: F811 def __lowercase ( snake_case, snake_case ): """simple docstring""" if os.environ.get('''TESTING_MOCKED_DATALOADERS''', snake_case ) == "1": __magic_name__ :Any = 2 # New Code # __magic_name__ :int = int(args.gradient_accumulation_steps ) # Initialize accelerator __magic_name__ :Tuple = Accelerator( cpu=args.cpu, mixed_precision=args.mixed_precision, gradient_accumulation_steps=snake_case ) if accelerator.distributed_type == DistributedType.TPU and gradient_accumulation_steps > 1: raise NotImplementedError( '''Gradient accumulation on TPUs is currently not supported. Pass `gradient_accumulation_steps=1`''' ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __magic_name__ :Union[str, Any] = config['''lr'''] __magic_name__ :Tuple = int(config['''num_epochs'''] ) __magic_name__ :Optional[int] = int(config['''seed'''] ) __magic_name__ :Union[str, Any] = int(config['''batch_size'''] ) __magic_name__ :int = evaluate.load('''glue''', '''mrpc''' ) set_seed(snake_case ) __magic_name__ , __magic_name__ :List[Any] = get_dataloaders(snake_case, snake_case ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __magic_name__ :List[str] = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''', return_dict=snake_case ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). __magic_name__ :str = model.to(accelerator.device ) # Instantiate optimizer __magic_name__ :str = AdamW(params=model.parameters(), lr=snake_case ) # Instantiate scheduler __magic_name__ :List[Any] = get_linear_schedule_with_warmup( optimizer=snake_case, num_warmup_steps=1_0_0, num_training_steps=(len(snake_case ) * num_epochs), ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ :str = accelerator.prepare( snake_case, snake_case, snake_case, snake_case, snake_case ) # Now we train the model for epoch in range(snake_case ): model.train() for step, batch in enumerate(snake_case ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(snake_case ): __magic_name__ :str = model(**snake_case ) __magic_name__ :Union[str, Any] = output.loss accelerator.backward(snake_case ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(snake_case ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __magic_name__ :Any = model(**snake_case ) __magic_name__ :Union[str, Any] = outputs.logits.argmax(dim=-1 ) __magic_name__ , __magic_name__ :Union[str, Any] = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=snake_case, references=snake_case, ) __magic_name__ :Dict = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'''epoch {epoch}:''', snake_case ) def __lowercase ( ): """simple docstring""" __magic_name__ :List[Any] = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''', type=snake_case, default=snake_case, choices=['''no''', '''fp16''', '''bf16''', '''fp8'''], help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''', ) # New Code # parser.add_argument( '''--gradient_accumulation_steps''', type=snake_case, default=1, help='''The number of minibatches to be ran before gradients are accumulated.''', ) parser.add_argument('''--cpu''', action='''store_true''', help='''If passed, will train on the CPU.''' ) __magic_name__ :int = parser.parse_args() __magic_name__ :Optional[int] = {'''lr''': 2E-5, '''num_epochs''': 3, '''seed''': 4_2, '''batch_size''': 1_6} training_function(snake_case, snake_case ) if __name__ == "__main__": main()
180
SCREAMING_SNAKE_CASE__ : str = """Alexander Joslin""" import operator as op from .stack import Stack def __lowercase ( snake_case ): """simple docstring""" __magic_name__ :Optional[int] = {'''*''': op.mul, '''/''': op.truediv, '''+''': op.add, '''-''': op.sub} __magic_name__ :Stack[int] = Stack() __magic_name__ :Stack[str] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(snake_case ) ) elif i in operators: # RULE 2 operator_stack.push(snake_case ) elif i == ")": # RULE 4 __magic_name__ :Optional[int] = operator_stack.peek() operator_stack.pop() __magic_name__ :List[str] = operand_stack.peek() operand_stack.pop() __magic_name__ :Optional[Any] = operand_stack.peek() operand_stack.pop() __magic_name__ :Optional[int] = operators[opr](snake_case, snake_case ) operand_stack.push(snake_case ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Any = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(f"{equation} = {dijkstras_two_stack_algorithm(equation)}")
180
1
import math import qiskit def _UpperCamelCase (a__ :int = 1 , a__ :int = 1 , a__ :int = 1 ): """simple docstring""" if ( isinstance(a__ , a__ ) or isinstance(a__ , a__ ) or isinstance(a__ , a__ ) ): raise TypeError("""inputs must be integers.""" ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError("""inputs must be positive.""" ) if ( (math.floor(a__ ) != input_a) or (math.floor(a__ ) != input_a) or (math.floor(a__ ) != carry_in) ): raise ValueError("""inputs must be exact integers.""" ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError("""inputs must be less or equal to 2.""" ) # build registers UpperCamelCase__ = qiskit.QuantumRegister(4 , """qr""" ) UpperCamelCase__ = qiskit.ClassicalRegister(2 , """cr""" ) # list the entries UpperCamelCase__ = [input_a, input_a, carry_in] UpperCamelCase__ = qiskit.QuantumCircuit(a__ , a__ ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(a__ ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(a__ ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(a__ ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , a__ ) # measure the last two qbits UpperCamelCase__ = qiskit.Aer.get_backend("""aer_simulator""" ) UpperCamelCase__ = qiskit.execute(a__ , a__ , shots=1000 ) return job.result().get_counts(a__ ) if __name__ == "__main__": print(f"""Total sum count for state is: {quantum_full_adder(1, 1, 1)}""")
619
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird import BigBirdTokenizer else: UpperCamelCase__ = None UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = {"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} UpperCamelCase__ = { "vocab_file": { "google/bigbird-roberta-base": "https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model", "google/bigbird-roberta-large": ( "https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model" ), "google/bigbird-base-trivia-itc": ( "https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model" ), }, "tokenizer_file": { "google/bigbird-roberta-base": ( "https://huggingface.co/google/bigbird-roberta-base/resolve/main/tokenizer.json" ), "google/bigbird-roberta-large": ( "https://huggingface.co/google/bigbird-roberta-large/resolve/main/tokenizer.json" ), "google/bigbird-base-trivia-itc": ( "https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/tokenizer.json" ), }, } UpperCamelCase__ = { "google/bigbird-roberta-base": 4_096, "google/bigbird-roberta-large": 4_096, "google/bigbird-base-trivia-itc": 4_096, } UpperCamelCase__ = "▁" class __SCREAMING_SNAKE_CASE ( _a ): snake_case : Optional[Any] = VOCAB_FILES_NAMES snake_case : List[str] = PRETRAINED_VOCAB_FILES_MAP snake_case : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case : int = BigBirdTokenizer snake_case : List[Any] = ["""input_ids""", """attention_mask"""] snake_case : List[int] = [] def __init__( self , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase="<unk>" , __lowerCAmelCase="<s>" , __lowerCAmelCase="</s>" , __lowerCAmelCase="<pad>" , __lowerCAmelCase="[SEP]" , __lowerCAmelCase="[MASK]" , __lowerCAmelCase="[CLS]" , **__lowerCAmelCase , ): UpperCamelCase__ = AddedToken(__lowerCAmelCase , lstrip=__lowerCAmelCase , rstrip=__lowerCAmelCase ) if isinstance(__lowerCAmelCase , __lowerCAmelCase ) else bos_token UpperCamelCase__ = AddedToken(__lowerCAmelCase , lstrip=__lowerCAmelCase , rstrip=__lowerCAmelCase ) if isinstance(__lowerCAmelCase , __lowerCAmelCase ) else eos_token UpperCamelCase__ = AddedToken(__lowerCAmelCase , lstrip=__lowerCAmelCase , rstrip=__lowerCAmelCase ) if isinstance(__lowerCAmelCase , __lowerCAmelCase ) else unk_token UpperCamelCase__ = AddedToken(__lowerCAmelCase , lstrip=__lowerCAmelCase , rstrip=__lowerCAmelCase ) if isinstance(__lowerCAmelCase , __lowerCAmelCase ) else pad_token UpperCamelCase__ = AddedToken(__lowerCAmelCase , lstrip=__lowerCAmelCase , rstrip=__lowerCAmelCase ) if isinstance(__lowerCAmelCase , __lowerCAmelCase ) else cls_token UpperCamelCase__ = AddedToken(__lowerCAmelCase , lstrip=__lowerCAmelCase , rstrip=__lowerCAmelCase ) if isinstance(__lowerCAmelCase , __lowerCAmelCase ) else sep_token # Mask token behave like a normal word, i.e. include the space before it UpperCamelCase__ = AddedToken(__lowerCAmelCase , lstrip=__lowerCAmelCase , rstrip=__lowerCAmelCase ) if isinstance(__lowerCAmelCase , __lowerCAmelCase ) else mask_token super().__init__( __lowerCAmelCase , tokenizer_file=__lowerCAmelCase , bos_token=__lowerCAmelCase , eos_token=__lowerCAmelCase , unk_token=__lowerCAmelCase , sep_token=__lowerCAmelCase , pad_token=__lowerCAmelCase , cls_token=__lowerCAmelCase , mask_token=__lowerCAmelCase , **__lowerCAmelCase , ) UpperCamelCase__ = vocab_file UpperCamelCase__ = False if not self.vocab_file else True def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase = None ): UpperCamelCase__ = [self.sep_token_id] UpperCamelCase__ = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase = None , __lowerCAmelCase = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( """You should not supply a second sequence if the provided sequence of """ """ids is already formatted with special tokens for the model.""" ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is None: return [1] + ([0] * len(__lowerCAmelCase )) + [1] return [1] + ([0] * len(__lowerCAmelCase )) + [1] + ([0] * len(__lowerCAmelCase )) + [1] def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase = None ): UpperCamelCase__ = [self.sep_token_id] UpperCamelCase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase = None ): if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(__lowerCAmelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCamelCase__ = os.path.join( __lowerCAmelCase , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowerCAmelCase ): copyfile(self.vocab_file , __lowerCAmelCase ) return (out_vocab_file,)
619
1
def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : Dict ): __lowerCAmelCase = (boundary[1] - boundary[0]) / steps __lowerCAmelCase = boundary[0] __lowerCAmelCase = boundary[1] __lowerCAmelCase = make_points(lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ) __lowerCAmelCase = 0.0 y += (h / 2.0) * f(lowerCamelCase__ ) for i in x_i: # print(i) y += h * f(lowerCamelCase__ ) y += (h / 2.0) * f(lowerCamelCase__ ) return y def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : Union[str, Any], lowerCAmelCase_ : Tuple ): __lowerCAmelCase = a + h while x < (b - h): yield x __lowerCAmelCase = x + h def a_ ( lowerCAmelCase_ : Any ): # enter your function here __lowerCAmelCase = (x - 0) * (x - 0) return y def a_ ( ): __lowerCAmelCase = 0.0 # Lower bound of integration __lowerCAmelCase = 1.0 # Upper bound of integration __lowerCAmelCase = 10.0 # define number of steps or resolution __lowerCAmelCase = [a, b] # define boundary of integration __lowerCAmelCase = method_a(lowerCamelCase__, lowerCamelCase__ ) print(F"""y = {y}""" ) if __name__ == "__main__": main()
710
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _snake_case : str = logging.get_logger(__name__) _snake_case : Optional[Any] = '▁' _snake_case : List[Any] = {'vocab_file': 'sentencepiece.bpe.model'} _snake_case : List[Any] = { 'vocab_file': { 'facebook/xglm-564M': 'https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model', } } _snake_case : Union[str, Any] = { 'facebook/xglm-564M': 2048, } class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = VOCAB_FILES_NAMES a_ = PRETRAINED_VOCAB_FILES_MAP a_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ = ["""input_ids""", """attention_mask"""] def __init__( self : Any , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : int="<s>" , lowerCAmelCase_ : Dict="</s>" , lowerCAmelCase_ : Tuple="</s>" , lowerCAmelCase_ : Union[str, Any]="<s>" , lowerCAmelCase_ : Optional[Any]="<unk>" , lowerCAmelCase_ : Optional[int]="<pad>" , lowerCAmelCase_ : Optional[Dict[str, Any]] = None , **lowerCAmelCase_ : int , ) -> None: __lowerCAmelCase = {} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer __lowerCAmelCase = 7 __lowerCAmelCase = [f"""<madeupword{i}>""" for i in range(self.num_madeup_words )] __lowerCAmelCase = kwargs.get('additional_special_tokens' , [] ) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=lowerCAmelCase_ , eos_token=lowerCAmelCase_ , unk_token=lowerCAmelCase_ , sep_token=lowerCAmelCase_ , cls_token=lowerCAmelCase_ , pad_token=lowerCAmelCase_ , sp_model_kwargs=self.sp_model_kwargs , **lowerCAmelCase_ , ) __lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(lowerCAmelCase_ ) ) __lowerCAmelCase = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab __lowerCAmelCase = 1 # Mimic fairseq token-to-id alignment for the first 4 token __lowerCAmelCase = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} __lowerCAmelCase = len(self.sp_model ) __lowerCAmelCase = {f"""<madeupword{i}>""": sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words )} self.fairseq_tokens_to_ids.update(lowerCAmelCase_ ) __lowerCAmelCase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : Any ) -> int: __lowerCAmelCase = self.__dict__.copy() __lowerCAmelCase = None __lowerCAmelCase = self.sp_model.serialized_model_proto() return state def __setstate__( self : Optional[Any] , lowerCAmelCase_ : Tuple ) -> Optional[Any]: __lowerCAmelCase = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): __lowerCAmelCase = {} __lowerCAmelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def lowercase ( self : str , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.sep_token_id] + token_ids_a __lowerCAmelCase = [self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def lowercase ( self : Optional[int] , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : Optional[List[int]] = None , lowerCAmelCase_ : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase_ , token_ids_a=lowerCAmelCase_ , already_has_special_tokens=lowerCAmelCase_ ) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase_ )) return [1] + ([0] * len(lowerCAmelCase_ )) + [1, 1] + ([0] * len(lowerCAmelCase_ )) def lowercase ( self : Optional[Any] , lowerCAmelCase_ : List[int] , lowerCAmelCase_ : Optional[List[int]] = None ) -> List[int]: __lowerCAmelCase = [self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a ) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a ) * [0] @property def lowercase ( self : List[str] ) -> List[str]: return len(self.sp_model ) + self.fairseq_offset + self.num_madeup_words def lowercase ( self : List[str] ) -> List[Any]: __lowerCAmelCase = {self.convert_ids_to_tokens(lowerCAmelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowercase ( self : List[Any] , lowerCAmelCase_ : str ) -> List[str]: return self.sp_model.encode(lowerCAmelCase_ , out_type=lowerCAmelCase_ ) def lowercase ( self : Union[str, Any] , lowerCAmelCase_ : str ) -> Dict: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] __lowerCAmelCase = self.sp_model.PieceToId(lowerCAmelCase_ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def lowercase ( self : Dict , lowerCAmelCase_ : str ) -> int: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def lowercase ( self : Any , lowerCAmelCase_ : Tuple ) -> Tuple: __lowerCAmelCase = ''.join(lowerCAmelCase_ ).replace(lowerCAmelCase_ , ' ' ).strip() return out_string def lowercase ( self : int , lowerCAmelCase_ : str , lowerCAmelCase_ : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(lowerCAmelCase_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __lowerCAmelCase = os.path.join( lowerCAmelCase_ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCAmelCase_ ) elif not os.path.isfile(self.vocab_file ): with open(lowerCAmelCase_ , 'wb' ) as fi: __lowerCAmelCase = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase_ ) return (out_vocab_file,)
421
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer a =logging.get_logger(__name__) a ={"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} a ={ """vocab_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt""" ), """distilbert-base-german-cased""": """https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt""", """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt""" ), }, """tokenizer_file""": { """distilbert-base-uncased""": """https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json""", """distilbert-base-uncased-distilled-squad""": ( """https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-cased""": """https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json""", """distilbert-base-cased-distilled-squad""": ( """https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json""" ), """distilbert-base-german-cased""": ( """https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json""" ), """distilbert-base-multilingual-cased""": ( """https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json""" ), }, } a ={ """distilbert-base-uncased""": 512, """distilbert-base-uncased-distilled-squad""": 512, """distilbert-base-cased""": 512, """distilbert-base-cased-distilled-squad""": 512, """distilbert-base-german-cased""": 512, """distilbert-base-multilingual-cased""": 512, } a ={ """distilbert-base-uncased""": {"""do_lower_case""": True}, """distilbert-base-uncased-distilled-squad""": {"""do_lower_case""": True}, """distilbert-base-cased""": {"""do_lower_case""": False}, """distilbert-base-cased-distilled-squad""": {"""do_lower_case""": False}, """distilbert-base-german-cased""": {"""do_lower_case""": False}, """distilbert-base-multilingual-cased""": {"""do_lower_case""": False}, } class A_ ( SCREAMING_SNAKE_CASE ): _UpperCAmelCase : Optional[int] = VOCAB_FILES_NAMES _UpperCAmelCase : int = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase : Any = PRETRAINED_INIT_CONFIGURATION _UpperCAmelCase : int = ['''input_ids''', '''attention_mask'''] _UpperCAmelCase : int = DistilBertTokenizer def __init__( self : Union[str, Any] ,SCREAMING_SNAKE_CASE__ : List[Any]=None ,SCREAMING_SNAKE_CASE__ : List[str]=None ,SCREAMING_SNAKE_CASE__ : str=True ,SCREAMING_SNAKE_CASE__ : Tuple="[UNK]" ,SCREAMING_SNAKE_CASE__ : Tuple="[SEP]" ,SCREAMING_SNAKE_CASE__ : Optional[int]="[PAD]" ,SCREAMING_SNAKE_CASE__ : Union[str, Any]="[CLS]" ,SCREAMING_SNAKE_CASE__ : Union[str, Any]="[MASK]" ,SCREAMING_SNAKE_CASE__ : str=True ,SCREAMING_SNAKE_CASE__ : Optional[Any]=None ,**SCREAMING_SNAKE_CASE__ : Any ,): super().__init__( SCREAMING_SNAKE_CASE__ ,tokenizer_file=SCREAMING_SNAKE_CASE__ ,do_lower_case=SCREAMING_SNAKE_CASE__ ,unk_token=SCREAMING_SNAKE_CASE__ ,sep_token=SCREAMING_SNAKE_CASE__ ,pad_token=SCREAMING_SNAKE_CASE__ ,cls_token=SCREAMING_SNAKE_CASE__ ,mask_token=SCREAMING_SNAKE_CASE__ ,tokenize_chinese_chars=SCREAMING_SNAKE_CASE__ ,strip_accents=SCREAMING_SNAKE_CASE__ ,**SCREAMING_SNAKE_CASE__ ,) __lowerCamelCase : Optional[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( normalizer_state.get('lowercase' ,SCREAMING_SNAKE_CASE__) != do_lower_case or normalizer_state.get('strip_accents' ,SCREAMING_SNAKE_CASE__) != strip_accents or normalizer_state.get('handle_chinese_chars' ,SCREAMING_SNAKE_CASE__) != tokenize_chinese_chars ): __lowerCamelCase : str = getattr(SCREAMING_SNAKE_CASE__ ,normalizer_state.pop('type')) __lowerCamelCase : Union[str, Any] = do_lower_case __lowerCamelCase : str = strip_accents __lowerCamelCase : Optional[Any] = tokenize_chinese_chars __lowerCamelCase : int = normalizer_class(**SCREAMING_SNAKE_CASE__) __lowerCamelCase : int = do_lower_case def lowerCAmelCase ( self : Dict ,SCREAMING_SNAKE_CASE__ : Dict ,SCREAMING_SNAKE_CASE__ : Tuple=None): __lowerCamelCase : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowerCAmelCase ( self : Dict ,SCREAMING_SNAKE_CASE__ : List[int] ,SCREAMING_SNAKE_CASE__ : Optional[List[int]] = None): __lowerCamelCase : List[str] = [self.sep_token_id] __lowerCamelCase : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def lowerCAmelCase ( self : Optional[int] ,SCREAMING_SNAKE_CASE__ : str ,SCREAMING_SNAKE_CASE__ : Optional[str] = None): __lowerCamelCase : Any = self._tokenizer.model.save(SCREAMING_SNAKE_CASE__ ,name=SCREAMING_SNAKE_CASE__) return tuple(SCREAMING_SNAKE_CASE__)
652
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging a =logging.get_logger(__name__) a ="""▁""" a ={"""vocab_file""": """sentencepiece.bpe.model""", """monolingual_vocab_file""": """dict.txt"""} a ={ """vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model""", }, """monolingual_vocab_file""": { """vinai/bartpho-syllable""": """https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt""", }, } a ={"""vinai/bartpho-syllable""": 1024} class A_ ( SCREAMING_SNAKE_CASE ): _UpperCAmelCase : List[str] = VOCAB_FILES_NAMES _UpperCAmelCase : Optional[int] = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase : Dict = ['''input_ids''', '''attention_mask'''] def __init__( self : Dict ,SCREAMING_SNAKE_CASE__ : Union[str, Any] ,SCREAMING_SNAKE_CASE__ : List[Any] ,SCREAMING_SNAKE_CASE__ : List[Any]="<s>" ,SCREAMING_SNAKE_CASE__ : Any="</s>" ,SCREAMING_SNAKE_CASE__ : List[str]="</s>" ,SCREAMING_SNAKE_CASE__ : List[str]="<s>" ,SCREAMING_SNAKE_CASE__ : int="<unk>" ,SCREAMING_SNAKE_CASE__ : Dict="<pad>" ,SCREAMING_SNAKE_CASE__ : List[str]="<mask>" ,SCREAMING_SNAKE_CASE__ : Optional[Dict[str, Any]] = None ,**SCREAMING_SNAKE_CASE__ : Dict ,): # Mask token behave like a normal word, i.e. include the space before it __lowerCamelCase : Union[str, Any] = AddedToken(SCREAMING_SNAKE_CASE__ ,lstrip=SCREAMING_SNAKE_CASE__ ,rstrip=SCREAMING_SNAKE_CASE__) if isinstance(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__) else mask_token __lowerCamelCase : str = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=SCREAMING_SNAKE_CASE__ ,eos_token=SCREAMING_SNAKE_CASE__ ,unk_token=SCREAMING_SNAKE_CASE__ ,sep_token=SCREAMING_SNAKE_CASE__ ,cls_token=SCREAMING_SNAKE_CASE__ ,pad_token=SCREAMING_SNAKE_CASE__ ,mask_token=SCREAMING_SNAKE_CASE__ ,sp_model_kwargs=self.sp_model_kwargs ,**SCREAMING_SNAKE_CASE__ ,) __lowerCamelCase : int = vocab_file __lowerCamelCase : Tuple = monolingual_vocab_file __lowerCamelCase : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(str(SCREAMING_SNAKE_CASE__)) # Load the reduced vocab # Keep order of special tokens for backward compatibility __lowerCamelCase : Optional[int] = {} __lowerCamelCase : List[Any] = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(SCREAMING_SNAKE_CASE__) not in self.fairseq_tokens_to_ids: __lowerCamelCase : Any = cnt cnt += 1 with open(SCREAMING_SNAKE_CASE__ ,'r' ,encoding='utf-8') as f: for line in f.readlines(): __lowerCamelCase : Any = line.strip().split()[0] __lowerCamelCase : List[str] = len(self.fairseq_tokens_to_ids) if str(SCREAMING_SNAKE_CASE__) not in self.fairseq_tokens_to_ids: __lowerCamelCase : Dict = len(self.fairseq_tokens_to_ids) __lowerCamelCase : str = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self : int): __lowerCamelCase : Tuple = self.__dict__.copy() __lowerCamelCase : Optional[int] = None __lowerCamelCase : Optional[int] = self.sp_model.serialized_model_proto() return state def __setstate__( self : Optional[Any] ,SCREAMING_SNAKE_CASE__ : Optional[Any]): __lowerCamelCase : List[str] = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs'): __lowerCamelCase : str = {} __lowerCamelCase : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.LoadFromSerializedProto(self.sp_model_proto) def lowerCAmelCase ( self : Tuple ,SCREAMING_SNAKE_CASE__ : List[int] ,SCREAMING_SNAKE_CASE__ : Optional[List[int]] = None): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __lowerCamelCase : Tuple = [self.cls_token_id] __lowerCamelCase : int = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowerCAmelCase ( self : Dict ,SCREAMING_SNAKE_CASE__ : List[int] ,SCREAMING_SNAKE_CASE__ : Optional[List[int]] = None ,SCREAMING_SNAKE_CASE__ : bool = False): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=SCREAMING_SNAKE_CASE__ ,token_ids_a=SCREAMING_SNAKE_CASE__ ,already_has_special_tokens=SCREAMING_SNAKE_CASE__) if token_ids_a is None: return [1] + ([0] * len(SCREAMING_SNAKE_CASE__)) + [1] return [1] + ([0] * len(SCREAMING_SNAKE_CASE__)) + [1, 1] + ([0] * len(SCREAMING_SNAKE_CASE__)) + [1] def lowerCAmelCase ( self : Tuple ,SCREAMING_SNAKE_CASE__ : List[int] ,SCREAMING_SNAKE_CASE__ : Optional[List[int]] = None): __lowerCamelCase : Dict = [self.sep_token_id] __lowerCamelCase : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] @property def lowerCAmelCase ( self : List[str]): return len(self.fairseq_ids_to_tokens) def lowerCAmelCase ( self : Dict): __lowerCamelCase : Optional[int] = {self.convert_ids_to_tokens(SCREAMING_SNAKE_CASE__): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def lowerCAmelCase ( self : Dict ,SCREAMING_SNAKE_CASE__ : str): return self.sp_model.encode(SCREAMING_SNAKE_CASE__ ,out_type=SCREAMING_SNAKE_CASE__) def lowerCAmelCase ( self : Optional[int] ,SCREAMING_SNAKE_CASE__ : List[Any]): if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def lowerCAmelCase ( self : Union[str, Any] ,SCREAMING_SNAKE_CASE__ : Dict): return self.fairseq_ids_to_tokens[index] def lowerCAmelCase ( self : Tuple ,SCREAMING_SNAKE_CASE__ : str): __lowerCamelCase : Any = ''.join(SCREAMING_SNAKE_CASE__).replace(SCREAMING_SNAKE_CASE__ ,' ').strip() return out_string def lowerCAmelCase ( self : Union[str, Any] ,SCREAMING_SNAKE_CASE__ : str ,SCREAMING_SNAKE_CASE__ : Optional[str] = None): if not os.path.isdir(SCREAMING_SNAKE_CASE__): logger.error(F"Vocabulary path ({save_directory}) should be a directory") return __lowerCamelCase : Union[str, Any] = os.path.join( SCREAMING_SNAKE_CASE__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file']) __lowerCamelCase : Union[str, Any] = os.path.join( SCREAMING_SNAKE_CASE__ ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['monolingual_vocab_file'] ,) if os.path.abspath(self.vocab_file) != os.path.abspath(SCREAMING_SNAKE_CASE__) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file ,SCREAMING_SNAKE_CASE__) elif not os.path.isfile(self.vocab_file): with open(SCREAMING_SNAKE_CASE__ ,'wb') as fi: __lowerCamelCase : List[str] = self.sp_model.serialized_model_proto() fi.write(SCREAMING_SNAKE_CASE__) if os.path.abspath(self.monolingual_vocab_file) != os.path.abspath( SCREAMING_SNAKE_CASE__) and os.path.isfile(self.monolingual_vocab_file): copyfile(self.monolingual_vocab_file ,SCREAMING_SNAKE_CASE__) elif not os.path.isfile(self.monolingual_vocab_file): with open(SCREAMING_SNAKE_CASE__ ,'w' ,encoding='utf-8') as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(F"{str(SCREAMING_SNAKE_CASE__)} \n") return out_vocab_file, out_monolingual_vocab_file
652
1
'''simple docstring''' import argparse import math import traceback import dateutil.parser as date_parser import requests def SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE ): SCREAMING_SNAKE_CASE_ :int = {} SCREAMING_SNAKE_CASE_ :Optional[int] = job['started_at'] SCREAMING_SNAKE_CASE_ :Union[str, Any] = job['completed_at'] SCREAMING_SNAKE_CASE_ :Any = date_parser.parse(SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ :int = date_parser.parse(SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ :Dict = round((end_datetime - start_datetime).total_seconds() / 6_0.0 ) SCREAMING_SNAKE_CASE_ :str = start SCREAMING_SNAKE_CASE_ :Optional[int] = end SCREAMING_SNAKE_CASE_ :List[str] = duration_in_min return job_info def SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None ): SCREAMING_SNAKE_CASE_ :List[str] = None if token is not None: SCREAMING_SNAKE_CASE_ :List[str] = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} SCREAMING_SNAKE_CASE_ :Any = F'https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100' SCREAMING_SNAKE_CASE_ :Any = requests.get(SCREAMING_SNAKE_CASE , headers=SCREAMING_SNAKE_CASE ).json() SCREAMING_SNAKE_CASE_ :Dict = {} try: job_time.update({job['name']: extract_time_from_single_job(SCREAMING_SNAKE_CASE ) for job in result['jobs']} ) SCREAMING_SNAKE_CASE_ :Any = math.ceil((result['total_count'] - 100) / 100 ) for i in range(SCREAMING_SNAKE_CASE ): SCREAMING_SNAKE_CASE_ :List[Any] = requests.get(url + F'&page={i + 2}' , headers=SCREAMING_SNAKE_CASE ).json() job_time.update({job['name']: extract_time_from_single_job(SCREAMING_SNAKE_CASE ) for job in result['jobs']} ) return job_time except Exception: print(F'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument("""--workflow_run_id""", type=str, required=True, help="""A GitHub Actions workflow run id.""") SCREAMING_SNAKE_CASE__ : Optional[Any] = parser.parse_args() SCREAMING_SNAKE_CASE__ : Union[str, Any] = get_job_time(args.workflow_run_id) SCREAMING_SNAKE_CASE__ : List[str] = dict(sorted(job_time.items(), key=lambda item: item[1]["duration"], reverse=True)) for k, v in job_time.items(): print(f"""{k}: {v["duration"]}""")
233
'''simple docstring''' from math import sqrt def SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE = 100_0000 ): SCREAMING_SNAKE_CASE_ :int = 0 SCREAMING_SNAKE_CASE_ :int = 0 SCREAMING_SNAKE_CASE_ :int while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(SCREAMING_SNAKE_CASE , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f"""{solution() = }""")
233
1
'''simple docstring''' from __future__ import annotations def __lowercase ( __lowercase ) -> int: '''simple docstring''' _A = len(UpperCamelCase__ ) // 2 # choose the middle 3 elements _A = lst[m - 1 : m + 2] # if middle element is peak if three[1] > three[0] and three[1] > three[2]: return three[1] # if increasing, recurse on right elif three[0] < three[2]: if len(lst[:m] ) == 2: m -= 1 return peak(lst[m:] ) # decreasing else: if len(lst[:m] ) == 2: m += 1 return peak(lst[:m] ) if __name__ == "__main__": import doctest doctest.testmod()
330
'''simple docstring''' def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> int: def count_of_possible_combinations(UpperCamelCase__ ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item ) for item in array ) return count_of_possible_combinations(UpperCamelCase__ ) def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> int: def count_of_possible_combinations_with_dp_array( UpperCamelCase__ , UpperCamelCase__ ) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] __lowerCamelCase = sum( count_of_possible_combinations_with_dp_array(target - item , UpperCamelCase__ ) for item in array ) __lowerCamelCase = answer return answer __lowerCamelCase = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(UpperCamelCase__ , UpperCamelCase__ ) def __lowerCAmelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> int: __lowerCamelCase = [0] * (target + 1) __lowerCamelCase = 1 for i in range(1 , target + 1 ): for j in range(UpperCamelCase__ ): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() __UpperCAmelCase =3 __UpperCAmelCase =5 __UpperCAmelCase =[1, 2, 5] print(combination_sum_iv(n, array, target))
546
0
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_xlnet import XLNetTokenizer else: _lowercase = None _lowercase = logging.get_logger(__name__) _lowercase = {'''vocab_file''': '''spiece.model''', '''tokenizer_file''': '''tokenizer.json'''} _lowercase = { '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', }, '''tokenizer_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/tokenizer.json''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/tokenizer.json''', }, } _lowercase = { '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } _lowercase = '''▁''' # Segments (not really needed) _lowercase = 0 _lowercase = 1 _lowercase = 2 _lowercase = 3 _lowercase = 4 class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = 'left' UpperCamelCase_ = XLNetTokenizer def __init__( self : Dict ,lowerCAmelCase__ : int=None ,lowerCAmelCase__ : int=None ,lowerCAmelCase__ : int=False ,lowerCAmelCase__ : List[Any]=True ,lowerCAmelCase__ : int=False ,lowerCAmelCase__ : str="<s>" ,lowerCAmelCase__ : List[str]="</s>" ,lowerCAmelCase__ : Dict="<unk>" ,lowerCAmelCase__ : Optional[int]="<sep>" ,lowerCAmelCase__ : int="<pad>" ,lowerCAmelCase__ : Optional[Any]="<cls>" ,lowerCAmelCase__ : Union[str, Any]="<mask>" ,lowerCAmelCase__ : Tuple=["<eop>", "<eod>"] ,**lowerCAmelCase__ : Dict ,) -> Tuple: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else mask_token super().__init__( vocab_file=lowerCAmelCase__ ,tokenizer_file=lowerCAmelCase__ ,do_lower_case=lowerCAmelCase__ ,remove_space=lowerCAmelCase__ ,keep_accents=lowerCAmelCase__ ,bos_token=lowerCAmelCase__ ,eos_token=lowerCAmelCase__ ,unk_token=lowerCAmelCase__ ,sep_token=lowerCAmelCase__ ,pad_token=lowerCAmelCase__ ,cls_token=lowerCAmelCase__ ,mask_token=lowerCAmelCase__ ,additional_special_tokens=lowerCAmelCase__ ,**lowerCAmelCase__ ,) lowerCAmelCase_ : Union[str, Any] = 3 lowerCAmelCase_ : List[str] = do_lower_case lowerCAmelCase_ : Tuple = remove_space lowerCAmelCase_ : List[str] = keep_accents lowerCAmelCase_ : Optional[Any] = vocab_file lowerCAmelCase_ : Tuple = False if not self.vocab_file else True def UpperCAmelCase_ ( self : Dict ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : str = [self.sep_token_id] lowerCAmelCase_ : Tuple = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : Any = [self.sep_token_id] lowerCAmelCase_ : Any = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCAmelCase_ : Optional[Any] = os.path.join( lowerCAmelCase__ ,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file ,lowerCAmelCase__ ) return (out_vocab_file,)
683
from __future__ import annotations from collections.abc import Sequence from typing import Literal def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : Optional[int] = list(snake_case__) lowerCAmelCase_ : Tuple = list(snake_case__) lowerCAmelCase_ : List[str] = 0 for i in range(len(snake_case__)): if lista[i] != lista[i]: count += 1 lowerCAmelCase_ : Dict = "_" if count > 1: return False else: return "".join(snake_case__) def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Union[str, Any] = [] while True: lowerCAmelCase_ : Tuple = ["$"] * len(snake_case__) lowerCAmelCase_ : Tuple = [] for i in range(len(snake_case__)): for j in range(i + 1 , len(snake_case__)): lowerCAmelCase_ : Optional[int] = compare_string(binary[i] , binary[j]) if k is False: lowerCAmelCase_ : str = "*" lowerCAmelCase_ : Tuple = "*" temp.append("X") for i in range(len(snake_case__)): if checka[i] == "$": pi.append(binary[i]) if len(snake_case__) == 0: return pi lowerCAmelCase_ : List[Any] = list(set(snake_case__)) def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : Optional[int] = [] for minterm in minterms: lowerCAmelCase_ : Dict = "" for _ in range(snake_case__): lowerCAmelCase_ : Dict = str(minterm % 2) + string minterm //= 2 temp.append(snake_case__) return temp def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : Optional[Any] = list(snake_case__) lowerCAmelCase_ : Dict = list(snake_case__) lowerCAmelCase_ : Dict = 0 for i in range(len(snake_case__)): if lista[i] != lista[i]: count_n += 1 return count_n == count def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : Optional[Any] = [] lowerCAmelCase_ : Dict = [0] * len(snake_case__) for i in range(len(chart[0])): lowerCAmelCase_ : List[Any] = 0 lowerCAmelCase_ : int = -1 for j in range(len(snake_case__)): if chart[j][i] == 1: count += 1 lowerCAmelCase_ : Optional[int] = j if count == 1: lowerCAmelCase_ : Union[str, Any] = 1 for i in range(len(snake_case__)): if select[i] == 1: for j in range(len(chart[0])): if chart[i][j] == 1: for k in range(len(snake_case__)): lowerCAmelCase_ : Tuple = 0 temp.append(prime_implicants[i]) while True: lowerCAmelCase_ : Optional[Any] = 0 lowerCAmelCase_ : Dict = -1 lowerCAmelCase_ : Tuple = 0 for i in range(len(snake_case__)): lowerCAmelCase_ : Dict = chart[i].count(1) if count_n > max_n: lowerCAmelCase_ : Optional[int] = count_n lowerCAmelCase_ : Optional[Any] = i if max_n == 0: return temp temp.append(prime_implicants[rem]) for i in range(len(chart[0])): if chart[rem][i] == 1: for j in range(len(snake_case__)): lowerCAmelCase_ : Any = 0 def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : str = [[0 for x in range(len(snake_case__))] for x in range(len(snake_case__))] for i in range(len(snake_case__)): lowerCAmelCase_ : Optional[Any] = prime_implicants[i].count("_") for j in range(len(snake_case__)): if is_for_table(prime_implicants[i] , binary[j] , snake_case__): lowerCAmelCase_ : Dict = 1 return chart def UpperCamelCase ( ): lowerCAmelCase_ : Optional[Any] = int(input("Enter the no. of variables\n")) lowerCAmelCase_ : Tuple = [ float(snake_case__) for x in input( "Enter the decimal representation of Minterms 'Spaces Separated'\n").split() ] lowerCAmelCase_ : Any = decimal_to_binary(snake_case__ , snake_case__) lowerCAmelCase_ : Dict = check(snake_case__) print("Prime Implicants are:") print(snake_case__) lowerCAmelCase_ : int = prime_implicant_chart(snake_case__ , snake_case__) lowerCAmelCase_ : List[str] = selection(snake_case__ , snake_case__) print("Essential Prime Implicants are:") print(snake_case__) if __name__ == "__main__": import doctest doctest.testmod() main()
683
1
'''simple docstring''' import json import os import sys import tempfile import unittest from pathlib import Path from shutil import copyfile from huggingface_hub import HfFolder, Repository, create_repo, delete_repo from requests.exceptions import HTTPError import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, PROCESSOR_MAPPING, TOKENIZER_MAPPING, AutoConfig, AutoFeatureExtractor, AutoProcessor, AutoTokenizer, BertTokenizer, ProcessorMixin, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaProcessor, ) from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test from transformers.tokenization_utils import TOKENIZER_CONFIG_FILE from transformers.utils import FEATURE_EXTRACTOR_NAME, is_tokenizers_available sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 from test_module.custom_processing import CustomProcessor # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 a : Optional[Any] = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') a : Optional[Any] = get_tests_dir('''fixtures/vocab.json''') a : Optional[int] = get_tests_dir('''fixtures''') class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): __SCREAMING_SNAKE_CASE = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """bla""", """blou"""] def A ( self : Dict ): """simple docstring""" __snake_case = 0 def A ( self : Optional[int] ): """simple docstring""" __snake_case = AutoProcessor.from_pretrained("facebook/wav2vec2-base-960h" ) self.assertIsInstance(a_ , a_ ) def A ( self : Optional[Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: __snake_case = WavaVecaConfig() __snake_case = AutoProcessor.from_pretrained("facebook/wav2vec2-base-960h" ) # save in new folder model_config.save_pretrained(a_ ) processor.save_pretrained(a_ ) __snake_case = AutoProcessor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : Union[str, Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # copy relevant files copyfile(a_ , os.path.join(a_ , a_ ) ) copyfile(a_ , os.path.join(a_ , "vocab.json" ) ) __snake_case = AutoProcessor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : List[str] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: __snake_case = WavaVecaFeatureExtractor() __snake_case = AutoTokenizer.from_pretrained("facebook/wav2vec2-base-960h" ) __snake_case = WavaVecaProcessor(a_ , a_ ) # save in new folder processor.save_pretrained(a_ ) # drop `processor_class` in tokenizer with open(os.path.join(a_ , a_ ) , "r" ) as f: __snake_case = json.load(a_ ) config_dict.pop("processor_class" ) with open(os.path.join(a_ , a_ ) , "w" ) as f: f.write(json.dumps(a_ ) ) __snake_case = AutoProcessor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : Optional[Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: __snake_case = WavaVecaFeatureExtractor() __snake_case = AutoTokenizer.from_pretrained("facebook/wav2vec2-base-960h" ) __snake_case = WavaVecaProcessor(a_ , a_ ) # save in new folder processor.save_pretrained(a_ ) # drop `processor_class` in feature extractor with open(os.path.join(a_ , a_ ) , "r" ) as f: __snake_case = json.load(a_ ) config_dict.pop("processor_class" ) with open(os.path.join(a_ , a_ ) , "w" ) as f: f.write(json.dumps(a_ ) ) __snake_case = AutoProcessor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : Optional[Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: __snake_case = WavaVecaConfig(processor_class="Wav2Vec2Processor" ) model_config.save_pretrained(a_ ) # copy relevant files copyfile(a_ , os.path.join(a_ , "vocab.json" ) ) # create emtpy sample processor with open(os.path.join(a_ , a_ ) , "w" ) as f: f.write("{}" ) __snake_case = AutoProcessor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : List[Any] ): """simple docstring""" with self.assertRaises(a_ ): __snake_case = AutoProcessor.from_pretrained("hf-internal-testing/test_dynamic_processor" ) # If remote code is disabled, we can't load this config. with self.assertRaises(a_ ): __snake_case = AutoProcessor.from_pretrained( "hf-internal-testing/test_dynamic_processor" , trust_remote_code=a_ ) __snake_case = AutoProcessor.from_pretrained("hf-internal-testing/test_dynamic_processor" , trust_remote_code=a_ ) self.assertTrue(processor.special_attribute_present ) self.assertEqual(processor.__class__.__name__ , "NewProcessor" ) __snake_case = processor.feature_extractor self.assertTrue(feature_extractor.special_attribute_present ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) __snake_case = processor.tokenizer self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizerFast" ) # Test we can also load the slow version __snake_case = AutoProcessor.from_pretrained( "hf-internal-testing/test_dynamic_processor" , trust_remote_code=a_ , use_fast=a_ ) __snake_case = new_processor.tokenizer self.assertTrue(new_tokenizer.special_attribute_present ) self.assertEqual(new_tokenizer.__class__.__name__ , "NewTokenizer" ) else: self.assertEqual(tokenizer.__class__.__name__ , "NewTokenizer" ) def A ( self : List[str] ): """simple docstring""" try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) AutoTokenizer.register(a_ , slow_tokenizer_class=a_ ) AutoProcessor.register(a_ , a_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a_ ): AutoProcessor.register(a_ , a_ ) # Now that the config is registered, it can be used as any other config with the auto-API __snake_case = CustomFeatureExtractor.from_pretrained(a_ ) with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = CustomTokenizer(a_ ) __snake_case = CustomProcessor(a_ , a_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained(a_ ) __snake_case = AutoProcessor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def A ( self : Union[str, Any] ): """simple docstring""" class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = False class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = False class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """AutoFeatureExtractor""" __SCREAMING_SNAKE_CASE = """AutoTokenizer""" __SCREAMING_SNAKE_CASE = False try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) AutoTokenizer.register(a_ , slow_tokenizer_class=a_ ) AutoProcessor.register(a_ , a_ ) # If remote code is not set, the default is to use local classes. __snake_case = AutoProcessor.from_pretrained("hf-internal-testing/test_dynamic_processor" ) self.assertEqual(processor.__class__.__name__ , "NewProcessor" ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote code is disabled, we load the local ones. __snake_case = AutoProcessor.from_pretrained( "hf-internal-testing/test_dynamic_processor" , trust_remote_code=a_ ) self.assertEqual(processor.__class__.__name__ , "NewProcessor" ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub. __snake_case = AutoProcessor.from_pretrained( "hf-internal-testing/test_dynamic_processor" , trust_remote_code=a_ ) self.assertEqual(processor.__class__.__name__ , "NewProcessor" ) self.assertTrue(processor.special_attribute_present ) self.assertTrue(processor.feature_extractor.special_attribute_present ) self.assertTrue(processor.tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = AutoProcessor.from_pretrained("hf-internal-testing/tiny-random-bert" ) self.assertEqual(processor.__class__.__name__ , "BertTokenizerFast" ) def A ( self : str ): """simple docstring""" __snake_case = AutoProcessor.from_pretrained("hf-internal-testing/tiny-random-convnext" ) self.assertEqual(processor.__class__.__name__ , "ConvNextImageProcessor" ) @is_staging_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): __SCREAMING_SNAKE_CASE = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """bla""", """blou"""] @classmethod def A ( cls : Optional[Any] ): """simple docstring""" __snake_case = TOKEN HfFolder.save_token(a_ ) @classmethod def A ( cls : str ): """simple docstring""" try: delete_repo(token=cls._token , repo_id="test-processor" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-processor-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-processor" ) except HTTPError: pass def A ( self : List[str] ): """simple docstring""" __snake_case = WavaVecaProcessor.from_pretrained(a_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(a_ , "test-processor" ) , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = WavaVecaProcessor.from_pretrained(f'''{USER}/test-processor''' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(a_ , getattr(new_processor.feature_extractor , a_ ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def A ( self : Tuple ): """simple docstring""" __snake_case = WavaVecaProcessor.from_pretrained(a_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(a_ , "test-processor-org" ) , push_to_hub=a_ , use_auth_token=self._token , organization="valid_org" , ) __snake_case = WavaVecaProcessor.from_pretrained("valid_org/test-processor-org" ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(a_ , getattr(new_processor.feature_extractor , a_ ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def A ( self : Any ): """simple docstring""" CustomFeatureExtractor.register_for_auto_class() CustomTokenizer.register_for_auto_class() CustomProcessor.register_for_auto_class() __snake_case = CustomFeatureExtractor.from_pretrained(a_ ) with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = CustomTokenizer(a_ ) __snake_case = CustomProcessor(a_ , a_ ) with tempfile.TemporaryDirectory() as tmp_dir: create_repo(f'''{USER}/test-dynamic-processor''' , token=self._token ) __snake_case = Repository(a_ , clone_from=f'''{USER}/test-dynamic-processor''' , token=self._token ) processor.save_pretrained(a_ ) # This has added the proper auto_map field to the feature extractor config self.assertDictEqual( processor.feature_extractor.auto_map , { "AutoFeatureExtractor": "custom_feature_extraction.CustomFeatureExtractor", "AutoProcessor": "custom_processing.CustomProcessor", } , ) # This has added the proper auto_map field to the tokenizer config with open(os.path.join(a_ , "tokenizer_config.json" ) ) as f: __snake_case = json.load(a_ ) self.assertDictEqual( tokenizer_config["auto_map"] , { "AutoTokenizer": ["custom_tokenization.CustomTokenizer", None], "AutoProcessor": "custom_processing.CustomProcessor", } , ) # The code has been copied from fixtures self.assertTrue(os.path.isfile(os.path.join(a_ , "custom_feature_extraction.py" ) ) ) self.assertTrue(os.path.isfile(os.path.join(a_ , "custom_tokenization.py" ) ) ) self.assertTrue(os.path.isfile(os.path.join(a_ , "custom_processing.py" ) ) ) repo.push_to_hub() __snake_case = AutoProcessor.from_pretrained(f'''{USER}/test-dynamic-processor''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_processor is from the CustomProcessor class of a dynamic module self.assertEqual(new_processor.__class__.__name__ , "CustomProcessor" )
69
"""simple docstring""" import re def lowerCamelCase_( _lowerCamelCase ) -> str: '''simple docstring''' if len(re.findall("[ATCG]" , _lowerCamelCase ) ) != len(_lowerCamelCase ): raise ValueError("Invalid Strand" ) return dna.translate(dna.maketrans("ATCG" , "TAGC" ) ) if __name__ == "__main__": import doctest doctest.testmod()
46
0
import torch import torch.nn as nn from transformers.modeling_utils import ModuleUtilsMixin from transformers.models.ta.modeling_ta import TaBlock, TaConfig, TaLayerNorm from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class UpperCamelCase__ ( __magic_name__ , __magic_name__ , __magic_name__ ): @register_to_config def __init__( self : int , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : float , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : str , UpperCamelCase__ : bool = False , ): '''simple docstring''' super().__init__() lowercase_ = nn.Embedding(UpperCamelCase__ , UpperCamelCase__ ) lowercase_ = nn.Embedding(UpperCamelCase__ , UpperCamelCase__ ) lowercase_ = False lowercase_ = nn.Dropout(p=UpperCamelCase__ ) lowercase_ = TaConfig( vocab_size=UpperCamelCase__ , d_model=UpperCamelCase__ , num_heads=UpperCamelCase__ , d_kv=UpperCamelCase__ , d_ff=UpperCamelCase__ , dropout_rate=UpperCamelCase__ , feed_forward_proj=UpperCamelCase__ , is_decoder=UpperCamelCase__ , is_encoder_decoder=UpperCamelCase__ , ) lowercase_ = nn.ModuleList() for lyr_num in range(UpperCamelCase__ ): lowercase_ = TaBlock(UpperCamelCase__ ) self.encoders.append(UpperCamelCase__ ) lowercase_ = TaLayerNorm(UpperCamelCase__ ) lowercase_ = nn.Dropout(p=UpperCamelCase__ ) def UpperCAmelCase__ ( self : Optional[Any] , UpperCamelCase__ : Optional[int] , UpperCamelCase__ : str ): '''simple docstring''' lowercase_ = self.token_embedder(UpperCamelCase__ ) lowercase_ = encoder_input_tokens.shape[1] lowercase_ = torch.arange(UpperCamelCase__ , device=encoder_input_tokens.device ) x += self.position_encoding(UpperCamelCase__ ) lowercase_ = self.dropout_pre(UpperCamelCase__ ) # inverted the attention mask lowercase_ = encoder_input_tokens.size() lowercase_ = self.get_extended_attention_mask(UpperCamelCase__ , UpperCamelCase__ ) for lyr in self.encoders: lowercase_ = lyr(UpperCamelCase__ , UpperCamelCase__ )[0] lowercase_ = self.layer_norm(UpperCamelCase__ ) return self.dropout_post(UpperCamelCase__ ), encoder_inputs_mask
650
import os import re import warnings from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_ta import TaTokenizer else: a = None a = logging.get_logger(__name__) a = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} a = { 'vocab_file': { 't5-small': 'https://huggingface.co/t5-small/resolve/main/spiece.model', 't5-base': 'https://huggingface.co/t5-base/resolve/main/spiece.model', 't5-large': 'https://huggingface.co/t5-large/resolve/main/spiece.model', 't5-3b': 'https://huggingface.co/t5-3b/resolve/main/spiece.model', 't5-11b': 'https://huggingface.co/t5-11b/resolve/main/spiece.model', }, 'tokenizer_file': { 't5-small': 'https://huggingface.co/t5-small/resolve/main/tokenizer.json', 't5-base': 'https://huggingface.co/t5-base/resolve/main/tokenizer.json', 't5-large': 'https://huggingface.co/t5-large/resolve/main/tokenizer.json', 't5-3b': 'https://huggingface.co/t5-3b/resolve/main/tokenizer.json', 't5-11b': 'https://huggingface.co/t5-11b/resolve/main/tokenizer.json', }, } # TODO(PVP) - this should be removed in Transformers v5 a = { 't5-small': 5_1_2, 't5-base': 5_1_2, 't5-large': 5_1_2, 't5-3b': 5_1_2, 't5-11b': 5_1_2, } class UpperCamelCase__ ( __magic_name__ ): __SCREAMING_SNAKE_CASE : Union[str, Any] = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE : int = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __SCREAMING_SNAKE_CASE : str = ['input_ids', 'attention_mask'] __SCREAMING_SNAKE_CASE : Dict = TaTokenizer __SCREAMING_SNAKE_CASE : List[int] = [] def __init__( self : int , UpperCamelCase__ : Union[str, Any]=None , UpperCamelCase__ : Any=None , UpperCamelCase__ : Dict="</s>" , UpperCamelCase__ : Tuple="<unk>" , UpperCamelCase__ : Optional[Any]="<pad>" , UpperCamelCase__ : Union[str, Any]=100 , UpperCamelCase__ : Optional[Any]=None , **UpperCamelCase__ : List[str] , ): '''simple docstring''' if extra_ids > 0 and additional_special_tokens is None: lowercase_ = [F'''<extra_id_{i}>''' for i in range(UpperCamelCase__ )] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra special tokens lowercase_ = len(set(filter(lambda UpperCamelCase__ : bool("""extra_id_""" in str(UpperCamelCase__ ) ) , UpperCamelCase__ ) ) ) if extra_tokens != extra_ids: raise ValueError( F'''Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are''' """ provided to T5Tokenizer. In this case the additional_special_tokens must include the extra_ids""" """ tokens""" ) super().__init__( UpperCamelCase__ , tokenizer_file=UpperCamelCase__ , eos_token=UpperCamelCase__ , unk_token=UpperCamelCase__ , pad_token=UpperCamelCase__ , extra_ids=UpperCamelCase__ , additional_special_tokens=UpperCamelCase__ , **UpperCamelCase__ , ) lowercase_ = vocab_file lowercase_ = False if not self.vocab_file else True lowercase_ = extra_ids @staticmethod def UpperCAmelCase__ ( UpperCamelCase__ : Tuple , UpperCamelCase__ : str , UpperCamelCase__ : int ): '''simple docstring''' if pretrained_model_name_or_path in TaTokenizerFast.max_model_input_sizes: lowercase_ = TaTokenizerFast.max_model_input_sizes[pretrained_model_name_or_path] if init_max_model_length is not None and init_max_model_length != max_model_length: return init_max_model_length elif init_max_model_length is None: warnings.warn( """This tokenizer was incorrectly instantiated with a model max length of""" F''' {deprecated_max_model_length} which will be corrected in Transformers v5.\nFor now, this''' """ behavior is kept to avoid breaking backwards compatibility when padding/encoding with""" """ `truncation is True`.\n- Be aware that you SHOULD NOT rely on""" F''' {pretrained_model_name_or_path} automatically truncating your input to''' F''' {deprecated_max_model_length} when padding/encoding.\n- If you want to encode/pad to sequences''' F''' longer than {deprecated_max_model_length} you can either instantiate this tokenizer with''' """ `model_max_length` or pass `max_length` when encoding/padding.\n- To avoid this warning, please""" """ instantiate this tokenizer with `model_max_length` set to your preferred value.""" , UpperCamelCase__ , ) return max_model_length def UpperCAmelCase__ ( self : int , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None ): '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(UpperCamelCase__ ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase_ = os.path.join( UpperCamelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ): copyfile(self.vocab_file , UpperCamelCase__ ) logger.info(F'''Copy vocab file to {out_vocab_file}''' ) return (out_vocab_file,) def UpperCAmelCase__ ( self : int , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' lowercase_ = token_ids_a + [self.eos_token_id] if token_ids_a is None: return self.prefix_tokens + token_ids_a else: lowercase_ = token_ids_a + [self.eos_token_id] return self.prefix_tokens + token_ids_a + token_ids_a def UpperCAmelCase__ ( self : str , UpperCamelCase__ : List[int] , UpperCamelCase__ : Optional[List[int]] = None ): '''simple docstring''' lowercase_ = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos ) * [0] return len(token_ids_a + eos + token_ids_a + eos ) * [0] def UpperCAmelCase__ ( self : Dict ): '''simple docstring''' return list( set(filter(lambda UpperCamelCase__ : bool(re.search(R"""<extra_id_\d+>""" , UpperCamelCase__ ) ) is not None , self.additional_special_tokens ) ) ) def UpperCAmelCase__ ( self : str ): '''simple docstring''' return [self.convert_tokens_to_ids(UpperCamelCase__ ) for token in self.get_sentinel_tokens()]
650
1
from itertools import permutations def _a ( __UpperCamelCase : tuple ): if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False lowerCAmelCase__ : List[str] = [7, 11, 13, 17] for i, test in enumerate(__UpperCamelCase ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def _a ( __UpperCamelCase : int = 10 ): return sum( int(''''''.join(map(__UpperCamelCase ,__UpperCamelCase ) ) ) for num in permutations(range(__UpperCamelCase ) ) if is_substring_divisible(__UpperCamelCase ) ) if __name__ == "__main__": print(f"""{solution() = }""")
233
from __future__ import annotations from sys import maxsize from typing import Generic, TypeVar A__ : Dict = TypeVar("""T""") def _a ( __UpperCamelCase : int ): return (position - 1) // 2 def _a ( __UpperCamelCase : int ): return (2 * position) + 1 def _a ( __UpperCamelCase : int ): return (2 * position) + 2 class lowercase ( Generic[T] ): def __init__( self ): """simple docstring""" lowerCAmelCase__ : list[tuple[T, int]] = [] lowerCAmelCase__ : dict[T, int] = {} lowerCAmelCase__ : int = 0 def __len__( self ): """simple docstring""" return self.elements def __repr__( self ): """simple docstring""" return str(self.heap ) def lowercase_ ( self ): """simple docstring""" return self.elements == 0 def lowercase_ ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): """simple docstring""" self.heap.append((elem, weight) ) lowerCAmelCase__ : Dict = self.elements self.elements += 1 self._bubble_up(SCREAMING_SNAKE_CASE__ ) def lowercase_ ( self ): """simple docstring""" if self.elements > 1: self._swap_nodes(0 , self.elements - 1 ) lowerCAmelCase__ , lowerCAmelCase__ : List[str] = self.heap.pop() del self.position_map[elem] self.elements -= 1 if self.elements > 0: lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = self.heap[0] self._bubble_down(SCREAMING_SNAKE_CASE__ ) return elem def lowercase_ ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCAmelCase__ : Optional[Any] = self.position_map[elem] lowerCAmelCase__ : str = (elem, weight) if position > 0: lowerCAmelCase__ : Any = get_parent_position(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.heap[parent_position] if parent_weight > weight: self._bubble_up(SCREAMING_SNAKE_CASE__ ) else: self._bubble_down(SCREAMING_SNAKE_CASE__ ) else: self._bubble_down(SCREAMING_SNAKE_CASE__ ) def lowercase_ ( self , SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCAmelCase__ : Union[str, Any] = self.position_map[elem] if curr_pos == 0: return None lowerCAmelCase__ : Optional[int] = get_parent_position(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ , lowerCAmelCase__ : List[str] = self.heap[curr_pos] lowerCAmelCase__ , lowerCAmelCase__ : Union[str, Any] = self.heap[parent_position] if parent_weight > weight: self._swap_nodes(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return self._bubble_up(SCREAMING_SNAKE_CASE__ ) return None def lowercase_ ( self , SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCAmelCase__ : Optional[int] = self.position_map[elem] lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = self.heap[curr_pos] lowerCAmelCase__ : Tuple = get_child_left_position(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ : Any = get_child_right_position(SCREAMING_SNAKE_CASE__ ) if child_left_position < self.elements and child_right_position < self.elements: lowerCAmelCase__ , lowerCAmelCase__ : Dict = self.heap[child_left_position] lowerCAmelCase__ , lowerCAmelCase__ : Any = self.heap[child_right_position] if child_right_weight < child_left_weight and child_right_weight < weight: self._swap_nodes(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return self._bubble_down(SCREAMING_SNAKE_CASE__ ) if child_left_position < self.elements: lowerCAmelCase__ , lowerCAmelCase__ : Any = self.heap[child_left_position] if child_left_weight < weight: self._swap_nodes(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return self._bubble_down(SCREAMING_SNAKE_CASE__ ) else: return None if child_right_position < self.elements: lowerCAmelCase__ , lowerCAmelCase__ : Optional[Any] = self.heap[child_right_position] if child_right_weight < weight: self._swap_nodes(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) return self._bubble_down(SCREAMING_SNAKE_CASE__ ) return None def lowercase_ ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): """simple docstring""" lowerCAmelCase__ : Dict = self.heap[nodea_pos][0] lowerCAmelCase__ : Dict = self.heap[nodea_pos][0] lowerCAmelCase__ , lowerCAmelCase__ : List[Any] = ( self.heap[nodea_pos], self.heap[nodea_pos], ) lowerCAmelCase__ : Tuple = nodea_pos lowerCAmelCase__ : int = nodea_pos class lowercase ( Generic[T] ): def __init__( self ): """simple docstring""" lowerCAmelCase__ : dict[T, dict[T, int]] = {} lowerCAmelCase__ : int = 0 def __repr__( self ): """simple docstring""" return str(self.connections ) def __len__( self ): """simple docstring""" return self.nodes def lowercase_ ( self , SCREAMING_SNAKE_CASE__ ): """simple docstring""" if node not in self.connections: lowerCAmelCase__ : Union[str, Any] = {} self.nodes += 1 def lowercase_ ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): """simple docstring""" self.add_node(SCREAMING_SNAKE_CASE__ ) self.add_node(SCREAMING_SNAKE_CASE__ ) lowerCAmelCase__ : Tuple = weight lowerCAmelCase__ : Tuple = weight def _a ( __UpperCamelCase : GraphUndirectedWeighted[T] ,): lowerCAmelCase__ : dict[T, int] = {node: maxsize for node in graph.connections} lowerCAmelCase__ : dict[T, T | None] = {node: None for node in graph.connections} lowerCAmelCase__ : MinPriorityQueue[T] = MinPriorityQueue() for node, weight in dist.items(): priority_queue.push(__UpperCamelCase ,__UpperCamelCase ) if priority_queue.is_empty(): return dist, parent # initialization lowerCAmelCase__ : List[Any] = priority_queue.extract_min() lowerCAmelCase__ : str = 0 for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: lowerCAmelCase__ : List[str] = dist[node] + graph.connections[node][neighbour] priority_queue.update_key(__UpperCamelCase ,dist[neighbour] ) lowerCAmelCase__ : Optional[Any] = node # running prim's algorithm while not priority_queue.is_empty(): lowerCAmelCase__ : List[str] = priority_queue.extract_min() for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: lowerCAmelCase__ : Optional[Any] = dist[node] + graph.connections[node][neighbour] priority_queue.update_key(__UpperCamelCase ,dist[neighbour] ) lowerCAmelCase__ : Optional[int] = node return dist, parent
233
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor _A = logging.get_logger(__name__) class _lowerCamelCase ( a_ ): def __init__( self : str , *UpperCamelCase : Dict , **UpperCamelCase : Union[str, Any] ) -> None: """simple docstring""" warnings.warn( """The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use PerceiverImageProcessor instead.""" , UpperCamelCase , ) super().__init__(*UpperCamelCase , **UpperCamelCase )
708
"""simple docstring""" def lowercase_ ( __UpperCAmelCase = 100_0000 ) -> int: lowerCAmelCase__ : int = [i - 1 for i in range(limit + 1 )] for i in range(2 , limit + 1 ): if phi[i] == i - 1: for j in range(2 * i , limit + 1 , __UpperCAmelCase ): phi[j] -= phi[j] // i return sum(phi[2 : limit + 1] ) if __name__ == "__main__": print(solution())
507
0
import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class A__ ( unittest.TestCase ): """simple docstring""" def __init__( self : Tuple , lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Union[str, Any]=13 , lowerCamelCase__ : str=7 , lowerCamelCase__ : List[str]=True , lowerCamelCase__ : Optional[int]=True , lowerCamelCase__ : Union[str, Any]=True , lowerCamelCase__ : List[Any]=True , lowerCamelCase__ : Tuple=99 , lowerCamelCase__ : int=32 , lowerCamelCase__ : List[Any]=5 , lowerCamelCase__ : str=4 , lowerCamelCase__ : str=37 , lowerCamelCase__ : Dict="gelu" , lowerCamelCase__ : Tuple=0.1 , lowerCamelCase__ : Tuple=0.1 , lowerCamelCase__ : Tuple=512 , lowerCamelCase__ : Optional[int]=16 , lowerCamelCase__ : List[str]=2 , lowerCamelCase__ : Union[str, Any]=0.02 , lowerCamelCase__ : List[Any]=4 , ): a__ : Optional[Any] = parent a__ : List[str] = batch_size a__ : str = seq_length a__ : Optional[int] = is_training a__ : Optional[Any] = use_attention_mask a__ : str = use_token_type_ids a__ : Optional[Any] = use_labels a__ : str = vocab_size a__ : Optional[Any] = hidden_size a__ : str = num_hidden_layers a__ : int = num_attention_heads a__ : Tuple = intermediate_size a__ : List[Any] = hidden_act a__ : Optional[int] = hidden_dropout_prob a__ : Any = attention_probs_dropout_prob a__ : str = max_position_embeddings a__ : Optional[int] = type_vocab_size a__ : Tuple = type_sequence_label_size a__ : Optional[int] = initializer_range a__ : List[Any] = num_choices def _UpperCamelCase( self : Optional[int] ): a__ : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a__ : int = None if self.use_attention_mask: a__ : str = random_attention_mask([self.batch_size, self.seq_length] ) a__ : Dict = None if self.use_token_type_ids: a__ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) a__ : int = RobertaPreLayerNormConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _UpperCamelCase( self : Tuple ): a__ : str = self.prepare_config_and_inputs() a__, a__, a__, a__ : Optional[Any] = config_and_inputs a__ : Optional[int] = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict def _UpperCamelCase( self : str ): a__ : Optional[Any] = self.prepare_config_and_inputs() a__, a__, a__, a__ : Union[str, Any] = config_and_inputs a__ : Optional[int] = True a__ : Dict = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) a__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class A__ ( A__ , unittest.TestCase ): """simple docstring""" _lowercase = True _lowercase = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def _UpperCamelCase( self : List[str] ): a__ : Dict = FlaxRobertaPreLayerNormModelTester(self ) @slow def _UpperCamelCase( self : Dict ): for model_class_name in self.all_model_classes: a__ : List[str] = model_class_name.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=lowerCamelCase__ ) a__ : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class A__ ( unittest.TestCase ): """simple docstring""" @slow def _UpperCamelCase( self : Dict ): a__ : Any = FlaxRobertaPreLayerNormForMaskedLM.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=lowerCamelCase__ ) a__ : str = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) a__ : int = model(lowerCamelCase__ )[0] a__ : int = [1, 11, 50_265] self.assertEqual(list(output.shape ) , lowerCamelCase__ ) # compare the actual values for a slice. a__ : str = np.array( [[[40.4880, 18.0199, -5.2367], [-1.8877, -4.0885, 10.7085], [-2.2613, -5.6110, 7.2665]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , lowerCamelCase__ , atol=1E-4 ) ) @slow def _UpperCamelCase( self : Optional[Any] ): a__ : int = FlaxRobertaPreLayerNormModel.from_pretrained("andreasmadsen/efficient_mlm_m0.40" , from_pt=lowerCamelCase__ ) a__ : Any = np.array([[0, 31_414, 232, 328, 740, 1_140, 12_695, 69, 46_078, 1_588, 2]] , dtype=jnp.intaa ) a__ : str = model(lowerCamelCase__ )[0] # compare the actual values for a slice. a__ : Optional[Any] = np.array( [[[0.0208, -0.0356, 0.0237], [-0.1569, -0.0411, -0.2626], [0.1879, 0.0125, -0.0089]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , lowerCamelCase__ , atol=1E-4 ) )
37
import unittest from datasets import load_dataset from transformers import BloomTokenizerFast from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCamelCase_ ( lowercase , unittest.TestCase ): __lowercase : Optional[Any] = None __lowercase : Tuple = BloomTokenizerFast __lowercase : Tuple = BloomTokenizerFast __lowercase : int = True __lowercase : str = False __lowercase : int = "tokenizer_file" __lowercase : List[str] = {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "<unk>", "pad_token": "<pad>"} def lowercase ( self ) -> str: """simple docstring""" super().setUp() _UpperCamelCase = BloomTokenizerFast.from_pretrained("bigscience/tokenizer" ) tokenizer.save_pretrained(self.tmpdirname ) def lowercase ( self , **lowerCamelCase_ ) -> Any: """simple docstring""" kwargs.update(self.special_tokens_map ) return BloomTokenizerFast.from_pretrained(self.tmpdirname , **lowerCamelCase_ ) def lowercase ( self ) -> Optional[int]: """simple docstring""" _UpperCamelCase = self.get_rust_tokenizer() _UpperCamelCase = ["The quick brown fox</s>", "jumps over the lazy dog</s>"] _UpperCamelCase = [[21_75, 2_37_14, 7_31_73, 14_42_52, 2], [77, 13_26_19, 34_78, 3_68, 10_95_86, 3_54_33, 2]] _UpperCamelCase = tokenizer.batch_encode_plus(lowerCamelCase_ )["input_ids"] self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) _UpperCamelCase = tokenizer.batch_decode(lowerCamelCase_ ) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) def lowercase ( self , lowerCamelCase_=6 ) -> Dict: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): _UpperCamelCase = self.rust_tokenizer_class.from_pretrained(lowerCamelCase_ , **lowerCamelCase_ ) # tokenizer_r.pad_token = None # Hotfixing padding = None # Simple input _UpperCamelCase = "This is a simple input" _UpperCamelCase = ["This is a simple input 1", "This is a simple input 2"] _UpperCamelCase = ("This is a simple input", "This is a pair") _UpperCamelCase = [ ("This is a simple input 1", "This is a simple input 2"), ("This is a simple pair 1", "This is a simple pair 2"), ] # Simple input tests try: tokenizer_r.encode(lowerCamelCase_ , max_length=lowerCamelCase_ ) tokenizer_r.encode_plus(lowerCamelCase_ , max_length=lowerCamelCase_ ) tokenizer_r.batch_encode_plus(lowerCamelCase_ , max_length=lowerCamelCase_ ) tokenizer_r.encode(lowerCamelCase_ , max_length=lowerCamelCase_ ) tokenizer_r.batch_encode_plus(lowerCamelCase_ , max_length=lowerCamelCase_ ) except ValueError: self.fail("Bloom Tokenizer should be able to deal with padding" ) _UpperCamelCase = None # Hotfixing padding = None self.assertRaises(lowerCamelCase_ , tokenizer_r.encode , lowerCamelCase_ , max_length=lowerCamelCase_ , padding="max_length" ) # Simple input self.assertRaises(lowerCamelCase_ , tokenizer_r.encode_plus , lowerCamelCase_ , max_length=lowerCamelCase_ , padding="max_length" ) # Simple input self.assertRaises( lowerCamelCase_ , tokenizer_r.batch_encode_plus , lowerCamelCase_ , max_length=lowerCamelCase_ , padding="max_length" , ) # Pair input self.assertRaises(lowerCamelCase_ , tokenizer_r.encode , lowerCamelCase_ , max_length=lowerCamelCase_ , padding="max_length" ) # Pair input self.assertRaises(lowerCamelCase_ , tokenizer_r.encode_plus , lowerCamelCase_ , max_length=lowerCamelCase_ , padding="max_length" ) # Pair input self.assertRaises( lowerCamelCase_ , tokenizer_r.batch_encode_plus , lowerCamelCase_ , max_length=lowerCamelCase_ , padding="max_length" , ) def lowercase ( self ) -> List[str]: """simple docstring""" _UpperCamelCase = self.get_rust_tokenizer() _UpperCamelCase = load_dataset("xnli" , "all_languages" , split="test" , streaming=lowerCamelCase_ ) _UpperCamelCase = next(iter(lowerCamelCase_ ) )["premise"] # pick up one data _UpperCamelCase = list(sample_data.values() ) _UpperCamelCase = list(map(tokenizer.encode , lowerCamelCase_ ) ) _UpperCamelCase = [tokenizer.decode(lowerCamelCase_ , clean_up_tokenization_spaces=lowerCamelCase_ ) for x in output_tokens] self.assertListEqual(lowerCamelCase_ , lowerCamelCase_ ) def lowercase ( self ) -> Dict: """simple docstring""" self.assertGreaterEqual(len(self.tokenizer_class.pretrained_vocab_files_map ) , 1 ) self.assertGreaterEqual(len(list(self.tokenizer_class.pretrained_vocab_files_map.values() )[0] ) , 1 )
147
0
from ..utils import DummyObject, requires_backends class A__ ( metaclass=__lowerCamelCase): A_ : Any = ['flax'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) class A__ ( metaclass=__lowerCamelCase): A_ : Optional[Any] = ['flax'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) class A__ ( metaclass=__lowerCamelCase): A_ : Optional[int] = ['flax'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) class A__ ( metaclass=__lowerCamelCase): A_ : str = ['flax'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) class A__ ( metaclass=__lowerCamelCase): A_ : List[str] = ['flax'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) class A__ ( metaclass=__lowerCamelCase): A_ : str = ['flax'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) class A__ ( metaclass=__lowerCamelCase): A_ : Tuple = ['flax'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) class A__ ( metaclass=__lowerCamelCase): A_ : str = ['flax'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) class A__ ( metaclass=__lowerCamelCase): A_ : Union[str, Any] = ['flax'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) class A__ ( metaclass=__lowerCamelCase): A_ : int = ['flax'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) class A__ ( metaclass=__lowerCamelCase): A_ : Any = ['flax'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) class A__ ( metaclass=__lowerCamelCase): A_ : List[str] = ['flax'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) class A__ ( metaclass=__lowerCamelCase): A_ : Optional[int] = ['flax'] def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(self , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] ) @classmethod def __lowerCamelCase ( cls , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): requires_backends(cls , ['flax'] )
706
"""simple docstring""" def __lowerCAmelCase (_UpperCamelCase , _UpperCamelCase = False ): if not isinstance(_UpperCamelCase , _UpperCamelCase ): __lowerCAmelCase : str = F"Expected string as input, found {type(_UpperCamelCase )}" raise ValueError(_UpperCamelCase ) if not isinstance(_UpperCamelCase , _UpperCamelCase ): __lowerCAmelCase : Union[str, Any] = F"Expected boolean as use_pascal parameter, found {type(_UpperCamelCase )}" raise ValueError(_UpperCamelCase ) __lowerCAmelCase : Tuple = input_str.split('_' ) __lowerCAmelCase : int = 0 if use_pascal else 1 __lowerCAmelCase : Any = words[start_index:] __lowerCAmelCase : Any = [word[0].upper() + word[1:] for word in words_to_capitalize] __lowerCAmelCase : Any = '' if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
549
0
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING _A : List[str] = logging.get_logger(__name__) @add_end_docstrings(__SCREAMING_SNAKE_CASE ) class __snake_case ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self , **A_ ): '''simple docstring''' super().__init__(**A_ ) if self.framework == "tf": raise ValueError(f'''The {self.__class__} is only available in PyTorch.''' ) requires_backends(self , '''vision''' ) self.check_model_type(A_ ) def __call__( self , A_ , A_ = None , **A_ , ): '''simple docstring''' if "text_queries" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs.pop('''text_queries''' ) if isinstance(A_ , (str, Image.Image) ): SCREAMING_SNAKE_CASE__ = {'''image''': image, '''candidate_labels''': candidate_labels} else: SCREAMING_SNAKE_CASE__ = image SCREAMING_SNAKE_CASE__ = super().__call__(A_ , **A_ ) return results def lowercase_ ( self , **A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = {} if "threshold" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs['''threshold'''] if "top_k" in kwargs: SCREAMING_SNAKE_CASE__ = kwargs['''top_k'''] return {}, {}, postprocess_params def lowercase_ ( self , A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = load_image(inputs['''image'''] ) SCREAMING_SNAKE_CASE__ = inputs['''candidate_labels'''] if isinstance(A_ , A_ ): SCREAMING_SNAKE_CASE__ = candidate_labels.split(''',''' ) SCREAMING_SNAKE_CASE__ = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(A_ ): SCREAMING_SNAKE_CASE__ = self.tokenizer(A_ , return_tensors=self.framework ) SCREAMING_SNAKE_CASE__ = self.image_processor(A_ , return_tensors=self.framework ) yield { "is_last": i == len(A_ ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def lowercase_ ( self , A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = model_inputs.pop('''target_size''' ) SCREAMING_SNAKE_CASE__ = model_inputs.pop('''candidate_label''' ) SCREAMING_SNAKE_CASE__ = model_inputs.pop('''is_last''' ) SCREAMING_SNAKE_CASE__ = self.model(**A_ ) SCREAMING_SNAKE_CASE__ = {'''target_size''': target_size, '''candidate_label''': candidate_label, '''is_last''': is_last, **outputs} return model_outputs def lowercase_ ( self , A_ , A_=0.1 , A_=None ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = [] for model_output in model_outputs: SCREAMING_SNAKE_CASE__ = model_output['''candidate_label'''] SCREAMING_SNAKE_CASE__ = BaseModelOutput(A_ ) SCREAMING_SNAKE_CASE__ = self.image_processor.post_process_object_detection( outputs=A_ , threshold=A_ , target_sizes=model_output['''target_size'''] )[0] for index in outputs["scores"].nonzero(): SCREAMING_SNAKE_CASE__ = outputs['''scores'''][index].item() SCREAMING_SNAKE_CASE__ = self._get_bounding_box(outputs['''boxes'''][index][0] ) SCREAMING_SNAKE_CASE__ = {'''score''': score, '''label''': label, '''box''': box} results.append(A_ ) SCREAMING_SNAKE_CASE__ = sorted(A_ , key=lambda A_ : x["score"] , reverse=A_ ) if top_k: SCREAMING_SNAKE_CASE__ = results[:top_k] return results def lowercase_ ( self , A_ ): '''simple docstring''' if self.framework != "pt": raise ValueError('''The ZeroShotObjectDetectionPipeline is only available in PyTorch.''' ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = box.int().tolist() SCREAMING_SNAKE_CASE__ = { '''xmin''': xmin, '''ymin''': ymin, '''xmax''': xmax, '''ymax''': ymax, } return bbox
100
"""simple docstring""" # Usage: # ./gen-card-facebook-wmt19.py import os from pathlib import Path def _lowerCAmelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): """simple docstring""" _lowercase: int = { '''en''': '''Machine learning is great, isn\'t it?''', '''ru''': '''Машинное обучение - это здорово, не так ли?''', '''de''': '''Maschinelles Lernen ist großartig, oder?''', } # BLUE scores as follows: # "pair": [fairseq, transformers] _lowercase: Dict = { '''ru-en''': ['''[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)''', '''39.20'''], '''en-ru''': ['''[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)''', '''33.47'''], '''en-de''': ['''[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)''', '''42.83'''], '''de-en''': ['''[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)''', '''41.35'''], } _lowercase: Optional[Any] = f'''{src_lang}-{tgt_lang}''' _lowercase: Tuple = f''' --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt19 - facebook license: apache-2.0 datasets: - wmt19 metrics: - bleu --- # FSMT ## Model description This is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}. For more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616). The abbreviation FSMT stands for FairSeqMachineTranslation All four models are available: * [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru) * [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en) * [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de) * [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = "facebook/wmt19-{src_lang}-{tgt_lang}" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = "{texts[src_lang]}" input_ids = tokenizer.encode(input, return_tensors="pt") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias - The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981) ## Training data Pretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616). ## Eval results pair | fairseq | transformers -------|---------|---------- {pair} | {scores[pair][0]} | {scores[pair][1]} The score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support: - model ensemble, therefore the best performing checkpoint was ported (``model4.pt``). - re-ranking The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=15 mkdir -p $DATA_DIR sacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH="src:examples/seq2seq" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` note: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`. ## Data Sources - [training, etc.](http://www.statmt.org/wmt19/) - [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561) ### BibTeX entry and citation info ```bibtex @inproceedings{{..., year={{2020}}, title={{Facebook FAIR\'s WMT19 News Translation Task Submission}}, author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}}, booktitle={{Proc. of WMT}}, }} ``` ## TODO - port model ensemble (fairseq uses 4 model checkpoints) ''' os.makedirs(_UpperCamelCase , exist_ok=_UpperCamelCase ) _lowercase: List[str] = os.path.join(_UpperCamelCase , '''README.md''' ) print(f'''Generating {path}''' ) with open(_UpperCamelCase , '''w''' , encoding='''utf-8''' ) as f: f.write(_UpperCamelCase ) # make sure we are under the root of the project A__ : List[str] = Path(__file__).resolve().parent.parent.parent A__ : Optional[int] = repo_dir / 'model_cards' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: A__ , A__ , A__ : int = model_name.split('-') A__ : Optional[int] = model_cards_dir / 'facebook' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
353
0
def __lowercase ( snake_case ): """simple docstring""" __magic_name__ :Tuple = 1 __magic_name__ :Any = 2 while i * i <= n: __magic_name__ :str = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def __lowercase ( ): """simple docstring""" __magic_name__ :List[str] = 1 __magic_name__ :Optional[int] = 1 while True: i += 1 t_num += i if count_divisors(snake_case ) > 5_0_0: break return t_num if __name__ == "__main__": print(solution())
180
from __future__ import annotations from typing import Any class lowerCamelCase_ : def __init__( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = 0 ): """simple docstring""" __magic_name__ , __magic_name__ :Any = row, column __magic_name__ :Union[str, Any] = [[default_value for c in range(__lowerCAmelCase )] for r in range(__lowerCAmelCase )] def __str__( self ): """simple docstring""" __magic_name__ :str = F'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier __magic_name__ :Dict = 0 for row_vector in self.array: for obj in row_vector: __magic_name__ :Dict = max(__lowerCAmelCase , len(str(__lowerCAmelCase ) ) ) __magic_name__ :List[str] = F'''%{max_element_length}s''' # Make string and return def single_line(__lowerCAmelCase ) -> str: nonlocal string_format_identifier __magic_name__ :int = '''[''' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(__lowerCAmelCase ) for row_vector in self.array ) return s def __repr__( self ): """simple docstring""" return str(self ) def A ( self , __lowerCAmelCase ): """simple docstring""" if not (isinstance(__lowerCAmelCase , (list, tuple) ) and len(__lowerCAmelCase ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self , __lowerCAmelCase ): """simple docstring""" assert self.validate_indicies(__lowerCAmelCase ) return self.array[loc[0]][loc[1]] def __setitem__( self , __lowerCAmelCase , __lowerCAmelCase ): """simple docstring""" assert self.validate_indicies(__lowerCAmelCase ) __magic_name__ :List[Any] = value def __add__( self , __lowerCAmelCase ): """simple docstring""" assert isinstance(__lowerCAmelCase , __lowerCAmelCase ) assert self.row == another.row and self.column == another.column # Add __magic_name__ :List[Any] = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __magic_name__ :Any = self[r, c] + another[r, c] return result def __neg__( self ): """simple docstring""" __magic_name__ :Dict = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __magic_name__ :Any = -self[r, c] return result def __sub__( self , __lowerCAmelCase ): """simple docstring""" return self + (-another) def __mul__( self , __lowerCAmelCase ): """simple docstring""" if isinstance(__lowerCAmelCase , (int, float) ): # Scalar multiplication __magic_name__ :List[str] = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __magic_name__ :Tuple = self[r, c] * another return result elif isinstance(__lowerCAmelCase , __lowerCAmelCase ): # Matrix multiplication assert self.column == another.row __magic_name__ :List[Any] = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __magic_name__ :List[str] = F'''Unsupported type given for another ({type(__lowerCAmelCase )})''' raise TypeError(__lowerCAmelCase ) def A ( self ): """simple docstring""" __magic_name__ :int = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): __magic_name__ :List[str] = self[r, c] return result def A ( self , __lowerCAmelCase , __lowerCAmelCase ): """simple docstring""" assert isinstance(__lowerCAmelCase , __lowerCAmelCase ) and isinstance(__lowerCAmelCase , __lowerCAmelCase ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __magic_name__ :int = v.transpose() __magic_name__ :str = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def __lowercase ( ): """simple docstring""" __magic_name__ :Optional[int] = Matrix(3, 3, 0 ) for i in range(3 ): __magic_name__ :Tuple = 1 print(f'''a^(-1) is {ainv}''' ) # u, v __magic_name__ :Optional[int] = Matrix(3, 1, 0 ) __magic_name__ , __magic_name__ , __magic_name__ :Optional[int] = 1, 2, -3 __magic_name__ :int = Matrix(3, 1, 0 ) __magic_name__ , __magic_name__ , __magic_name__ :str = 4, -2, 5 print(f'''u is {u}''' ) print(f'''v is {v}''' ) print(f'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(f'''(a + uv^T)^(-1) is {ainv.sherman_morrison(snake_case, snake_case )}''' ) def __lowercase ( ): """simple docstring""" import doctest doctest.testmod() testa()
180
1
import os from argparse import ArgumentParser, Namespace from ..data import SingleSentenceClassificationProcessor as Processor from ..pipelines import TextClassificationPipeline from ..utils import is_tf_available, is_torch_available, logging from . import BaseTransformersCLICommand if not is_tf_available() and not is_torch_available(): raise RuntimeError('At least one of PyTorch or TensorFlow 2.0+ should be installed to use CLI training') # TF training parameters a_ : Dict = False a_ : Tuple = False def __lowercase( UpperCAmelCase__ ): """simple docstring""" return TrainCommand(UpperCAmelCase__ ) class lowerCamelCase__ ( UpperCAmelCase_): """simple docstring""" @staticmethod def _a (__a ): '''simple docstring''' lowerCamelCase = parser.add_parser("train" , help="CLI tool to train a model on a task." ) train_parser.add_argument( "--train_data" , type=__a , required=__a , help="path to train (and optionally evaluation) dataset as a csv with tab separated labels and sentences." , ) train_parser.add_argument( "--column_label" , type=__a , default=0 , help="Column of the dataset csv file with example labels." ) train_parser.add_argument( "--column_text" , type=__a , default=1 , help="Column of the dataset csv file with example texts." ) train_parser.add_argument( "--column_id" , type=__a , default=2 , help="Column of the dataset csv file with example ids." ) train_parser.add_argument( "--skip_first_row" , action="store_true" , help="Skip the first row of the csv file (headers)." ) train_parser.add_argument("--validation_data" , type=__a , default="" , help="path to validation dataset." ) train_parser.add_argument( "--validation_split" , type=__a , default=0.1 , help="if validation dataset is not provided, fraction of train dataset to use as validation dataset." , ) train_parser.add_argument("--output" , type=__a , default="./" , help="path to saved the trained model." ) train_parser.add_argument( "--task" , type=__a , default="text_classification" , help="Task to train the model on." ) train_parser.add_argument( "--model" , type=__a , default="bert-base-uncased" , help="Model's name or path to stored model." ) train_parser.add_argument("--train_batch_size" , type=__a , default=32 , help="Batch size for training." ) train_parser.add_argument("--valid_batch_size" , type=__a , default=64 , help="Batch size for validation." ) train_parser.add_argument("--learning_rate" , type=__a , default=3E-5 , help="Learning rate." ) train_parser.add_argument("--adam_epsilon" , type=__a , default=1E-08 , help="Epsilon for Adam optimizer." ) train_parser.set_defaults(func=__a ) def __init__(self , __a ): '''simple docstring''' lowerCamelCase = logging.get_logger("transformers-cli/training" ) lowerCamelCase = "tf" if is_tf_available() else "torch" os.makedirs(args.output , exist_ok=__a ) lowerCamelCase = args.output lowerCamelCase = args.column_label lowerCamelCase = args.column_text lowerCamelCase = args.column_id self.logger.info(F"""Loading {args.task} pipeline for {args.model}""" ) if args.task == "text_classification": lowerCamelCase = TextClassificationPipeline.from_pretrained(args.model ) elif args.task == "token_classification": raise NotImplementedError elif args.task == "question_answering": raise NotImplementedError self.logger.info(F"""Loading dataset from {args.train_data}""" ) lowerCamelCase = Processor.create_from_csv( args.train_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) lowerCamelCase = None if args.validation_data: self.logger.info(F"""Loading validation dataset from {args.validation_data}""" ) lowerCamelCase = Processor.create_from_csv( args.validation_data , column_label=args.column_label , column_text=args.column_text , column_id=args.column_id , skip_first_row=args.skip_first_row , ) lowerCamelCase = args.validation_split lowerCamelCase = args.train_batch_size lowerCamelCase = args.valid_batch_size lowerCamelCase = args.learning_rate lowerCamelCase = args.adam_epsilon def _a (self ): '''simple docstring''' if self.framework == "tf": return self.run_tf() return self.run_torch() def _a (self ): '''simple docstring''' raise NotImplementedError def _a (self ): '''simple docstring''' self.pipeline.fit( self.train_dataset , validation_data=self.valid_dataset , validation_split=self.validation_split , learning_rate=self.learning_rate , adam_epsilon=self.adam_epsilon , train_batch_size=self.train_batch_size , valid_batch_size=self.valid_batch_size , ) # Save trained pipeline self.pipeline.save_pretrained(self.output )
623
import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeline, AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_AUDIO_BATCH_PARAMS, TEXT_TO_AUDIO_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class lowerCamelCase__ ( UpperCAmelCase_ , unittest.TestCase): """simple docstring""" _A = AudioLDMPipeline _A = TEXT_TO_AUDIO_PARAMS _A = TEXT_TO_AUDIO_BATCH_PARAMS _A = frozenset( [ 'num_inference_steps', 'num_waveforms_per_prompt', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ]) def _a (self ): '''simple docstring''' torch.manual_seed(0 ) lowerCamelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=(32, 64) , class_embed_type="simple_projection" , projection_class_embeddings_input_dim=32 , class_embeddings_concat=__a , ) lowerCamelCase = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=__a , set_alpha_to_one=__a , ) torch.manual_seed(0 ) lowerCamelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=1 , out_channels=1 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0 ) lowerCamelCase = ClapTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , projection_dim=32 , ) lowerCamelCase = ClapTextModelWithProjection(__a ) lowerCamelCase = RobertaTokenizer.from_pretrained("hf-internal-testing/tiny-random-roberta" , model_max_length=77 ) lowerCamelCase = SpeechTaHifiGanConfig( model_in_dim=8 , sampling_rate=1_60_00 , upsample_initial_channel=16 , upsample_rates=[2, 2] , upsample_kernel_sizes=[4, 4] , resblock_kernel_sizes=[3, 7] , resblock_dilation_sizes=[[1, 3, 5], [1, 3, 5]] , normalize_before=__a , ) lowerCamelCase = SpeechTaHifiGan(__a ) lowerCamelCase = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "vocoder": vocoder, } return components def _a (self , __a , __a=0 ): '''simple docstring''' if str(__a ).startswith("mps" ): lowerCamelCase = torch.manual_seed(__a ) else: lowerCamelCase = torch.Generator(device=__a ).manual_seed(__a ) lowerCamelCase = { "prompt": "A hammer hitting a wooden surface", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, } return inputs def _a (self ): '''simple docstring''' lowerCamelCase = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCamelCase = self.get_dummy_components() lowerCamelCase = AudioLDMPipeline(**__a ) lowerCamelCase = audioldm_pipe.to(__a ) audioldm_pipe.set_progress_bar_config(disable=__a ) lowerCamelCase = self.get_dummy_inputs(__a ) lowerCamelCase = audioldm_pipe(**__a ) lowerCamelCase = output.audios[0] assert audio.ndim == 1 assert len(__a ) == 2_56 lowerCamelCase = audio[:10] lowerCamelCase = np.array( [-0.0050, 0.0050, -0.0060, 0.0033, -0.0026, 0.0033, -0.0027, 0.0033, -0.0028, 0.0033] ) assert np.abs(audio_slice - expected_slice ).max() < 1E-2 def _a (self ): '''simple docstring''' lowerCamelCase = self.get_dummy_components() lowerCamelCase = AudioLDMPipeline(**__a ) lowerCamelCase = audioldm_pipe.to(__a ) lowerCamelCase = audioldm_pipe.to(__a ) audioldm_pipe.set_progress_bar_config(disable=__a ) lowerCamelCase = self.get_dummy_inputs(__a ) lowerCamelCase = 3 * [inputs["prompt"]] # forward lowerCamelCase = audioldm_pipe(**__a ) lowerCamelCase = output.audios[0] lowerCamelCase = self.get_dummy_inputs(__a ) lowerCamelCase = 3 * [inputs.pop("prompt" )] lowerCamelCase = audioldm_pipe.tokenizer( __a , padding="max_length" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=__a , return_tensors="pt" , ) lowerCamelCase = text_inputs["input_ids"].to(__a ) lowerCamelCase = audioldm_pipe.text_encoder( __a , ) lowerCamelCase = prompt_embeds.text_embeds # additional L_2 normalization over each hidden-state lowerCamelCase = F.normalize(__a , dim=-1 ) lowerCamelCase = prompt_embeds # forward lowerCamelCase = audioldm_pipe(**__a ) lowerCamelCase = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1E-2 def _a (self ): '''simple docstring''' lowerCamelCase = self.get_dummy_components() lowerCamelCase = AudioLDMPipeline(**__a ) lowerCamelCase = audioldm_pipe.to(__a ) lowerCamelCase = audioldm_pipe.to(__a ) audioldm_pipe.set_progress_bar_config(disable=__a ) lowerCamelCase = self.get_dummy_inputs(__a ) lowerCamelCase = 3 * ["this is a negative prompt"] lowerCamelCase = negative_prompt lowerCamelCase = 3 * [inputs["prompt"]] # forward lowerCamelCase = audioldm_pipe(**__a ) lowerCamelCase = output.audios[0] lowerCamelCase = self.get_dummy_inputs(__a ) lowerCamelCase = 3 * [inputs.pop("prompt" )] lowerCamelCase = [] for p in [prompt, negative_prompt]: lowerCamelCase = audioldm_pipe.tokenizer( __a , padding="max_length" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=__a , return_tensors="pt" , ) lowerCamelCase = text_inputs["input_ids"].to(__a ) lowerCamelCase = audioldm_pipe.text_encoder( __a , ) lowerCamelCase = text_embeds.text_embeds # additional L_2 normalization over each hidden-state lowerCamelCase = F.normalize(__a , dim=-1 ) embeds.append(__a ) lowerCamelCase , lowerCamelCase = embeds # forward lowerCamelCase = audioldm_pipe(**__a ) lowerCamelCase = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1E-2 def _a (self ): '''simple docstring''' lowerCamelCase = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCamelCase = self.get_dummy_components() lowerCamelCase = PNDMScheduler(skip_prk_steps=__a ) lowerCamelCase = AudioLDMPipeline(**__a ) lowerCamelCase = audioldm_pipe.to(__a ) audioldm_pipe.set_progress_bar_config(disable=__a ) lowerCamelCase = self.get_dummy_inputs(__a ) lowerCamelCase = "egg cracking" lowerCamelCase = audioldm_pipe(**__a , negative_prompt=__a ) lowerCamelCase = output.audios[0] assert audio.ndim == 1 assert len(__a ) == 2_56 lowerCamelCase = audio[:10] lowerCamelCase = np.array( [-0.0051, 0.0050, -0.0060, 0.0034, -0.0026, 0.0033, -0.0027, 0.0033, -0.0028, 0.0032] ) assert np.abs(audio_slice - expected_slice ).max() < 1E-2 def _a (self ): '''simple docstring''' lowerCamelCase = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCamelCase = self.get_dummy_components() lowerCamelCase = PNDMScheduler(skip_prk_steps=__a ) lowerCamelCase = AudioLDMPipeline(**__a ) lowerCamelCase = audioldm_pipe.to(__a ) audioldm_pipe.set_progress_bar_config(disable=__a ) lowerCamelCase = "A hammer hitting a wooden surface" # test num_waveforms_per_prompt=1 (default) lowerCamelCase = audioldm_pipe(__a , num_inference_steps=2 ).audios assert audios.shape == (1, 2_56) # test num_waveforms_per_prompt=1 (default) for batch of prompts lowerCamelCase = 2 lowerCamelCase = audioldm_pipe([prompt] * batch_size , num_inference_steps=2 ).audios assert audios.shape == (batch_size, 2_56) # test num_waveforms_per_prompt for single prompt lowerCamelCase = 2 lowerCamelCase = audioldm_pipe(__a , num_inference_steps=2 , num_waveforms_per_prompt=__a ).audios assert audios.shape == (num_waveforms_per_prompt, 2_56) # test num_waveforms_per_prompt for batch of prompts lowerCamelCase = 2 lowerCamelCase = audioldm_pipe( [prompt] * batch_size , num_inference_steps=2 , num_waveforms_per_prompt=__a ).audios assert audios.shape == (batch_size * num_waveforms_per_prompt, 2_56) def _a (self ): '''simple docstring''' lowerCamelCase = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCamelCase = self.get_dummy_components() lowerCamelCase = AudioLDMPipeline(**__a ) lowerCamelCase = audioldm_pipe.to(__a ) audioldm_pipe.set_progress_bar_config(disable=__a ) lowerCamelCase = audioldm_pipe.vocoder.config.sampling_rate lowerCamelCase = self.get_dummy_inputs(__a ) lowerCamelCase = audioldm_pipe(audio_length_in_s=0.016 , **__a ) lowerCamelCase = output.audios[0] assert audio.ndim == 1 assert len(__a ) / vocoder_sampling_rate == 0.016 lowerCamelCase = audioldm_pipe(audio_length_in_s=0.032 , **__a ) lowerCamelCase = output.audios[0] assert audio.ndim == 1 assert len(__a ) / vocoder_sampling_rate == 0.032 def _a (self ): '''simple docstring''' lowerCamelCase = self.get_dummy_components() lowerCamelCase = AudioLDMPipeline(**__a ) lowerCamelCase = audioldm_pipe.to(__a ) audioldm_pipe.set_progress_bar_config(disable=__a ) lowerCamelCase = ["hey"] lowerCamelCase = audioldm_pipe(__a , num_inference_steps=1 ) lowerCamelCase = output.audios.shape assert audio_shape == (1, 2_56) lowerCamelCase = audioldm_pipe.vocoder.config config.model_in_dim *= 2 lowerCamelCase = SpeechTaHifiGan(__a ).to(__a ) lowerCamelCase = audioldm_pipe(__a , num_inference_steps=1 ) lowerCamelCase = output.audios.shape # waveform shape is unchanged, we just have 2x the number of mel channels in the spectrogram assert audio_shape == (1, 2_56) def _a (self ): '''simple docstring''' self._test_attention_slicing_forward_pass(test_mean_pixel_difference=__a ) def _a (self ): '''simple docstring''' self._test_inference_batch_single_identical(test_mean_pixel_difference=__a ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def _a (self ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__a ) @slow class lowerCamelCase__ ( unittest.TestCase): """simple docstring""" def _a (self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _a (self , __a , __a="cpu" , __a=torch.floataa , __a=0 ): '''simple docstring''' lowerCamelCase = torch.Generator(device=__a ).manual_seed(__a ) lowerCamelCase = np.random.RandomState(__a ).standard_normal((1, 8, 1_28, 16) ) lowerCamelCase = torch.from_numpy(__a ).to(device=__a , dtype=__a ) lowerCamelCase = { "prompt": "A hammer hitting a wooden surface", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 2.5, } return inputs def _a (self ): '''simple docstring''' lowerCamelCase = AudioLDMPipeline.from_pretrained("cvssp/audioldm" ) lowerCamelCase = audioldm_pipe.to(__a ) audioldm_pipe.set_progress_bar_config(disable=__a ) lowerCamelCase = self.get_inputs(__a ) lowerCamelCase = 25 lowerCamelCase = audioldm_pipe(**__a ).audios[0] assert audio.ndim == 1 assert len(__a ) == 8_19_20 lowerCamelCase = audio[7_72_30:7_72_40] lowerCamelCase = np.array( [-0.4884, -0.4607, 0.0023, 0.5007, 0.5896, 0.5151, 0.3813, -0.0208, -0.3687, -0.4315] ) lowerCamelCase = np.abs(expected_slice - audio_slice ).max() assert max_diff < 1E-2 def _a (self ): '''simple docstring''' lowerCamelCase = AudioLDMPipeline.from_pretrained("cvssp/audioldm" ) lowerCamelCase = LMSDiscreteScheduler.from_config(audioldm_pipe.scheduler.config ) lowerCamelCase = audioldm_pipe.to(__a ) audioldm_pipe.set_progress_bar_config(disable=__a ) lowerCamelCase = self.get_inputs(__a ) lowerCamelCase = audioldm_pipe(**__a ).audios[0] assert audio.ndim == 1 assert len(__a ) == 8_19_20 lowerCamelCase = audio[2_77_80:2_77_90] lowerCamelCase = np.array([-0.2131, -0.0873, -0.0124, -0.0189, 0.0569, 0.1373, 0.1883, 0.2886, 0.3297, 0.2212] ) lowerCamelCase = np.abs(expected_slice - audio_slice ).max() assert max_diff < 3E-2
623
1
import math import random from typing import Any from .hill_climbing import SearchProblem def a_ ( lowerCAmelCase_ : str, lowerCAmelCase_ : bool = True, lowerCAmelCase_ : float = math.inf, lowerCAmelCase_ : float = -math.inf, lowerCAmelCase_ : float = math.inf, lowerCAmelCase_ : float = -math.inf, lowerCAmelCase_ : bool = False, lowerCAmelCase_ : float = 100, lowerCAmelCase_ : float = 0.01, lowerCAmelCase_ : float = 1, ): __lowerCAmelCase = False __lowerCAmelCase = search_prob __lowerCAmelCase = start_temperate __lowerCAmelCase = [] __lowerCAmelCase = 0 __lowerCAmelCase = None while not search_end: __lowerCAmelCase = current_state.score() if best_state is None or current_score > best_state.score(): __lowerCAmelCase = current_state scores.append(lowerCAmelCase_ ) iterations += 1 __lowerCAmelCase = None __lowerCAmelCase = current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to __lowerCAmelCase = random.randint(0, len(lowerCAmelCase_ ) - 1 ) # picking a random neighbor __lowerCAmelCase = neighbors.pop(lowerCAmelCase_ ) __lowerCAmelCase = picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: __lowerCAmelCase = change * -1 # in case we are finding minimum if change > 0: # improves the solution __lowerCAmelCase = picked_neighbor else: __lowerCAmelCase = (math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability __lowerCAmelCase = picked_neighbor __lowerCAmelCase = current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor __lowerCAmelCase = True else: __lowerCAmelCase = next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(lowerCAmelCase_ ), lowerCAmelCase_ ) plt.xlabel('Iterations' ) plt.ylabel('Function values' ) plt.show() return best_state if __name__ == "__main__": def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : List[str] ): return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) _snake_case : Dict = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) _snake_case : str = simulated_annealing( prob, find_max=False, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( 'The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ' F"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) # starting the problem with initial coordinates (12, 47) _snake_case : int = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) _snake_case : Dict = simulated_annealing( prob, find_max=True, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( 'The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ' F"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) def a_ ( lowerCAmelCase_ : List[Any], lowerCAmelCase_ : Dict ): return (3 * x**2) - (6 * y) _snake_case : Optional[int] = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) _snake_case : Optional[int] = simulated_annealing(prob, find_max=False, visualization=True) print( 'The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ' F"""{local_min.score()}""" ) _snake_case : str = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) _snake_case : Any = simulated_annealing(prob, find_max=True, visualization=True) print( 'The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ' F"""{local_min.score()}""" )
421
import unittest import numpy as np from transformers import RobertaPreLayerNormConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roberta_prelayernorm.modeling_flax_roberta_prelayernorm import ( FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormModel, ) class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self : Dict , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Union[str, Any]=1_3 , lowerCAmelCase_ : List[str]=7 , lowerCAmelCase_ : Union[str, Any]=True , lowerCAmelCase_ : str=True , lowerCAmelCase_ : Optional[Any]=True , lowerCAmelCase_ : Union[str, Any]=True , lowerCAmelCase_ : List[Any]=9_9 , lowerCAmelCase_ : Tuple=3_2 , lowerCAmelCase_ : Any=5 , lowerCAmelCase_ : Any=4 , lowerCAmelCase_ : Dict=3_7 , lowerCAmelCase_ : Tuple="gelu" , lowerCAmelCase_ : List[Any]=0.1 , lowerCAmelCase_ : str=0.1 , lowerCAmelCase_ : str=5_1_2 , lowerCAmelCase_ : Dict=1_6 , lowerCAmelCase_ : Tuple=2 , lowerCAmelCase_ : Optional[int]=0.02 , lowerCAmelCase_ : Optional[int]=4 , ) -> List[Any]: __lowerCAmelCase = parent __lowerCAmelCase = batch_size __lowerCAmelCase = seq_length __lowerCAmelCase = is_training __lowerCAmelCase = use_attention_mask __lowerCAmelCase = use_token_type_ids __lowerCAmelCase = use_labels __lowerCAmelCase = vocab_size __lowerCAmelCase = hidden_size __lowerCAmelCase = num_hidden_layers __lowerCAmelCase = num_attention_heads __lowerCAmelCase = intermediate_size __lowerCAmelCase = hidden_act __lowerCAmelCase = hidden_dropout_prob __lowerCAmelCase = attention_probs_dropout_prob __lowerCAmelCase = max_position_embeddings __lowerCAmelCase = type_vocab_size __lowerCAmelCase = type_sequence_label_size __lowerCAmelCase = initializer_range __lowerCAmelCase = num_choices def lowercase ( self : List[str] ) -> Optional[int]: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase = None if self.use_attention_mask: __lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCAmelCase = None if self.use_token_type_ids: __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __lowerCAmelCase = RobertaPreLayerNormConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCAmelCase_ , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def lowercase ( self : Dict ) -> Dict: __lowerCAmelCase = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = config_and_inputs __lowerCAmelCase = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict def lowercase ( self : Tuple ) -> int: __lowerCAmelCase = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase = config_and_inputs __lowerCAmelCase = True __lowerCAmelCase = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax # Copied from tests.models.roberta.test_modelling_flax_roberta.FlaxRobertaPreLayerNormModelTest with ROBERTA->ROBERTA_PRELAYERNORM,Roberta->RobertaPreLayerNorm,roberta-base->andreasmadsen/efficient_mlm_m0.40 class _UpperCAmelCase ( _UpperCamelCase , unittest.TestCase ): """simple docstring""" a_ = True a_ = ( ( FlaxRobertaPreLayerNormModel, FlaxRobertaPreLayerNormForCausalLM, FlaxRobertaPreLayerNormForMaskedLM, FlaxRobertaPreLayerNormForSequenceClassification, FlaxRobertaPreLayerNormForTokenClassification, FlaxRobertaPreLayerNormForMultipleChoice, FlaxRobertaPreLayerNormForQuestionAnswering, ) if is_flax_available() else () ) def lowercase ( self : Any ) -> Dict: __lowerCAmelCase = FlaxRobertaPreLayerNormModelTester(self ) @slow def lowercase ( self : Tuple ) -> List[str]: for model_class_name in self.all_model_classes: __lowerCAmelCase = model_class_name.from_pretrained('andreasmadsen/efficient_mlm_m0.40' , from_pt=lowerCAmelCase_ ) __lowerCAmelCase = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCAmelCase_ ) @require_flax class _UpperCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def lowercase ( self : int ) -> int: __lowerCAmelCase = FlaxRobertaPreLayerNormForMaskedLM.from_pretrained('andreasmadsen/efficient_mlm_m0.40' , from_pt=lowerCAmelCase_ ) __lowerCAmelCase = np.array([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] , dtype=jnp.intaa ) __lowerCAmelCase = model(lowerCAmelCase_ )[0] __lowerCAmelCase = [1, 1_1, 5_0_2_6_5] self.assertEqual(list(output.shape ) , lowerCAmelCase_ ) # compare the actual values for a slice. __lowerCAmelCase = np.array( [[[40.48_80, 18.01_99, -5.23_67], [-1.88_77, -4.08_85, 10.70_85], [-2.26_13, -5.61_10, 7.26_65]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase_ , atol=1e-4 ) ) @slow def lowercase ( self : Union[str, Any] ) -> Tuple: __lowerCAmelCase = FlaxRobertaPreLayerNormModel.from_pretrained('andreasmadsen/efficient_mlm_m0.40' , from_pt=lowerCAmelCase_ ) __lowerCAmelCase = np.array([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] , dtype=jnp.intaa ) __lowerCAmelCase = model(lowerCAmelCase_ )[0] # compare the actual values for a slice. __lowerCAmelCase = np.array( [[[0.02_08, -0.03_56, 0.02_37], [-0.15_69, -0.04_11, -0.26_26], [0.18_79, 0.01_25, -0.00_89]]] , dtype=np.floataa ) self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase_ , atol=1e-4 ) )
421
1
import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = { """microsoft/conditional-detr-resnet-50""": ( """https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json""" ), } class _UpperCAmelCase ( _SCREAMING_SNAKE_CASE ): a = "conditional_detr" a = ["past_key_values"] a = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self , a__=True , a__=None , a__=3 , a__=300 , a__=6 , a__=2048 , a__=8 , a__=6 , a__=2048 , a__=8 , a__=0.0 , a__=0.0 , a__=True , a__="relu" , a__=256 , a__=0.1 , a__=0.0 , a__=0.0 , a__=0.02 , a__=1.0 , a__=False , a__="sine" , a__="resnet50" , a__=True , a__=False , a__=2 , a__=5 , a__=2 , a__=1 , a__=1 , a__=2 , a__=5 , a__=2 , a__=0.25 , **a__ , ): if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) A_ : List[str] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(snake_case_ , snake_case_ ): A_ : Optional[Any] = backbone_config.get("""model_type""" ) A_ : Dict = CONFIG_MAPPING[backbone_model_type] A_ : List[Any] = config_class.from_dict(snake_case_ ) A_ : Optional[Any] = use_timm_backbone A_ : Dict = backbone_config A_ : List[Any] = num_channels A_ : Tuple = num_queries A_ : Union[str, Any] = d_model A_ : Dict = encoder_ffn_dim A_ : Union[str, Any] = encoder_layers A_ : int = encoder_attention_heads A_ : Optional[int] = decoder_ffn_dim A_ : Tuple = decoder_layers A_ : int = decoder_attention_heads A_ : int = dropout A_ : str = attention_dropout A_ : Tuple = activation_dropout A_ : List[Any] = activation_function A_ : Optional[int] = init_std A_ : Optional[Any] = init_xavier_std A_ : Any = encoder_layerdrop A_ : Optional[int] = decoder_layerdrop A_ : Optional[Any] = encoder_layers A_ : Any = auxiliary_loss A_ : List[str] = position_embedding_type A_ : Optional[int] = backbone A_ : Tuple = use_pretrained_backbone A_ : int = dilation # Hungarian matcher A_ : int = class_cost A_ : Tuple = bbox_cost A_ : Optional[int] = giou_cost # Loss coefficients A_ : List[Any] = mask_loss_coefficient A_ : Any = dice_loss_coefficient A_ : int = cls_loss_coefficient A_ : Any = bbox_loss_coefficient A_ : Union[str, Any] = giou_loss_coefficient A_ : int = focal_alpha super().__init__(is_encoder_decoder=snake_case_ , **snake_case_ ) @property def _lowerCamelCase ( self ): return self.encoder_attention_heads @property def _lowerCamelCase ( self ): return self.d_model def _lowerCamelCase ( self ): A_ : List[Any] = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: A_ : Union[str, Any] = self.backbone_config.to_dict() A_ : Dict = self.__class__.model_type return output class _UpperCAmelCase ( _SCREAMING_SNAKE_CASE ): a = version.parse('''1.11''' ) @property def _lowerCamelCase ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def _lowerCamelCase ( self ): return 1E-5 @property def _lowerCamelCase ( self ): return 12
569
"""simple docstring""" import inspect import unittest from transformers import ConvNextVaConfig from transformers.models.auto import get_values from transformers.models.auto.modeling_auto import MODEL_FOR_BACKBONE_MAPPING_NAMES, MODEL_MAPPING_NAMES from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextVaBackbone, ConvNextVaForImageClassification, ConvNextVaModel from transformers.models.convnextva.modeling_convnextva import CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a : """simple docstring""" def __init__( self , snake_case_ , snake_case_=13 , snake_case_=32 , snake_case_=3 , snake_case_=4 , snake_case_=[10, 20, 30, 40] , snake_case_=[2, 2, 3, 2] , snake_case_=True , snake_case_=True , snake_case_=37 , snake_case_="gelu" , snake_case_=10 , snake_case_=0.02 , snake_case_=["stage2", "stage3", "stage4"] , snake_case_=[2, 3, 4] , snake_case_=None , ) -> Tuple: _UpperCAmelCase = parent _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = num_channels _UpperCAmelCase = num_stages _UpperCAmelCase = hidden_sizes _UpperCAmelCase = depths _UpperCAmelCase = is_training _UpperCAmelCase = use_labels _UpperCAmelCase = intermediate_size _UpperCAmelCase = hidden_act _UpperCAmelCase = num_labels _UpperCAmelCase = initializer_range _UpperCAmelCase = out_features _UpperCAmelCase = out_indices _UpperCAmelCase = scope def __A ( self ) -> Tuple: _UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) _UpperCAmelCase = self.get_config() return config, pixel_values, labels def __A ( self ) -> int: return ConvNextVaConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=snake_case_ , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def __A ( self , snake_case_ , snake_case_ , snake_case_ ) -> str: _UpperCAmelCase = ConvNextVaModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() _UpperCAmelCase = model(snake_case_ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def __A ( self , snake_case_ , snake_case_ , snake_case_ ) -> Union[str, Any]: _UpperCAmelCase = ConvNextVaForImageClassification(snake_case_ ) model.to(snake_case_ ) model.eval() _UpperCAmelCase = model(snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __A ( self , snake_case_ , snake_case_ , snake_case_ ) -> str: _UpperCAmelCase = ConvNextVaBackbone(config=snake_case_ ) model.to(snake_case_ ) model.eval() _UpperCAmelCase = model(snake_case_ ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None _UpperCAmelCase = None _UpperCAmelCase = ConvNextVaBackbone(config=snake_case_ ) model.to(snake_case_ ) model.eval() _UpperCAmelCase = model(snake_case_ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def __A ( self ) -> List[Any]: _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {"pixel_values": pixel_values} return config, inputs_dict def __A ( self ) -> Tuple: _UpperCAmelCase = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = config_and_inputs _UpperCAmelCase = {"pixel_values": pixel_values, "labels": labels} return config, inputs_dict @require_torch class a ( _SCREAMING_SNAKE_CASE, _SCREAMING_SNAKE_CASE, unittest.TestCase ): """simple docstring""" A__ : Dict = ( ( ConvNextVaModel, ConvNextVaForImageClassification, ConvNextVaBackbone, ) if is_torch_available() else () ) A__ : Optional[Any] = ( {"feature-extraction": ConvNextVaModel, "image-classification": ConvNextVaForImageClassification} if is_torch_available() else {} ) A__ : Any = False A__ : str = False A__ : List[str] = False A__ : Optional[int] = False A__ : Dict = False def __A ( self ) -> List[Any]: _UpperCAmelCase = ConvNextVaModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ , hidden_size=37 ) def __A ( self ) -> int: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __A ( self ) -> Any: return @unittest.skip(reason="ConvNextV2 does not use inputs_embeds" ) def __A ( self ) -> int: pass @unittest.skip(reason="ConvNextV2 does not support input and output embeddings" ) def __A ( self ) -> Union[str, Any]: pass @unittest.skip(reason="ConvNextV2 does not use feedforward chunking" ) def __A ( self ) -> Optional[int]: pass def __A ( self ) -> Optional[Any]: if not self.model_tester.is_training: return for model_class in self.all_model_classes: _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_with_labels() _UpperCAmelCase = True if model_class.__name__ in [ *get_values(snake_case_ ), *get_values(snake_case_ ), ]: continue _UpperCAmelCase = model_class(snake_case_ ) model.to(snake_case_ ) model.train() _UpperCAmelCase = self._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) _UpperCAmelCase = model(**snake_case_ ).loss loss.backward() def __A ( self ) -> str: if not self.model_tester.is_training: return for model_class in self.all_model_classes: _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_with_labels() _UpperCAmelCase = False _UpperCAmelCase = True if ( model_class.__name__ in [*get_values(snake_case_ ), *get_values(snake_case_ )] or not model_class.supports_gradient_checkpointing ): continue _UpperCAmelCase = model_class(snake_case_ ) model.to(snake_case_ ) model.gradient_checkpointing_enable() model.train() _UpperCAmelCase = self._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) _UpperCAmelCase = model(**snake_case_ ).loss loss.backward() def __A ( self ) -> Tuple: _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(snake_case_ ) _UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ["pixel_values"] self.assertListEqual(arg_names[:1] , snake_case_ ) def __A ( self ) -> int: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def __A ( self ) -> Optional[int]: def check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ): _UpperCAmelCase = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): _UpperCAmelCase = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) _UpperCAmelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _UpperCAmelCase = self.model_tester.num_stages self.assertEqual(len(snake_case_ ) , expected_num_stages + 1 ) # ConvNextV2's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = True check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _UpperCAmelCase = True check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ) def __A ( self ) -> Union[str, Any]: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case_ ) @slow def __A ( self ) -> Dict: for model_name in CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase = ConvNextVaModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def A__ ( ) -> Optional[int]: '''simple docstring''' _UpperCAmelCase = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class a ( unittest.TestCase ): """simple docstring""" @cached_property def __A ( self ) -> Dict: return AutoImageProcessor.from_pretrained("facebook/convnextv2-tiny-1k-224" ) if is_vision_available() else None @slow def __A ( self ) -> int: _UpperCAmelCase = ConvNextVaForImageClassification.from_pretrained("facebook/convnextv2-tiny-1k-224" ).to(snake_case_ ) _UpperCAmelCase = self.default_image_processor _UpperCAmelCase = prepare_img() _UpperCAmelCase = preprocessor(images=snake_case_ , return_tensors="pt" ).to(snake_case_ ) # forward pass with torch.no_grad(): _UpperCAmelCase = model(**snake_case_ ) # verify the logits _UpperCAmelCase = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , snake_case_ ) _UpperCAmelCase = torch.tensor([0.99_96, 0.19_66, -0.43_86] ).to(snake_case_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case_ , atol=1e-4 ) )
426
0
"""simple docstring""" import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def UpperCamelCase ( _lowerCAmelCase : Optional[Any] ): __a = filter(lambda _lowerCAmelCase : p.requires_grad , model.parameters() ) __a = sum([np.prod(p.size() ) for p in model_parameters] ) return params __A = logging.getLogger(__name__) def UpperCamelCase ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ): if metric == "rouge2": __a = """{val_avg_rouge2:.4f}-{step_count}""" elif metric == "bleu": __a = """{val_avg_bleu:.4f}-{step_count}""" elif metric == "em": __a = """{val_avg_em:.4f}-{step_count}""" else: raise NotImplementedError( f"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" """ function.""" ) __a = ModelCheckpoint( dirpath=_SCREAMING_SNAKE_CASE , filename=_SCREAMING_SNAKE_CASE , monitor=f"""val_{metric}""" , mode="""max""" , save_top_k=3 , every_n_epochs=1 , ) return checkpoint_callback def UpperCamelCase ( _lowerCAmelCase : str , _lowerCAmelCase : List[str] ): return EarlyStopping( monitor=f"""val_{metric}""" , mode="""min""" if """loss""" in metric else """max""" , patience=_SCREAMING_SNAKE_CASE , verbose=_SCREAMING_SNAKE_CASE , ) class a ( pl.Callback ): def lowerCAmelCase_ ( self : List[str] , lowerCamelCase_ : str , lowerCamelCase_ : str ) -> Tuple: __a = {F"""lr_group_{i}""": param["""lr"""] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(__UpperCamelCase ) @rank_zero_only def lowerCAmelCase_ ( self : Union[str, Any] , lowerCamelCase_ : pl.Trainer , lowerCamelCase_ : pl.LightningModule , lowerCamelCase_ : str , lowerCamelCase_ : Tuple=True ) -> None: logger.info(F"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) __a = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ["""log""", """progress_bar""", """preds"""]} ) # Log results __a = Path(pl_module.hparams.output_dir ) if type_path == "test": __a = od / """test_results.txt""" __a = od / """test_generations.txt""" else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __a = od / F"""{type_path}_results/{trainer.global_step:05d}.txt""" __a = od / F"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=__UpperCamelCase ) generations_file.parent.mkdir(exist_ok=__UpperCamelCase ) with open(__UpperCamelCase , """a+""" ) as writer: for key in sorted(__UpperCamelCase ): if key in ["log", "progress_bar", "preds"]: continue __a = metrics[key] if isinstance(__UpperCamelCase , torch.Tensor ): __a = val.item() __a = F"""{key}: {val:.6f}\n""" writer.write(__UpperCamelCase ) if not save_generations: return if "preds" in metrics: __a = """\n""".join(metrics["""preds"""] ) generations_file.open("""w+""" ).write(__UpperCamelCase ) @rank_zero_only def lowerCAmelCase_ ( self : Optional[int] , lowerCamelCase_ : int , lowerCamelCase_ : Dict ) -> Union[str, Any]: try: __a = pl_module.model.model.num_parameters() except AttributeError: __a = pl_module.model.num_parameters() __a = count_trainable_parameters(__UpperCamelCase ) # mp stands for million parameters trainer.logger.log_metrics({"""n_params""": npars, """mp""": npars / 1E6, """grad_mp""": n_trainable_pars / 1E6} ) @rank_zero_only def lowerCAmelCase_ ( self : str , lowerCamelCase_ : pl.Trainer , lowerCamelCase_ : pl.LightningModule ) -> List[Any]: save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(__UpperCamelCase , __UpperCamelCase , """test""" ) @rank_zero_only def lowerCAmelCase_ ( self : Optional[Any] , lowerCamelCase_ : pl.Trainer , lowerCamelCase_ : List[str] ) -> Optional[Any]: save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
707
"""simple docstring""" from sklearn.metrics import matthews_corrcoef import datasets __A = """ Compute the Matthews correlation coefficient (MCC) The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary and multiclass classifications. It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient value between -1 and +1. A coefficient of +1 represents a perfect prediction, 0 an average random prediction and -1 an inverse prediction. The statistic is also known as the phi coefficient. [source: Wikipedia] """ __A = """ Args: predictions (list of int): Predicted labels, as returned by a model. references (list of int): Ground truth labels. sample_weight (list of int, float, or bool): Sample weights. Defaults to `None`. Returns: matthews_correlation (dict containing float): Matthews correlation. Examples: Example 1, a basic example with only predictions and references as inputs: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3]) >>> print(round(results['matthews_correlation'], 2)) 0.54 Example 2, the same example as above, but also including sample weights: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 3, 1, 1, 1, 2]) >>> print(round(results['matthews_correlation'], 2)) 0.1 Example 3, the same example as above, but with sample weights that cause a negative correlation: >>> matthews_metric = datasets.load_metric(\"matthews_correlation\") >>> results = matthews_metric.compute(references=[1, 3, 2, 0, 3, 2], ... predictions=[1, 2, 2, 0, 3, 3], ... sample_weight=[0.5, 1, 0, 0, 0, 1]) >>> print(round(results['matthews_correlation'], 2)) -0.25 """ __A = """\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a ( datasets.Metric ): def lowerCAmelCase_ ( self : List[Any] ) -> Optional[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int32""" ), """references""": datasets.Value("""int32""" ), } ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.matthews_corrcoef.html""" ] , ) def lowerCAmelCase_ ( self : Optional[int] , lowerCamelCase_ : str , lowerCamelCase_ : Any , lowerCamelCase_ : Tuple=None ) -> Dict: return { "matthews_correlation": float(matthews_corrcoef(lowerCamelCase_ , lowerCamelCase_ , sample_weight=lowerCamelCase_ ) ), }
173
0
'''simple docstring''' import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTokenizerFast from utils import get_device, get_timestamp, show_pil class A_ : '''simple docstring''' def __init__( self , A_ = "cpu" , A_ = "openai/clip-vit-large-patch14" ): _UpperCamelCase = device _UpperCamelCase = CLIPTokenizerFast.from_pretrained(A_ ) _UpperCamelCase = [0.4814_5466, 0.457_8275, 0.4082_1073] _UpperCamelCase = [0.2686_2954, 0.2613_0258, 0.2757_7711] _UpperCamelCase = torchvision.transforms.Normalize(self.image_mean , self.image_std ) _UpperCamelCase = torchvision.transforms.Resize(2_24 ) _UpperCamelCase = torchvision.transforms.CenterCrop(2_24 ) def a ( self , A_ ): _UpperCamelCase = self.resize(A_ ) _UpperCamelCase = self.center_crop(A_ ) _UpperCamelCase = self.normalize(A_ ) return images def __call__( self , A_=None , A_=None , **A_ ): _UpperCamelCase = self.tokenizer(text=A_ , **A_ ) _UpperCamelCase = self.preprocess_img(A_ ) _UpperCamelCase = {key: value.to(self.device ) for (key, value) in encoding.items()} return encoding class A_ ( nn.Module ): '''simple docstring''' def __init__( self , A_=10 , A_=0.01 , A_=None , A_=None , A_=None , A_=None , A_=None , A_=None , A_=False , A_=True , A_="image" , A_=True , A_=False , A_=False , A_=False , ): super().__init__() _UpperCamelCase = None _UpperCamelCase = device if device else get_device() if vqgan: _UpperCamelCase = vqgan else: _UpperCamelCase = load_vqgan(self.device , conf_path=A_ , ckpt_path=A_ ) self.vqgan.eval() if clip: _UpperCamelCase = clip else: _UpperCamelCase = CLIPModel.from_pretrained("openai/clip-vit-base-patch32" ) self.clip.to(self.device ) _UpperCamelCase = ProcessorGradientFlow(device=self.device ) _UpperCamelCase = iterations _UpperCamelCase = lr _UpperCamelCase = log _UpperCamelCase = make_grid _UpperCamelCase = return_val _UpperCamelCase = quantize _UpperCamelCase = self.vqgan.decoder.z_shape def a ( self , A_=None , A_=None , A_=5 , A_=True ): _UpperCamelCase = [] if output_path is None: _UpperCamelCase = "./animation.gif" if input_path is None: _UpperCamelCase = self.save_path _UpperCamelCase = sorted(glob(input_path + "/*" ) ) if not len(A_ ): raise ValueError( "No images found in save path, aborting (did you pass save_intermediate=True to the generate" " function?)" ) if len(A_ ) == 1: print("Only one image found in save path, (did you pass save_intermediate=True to the generate function?)" ) _UpperCamelCase = total_duration / len(A_ ) _UpperCamelCase = [frame_duration] * len(A_ ) if extend_frames: _UpperCamelCase = 1.5 _UpperCamelCase = 3 for file_name in paths: if file_name.endswith(".png" ): images.append(imageio.imread(A_ ) ) imageio.mimsave(A_ , A_ , duration=A_ ) print(F"gif saved to {output_path}" ) def a ( self , A_=None , A_=None ): if not (path or img): raise ValueError("Input either path or tensor" ) if img is not None: raise NotImplementedError _UpperCamelCase = preprocess(Image.open(A_ ) , target_image_size=2_56 ).to(self.device ) _UpperCamelCase = preprocess_vqgan(A_ ) _UpperCamelCase , *_UpperCamelCase = self.vqgan.encode(A_ ) return z def a ( self , A_ ): _UpperCamelCase = self.latent.detach().requires_grad_() _UpperCamelCase = base_latent + transform_vector if self.quantize: _UpperCamelCase , *_UpperCamelCase = self.vqgan.quantize(A_ ) else: _UpperCamelCase = trans_latent return self.vqgan.decode(A_ ) def a ( self , A_ , A_ , A_=None ): _UpperCamelCase = self.clip_preprocessor(text=A_ , images=A_ , return_tensors="pt" , padding=A_ ) _UpperCamelCase = self.clip(**A_ ) _UpperCamelCase = clip_outputs.logits_per_image if weights is not None: _UpperCamelCase = similarity_logits * weights return similarity_logits.sum() def a ( self , A_ , A_ , A_ ): _UpperCamelCase = self._get_clip_similarity(pos_prompts["prompts"] , A_ , weights=(1 / pos_prompts["weights"]) ) if neg_prompts: _UpperCamelCase = self._get_clip_similarity(neg_prompts["prompts"] , A_ , weights=neg_prompts["weights"] ) else: _UpperCamelCase = torch.tensor([1] , device=self.device ) _UpperCamelCase = -torch.log(A_ ) + torch.log(A_ ) return loss def a ( self , A_ , A_ , A_ ): _UpperCamelCase = torch.randn_like(self.latent , requires_grad=A_ , device=self.device ) _UpperCamelCase = torch.optim.Adam([vector] , lr=self.lr ) for i in range(self.iterations ): optim.zero_grad() _UpperCamelCase = self._add_vector(A_ ) _UpperCamelCase = loop_post_process(A_ ) _UpperCamelCase = self._get_CLIP_loss(A_ , A_ , A_ ) print("CLIP loss" , A_ ) if self.log: wandb.log({"CLIP Loss": clip_loss} ) clip_loss.backward(retain_graph=A_ ) optim.step() if self.return_val == "image": yield custom_to_pil(transformed_img[0] ) else: yield vector def a ( self , A_ , A_ , A_ ): wandb.init(reinit=A_ , project="face-editor" ) wandb.config.update({"Positive Prompts": positive_prompts} ) wandb.config.update({"Negative Prompts": negative_prompts} ) wandb.config.update({"lr": self.lr, "iterations": self.iterations} ) if image_path: _UpperCamelCase = Image.open(A_ ) _UpperCamelCase = image.resize((2_56, 2_56) ) wandb.log("Original Image" , wandb.Image(A_ ) ) def a ( self , A_ ): if not prompts: return [] _UpperCamelCase = [] _UpperCamelCase = [] if isinstance(A_ , A_ ): _UpperCamelCase = [prompt.strip() for prompt in prompts.split("|" )] for prompt in prompts: if isinstance(A_ , (tuple, list) ): _UpperCamelCase = prompt[0] _UpperCamelCase = float(prompt[1] ) elif ":" in prompt: _UpperCamelCase , _UpperCamelCase = prompt.split(":" ) _UpperCamelCase = float(A_ ) else: _UpperCamelCase = prompt _UpperCamelCase = 1.0 processed_prompts.append(A_ ) weights.append(A_ ) return { "prompts": processed_prompts, "weights": torch.tensor(A_ , device=self.device ), } def a ( self , A_ , A_=None , A_=None , A_=True , A_=False , A_=True , A_=True , A_=None , ): if image_path: _UpperCamelCase = self._get_latent(A_ ) else: _UpperCamelCase = torch.randn(self.latent_dim , device=self.device ) if self.log: self._init_logging(A_ , A_ , A_ ) assert pos_prompts, "You must provide at least one positive prompt." _UpperCamelCase = self.process_prompts(A_ ) _UpperCamelCase = self.process_prompts(A_ ) if save_final and save_path is None: _UpperCamelCase = os.path.join("./outputs/" , "_".join(pos_prompts["prompts"] ) ) if not os.path.exists(A_ ): os.makedirs(A_ ) else: _UpperCamelCase = save_path + "_" + get_timestamp() os.makedirs(A_ ) _UpperCamelCase = save_path _UpperCamelCase = self.vqgan.decode(self.latent )[0] if show_intermediate: print("Original Image" ) show_pil(custom_to_pil(A_ ) ) _UpperCamelCase = loop_post_process(A_ ) for iter, transformed_img in enumerate(self._optimize_CLIP(A_ , A_ , A_ ) ): if show_intermediate: show_pil(A_ ) if save_intermediate: transformed_img.save(os.path.join(self.save_path , F"iter_{iter:03d}.png" ) ) if self.log: wandb.log({"Image": wandb.Image(A_ )} ) if show_final: show_pil(A_ ) if save_final: transformed_img.save(os.path.join(self.save_path , F"iter_{iter:03d}_final.png" ) )
138
'''simple docstring''' import argparse from copy import deepcopy import numpy as np from datasets import ClassLabel, DatasetDict, load_dataset from evaluate import load from transformers import ( AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, Trainer, TrainerCallback, TrainingArguments, set_seed, ) def lowercase__( )-> Dict: """simple docstring""" _UpperCamelCase = argparse.ArgumentParser() parser.add_argument("--model_ckpt" , type=_UpperCamelCase , default="microsoft/unixcoder-base-nine" ) parser.add_argument("--num_epochs" , type=_UpperCamelCase , default=5 ) parser.add_argument("--batch_size" , type=_UpperCamelCase , default=6 ) parser.add_argument("--gradient_accumulation_steps" , type=_UpperCamelCase , default=1 ) parser.add_argument("--freeze" , type=_UpperCamelCase , default=_UpperCamelCase ) parser.add_argument("--learning_rate" , type=_UpperCamelCase , default=5E-4 ) parser.add_argument("--seed" , type=_UpperCamelCase , default=0 ) parser.add_argument("--lr_scheduler_type" , type=_UpperCamelCase , default="cosine" ) parser.add_argument("--num_warmup_steps" , type=_UpperCamelCase , default=10 ) parser.add_argument("--weight_decay" , type=_UpperCamelCase , default=0.01 ) parser.add_argument("--output_dir" , type=_UpperCamelCase , default="./results" ) return parser.parse_args() snake_case_ : Optional[Any] = load('''accuracy''') def lowercase__( _UpperCamelCase : str )-> Any: """simple docstring""" _UpperCamelCase , _UpperCamelCase = eval_pred _UpperCamelCase = np.argmax(_UpperCamelCase , axis=1 ) return metric.compute(predictions=_UpperCamelCase , references=_UpperCamelCase ) class A_ ( lowerCAmelCase_ ): '''simple docstring''' def __init__( self , A_ ): super().__init__() _UpperCamelCase = trainer def a ( self , A_ , A_ , A_ , **A_ ): if control.should_evaluate: _UpperCamelCase = deepcopy(A_ ) self._trainer.evaluate(eval_dataset=self._trainer.train_dataset , metric_key_prefix="train" ) return control_copy def lowercase__( )-> Any: """simple docstring""" _UpperCamelCase = get_args() set_seed(args.seed ) _UpperCamelCase = load_dataset("codeparrot/codecomplex" , split="train" ) _UpperCamelCase = dataset.train_test_split(test_size=0.2 ) _UpperCamelCase = train_test["test"].train_test_split(test_size=0.5 ) _UpperCamelCase = DatasetDict( { "train": train_test["train"], "test": test_validation["train"], "valid": test_validation["test"], } ) print("Loading tokenizer and model" ) _UpperCamelCase = AutoTokenizer.from_pretrained(args.model_ckpt ) _UpperCamelCase = tokenizer.eos_token _UpperCamelCase = AutoModelForSequenceClassification.from_pretrained(args.model_ckpt , num_labels=7 ) _UpperCamelCase = model.config.eos_token_id if args.freeze: for param in model.roberta.parameters(): _UpperCamelCase = False _UpperCamelCase = ClassLabel(num_classes=7 , names=list(set(train_test_validation["train"]["complexity"] ) ) ) def tokenize(_UpperCamelCase : str ): _UpperCamelCase = tokenizer(example["src"] , truncation=_UpperCamelCase , max_length=1024 ) _UpperCamelCase = labels.straint(example["complexity"] ) return { "input_ids": inputs["input_ids"], "attention_mask": inputs["attention_mask"], "label": label, } _UpperCamelCase = train_test_validation.map( _UpperCamelCase , batched=_UpperCamelCase , remove_columns=train_test_validation["train"].column_names , ) _UpperCamelCase = DataCollatorWithPadding(tokenizer=_UpperCamelCase ) _UpperCamelCase = TrainingArguments( output_dir=args.output_dir , learning_rate=args.learning_rate , lr_scheduler_type=args.lr_scheduler_type , evaluation_strategy="epoch" , save_strategy="epoch" , logging_strategy="epoch" , per_device_train_batch_size=args.batch_size , per_device_eval_batch_size=args.batch_size , num_train_epochs=args.num_epochs , gradient_accumulation_steps=args.gradient_accumulation_steps , weight_decay=0.01 , metric_for_best_model="accuracy" , run_name="complexity-java" , report_to="wandb" , ) _UpperCamelCase = Trainer( model=_UpperCamelCase , args=_UpperCamelCase , train_dataset=tokenized_datasets["train"] , eval_dataset=tokenized_datasets["valid"] , tokenizer=_UpperCamelCase , data_collator=_UpperCamelCase , compute_metrics=_UpperCamelCase , ) print("Training..." ) trainer.add_callback(CustomCallback(_UpperCamelCase ) ) trainer.train() if __name__ == "__main__": main()
138
1
'''simple docstring''' from __future__ import annotations class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase ) -> Dict: A_ : List[Any] = TypeError( """Matrices must be formed from a list of zero or more lists containing at """ """least one and the same number of values, each of which must be of type """ """int or float.""" ) if len(_lowerCamelCase ) != 0: A_ : List[str] = len(rows[0] ) if cols == 0: raise error for row in rows: if len(_lowerCamelCase ) != cols: raise error for value in row: if not isinstance(_lowerCamelCase , (int, float) ): raise error A_ : str = rows else: A_ : Optional[int] = [] def UpperCAmelCase_ ( self ) -> list[list[int]]: return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def UpperCAmelCase_ ( self ) -> int: return len(self.rows ) @property def UpperCAmelCase_ ( self ) -> int: return len(self.rows[0] ) @property def UpperCAmelCase_ ( self ) -> tuple[int, int]: return (self.num_rows, self.num_columns) @property def UpperCAmelCase_ ( self ) -> bool: return self.order[0] == self.order[1] def UpperCAmelCase_ ( self ) -> Matrix: A_ : Dict = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> int: if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def UpperCAmelCase_ ( self ) -> bool: return bool(self.determinant() ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> int: A_ : List[Any] = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(_lowerCamelCase ).determinant() def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> int: if (row + column) % 2 == 0: return self.get_minor(_lowerCamelCase , _lowerCamelCase ) return -1 * self.get_minor(_lowerCamelCase , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Matrix: return Matrix( [ [self.get_minor(_lowerCamelCase , _lowerCamelCase ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def UpperCAmelCase_ ( self ) -> Matrix: return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def UpperCAmelCase_ ( self ) -> Matrix: A_ : Union[str, Any] = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Matrix: A_ : Optional[Any] = self.determinant() if not determinant: raise TypeError("""Only matrices with a non-zero determinant have an inverse""" ) return self.adjugate() * (1 / determinant) def __repr__( self ) -> str: return str(self.rows ) def __str__( self ) -> str: if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ """[""" + """. """.join([str(_lowerCamelCase ) for value in row] ) + """.]""" for row in self.rows ] ) + "]" ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase = None ) -> None: A_ : str = TypeError("""Row must be a list containing all ints and/or floats""" ) if not isinstance(_lowerCamelCase , _lowerCamelCase ): raise type_error for value in row: if not isinstance(_lowerCamelCase , (int, float) ): raise type_error if len(_lowerCamelCase ) != self.num_columns: raise ValueError( """Row must be equal in length to the other rows in the matrix""" ) if position is None: self.rows.append(_lowerCamelCase ) else: A_ : Optional[Any] = self.rows[0:position] + [row] + self.rows[position:] def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase = None ) -> None: A_ : int = TypeError( """Column must be a list containing all ints and/or floats""" ) if not isinstance(_lowerCamelCase , _lowerCamelCase ): raise type_error for value in column: if not isinstance(_lowerCamelCase , (int, float) ): raise type_error if len(_lowerCamelCase ) != self.num_rows: raise ValueError( """Column must be equal in length to the other columns in the matrix""" ) if position is None: A_ : str = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: A_ : str = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self , _lowerCamelCase ) -> bool: if not isinstance(_lowerCamelCase , _lowerCamelCase ): return NotImplemented return self.rows == other.rows def __ne__( self , _lowerCamelCase ) -> bool: return not self == other def __neg__( self ) -> Matrix: return self * -1 def __add__( self , _lowerCamelCase ) -> Matrix: if self.order != other.order: raise ValueError("""Addition requires matrices of the same order""" ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self , _lowerCamelCase ) -> Matrix: if self.order != other.order: raise ValueError("""Subtraction requires matrices of the same order""" ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self , _lowerCamelCase ) -> Matrix: if isinstance(_lowerCamelCase , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(_lowerCamelCase , _lowerCamelCase ): if self.num_columns != other.num_rows: raise ValueError( """The number of columns in the first matrix must """ """be equal to the number of rows in the second""" ) return Matrix( [ [Matrix.dot_product(_lowerCamelCase , _lowerCamelCase ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( """A Matrix can only be multiplied by an int, float, or another matrix""" ) def __pow__( self , _lowerCamelCase ) -> Matrix: if not isinstance(_lowerCamelCase , _lowerCamelCase ): raise TypeError("""A Matrix can only be raised to the power of an int""" ) if not self.is_square: raise ValueError("""Only square matrices can be raised to a power""" ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( """Only invertable matrices can be raised to a negative power""" ) A_ : Optional[int] = self for _ in range(other - 1 ): result *= self return result @classmethod def UpperCAmelCase_ ( cls , _lowerCamelCase , _lowerCamelCase ) -> int: return sum(row[i] * column[i] for i in range(len(_lowerCamelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
385
'''simple docstring''' import importlib.util import json import os import warnings from dataclasses import dataclass, field import torch from ..training_args import TrainingArguments from ..utils import cached_property, is_sagemaker_dp_enabled, logging UpperCamelCase__ : str = logging.get_logger(__name__) def UpperCAmelCase ( ) -> Optional[int]: """simple docstring""" A_ : Optional[Any] = os.getenv("""SM_HP_MP_PARAMETERS""" , """{}""" ) try: # Parse it and check the field "partitions" is included, it is required for model parallel. A_ : Any = json.loads(a_ ) if "partitions" not in smp_options: return False except json.JSONDecodeError: return False # Get the sagemaker specific framework parameters from mpi_options variable. A_ : Dict = os.getenv("""SM_FRAMEWORK_PARAMS""" , """{}""" ) try: # Parse it and check the field "sagemaker_distributed_dataparallel_enabled". A_ : int = json.loads(a_ ) if not mpi_options.get("""sagemaker_mpi_enabled""" , a_ ): return False except json.JSONDecodeError: return False # Lastly, check if the `smdistributed` module is present. return importlib.util.find_spec("""smdistributed""" ) is not None if is_sagemaker_model_parallel_available(): import smdistributed.modelparallel.torch as smp smp.init() @dataclass class _lowerCAmelCase ( __A ): """simple docstring""" lowerCamelCase = field( default='''''', metadata={'''help''': '''Used by the SageMaker launcher to send mp-specific args. Ignored in SageMakerTrainer'''}, ) def UpperCAmelCase_ ( self ) -> Optional[int]: super().__post_init__() warnings.warn( """`SageMakerTrainingArguments` is deprecated and will be removed in v5 of Transformers. You can use """ """`TrainingArguments` instead.""" , _lowerCamelCase , ) @cached_property def UpperCAmelCase_ ( self ) -> "torch.device": logger.info("""PyTorch: setting up devices""" ) if torch.distributed.is_available() and torch.distributed.is_initialized() and self.local_rank == -1: logger.warning( """torch.distributed process group is initialized, but local_rank == -1. """ """In order to use Torch DDP, launch your script with `python -m torch.distributed.launch""" ) if self.no_cuda: A_ : Union[str, Any] = torch.device("""cpu""" ) A_ : Any = 0 elif is_sagemaker_model_parallel_available(): A_ : Dict = smp.local_rank() A_ : Union[str, Any] = torch.device("""cuda""" , _lowerCamelCase ) A_ : int = 1 elif is_sagemaker_dp_enabled(): import smdistributed.dataparallel.torch.torch_smddp # noqa: F401 torch.distributed.init_process_group(backend="""smddp""" , timeout=self.ddp_timeout_delta ) A_ : str = int(os.getenv("""SMDATAPARALLEL_LOCAL_RANK""" ) ) A_ : Any = torch.device("""cuda""" , self.local_rank ) A_ : List[str] = 1 elif self.local_rank == -1: # if n_gpu is > 1 we'll use nn.DataParallel. # If you only want to use a specific subset of GPUs use `CUDA_VISIBLE_DEVICES=0` # Explicitly set CUDA to the first (index 0) CUDA device, otherwise `set_device` will # trigger an error that a device index is missing. Index 0 takes into account the # GPUs available in the environment, so `CUDA_VISIBLE_DEVICES=1,2` with `cuda:0` # will use the first GPU in that env, i.e. GPU#1 A_ : Optional[Any] = torch.device("""cuda:0""" if torch.cuda.is_available() else """cpu""" ) # Sometimes the line in the postinit has not been run before we end up here, so just checking we're not at # the default value. A_ : Optional[int] = torch.cuda.device_count() else: # Here, we'll use torch.distributed. # Initializes the distributed backend which will take care of synchronizing nodes/GPUs if not torch.distributed.is_initialized(): torch.distributed.init_process_group(backend="""nccl""" , timeout=self.ddp_timeout_delta ) A_ : Any = torch.device("""cuda""" , self.local_rank ) A_ : str = 1 if device.type == "cuda": torch.cuda.set_device(_lowerCamelCase ) return device @property def UpperCAmelCase_ ( self ) -> Optional[int]: if is_sagemaker_model_parallel_available(): return smp.dp_size() return super().world_size @property def UpperCAmelCase_ ( self ) -> int: return not is_sagemaker_model_parallel_available() @property def UpperCAmelCase_ ( self ) -> Tuple: return False
385
1
'''simple docstring''' import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy as np from .import_utils import is_flax_available, is_tf_available, is_torch_available, is_torch_fx_proxy if is_flax_available(): import jax.numpy as jnp class lowerCAmelCase ( A ): def __get__( self : Union[str, Any] , __lowercase : List[str] , __lowercase : int=None ): """simple docstring""" if obj is None: return self if self.fget is None: raise AttributeError('unreadable attribute' ) __lowercase ='__cached_' + self.fget.__name__ __lowercase =getattr(__lowercase , __lowercase , __lowercase ) if cached is None: __lowercase =self.fget(__lowercase ) setattr(__lowercase , __lowercase , __lowercase ) return cached def __UpperCamelCase ( lowercase__ : str ): '''simple docstring''' __lowercase =val.lower() if val in {"y", "yes", "t", "true", "on", "1"}: return 1 if val in {"n", "no", "f", "false", "off", "0"}: return 0 raise ValueError(F'''invalid truth value {val!r}''' ) def __UpperCamelCase ( lowercase__ : Optional[int] ): '''simple docstring''' if is_torch_fx_proxy(lowercase__ ): return True if is_torch_available(): import torch if isinstance(lowercase__, torch.Tensor ): return True if is_tf_available(): import tensorflow as tf if isinstance(lowercase__, tf.Tensor ): return True if is_flax_available(): import jax.numpy as jnp from jax.core import Tracer if isinstance(lowercase__, (jnp.ndarray, Tracer) ): return True return isinstance(lowercase__, np.ndarray ) def __UpperCamelCase ( lowercase__ : Union[str, Any] ): '''simple docstring''' return isinstance(lowercase__, np.ndarray ) def __UpperCamelCase ( lowercase__ : Dict ): '''simple docstring''' return _is_numpy(lowercase__ ) def __UpperCamelCase ( lowercase__ : Optional[int] ): '''simple docstring''' import torch return isinstance(lowercase__, torch.Tensor ) def __UpperCamelCase ( lowercase__ : Any ): '''simple docstring''' return False if not is_torch_available() else _is_torch(lowercase__ ) def __UpperCamelCase ( lowercase__ : Optional[int] ): '''simple docstring''' import torch return isinstance(lowercase__, torch.device ) def __UpperCamelCase ( lowercase__ : str ): '''simple docstring''' return False if not is_torch_available() else _is_torch_device(lowercase__ ) def __UpperCamelCase ( lowercase__ : List[Any] ): '''simple docstring''' import torch if isinstance(lowercase__, lowercase__ ): if hasattr(lowercase__, lowercase__ ): __lowercase =getattr(lowercase__, lowercase__ ) else: return False return isinstance(lowercase__, torch.dtype ) def __UpperCamelCase ( lowercase__ : Optional[Any] ): '''simple docstring''' return False if not is_torch_available() else _is_torch_dtype(lowercase__ ) def __UpperCamelCase ( lowercase__ : int ): '''simple docstring''' import tensorflow as tf return isinstance(lowercase__, tf.Tensor ) def __UpperCamelCase ( lowercase__ : Optional[int] ): '''simple docstring''' return False if not is_tf_available() else _is_tensorflow(lowercase__ ) def __UpperCamelCase ( lowercase__ : Any ): '''simple docstring''' import tensorflow as tf # the `is_symbolic_tensor` predicate is only available starting with TF 2.14 if hasattr(lowercase__, 'is_symbolic_tensor' ): return tf.is_symbolic_tensor(lowercase__ ) return type(lowercase__ ) == tf.Tensor def __UpperCamelCase ( lowercase__ : Dict ): '''simple docstring''' return False if not is_tf_available() else _is_tf_symbolic_tensor(lowercase__ ) def __UpperCamelCase ( lowercase__ : str ): '''simple docstring''' import jax.numpy as jnp # noqa: F811 return isinstance(lowercase__, jnp.ndarray ) def __UpperCamelCase ( lowercase__ : List[str] ): '''simple docstring''' return False if not is_flax_available() else _is_jax(lowercase__ ) def __UpperCamelCase ( lowercase__ : Optional[Any] ): '''simple docstring''' if isinstance(lowercase__, (dict, UserDict) ): return {k: to_py_obj(lowercase__ ) for k, v in obj.items()} elif isinstance(lowercase__, (list, tuple) ): return [to_py_obj(lowercase__ ) for o in obj] elif is_tf_tensor(lowercase__ ): return obj.numpy().tolist() elif is_torch_tensor(lowercase__ ): return obj.detach().cpu().tolist() elif is_jax_tensor(lowercase__ ): return np.asarray(lowercase__ ).tolist() elif isinstance(lowercase__, (np.ndarray, np.number) ): # tolist also works on 0d np arrays return obj.tolist() else: return obj def __UpperCamelCase ( lowercase__ : Any ): '''simple docstring''' if isinstance(lowercase__, (dict, UserDict) ): return {k: to_numpy(lowercase__ ) for k, v in obj.items()} elif isinstance(lowercase__, (list, tuple) ): return np.array(lowercase__ ) elif is_tf_tensor(lowercase__ ): return obj.numpy() elif is_torch_tensor(lowercase__ ): return obj.detach().cpu().numpy() elif is_jax_tensor(lowercase__ ): return np.asarray(lowercase__ ) else: return obj class lowerCAmelCase ( A ): def snake_case ( self : Tuple ): """simple docstring""" __lowercase =fields(self ) # Safety and consistency checks if not len(__lowercase ): raise ValueError(f'''{self.__class__.__name__} has no fields.''' ) if not all(field.default is None for field in class_fields[1:] ): raise ValueError(f'''{self.__class__.__name__} should not have more than one required field.''' ) __lowercase =getattr(self , class_fields[0].name ) __lowercase =all(getattr(self , field.name ) is None for field in class_fields[1:] ) if other_fields_are_none and not is_tensor(__lowercase ): if isinstance(__lowercase , __lowercase ): __lowercase =first_field.items() __lowercase =True else: try: __lowercase =iter(__lowercase ) __lowercase =True except TypeError: __lowercase =False # if we provided an iterator as first field and the iterator is a (key, value) iterator # set the associated fields if first_field_iterator: for idx, element in enumerate(__lowercase ): if ( not isinstance(__lowercase , (list, tuple) ) or not len(__lowercase ) == 2 or not isinstance(element[0] , __lowercase ) ): if idx == 0: # If we do not have an iterator of key/values, set it as attribute __lowercase =first_field else: # If we have a mixed iterator, raise an error raise ValueError( f'''Cannot set key/value for {element}. It needs to be a tuple (key, value).''' ) break setattr(self , element[0] , element[1] ) if element[1] is not None: __lowercase =element[1] elif first_field is not None: __lowercase =first_field else: for field in class_fields: __lowercase =getattr(self , field.name ) if v is not None: __lowercase =v def __delitem__( self : Optional[int] , *__lowercase : List[str] , **__lowercase : str ): """simple docstring""" raise Exception(f'''You cannot use ``__delitem__`` on a {self.__class__.__name__} instance.''' ) def snake_case ( self : Any , *__lowercase : int , **__lowercase : str ): """simple docstring""" raise Exception(f'''You cannot use ``setdefault`` on a {self.__class__.__name__} instance.''' ) def snake_case ( self : Optional[Any] , *__lowercase : List[str] , **__lowercase : List[Any] ): """simple docstring""" raise Exception(f'''You cannot use ``pop`` on a {self.__class__.__name__} instance.''' ) def snake_case ( self : List[Any] , *__lowercase : Any , **__lowercase : str ): """simple docstring""" raise Exception(f'''You cannot use ``update`` on a {self.__class__.__name__} instance.''' ) def __getitem__( self : Optional[int] , __lowercase : int ): """simple docstring""" if isinstance(__lowercase , __lowercase ): __lowercase =dict(self.items() ) return inner_dict[k] else: return self.to_tuple()[k] def __setattr__( self : int , __lowercase : Optional[int] , __lowercase : int ): """simple docstring""" if name in self.keys() and value is not None: # Don't call self.__setitem__ to avoid recursion errors super().__setitem__(__lowercase , __lowercase ) super().__setattr__(__lowercase , __lowercase ) def __setitem__( self : Optional[int] , __lowercase : Union[str, Any] , __lowercase : Optional[Any] ): """simple docstring""" super().__setitem__(__lowercase , __lowercase ) # Don't call self.__setattr__ to avoid recursion errors super().__setattr__(__lowercase , __lowercase ) def snake_case ( self : Any ): """simple docstring""" return tuple(self[k] for k in self.keys() ) class lowerCAmelCase ( A , A ): @classmethod def snake_case ( cls : Dict , __lowercase : str ): """simple docstring""" raise ValueError( f'''{value} is not a valid {cls.__name__}, please select one of {list(cls._valueamember_map_.keys() )}''' ) class lowerCAmelCase ( A ): lowerCAmelCase_ = "longest" lowerCAmelCase_ = "max_length" lowerCAmelCase_ = "do_not_pad" class lowerCAmelCase ( A ): lowerCAmelCase_ = "pt" lowerCAmelCase_ = "tf" lowerCAmelCase_ = "np" lowerCAmelCase_ = "jax" class lowerCAmelCase : def __init__( self : List[Any] , __lowercase : List[ContextManager] ): """simple docstring""" __lowercase =context_managers __lowercase =ExitStack() def __enter__( self : Tuple ): """simple docstring""" for context_manager in self.context_managers: self.stack.enter_context(__lowercase ) def __exit__( self : str , *__lowercase : Optional[Any] , **__lowercase : str ): """simple docstring""" self.stack.__exit__(*__lowercase , **__lowercase ) def __UpperCamelCase ( lowercase__ : Any ): '''simple docstring''' __lowercase =infer_framework(lowercase__ ) if framework == "tf": __lowercase =inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": __lowercase =inspect.signature(model_class.forward ) # PyTorch models else: __lowercase =inspect.signature(model_class.__call__ ) # Flax models for p in signature.parameters: if p == "return_loss" and signature.parameters[p].default is True: return True return False def __UpperCamelCase ( lowercase__ : str ): '''simple docstring''' __lowercase =model_class.__name__ __lowercase =infer_framework(lowercase__ ) if framework == "tf": __lowercase =inspect.signature(model_class.call ) # TensorFlow models elif framework == "pt": __lowercase =inspect.signature(model_class.forward ) # PyTorch models else: __lowercase =inspect.signature(model_class.__call__ ) # Flax models if "QuestionAnswering" in model_name: return [p for p in signature.parameters if "label" in p or p in ("start_positions", "end_positions")] else: return [p for p in signature.parameters if "label" in p] def __UpperCamelCase ( lowercase__ : MutableMapping, lowercase__ : str = "", lowercase__ : str = "." ): '''simple docstring''' def _flatten_dict(lowercase__ : str, lowercase__ : Optional[Any]="", lowercase__ : Optional[Any]="." ): for k, v in d.items(): __lowercase =str(lowercase__ ) + delimiter + str(lowercase__ ) if parent_key else k if v and isinstance(lowercase__, lowercase__ ): yield from flatten_dict(lowercase__, lowercase__, delimiter=lowercase__ ).items() else: yield key, v return dict(_flatten_dict(lowercase__, lowercase__, lowercase__ ) ) @contextmanager def __UpperCamelCase ( lowercase__ : List[str], lowercase__ : bool = False ): '''simple docstring''' if use_temp_dir: with tempfile.TemporaryDirectory() as tmp_dir: yield tmp_dir else: yield working_dir def __UpperCamelCase ( lowercase__ : Tuple, lowercase__ : List[Any]=None ): '''simple docstring''' if is_numpy_array(lowercase__ ): return np.transpose(lowercase__, axes=lowercase__ ) elif is_torch_tensor(lowercase__ ): return array.T if axes is None else array.permute(*lowercase__ ) elif is_tf_tensor(lowercase__ ): import tensorflow as tf return tf.transpose(lowercase__, perm=lowercase__ ) elif is_jax_tensor(lowercase__ ): return jnp.transpose(lowercase__, axes=lowercase__ ) else: raise ValueError(F'''Type not supported for transpose: {type(lowercase__ )}.''' ) def __UpperCamelCase ( lowercase__ : List[str], lowercase__ : str ): '''simple docstring''' if is_numpy_array(lowercase__ ): return np.reshape(lowercase__, lowercase__ ) elif is_torch_tensor(lowercase__ ): return array.reshape(*lowercase__ ) elif is_tf_tensor(lowercase__ ): import tensorflow as tf return tf.reshape(lowercase__, lowercase__ ) elif is_jax_tensor(lowercase__ ): return jnp.reshape(lowercase__, lowercase__ ) else: raise ValueError(F'''Type not supported for reshape: {type(lowercase__ )}.''' ) def __UpperCamelCase ( lowercase__ : List[Any], lowercase__ : str=None ): '''simple docstring''' if is_numpy_array(lowercase__ ): return np.squeeze(lowercase__, axis=lowercase__ ) elif is_torch_tensor(lowercase__ ): return array.squeeze() if axis is None else array.squeeze(dim=lowercase__ ) elif is_tf_tensor(lowercase__ ): import tensorflow as tf return tf.squeeze(lowercase__, axis=lowercase__ ) elif is_jax_tensor(lowercase__ ): return jnp.squeeze(lowercase__, axis=lowercase__ ) else: raise ValueError(F'''Type not supported for squeeze: {type(lowercase__ )}.''' ) def __UpperCamelCase ( lowercase__ : Tuple, lowercase__ : Optional[Any] ): '''simple docstring''' if is_numpy_array(lowercase__ ): return np.expand_dims(lowercase__, lowercase__ ) elif is_torch_tensor(lowercase__ ): return array.unsqueeze(dim=lowercase__ ) elif is_tf_tensor(lowercase__ ): import tensorflow as tf return tf.expand_dims(lowercase__, axis=lowercase__ ) elif is_jax_tensor(lowercase__ ): return jnp.expand_dims(lowercase__, axis=lowercase__ ) else: raise ValueError(F'''Type not supported for expand_dims: {type(lowercase__ )}.''' ) def __UpperCamelCase ( lowercase__ : Tuple ): '''simple docstring''' if is_numpy_array(lowercase__ ): return np.size(lowercase__ ) elif is_torch_tensor(lowercase__ ): return array.numel() elif is_tf_tensor(lowercase__ ): import tensorflow as tf return tf.size(lowercase__ ) elif is_jax_tensor(lowercase__ ): return array.size else: raise ValueError(F'''Type not supported for expand_dims: {type(lowercase__ )}.''' ) def __UpperCamelCase ( lowercase__ : str, lowercase__ : Tuple ): '''simple docstring''' for key, value in auto_map.items(): if isinstance(lowercase__, (tuple, list) ): __lowercase =[F'''{repo_id}--{v}''' if (v is not None and '--' not in v) else v for v in value] elif value is not None and "--" not in value: __lowercase =F'''{repo_id}--{value}''' return auto_map def __UpperCamelCase ( lowercase__ : Optional[int] ): '''simple docstring''' for base_class in inspect.getmro(lowercase__ ): __lowercase =base_class.__module__ __lowercase =base_class.__name__ if module.startswith('tensorflow' ) or module.startswith('keras' ) or name == "TFPreTrainedModel": return "tf" elif module.startswith('torch' ) or name == "PreTrainedModel": return "pt" elif module.startswith('flax' ) or module.startswith('jax' ) or name == "FlaxPreTrainedModel": return "flax" else: raise TypeError(F'''Could not infer framework from class {model_class}.''' )
119
'''simple docstring''' import importlib.util import os import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import ( is_accelerate_available, is_flax_available, is_safetensors_available, is_tf_available, is_torch_available, ) from . import BaseTransformersCLICommand def __UpperCamelCase ( lowercase__ : Union[str, Any] ): '''simple docstring''' return EnvironmentCommand() def __UpperCamelCase ( lowercase__ : Any ): '''simple docstring''' return EnvironmentCommand(args.accelerate_config_file ) class lowerCAmelCase ( A ): @staticmethod def snake_case ( __lowercase : ArgumentParser ): """simple docstring""" __lowercase =parser.add_parser('env' ) download_parser.set_defaults(func=__lowercase ) download_parser.add_argument( '--accelerate-config_file' , default=__lowercase , help='The accelerate config file to use for the default values in the launching script.' , ) download_parser.set_defaults(func=__lowercase ) def __init__( self : int , __lowercase : Any , *__lowercase : str ): """simple docstring""" __lowercase =accelerate_config_file def snake_case ( self : Optional[int] ): """simple docstring""" __lowercase ='not installed' if is_safetensors_available(): import safetensors __lowercase =safetensors.__version__ elif importlib.util.find_spec('safetensors' ) is not None: import safetensors __lowercase =f'''{safetensors.__version__} but is ignored because of PyTorch version too old.''' __lowercase ='not installed' __lowercase =__lowercase ='not found' if is_accelerate_available(): import accelerate from accelerate.commands.config import default_config_file, load_config_from_file __lowercase =accelerate.__version__ # Get the default from the config file. if self._accelerate_config_file is not None or os.path.isfile(__lowercase ): __lowercase =load_config_from_file(self._accelerate_config_file ).to_dict() __lowercase =( '\n'.join([f'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(__lowercase , __lowercase ) else f'''\t{accelerate_config}''' ) __lowercase ='not installed' __lowercase ='NA' if is_torch_available(): import torch __lowercase =torch.__version__ __lowercase =torch.cuda.is_available() __lowercase ='not installed' __lowercase ='NA' if is_tf_available(): import tensorflow as tf __lowercase =tf.__version__ try: # deprecated in v2.1 __lowercase =tf.test.is_gpu_available() except AttributeError: # returns list of devices, convert to bool __lowercase =bool(tf.config.list_physical_devices('GPU' ) ) __lowercase ='not installed' __lowercase ='not installed' __lowercase ='not installed' __lowercase ='NA' if is_flax_available(): import flax import jax import jaxlib __lowercase =flax.__version__ __lowercase =jax.__version__ __lowercase =jaxlib.__version__ __lowercase =jax.lib.xla_bridge.get_backend().platform __lowercase ={ '`transformers` version': version, 'Platform': platform.platform(), 'Python version': platform.python_version(), 'Huggingface_hub version': huggingface_hub.__version__, 'Safetensors version': f'''{safetensors_version}''', 'Accelerate version': f'''{accelerate_version}''', 'Accelerate config': f'''{accelerate_config_str}''', 'PyTorch version (GPU?)': f'''{pt_version} ({pt_cuda_available})''', 'Tensorflow version (GPU?)': f'''{tf_version} ({tf_cuda_available})''', 'Flax version (CPU?/GPU?/TPU?)': f'''{flax_version} ({jax_backend})''', 'Jax version': f'''{jax_version}''', 'JaxLib version': f'''{jaxlib_version}''', 'Using GPU in script?': '<fill in>', 'Using distributed or parallel set-up in script?': '<fill in>', } print('\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n' ) print(self.format_dict(__lowercase ) ) return info @staticmethod def snake_case ( __lowercase : Optional[Any] ): """simple docstring""" return "\n".join([f'''- {prop}: {val}''' for prop, val in d.items()] ) + "\n"
119
1
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): lowerCAmelCase_ : Any = get_tests_dir('fixtures/test_sentencepiece.model') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right lowerCAmelCase_ : Optional[Any] = 12_80_22 lowerCAmelCase_ : Union[str, Any] = 12_80_28 @require_sentencepiece class __SCREAMING_SNAKE_CASE (lowerCamelCase_ , unittest.TestCase ): """simple docstring""" __a =MaMaaaTokenizer __a =False __a =False __a =True def UpperCamelCase__ ( self : str ): super().setUp() _a = ["</s>", "<unk>", "▁This", "▁is", "▁a", "▁t", "est", "\u0120", "<pad>"] _a = dict(zip(__a , range(len(__a ) ) ) ) _a = Path(self.tmpdirname ) save_json(__a , save_dir / VOCAB_FILES_NAMES["vocab_file"] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(__a , save_dir / VOCAB_FILES_NAMES["spm_file"] ) _a = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase__ ( self : Any , **__a : str ): return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **__a ) def UpperCamelCase__ ( self : List[str] , __a : Tuple ): return ( "This is a test", "This is a test", ) def UpperCamelCase__ ( self : Optional[Any] ): _a = "</s>" _a = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__a ) , __a ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__a ) , __a ) def UpperCamelCase__ ( self : Optional[int] ): _a = self.get_tokenizer() _a = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "</s>" ) self.assertEqual(vocab_keys[1] , "<unk>" ) self.assertEqual(vocab_keys[-1] , "<s>" ) self.assertEqual(len(__a ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip("Skip this test while all models are still to be uploaded." ) def UpperCamelCase__ ( self : Dict ): pass def UpperCamelCase__ ( self : int ): _a = self.get_tokenizer() _a = tokenizer.tokenize("This is a test" ) self.assertListEqual(__a , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__a ) , [2, 3, 4, 5, 6] , ) _a = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(__a , ["▁This", "▁is", "▁a", "▁t", "est"] ) _a = tokenizer.convert_tokens_to_string(__a ) self.assertEqual(__a , "This is a test" ) @slow def UpperCamelCase__ ( self : Tuple ): # fmt: off _a = {"input_ids": [[12_80_22, 11_01_08, 3_97, 11, 3_82_72, 22_47, 12_48_11, 2_85, 1_81_05, 15_86, 2_07, 7, 3_95_34, 44_28, 3_97, 10_19, 1_81_05, 15_86, 2_07, 7, 4_13_37, 1_67_86, 2_41, 7, 2_02_14, 17, 12_56_90, 1_03_98, 7, 4_43_78, 5_80_69, 6_83_42, 77_98, 73_43, 11, 2_99, 3_33_10, 4, 1_58, 3_73_50, 9_40_77, 45_69, 2_99, 3_33_10, 90, 4, 5_28_40, 2_90, 4, 3_12_70, 1_12, 2_99, 6_82, 4, 5_28_40, 3_99_53, 1_40_79, 1_93, 5_25_19, 9_08_94, 1_78_94, 12_06_97, 11, 4_04_45, 5_51, 17, 10_19, 5_25_19, 9_08_94, 1_77_56, 9_63, 11, 4_04_45, 4_80, 17, 97_92, 11_20, 51_73, 13_93, 62_40, 1_67_86, 2_41, 12_09_96, 28, 12_45, 13_93, 11_82_40, 1_11_23, 10_19, 9_36_12, 26_91, 1_06_18, 9_80_58, 12_04_09, 19_28, 2_79, 4, 4_06_83, 3_67, 1_78, 2_07, 10_19, 1_03, 10_31_21, 5_06, 6_52_96, 5, 2], [12_80_22, 2_12_17, 3_67, 1_17, 12_54_50, 1_28, 7_19, 7, 73_08, 40, 9_36_12, 1_26_69, 11_16, 1_67_04, 71, 1_77_85, 36_99, 1_55_92, 35, 1_44, 95_84, 2_41, 1_19_43, 7_13, 9_50, 7_99, 22_47, 8_84_27, 1_50, 1_49, 11_88_13, 12_07_06, 10_19, 10_69_06, 8_15_18, 28, 12_24, 2_27_99, 3_97, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [12_80_22, 16_58, 12_33_11, 51_55, 55_78, 47_22, 2_79, 1_49_47, 23_66, 11_20, 11_97, 14, 13_48, 92_32, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__a , model_name="facebook/m2m100_418M" , revision="c168bae485c864188cf9aa0e4108b0b6934dc91e" , ) @require_torch @require_sentencepiece @require_tokenizers class __SCREAMING_SNAKE_CASE (unittest.TestCase ): """simple docstring""" __a ='facebook/m2m100_418M' __a =[ 'In my opinion, there are two levels of response from the French government.', 'NSA Affair Emphasizes Complete Lack of Debate on Intelligence', ] __a =[ 'Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.', 'L\'affaire NSA souligne l\'absence totale de débat sur le renseignement', ] # fmt: off __a =[EN_CODE, 593, 1949, 11_5781, 4, 7_1586, 4234, 6_0633, 12_6233, 432, 12_3808, 1_5592, 1197, 11_7132, 12_0618, 5, 2] @classmethod def UpperCamelCase__ ( cls : Optional[Any] ): _a = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en" , tgt_lang="fr" ) _a = 1 return cls def UpperCamelCase__ ( self : Tuple ): self.assertEqual(self.tokenizer.get_lang_id("ar" ) , 12_80_06 ) self.assertEqual(self.tokenizer.get_lang_id("en" ) , 12_80_22 ) self.assertEqual(self.tokenizer.get_lang_id("ro" ) , 12_80_76 ) self.assertEqual(self.tokenizer.get_lang_id("mr" ) , 12_80_63 ) def UpperCamelCase__ ( self : List[Any] ): _a = self.tokenizer.get_vocab() self.assertEqual(len(__a ) , self.tokenizer.vocab_size ) self.assertEqual(vocab["<unk>"] , 3 ) self.assertIn(self.tokenizer.get_lang_token("en" ) , __a ) def UpperCamelCase__ ( self : str ): _a = "en" _a = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , __a ) def UpperCamelCase__ ( self : Union[str, Any] ): self.assertIn(__a , self.tokenizer.all_special_ids ) # fmt: off _a = [FR_CODE, 53_64, 82, 86_42, 4, 2_94, 47, 8, 1_40_28, 1_36, 32_86, 97_06, 6, 9_07_97, 6, 14_40_12, 1_62, 8_81_28, 3_00_61, 5, 2] # fmt: on _a = self.tokenizer.decode(__a , skip_special_tokens=__a ) _a = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=__a ) self.assertEqual(__a , __a ) self.assertNotIn(self.tokenizer.eos_token , __a ) def UpperCamelCase__ ( self : List[str] ): _a = tempfile.mkdtemp() _a = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(__a ) _a = MaMaaaTokenizer.from_pretrained(__a ) self.assertDictEqual(new_tok.lang_token_to_id , __a ) @require_torch def UpperCamelCase__ ( self : Tuple ): _a = "en" _a = "fr" _a = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=__a , return_tensors="pt" ) _a = shift_tokens_right( batch["labels"] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: _a = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def UpperCamelCase__ ( self : Optional[int] ): _a = "mr" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) _a = "zh" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def UpperCamelCase__ ( self : List[Any] ): _a = "mr" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("mr" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) _a = "zh" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id("zh" )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def UpperCamelCase__ ( self : Dict ): _a = self.tokenizer._build_translation_inputs("A test" , return_tensors="pt" , src_lang="en" , tgt_lang="ar" ) self.assertEqual( nested_simplify(__a ) , { # en_XX, A, test, EOS "input_ids": [[12_80_22, 58, 41_83, 2]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 12_80_06, } , )
521
'''simple docstring''' import math import sys def _lowerCamelCase ( lowercase : str ) -> str: _a = "" try: with open(lowercase , "rb" ) as binary_file: _a = binary_file.read() for dat in data: _a = F'{dat:08b}' result += curr_byte return result except OSError: print("File not accessible" ) sys.exit() def _lowerCamelCase ( lowercase : str ) -> str: _a = {"0": "0", "1": "1"} _a , _a = "", "" _a = len(lowercase ) for i in range(len(lowercase ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue _a = lexicon[curr_string] result += last_match_id _a = last_match_id + "0" if math.loga(lowercase ).is_integer(): _a = {} for curr_key in list(lowercase ): _a = lexicon.pop(lowercase ) _a = new_lex _a = last_match_id + "1" index += 1 _a = "" return result def _lowerCamelCase ( lowercase : str , lowercase : str ) -> None: _a = 8 try: with open(lowercase , "wb" ) as opened_file: _a = [ to_write[i : i + byte_length] for i in range(0 , len(lowercase ) , lowercase ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append("10000000" ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array[:-1]: opened_file.write(int(lowercase , 2 ).to_bytes(1 , byteorder="big" ) ) except OSError: print("File not accessible" ) sys.exit() def _lowerCamelCase ( lowercase : str ) -> str: _a = 0 for letter in data_bits: if letter == "1": break counter += 1 _a = data_bits[counter:] _a = data_bits[counter + 1 :] return data_bits def _lowerCamelCase ( lowercase : str , lowercase : str ) -> None: _a = read_file_binary(lowercase ) _a = remove_prefix(lowercase ) _a = decompress_data(lowercase ) write_file_binary(lowercase , lowercase ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
521
1
'''simple docstring''' import os import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers.models.realm.configuration_realm import RealmConfig from transformers.models.realm.retrieval_realm import _REALM_BLOCK_RECORDS_FILENAME, RealmRetriever from transformers.models.realm.tokenization_realm import VOCAB_FILES_NAMES, RealmTokenizer class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def _a ( self : List[str] ): '''simple docstring''' A_ : Tuple = tempfile.mkdtemp() A_ : Any = 5 # Realm tok A_ : Optional[Any] = [ """[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """test""", """question""", """this""", """is""", """the""", """first""", """second""", """third""", """fourth""", """fifth""", """record""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] A_ : str = os.path.join(self.tmpdirname ,"""realm_tokenizer""" ) os.makedirs(_a ,exist_ok=_a ) A_ : int = os.path.join(_a ,VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) A_ : Optional[Any] = os.path.join(self.tmpdirname ,"""realm_block_records""" ) os.makedirs(_a ,exist_ok=_a ) def _a ( self : Any ): '''simple docstring''' return RealmTokenizer.from_pretrained(os.path.join(self.tmpdirname ,"""realm_tokenizer""" ) ) def _a ( self : Dict ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def _a ( self : Optional[int] ): '''simple docstring''' A_ : Tuple = RealmConfig(num_block_records=self.num_block_records ) return config def _a ( self : Optional[Any] ): '''simple docstring''' A_ : Any = Dataset.from_dict( { """id""": ["""0""", """1"""], """question""": ["""foo""", """bar"""], """answers""": [["""Foo""", """Bar"""], ["""Bar"""]], } ) return dataset def _a ( self : List[Any] ): '''simple docstring''' A_ : Tuple = np.array( [ B"""This is the first record""", B"""This is the second record""", B"""This is the third record""", B"""This is the fourth record""", B"""This is the fifth record""", B"""This is a longer longer longer record""", ] ,dtype=_a ,) return block_records def _a ( self : str ): '''simple docstring''' A_ : Optional[int] = RealmRetriever( block_records=self.get_dummy_block_records() ,tokenizer=self.get_tokenizer() ,) return retriever def _a ( self : List[Any] ): '''simple docstring''' A_ : int = self.get_config() A_ : str = self.get_dummy_retriever() A_ : str = retriever.tokenizer A_ : str = np.array([0, 3] ,dtype="""long""" ) A_ : Optional[int] = tokenizer(["""Test question"""] ).input_ids A_ : Optional[Any] = tokenizer( ["""the fourth"""] ,add_special_tokens=_a ,return_token_type_ids=_a ,return_attention_mask=_a ,).input_ids A_ : Dict = config.reader_seq_len A_ , A_ , A_ , A_ : Optional[Any] = retriever( _a ,_a ,answer_ids=_a ,max_length=_a ,return_tensors="""np""" ) self.assertEqual(len(_a ) ,2 ) self.assertEqual(len(_a ) ,2 ) self.assertEqual(len(_a ) ,2 ) self.assertEqual(concat_inputs.input_ids.shape ,(2, 10) ) self.assertEqual(concat_inputs.attention_mask.shape ,(2, 10) ) self.assertEqual(concat_inputs.token_type_ids.shape ,(2, 10) ) self.assertEqual(concat_inputs.special_tokens_mask.shape ,(2, 10) ) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[0] ) ,["""[CLS]""", """test""", """question""", """[SEP]""", """this""", """is""", """the""", """first""", """record""", """[SEP]"""] ,) self.assertEqual( tokenizer.convert_ids_to_tokens(concat_inputs.input_ids[1] ) ,["""[CLS]""", """test""", """question""", """[SEP]""", """this""", """is""", """the""", """fourth""", """record""", """[SEP]"""] ,) def _a ( self : Any ): '''simple docstring''' A_ : List[str] = self.get_config() A_ : str = self.get_dummy_retriever() A_ : Dict = retriever.tokenizer A_ : str = np.array([0, 3, 5] ,dtype="""long""" ) A_ : Any = tokenizer(["""Test question"""] ).input_ids A_ : Dict = tokenizer( ["""the fourth""", """longer longer"""] ,add_special_tokens=_a ,return_token_type_ids=_a ,return_attention_mask=_a ,).input_ids A_ : Union[str, Any] = config.reader_seq_len A_ , A_ , A_ , A_ : int = retriever( _a ,_a ,answer_ids=_a ,max_length=_a ,return_tensors="""np""" ) self.assertEqual([False, True, True] ,_a ) self.assertEqual([[-1, -1, -1], [6, -1, -1], [6, 7, 8]] ,_a ) self.assertEqual([[-1, -1, -1], [7, -1, -1], [7, 8, 9]] ,_a ) def _a ( self : Any ): '''simple docstring''' A_ : Optional[int] = self.get_dummy_retriever() retriever.save_pretrained(os.path.join(self.tmpdirname ,"""realm_block_records""" ) ) # Test local path A_ : Dict = retriever.from_pretrained(os.path.join(self.tmpdirname ,"""realm_block_records""" ) ) self.assertEqual(retriever.block_records[0] ,B"""This is the first record""" ) # Test mocked remote path with patch("""transformers.models.realm.retrieval_realm.hf_hub_download""" ) as mock_hf_hub_download: A_ : int = os.path.join( os.path.join(self.tmpdirname ,"""realm_block_records""" ) ,_REALM_BLOCK_RECORDS_FILENAME ) A_ : str = RealmRetriever.from_pretrained("""google/realm-cc-news-pretrained-openqa""" ) self.assertEqual(retriever.block_records[0] ,B"""This is the first record""" )
665
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_bart import BartTokenizer __magic_name__ = logging.get_logger(__name__) __magic_name__ = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} # See all BART models at https://huggingface.co/models?filter=bart __magic_name__ = { 'vocab_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/vocab.json', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/vocab.json', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/vocab.json', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/vocab.json', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/vocab.json', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/vocab.json', }, 'merges_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/merges.txt', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/merges.txt', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/merges.txt', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/merges.txt', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/merges.txt', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/merges.txt', }, 'tokenizer_file': { 'facebook/bart-base': 'https://huggingface.co/facebook/bart-base/resolve/main/tokenizer.json', 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/tokenizer.json', 'facebook/bart-large-mnli': 'https://huggingface.co/facebook/bart-large-mnli/resolve/main/tokenizer.json', 'facebook/bart-large-cnn': 'https://huggingface.co/facebook/bart-large-cnn/resolve/main/tokenizer.json', 'facebook/bart-large-xsum': 'https://huggingface.co/facebook/bart-large-xsum/resolve/main/tokenizer.json', 'yjernite/bart_eli5': 'https://huggingface.co/yjernite/bart_eli5/resolve/main/tokenizer.json', }, } __magic_name__ = { 'facebook/bart-base': 1_024, 'facebook/bart-large': 1_024, 'facebook/bart-large-mnli': 1_024, 'facebook/bart-large-cnn': 1_024, 'facebook/bart-large-xsum': 1_024, 'yjernite/bart_eli5': 1_024, } class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' a_ = VOCAB_FILES_NAMES a_ = PRETRAINED_VOCAB_FILES_MAP a_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a_ = ["""input_ids""", """attention_mask"""] a_ = BartTokenizer def __init__( self : str ,_a : Any=None ,_a : Optional[int]=None ,_a : int=None ,_a : Optional[int]="replace" ,_a : Dict="<s>" ,_a : Optional[Any]="</s>" ,_a : Dict="</s>" ,_a : Tuple="<s>" ,_a : Optional[Any]="<unk>" ,_a : List[str]="<pad>" ,_a : int="<mask>" ,_a : str=False ,_a : List[str]=True ,**_a : Dict ,): '''simple docstring''' super().__init__( _a ,_a ,tokenizer_file=_a ,errors=_a ,bos_token=_a ,eos_token=_a ,sep_token=_a ,cls_token=_a ,unk_token=_a ,pad_token=_a ,mask_token=_a ,add_prefix_space=_a ,trim_offsets=_a ,**_a ,) A_ : Dict = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" ,_a ) != add_prefix_space: A_ : List[str] = getattr(_a ,pre_tok_state.pop("""type""" ) ) A_ : Optional[int] = add_prefix_space A_ : int = pre_tok_class(**_a ) A_ : str = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` A_ : str = """post_processor""" A_ : List[Any] = getattr(self.backend_tokenizer ,_a ,_a ) if tokenizer_component_instance: A_ : Tuple = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: A_ : Tuple = tuple(state["""sep"""] ) if "cls" in state: A_ : Tuple = tuple(state["""cls"""] ) A_ : List[str] = False if state.get("""add_prefix_space""" ,_a ) != add_prefix_space: A_ : Dict = add_prefix_space A_ : Any = True if state.get("""trim_offsets""" ,_a ) != trim_offsets: A_ : Union[str, Any] = trim_offsets A_ : List[Any] = True if changes_to_apply: A_ : Optional[int] = getattr(_a ,state.pop("""type""" ) ) A_ : Tuple = component_class(**_a ) setattr(self.backend_tokenizer ,_a ,_a ) @property def _a ( self : List[str] ): '''simple docstring''' if self._mask_token is None: if self.verbose: logger.error("""Using mask_token, but it is not set yet.""" ) return None return str(self._mask_token ) @mask_token.setter def _a ( self : Union[str, Any] ,_a : Any ): '''simple docstring''' A_ : int = AddedToken(_a ,lstrip=_a ,rstrip=_a ) if isinstance(_a ,_a ) else value A_ : List[Any] = value def _a ( self : str ,*_a : str ,**_a : Optional[int] ): '''simple docstring''' A_ : Optional[Any] = kwargs.get("""is_split_into_words""" ,_a ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' """to use it with pretokenized inputs.""" ) return super()._batch_encode_plus(*_a ,**_a ) def _a ( self : str ,*_a : List[Any] ,**_a : str ): '''simple docstring''' A_ : List[str] = kwargs.get("""is_split_into_words""" ,_a ) if is_split_into_words and not self.add_prefix_space: raise ValueError( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' """to use it with pretokenized inputs.""" ) return super()._encode_plus(*_a ,**_a ) def _a ( self : Optional[int] ,_a : str ,_a : Optional[str] = None ): '''simple docstring''' A_ : str = self._tokenizer.model.save(_a ,name=_a ) return tuple(_a ) def _a ( self : str ,_a : Optional[int] ,_a : int=None ): '''simple docstring''' A_ : Optional[Any] = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def _a ( self : Optional[int] ,_a : List[int] ,_a : Optional[List[int]] = None ): '''simple docstring''' A_ : Dict = [self.sep_token_id] A_ : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
665
1
import random class SCREAMING_SNAKE_CASE__ : '''simple docstring''' @staticmethod def _lowerCAmelCase ( lowerCamelCase__ ): A : Any = [ord(lowerCamelCase__ ) for i in text] A : str = [] A : Dict = [] for i in plain: A : Dict = random.randint(1, 300 ) A : Dict = (i + k) * k cipher.append(lowerCamelCase__ ) key.append(lowerCamelCase__ ) return cipher, key @staticmethod def _lowerCAmelCase ( lowerCamelCase__, lowerCamelCase__ ): A : Optional[int] = [] for i in range(len(lowerCamelCase__ ) ): A : List[str] = int((cipher[i] - (key[i]) ** 2) / key[i] ) plain.append(chr(lowerCamelCase__ ) ) return "".join(lowerCamelCase__ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_:List[Any] = Onepad().encrypt("""Hello""") print(c, k) print(Onepad().decrypt(c, k))
520
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ , unittest.TestCase ): '''simple docstring''' __lowerCamelCase : Union[str, Any] = LongformerTokenizer __lowerCamelCase : List[Any] = True __lowerCamelCase : Optional[Any] = LongformerTokenizerFast __lowerCamelCase : Tuple = True def _lowerCAmelCase ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt A : List[Any] = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] A : Dict = dict(zip(lowerCamelCase__, range(len(lowerCamelCase__ ) ) ) ) A : Tuple = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] A : Optional[Any] = {"""unk_token""": """<unk>"""} A : Union[str, Any] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""vocab_file"""] ) A : Tuple = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file, """w""", encoding="""utf-8""" ) as fp: fp.write(json.dumps(lowerCamelCase__ ) + """\n""" ) with open(self.merges_file, """w""", encoding="""utf-8""" ) as fp: fp.write("""\n""".join(lowerCamelCase__ ) ) def _lowerCAmelCase ( self, **lowerCamelCase__ ): kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname, **lowerCamelCase__ ) def _lowerCAmelCase ( self, **lowerCamelCase__ ): kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname, **lowerCamelCase__ ) def _lowerCAmelCase ( self, lowerCamelCase__ ): A : Tuple = """lower newer""" A : List[Any] = """lower newer""" return input_text, output_text def _lowerCAmelCase ( self ): A : Tuple = self.tokenizer_class(self.vocab_file, self.merges_file, **self.special_tokens_map ) A : List[str] = """lower newer""" A : int = ["""l""", """o""", """w""", """er""", """\u0120""", """n""", """e""", """w""", """er"""] A : List[str] = tokenizer.tokenize(lowerCamelCase__ ) # , add_prefix_space=True) self.assertListEqual(lowerCamelCase__, lowerCamelCase__ ) A : Any = tokens + [tokenizer.unk_token] A : List[str] = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase__ ), lowerCamelCase__ ) def _lowerCAmelCase ( self ): A : Dict = self.get_tokenizer() self.assertListEqual(tokenizer.encode("""Hello world!""", add_special_tokens=lowerCamelCase__ ), [0, 3_1414, 232, 328, 2] ) self.assertListEqual( tokenizer.encode("""Hello world! cécé herlolip 418""", add_special_tokens=lowerCamelCase__ ), [0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2], ) @slow def _lowerCAmelCase ( self ): A : Dict = self.tokenizer_class.from_pretrained("""allenai/longformer-base-4096""" ) A : Tuple = tokenizer.encode("""sequence builders""", add_special_tokens=lowerCamelCase__ ) A : List[str] = tokenizer.encode("""multi-sequence build""", add_special_tokens=lowerCamelCase__ ) A : Any = tokenizer.encode( """sequence builders""", add_special_tokens=lowerCamelCase__, add_prefix_space=lowerCamelCase__ ) A : str = tokenizer.encode( """sequence builders""", """multi-sequence build""", add_special_tokens=lowerCamelCase__, add_prefix_space=lowerCamelCase__ ) A : str = tokenizer.build_inputs_with_special_tokens(lowerCamelCase__ ) A : Union[str, Any] = tokenizer.build_inputs_with_special_tokens(lowerCamelCase__, lowerCamelCase__ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def _lowerCAmelCase ( self ): A : str = self.get_tokenizer() A : List[Any] = """Encode this sequence.""" A : Tuple = tokenizer.byte_encoder[""" """.encode("""utf-8""" )[0]] # Testing encoder arguments A : Tuple = tokenizer.encode(lowerCamelCase__, add_special_tokens=lowerCamelCase__, add_prefix_space=lowerCamelCase__ ) A : Optional[Any] = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(lowerCamelCase__, lowerCamelCase__ ) A : Any = tokenizer.encode(lowerCamelCase__, add_special_tokens=lowerCamelCase__, add_prefix_space=lowerCamelCase__ ) A : List[Any] = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(lowerCamelCase__, lowerCamelCase__ ) tokenizer.add_special_tokens({"""bos_token""": """<s>"""} ) A : Optional[Any] = tokenizer.encode(lowerCamelCase__, add_special_tokens=lowerCamelCase__ ) A : Dict = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(lowerCamelCase__, lowerCamelCase__ ) # Testing spaces after special tokens A : List[str] = """<mask>""" tokenizer.add_special_tokens( {"""mask_token""": AddedToken(lowerCamelCase__, lstrip=lowerCamelCase__, rstrip=lowerCamelCase__ )} ) # mask token has a left space A : Union[str, Any] = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) A : str = """Encode <mask> sequence""" A : str = """Encode <mask>sequence""" A : List[Any] = tokenizer.encode(lowerCamelCase__ ) A : str = encoded.index(lowerCamelCase__ ) A : int = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(lowerCamelCase__, lowerCamelCase__ ) A : Any = tokenizer.encode(lowerCamelCase__ ) A : List[str] = encoded.index(lowerCamelCase__ ) A : Dict = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(lowerCamelCase__, lowerCamelCase__ ) def _lowerCAmelCase ( self ): pass def _lowerCAmelCase ( self ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): A : List[Any] = self.rust_tokenizer_class.from_pretrained(lowerCamelCase__, **lowerCamelCase__ ) A : List[Any] = self.tokenizer_class.from_pretrained(lowerCamelCase__, **lowerCamelCase__ ) A : str = """A, <mask> AllenNLP sentence.""" A : Dict = tokenizer_r.encode_plus(lowerCamelCase__, add_special_tokens=lowerCamelCase__, return_token_type_ids=lowerCamelCase__ ) A : List[str] = tokenizer_p.encode_plus(lowerCamelCase__, add_special_tokens=lowerCamelCase__, return_token_type_ids=lowerCamelCase__ ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["""token_type_ids"""] ), sum(tokens_p["""token_type_ids"""] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["""attention_mask"""] ) / len(tokens_r["""attention_mask"""] ), sum(tokens_p["""attention_mask"""] ) / len(tokens_p["""attention_mask"""] ), ) A : Union[str, Any] = tokenizer_r.convert_ids_to_tokens(tokens_r["""input_ids"""] ) A : int = tokenizer_p.convert_ids_to_tokens(tokens_p["""input_ids"""] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["""input_ids"""], [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual(tokens_r["""input_ids"""], [0, 250, 6, 5_0264, 3823, 487, 2_1992, 3645, 4, 2] ) self.assertSequenceEqual( lowerCamelCase__, ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) self.assertSequenceEqual( lowerCamelCase__, ["""<s>""", """A""", """,""", """<mask>""", """ĠAllen""", """N""", """LP""", """Ġsentence""", """.""", """</s>"""] ) def _lowerCAmelCase ( self ): for trim_offsets, add_prefix_space in itertools.product([True, False], repeat=2 ): A : Dict = self.rust_tokenizer_class.from_pretrained( self.tmpdirname, use_fast=lowerCamelCase__, add_prefix_space=lowerCamelCase__, trim_offsets=lowerCamelCase__ ) A : List[str] = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) A : Union[str, Any] = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state["""add_prefix_space"""], lowerCamelCase__ ) self.assertEqual(post_processor_state["""add_prefix_space"""], lowerCamelCase__ ) self.assertEqual(post_processor_state["""trim_offsets"""], lowerCamelCase__ ) def _lowerCAmelCase ( self ): # Test which aims to verify that the offsets are well adapted to the argument `add_prefix_space` and # `trim_offsets` for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): A : List[str] = """hello""" # `hello` is a token in the vocabulary of `pretrained_name` A : Union[str, Any] = f'''{text_of_1_token} {text_of_1_token}''' A : Optional[int] = self.rust_tokenizer_class.from_pretrained( lowerCamelCase__, use_fast=lowerCamelCase__, add_prefix_space=lowerCamelCase__, trim_offsets=lowerCamelCase__ ) A : Optional[Any] = tokenizer_r(lowerCamelCase__, return_offsets_mapping=lowerCamelCase__, add_special_tokens=lowerCamelCase__ ) self.assertEqual(encoding.offset_mapping[0], (0, len(lowerCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1], (len(lowerCamelCase__ ) + 1, len(lowerCamelCase__ ) + 1 + len(lowerCamelCase__ )), ) A : Optional[int] = self.rust_tokenizer_class.from_pretrained( lowerCamelCase__, use_fast=lowerCamelCase__, add_prefix_space=lowerCamelCase__, trim_offsets=lowerCamelCase__ ) A : int = tokenizer_r(lowerCamelCase__, return_offsets_mapping=lowerCamelCase__, add_special_tokens=lowerCamelCase__ ) self.assertEqual(encoding.offset_mapping[0], (0, len(lowerCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1], (len(lowerCamelCase__ ) + 1, len(lowerCamelCase__ ) + 1 + len(lowerCamelCase__ )), ) A : List[Any] = self.rust_tokenizer_class.from_pretrained( lowerCamelCase__, use_fast=lowerCamelCase__, add_prefix_space=lowerCamelCase__, trim_offsets=lowerCamelCase__ ) A : str = tokenizer_r(lowerCamelCase__, return_offsets_mapping=lowerCamelCase__, add_special_tokens=lowerCamelCase__ ) self.assertEqual(encoding.offset_mapping[0], (0, len(lowerCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1], (len(lowerCamelCase__ ), len(lowerCamelCase__ ) + 1 + len(lowerCamelCase__ )), ) A : Optional[Any] = self.rust_tokenizer_class.from_pretrained( lowerCamelCase__, use_fast=lowerCamelCase__, add_prefix_space=lowerCamelCase__, trim_offsets=lowerCamelCase__ ) A : List[Any] = tokenizer_r(lowerCamelCase__, return_offsets_mapping=lowerCamelCase__, add_special_tokens=lowerCamelCase__ ) self.assertEqual(encoding.offset_mapping[0], (0, len(lowerCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1], (len(lowerCamelCase__ ), len(lowerCamelCase__ ) + 1 + len(lowerCamelCase__ )), ) A : Optional[Any] = f''' {text}''' # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) A : str = self.rust_tokenizer_class.from_pretrained( lowerCamelCase__, use_fast=lowerCamelCase__, add_prefix_space=lowerCamelCase__, trim_offsets=lowerCamelCase__ ) A : str = tokenizer_r(lowerCamelCase__, return_offsets_mapping=lowerCamelCase__, add_special_tokens=lowerCamelCase__ ) self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(lowerCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1], (1 + len(lowerCamelCase__ ) + 1, 1 + len(lowerCamelCase__ ) + 1 + len(lowerCamelCase__ )), ) A : Any = self.rust_tokenizer_class.from_pretrained( lowerCamelCase__, use_fast=lowerCamelCase__, add_prefix_space=lowerCamelCase__, trim_offsets=lowerCamelCase__ ) A : Union[str, Any] = tokenizer_r(lowerCamelCase__, return_offsets_mapping=lowerCamelCase__, add_special_tokens=lowerCamelCase__ ) self.assertEqual(encoding.offset_mapping[0], (0, 1 + len(lowerCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1], (1 + len(lowerCamelCase__ ), 1 + len(lowerCamelCase__ ) + 1 + len(lowerCamelCase__ )), ) A : Any = self.rust_tokenizer_class.from_pretrained( lowerCamelCase__, use_fast=lowerCamelCase__, add_prefix_space=lowerCamelCase__, trim_offsets=lowerCamelCase__ ) A : Optional[int] = tokenizer_r(lowerCamelCase__, return_offsets_mapping=lowerCamelCase__, add_special_tokens=lowerCamelCase__ ) self.assertEqual(encoding.offset_mapping[0], (0, 1 + len(lowerCamelCase__ )) ) self.assertEqual( encoding.offset_mapping[1], (1 + len(lowerCamelCase__ ), 1 + len(lowerCamelCase__ ) + 1 + len(lowerCamelCase__ )), )
520
1
import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class __lowercase ( unittest.TestCase ): __magic_name__ : int = MODEL_FOR_MASKED_LM_MAPPING __magic_name__ : Dict = TF_MODEL_FOR_MASKED_LM_MAPPING def lowerCAmelCase_ ( self ) -> Dict: '''simple docstring''' super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def lowerCAmelCase_ ( self ) -> str: '''simple docstring''' A_ = pipeline(task='''fill-mask''' , model='''sshleifer/tiny-distilroberta-base''' , top_k=2 , framework='''tf''' ) A_ = unmasker('''My name is <mask>''' ) self.assertEqual( nested_simplify(a__ , decimals=6 ) , [ {'''sequence''': '''My name is grouped''', '''score''': 2.1E-05, '''token''': 3_8_0_1_5, '''token_str''': ''' grouped'''}, {'''sequence''': '''My name is accuser''', '''score''': 2.1E-05, '''token''': 2_5_5_0_6, '''token_str''': ''' accuser'''}, ] , ) A_ = unmasker('''The largest city in France is <mask>''' ) self.assertEqual( nested_simplify(a__ , decimals=6 ) , [ { '''sequence''': '''The largest city in France is grouped''', '''score''': 2.1E-05, '''token''': 3_8_0_1_5, '''token_str''': ''' grouped''', }, { '''sequence''': '''The largest city in France is accuser''', '''score''': 2.1E-05, '''token''': 2_5_5_0_6, '''token_str''': ''' accuser''', }, ] , ) A_ = unmasker('''My name is <mask>''' , targets=[''' Patrick''', ''' Clara''', ''' Teven'''] , top_k=3 ) self.assertEqual( nested_simplify(a__ , decimals=6 ) , [ {'''sequence''': '''My name is Clara''', '''score''': 2E-05, '''token''': 1_3_6_0_6, '''token_str''': ''' Clara'''}, {'''sequence''': '''My name is Patrick''', '''score''': 2E-05, '''token''': 3_4_9_9, '''token_str''': ''' Patrick'''}, {'''sequence''': '''My name is Te''', '''score''': 1.9E-05, '''token''': 2_9_4_1, '''token_str''': ''' Te'''}, ] , ) @require_torch def lowerCAmelCase_ ( self ) -> Optional[Any]: '''simple docstring''' A_ = pipeline(task='''fill-mask''' , model='''sshleifer/tiny-distilroberta-base''' , top_k=2 , framework='''pt''' ) A_ = unmasker('''My name is <mask>''' ) self.assertEqual( nested_simplify(a__ , decimals=6 ) , [ {'''sequence''': '''My name is Maul''', '''score''': 2.2E-05, '''token''': 3_5_6_7_6, '''token_str''': ''' Maul'''}, {'''sequence''': '''My name isELS''', '''score''': 2.2E-05, '''token''': 1_6_4_1_6, '''token_str''': '''ELS'''}, ] , ) A_ = unmasker('''The largest city in France is <mask>''' ) self.assertEqual( nested_simplify(a__ , decimals=6 ) , [ { '''sequence''': '''The largest city in France is Maul''', '''score''': 2.2E-05, '''token''': 3_5_6_7_6, '''token_str''': ''' Maul''', }, {'''sequence''': '''The largest city in France isELS''', '''score''': 2.2E-05, '''token''': 1_6_4_1_6, '''token_str''': '''ELS'''}, ] , ) A_ = unmasker('''My name is <mask>''' , targets=[''' Patrick''', ''' Clara''', ''' Teven'''] , top_k=3 ) self.assertEqual( nested_simplify(a__ , decimals=6 ) , [ {'''sequence''': '''My name is Patrick''', '''score''': 2.1E-05, '''token''': 3_4_9_9, '''token_str''': ''' Patrick'''}, {'''sequence''': '''My name is Te''', '''score''': 2E-05, '''token''': 2_9_4_1, '''token_str''': ''' Te'''}, {'''sequence''': '''My name is Clara''', '''score''': 2E-05, '''token''': 1_3_6_0_6, '''token_str''': ''' Clara'''}, ] , ) A_ = unmasker('''My name is <mask> <mask>''' , top_k=2 ) self.assertEqual( nested_simplify(a__ , decimals=6 ) , [ [ { '''score''': 2.2E-05, '''token''': 3_5_6_7_6, '''token_str''': ''' Maul''', '''sequence''': '''<s>My name is Maul<mask></s>''', }, {'''score''': 2.2E-05, '''token''': 1_6_4_1_6, '''token_str''': '''ELS''', '''sequence''': '''<s>My name isELS<mask></s>'''}, ], [ { '''score''': 2.2E-05, '''token''': 3_5_6_7_6, '''token_str''': ''' Maul''', '''sequence''': '''<s>My name is<mask> Maul</s>''', }, {'''score''': 2.2E-05, '''token''': 1_6_4_1_6, '''token_str''': '''ELS''', '''sequence''': '''<s>My name is<mask>ELS</s>'''}, ], ] , ) @require_torch_gpu def lowerCAmelCase_ ( self ) -> Union[str, Any]: '''simple docstring''' A_ = pipeline('''fill-mask''' , model='''hf-internal-testing/tiny-random-distilbert''' , device=0 , framework='''pt''' ) # convert model to fp16 pipe.model.half() A_ = pipe('''Paris is the [MASK] of France.''' ) # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(a__ , a__ ) @slow @require_torch def lowerCAmelCase_ ( self ) -> str: '''simple docstring''' A_ = pipeline(task='''fill-mask''' , model='''distilroberta-base''' , top_k=2 , framework='''pt''' ) self.run_large_test(a__ ) @slow @require_tf def lowerCAmelCase_ ( self ) -> List[Any]: '''simple docstring''' A_ = pipeline(task='''fill-mask''' , model='''distilroberta-base''' , top_k=2 , framework='''tf''' ) self.run_large_test(a__ ) def lowerCAmelCase_ ( self , a__ ) -> Tuple: '''simple docstring''' A_ = unmasker('''My name is <mask>''' ) self.assertEqual( nested_simplify(a__ ) , [ {'''sequence''': '''My name is John''', '''score''': 0.0_08, '''token''': 6_1_0, '''token_str''': ''' John'''}, {'''sequence''': '''My name is Chris''', '''score''': 0.0_07, '''token''': 1_5_7_3, '''token_str''': ''' Chris'''}, ] , ) A_ = unmasker('''The largest city in France is <mask>''' ) self.assertEqual( nested_simplify(a__ ) , [ { '''sequence''': '''The largest city in France is Paris''', '''score''': 0.2_51, '''token''': 2_2_0_1, '''token_str''': ''' Paris''', }, { '''sequence''': '''The largest city in France is Lyon''', '''score''': 0.2_14, '''token''': 1_2_7_9_0, '''token_str''': ''' Lyon''', }, ] , ) A_ = unmasker('''My name is <mask>''' , targets=[''' Patrick''', ''' Clara''', ''' Teven'''] , top_k=3 ) self.assertEqual( nested_simplify(a__ ) , [ {'''sequence''': '''My name is Patrick''', '''score''': 0.0_05, '''token''': 3_4_9_9, '''token_str''': ''' Patrick'''}, {'''sequence''': '''My name is Clara''', '''score''': 0.0_00, '''token''': 1_3_6_0_6, '''token_str''': ''' Clara'''}, {'''sequence''': '''My name is Te''', '''score''': 0.0_00, '''token''': 2_9_4_1, '''token_str''': ''' Te'''}, ] , ) @require_torch def lowerCAmelCase_ ( self ) -> List[Any]: '''simple docstring''' A_ = pipeline(task='''fill-mask''' , model='''sshleifer/tiny-distilroberta-base''' , framework='''pt''' ) A_ = None A_ = None self.run_pipeline_test(a__ , [] ) @require_tf def lowerCAmelCase_ ( self ) -> List[str]: '''simple docstring''' A_ = pipeline(task='''fill-mask''' , model='''sshleifer/tiny-distilroberta-base''' , framework='''tf''' ) A_ = None A_ = None self.run_pipeline_test(a__ , [] ) def lowerCAmelCase_ ( self , a__ , a__ , a__ ) -> Any: '''simple docstring''' if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest('''The provided tokenizer has no mask token, (probably reformer or wav2vec2)''' ) A_ = FillMaskPipeline(model=a__ , tokenizer=a__ ) A_ = [ F"This is another {tokenizer.mask_token} test", ] return fill_masker, examples def lowerCAmelCase_ ( self , a__ , a__ ) -> Any: '''simple docstring''' A_ = fill_masker.tokenizer A_ = fill_masker.model A_ = fill_masker( F"This is a {tokenizer.mask_token}" , ) self.assertEqual( a__ , [ {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, ] , ) A_ = fill_masker([F"This is a {tokenizer.mask_token}"] ) self.assertEqual( a__ , [ {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, ] , ) A_ = fill_masker([F"This is a {tokenizer.mask_token}", F"Another {tokenizer.mask_token} great test."] ) self.assertEqual( a__ , [ [ {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, ], [ {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, ], ] , ) with self.assertRaises(a__ ): fill_masker([None] ) # No mask_token is not supported with self.assertRaises(a__ ): fill_masker('''This is''' ) self.run_test_top_k(a__ , a__ ) self.run_test_targets(a__ , a__ ) self.run_test_top_k_targets(a__ , a__ ) self.fill_mask_with_duplicate_targets_and_top_k(a__ , a__ ) self.fill_mask_with_multiple_masks(a__ , a__ ) def lowerCAmelCase_ ( self , a__ , a__ ) -> List[Any]: '''simple docstring''' A_ = tokenizer.get_vocab() A_ = sorted(vocab.keys() )[:2] # Pipeline argument A_ = FillMaskPipeline(model=a__ , tokenizer=a__ , targets=a__ ) A_ = fill_masker(F"This is a {tokenizer.mask_token}" ) self.assertEqual( a__ , [ {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, ] , ) A_ = {vocab[el] for el in targets} self.assertEqual({el['''token'''] for el in outputs} , a__ ) A_ = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el['''token_str'''] for el in outputs} , set(a__ ) ) # Call argument A_ = FillMaskPipeline(model=a__ , tokenizer=a__ ) A_ = fill_masker(F"This is a {tokenizer.mask_token}" , targets=a__ ) self.assertEqual( a__ , [ {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, ] , ) A_ = {vocab[el] for el in targets} self.assertEqual({el['''token'''] for el in outputs} , a__ ) A_ = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el['''token_str'''] for el in outputs} , set(a__ ) ) # Score equivalence A_ = fill_masker(F"This is a {tokenizer.mask_token}" , targets=a__ ) A_ = [top_mask['''token_str'''] for top_mask in outputs] A_ = [top_mask['''score'''] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(a__ ) == set(a__ ): A_ = fill_masker(F"This is a {tokenizer.mask_token}" , targets=a__ ) A_ = [top_mask['''score'''] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(a__ ) , nested_simplify(a__ ) ) # Raises with invalid with self.assertRaises(a__ ): A_ = fill_masker(F"This is a {tokenizer.mask_token}" , targets=[] ) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(a__ ): A_ = fill_masker(F"This is a {tokenizer.mask_token}" , targets=[''''''] ) with self.assertRaises(a__ ): A_ = fill_masker(F"This is a {tokenizer.mask_token}" , targets='''''' ) def lowerCAmelCase_ ( self , a__ , a__ ) -> int: '''simple docstring''' A_ = FillMaskPipeline(model=a__ , tokenizer=a__ , top_k=2 ) A_ = fill_masker(F"This is a {tokenizer.mask_token}" ) self.assertEqual( a__ , [ {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, ] , ) A_ = FillMaskPipeline(model=a__ , tokenizer=a__ ) A_ = fill_masker(F"This is a {tokenizer.mask_token}" , top_k=2 ) self.assertEqual( a__ , [ {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, ] , ) self.assertEqual(nested_simplify(a__ ) , nested_simplify(a__ ) ) def lowerCAmelCase_ ( self , a__ , a__ ) -> str: '''simple docstring''' A_ = tokenizer.get_vocab() A_ = FillMaskPipeline(model=a__ , tokenizer=a__ ) # top_k=2, ntargets=3 A_ = sorted(vocab.keys() )[:3] A_ = fill_masker(F"This is a {tokenizer.mask_token}" , top_k=2 , targets=a__ ) # If we use the most probably targets, and filter differently, we should still # have the same results A_ = [el['''token_str'''] for el in sorted(a__ , key=lambda a__ : x["score"] , reverse=a__ )] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(a__ ).issubset(a__ ): A_ = fill_masker(F"This is a {tokenizer.mask_token}" , top_k=3 , targets=a__ ) # They should yield exactly the same result self.assertEqual(nested_simplify(a__ ) , nested_simplify(a__ ) ) def lowerCAmelCase_ ( self , a__ , a__ ) -> Union[str, Any]: '''simple docstring''' A_ = FillMaskPipeline(model=a__ , tokenizer=a__ ) A_ = tokenizer.get_vocab() # String duplicates + id duplicates A_ = sorted(vocab.keys() )[:3] A_ = [targets[0], targets[1], targets[0], targets[2], targets[1]] A_ = fill_masker(F"My name is {tokenizer.mask_token}" , targets=a__ , top_k=1_0 ) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(a__ ) , 3 ) def lowerCAmelCase_ ( self , a__ , a__ ) -> List[Any]: '''simple docstring''' A_ = FillMaskPipeline(model=a__ , tokenizer=a__ ) A_ = fill_masker( F"This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}" , top_k=2 ) self.assertEqual( a__ , [ [ {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, ], [ {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, ], [ {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, {'''sequence''': ANY(a__ ), '''score''': ANY(a__ ), '''token''': ANY(a__ ), '''token_str''': ANY(a__ )}, ], ] , )
141
from __future__ import annotations import csv import requests from bsa import BeautifulSoup def lowerCamelCase_ ( __UpperCamelCase = "" ): A_ = url or '''https://www.imdb.com/chart/top/?ref_=nv_mv_250''' A_ = BeautifulSoup(requests.get(__UpperCamelCase ).text , '''html.parser''' ) A_ = soup.find_all('''td''' , attrs='''titleColumn''' ) A_ = soup.find_all('''td''' , class_='''ratingColumn imdbRating''' ) return { title.a.text: float(rating.strong.text ) for title, rating in zip(__UpperCamelCase , __UpperCamelCase ) } def lowerCamelCase_ ( __UpperCamelCase = "IMDb_Top_250_Movies.csv" ): A_ = get_imdb_top_aaa_movies() with open(__UpperCamelCase , '''w''' , newline='''''' ) as out_file: A_ = csv.writer(__UpperCamelCase ) writer.writerow(['''Movie title''', '''IMDb rating'''] ) for title, rating in movies.items(): writer.writerow([title, rating] ) if __name__ == "__main__": write_movies()
141
1
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @property def _UpperCamelCase ( self ): '''simple docstring''' torch.manual_seed(0 ) snake_case: Union[str, Any] = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model @property def _UpperCamelCase ( self ): '''simple docstring''' torch.manual_seed(0 ) snake_case: str = VQModel( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=3 , ) return model @property def _UpperCamelCase ( self ): '''simple docstring''' torch.manual_seed(0 ) snake_case: Tuple = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModel(SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self ): '''simple docstring''' snake_case: Optional[int] = self.dummy_uncond_unet snake_case: Optional[Any] = DDIMScheduler() snake_case: Union[str, Any] = self.dummy_vq_model snake_case: Union[str, Any] = LDMPipeline(unet=SCREAMING_SNAKE_CASE__ , vqvae=SCREAMING_SNAKE_CASE__ , scheduler=SCREAMING_SNAKE_CASE__ ) ldm.to(SCREAMING_SNAKE_CASE__ ) ldm.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) snake_case: Tuple = torch.manual_seed(0 ) snake_case: Dict = ldm(generator=SCREAMING_SNAKE_CASE__ , num_inference_steps=2 , output_type='numpy' ).images snake_case: Union[str, Any] = torch.manual_seed(0 ) snake_case: List[Any] = ldm(generator=SCREAMING_SNAKE_CASE__ , num_inference_steps=2 , output_type='numpy' , return_dict=SCREAMING_SNAKE_CASE__ )[0] snake_case: Union[str, Any] = image[0, -3:, -3:, -1] snake_case: Tuple = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) snake_case: Optional[int] = np.array([0.85_12, 0.8_18, 0.64_11, 0.68_08, 0.44_65, 0.56_18, 0.46, 0.62_31, 0.51_72] ) snake_case: Union[str, Any] = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def _UpperCamelCase ( self ): '''simple docstring''' snake_case: Optional[int] = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(SCREAMING_SNAKE_CASE__ ) ldm.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) snake_case: int = torch.manual_seed(0 ) snake_case: Union[str, Any] = ldm(generator=SCREAMING_SNAKE_CASE__ , num_inference_steps=5 , output_type='numpy' ).images snake_case: Any = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) snake_case: Optional[int] = np.array([0.43_99, 0.4_49_75, 0.4_68_25, 0.4_74, 0.43_59, 0.45_81, 0.4_50_95, 0.43_41, 0.44_47] ) snake_case: List[Any] = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
692
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = "▁" __UpperCAmelCase = {"vocab_file": "sentencepiece.bpe.model"} __UpperCAmelCase = { "vocab_file": { "facebook/xglm-564M": "https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model", } } __UpperCAmelCase = { "facebook/xglm-564M": 2_048, } class SCREAMING_SNAKE_CASE ( snake_case ): '''simple docstring''' __UpperCamelCase = VOCAB_FILES_NAMES __UpperCamelCase = PRETRAINED_VOCAB_FILES_MAP __UpperCamelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCamelCase = ["input_ids", "attention_mask"] def __init__( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__="<s>" , SCREAMING_SNAKE_CASE__="</s>" , SCREAMING_SNAKE_CASE__="</s>" , SCREAMING_SNAKE_CASE__="<s>" , SCREAMING_SNAKE_CASE__="<unk>" , SCREAMING_SNAKE_CASE__="<pad>" , SCREAMING_SNAKE_CASE__ = None , **SCREAMING_SNAKE_CASE__ , ): '''simple docstring''' snake_case: Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer snake_case: Optional[Any] = 7 snake_case: List[str] = [F"""<madeupword{i}>""" for i in range(self.num_madeup_words )] snake_case: str = kwargs.get('additional_special_tokens' , [] ) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=SCREAMING_SNAKE_CASE__ , eos_token=SCREAMING_SNAKE_CASE__ , unk_token=SCREAMING_SNAKE_CASE__ , sep_token=SCREAMING_SNAKE_CASE__ , cls_token=SCREAMING_SNAKE_CASE__ , pad_token=SCREAMING_SNAKE_CASE__ , sp_model_kwargs=self.sp_model_kwargs , **SCREAMING_SNAKE_CASE__ , ) snake_case: int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(SCREAMING_SNAKE_CASE__ ) ) snake_case: int = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab snake_case: Tuple = 1 # Mimic fairseq token-to-id alignment for the first 4 token snake_case: Optional[Any] = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} snake_case: Union[str, Any] = len(self.sp_model ) snake_case: str = {F"""<madeupword{i}>""": sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words )} self.fairseq_tokens_to_ids.update(SCREAMING_SNAKE_CASE__ ) snake_case: Union[str, Any] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ): '''simple docstring''' snake_case: List[Any] = self.__dict__.copy() snake_case: Union[str, Any] = None snake_case: Union[str, Any] = self.sp_model.serialized_model_proto() return state def __setstate__( self , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' snake_case: Optional[int] = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): snake_case: Union[str, Any] = {} snake_case: Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def _UpperCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None ): '''simple docstring''' if token_ids_a is None: return [self.sep_token_id] + token_ids_a snake_case: Optional[Any] = [self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def _UpperCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None , SCREAMING_SNAKE_CASE__ = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=SCREAMING_SNAKE_CASE__ , token_ids_a=SCREAMING_SNAKE_CASE__ , already_has_special_tokens=SCREAMING_SNAKE_CASE__ ) if token_ids_a is None: return [1] + ([0] * len(SCREAMING_SNAKE_CASE__ )) return [1] + ([0] * len(SCREAMING_SNAKE_CASE__ )) + [1, 1] + ([0] * len(SCREAMING_SNAKE_CASE__ )) def _UpperCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None ): '''simple docstring''' snake_case: int = [self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a ) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a ) * [0] @property def _UpperCamelCase ( self ): '''simple docstring''' return len(self.sp_model ) + self.fairseq_offset + self.num_madeup_words def _UpperCamelCase ( self ): '''simple docstring''' snake_case: Optional[int] = {self.convert_ids_to_tokens(SCREAMING_SNAKE_CASE__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _UpperCamelCase ( self , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' return self.sp_model.encode(SCREAMING_SNAKE_CASE__ , out_type=SCREAMING_SNAKE_CASE__ ) def _UpperCamelCase ( self , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] snake_case: Dict = self.sp_model.PieceToId(SCREAMING_SNAKE_CASE__ ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def _UpperCamelCase ( self , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def _UpperCamelCase ( self , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' snake_case: Optional[Any] = ''.join(SCREAMING_SNAKE_CASE__ ).replace(SCREAMING_SNAKE_CASE__ , ' ' ).strip() return out_string def _UpperCamelCase ( self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = None ): '''simple docstring''' if not os.path.isdir(SCREAMING_SNAKE_CASE__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return snake_case: List[str] = os.path.join( SCREAMING_SNAKE_CASE__ , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(SCREAMING_SNAKE_CASE__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , SCREAMING_SNAKE_CASE__ ) elif not os.path.isfile(self.vocab_file ): with open(SCREAMING_SNAKE_CASE__ , 'wb' ) as fi: snake_case: int = self.sp_model.serialized_model_proto() fi.write(SCREAMING_SNAKE_CASE__ ) return (out_vocab_file,)
692
1
"""simple docstring""" from collections.abc import Iterable from typing import Generic, TypeVar a : Union[str, Any] = TypeVar('_T') class lowercase(Generic[_T] ): def __init__( self , __SCREAMING_SNAKE_CASE = None ) -> None: """simple docstring""" a__ = list(iterable or [] ) a__ = [] def __len__( self ) -> int: """simple docstring""" return len(self._stacka ) + len(self._stacka ) def __repr__( self ) -> str: """simple docstring""" return f'Queue({tuple(self._stacka[::-1] + self._stacka )})' def lowercase__ ( self , __SCREAMING_SNAKE_CASE ) -> None: """simple docstring""" self._stacka.append(__a ) def lowercase__ ( self ) -> _T: """simple docstring""" a__ = self._stacka.pop a__ = self._stacka.append if not self._stacka: while self._stacka: stacka_append(stacka_pop() ) if not self._stacka: raise IndexError('Queue is empty' ) return self._stacka.pop() if __name__ == "__main__": from doctest import testmod testmod()
273
import math_equivalence # From: git+https://github.com/hendrycks/math.git import datasets _UpperCamelCase = '''\ @article{hendrycksmath2021, title={Measuring Mathematical Problem Solving With the MATH Dataset}, author={Dan Hendrycks and Collin Burns and Saurav Kadavath and Akul Arora and Steven Basart and Eric Tang and Dawn Song and Jacob Steinhardt}, journal={arXiv preprint arXiv:2103.03874}, year={2021} } ''' _UpperCamelCase = '''\ This metric is used to assess performance on the Mathematics Aptitude Test of Heuristics (MATH) dataset. It first canonicalizes the inputs (e.g., converting "1/2" to "\\frac{1}{2}") and then computes accuracy. ''' _UpperCamelCase = R''' Calculates accuracy after canonicalizing inputs. Args: predictions: list of predictions to score. Each prediction is a string that contains natural language and LaTex. references: list of reference for each prediction. Each reference is a string that contains natural language and LaTex. Returns: accuracy: accuracy after canonicalizing inputs (e.g., converting "1/2" to "\\frac{1}{2}") Examples: >>> metric = datasets.load_metric("competition_math") >>> results = metric.compute(references=["\\frac{1}{2}"], predictions=["1/2"]) >>> print(results) {\'accuracy\': 1.0} ''' @datasets.utils.file_utils.add_end_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase ( datasets.Metric ): '''simple docstring''' def UpperCamelCase__ (self ) -> List[str]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' ), 'references': datasets.Value('string' ), } ) , homepage='https://github.com/hendrycks/math' , codebase_urls=['https://github.com/hendrycks/math'] , ) def UpperCamelCase__ (self , __a , __a ) -> int: """simple docstring""" UpperCAmelCase__ = 0.0 for i, j in zip(__a , __a ): n_correct += 1.0 if math_equivalence.is_equiv(__a , __a ) else 0.0 UpperCAmelCase__ = n_correct / len(__a ) return { "accuracy": accuracy, }
146
0
"""simple docstring""" from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def _snake_case ( ): A__ = HfArgumentParser(lowercase__ ) A__ = parser.parse_args_into_dataclasses()[0] A__ = TensorFlowBenchmark(args=lowercase__ ) try: A__ = parser.parse_args_into_dataclasses()[0] except ValueError as e: A__ = """Arg --no_{0} is no longer used, please use --no-{0} instead.""" A__ = """ """.join(str(lowercase__ ).split(""" """ )[:-1] ) A__ = """""" A__ = eval(str(lowercase__ ).split(""" """ )[-1] ) A__ = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:] ) else: wrong_args.append(lowercase__ ) if len(lowercase__ ) > 0: A__ = full_error_msg + begin_error_msg + str(lowercase__ ) raise ValueError(lowercase__ ) benchmark.run() if __name__ == "__main__": main()
719
"""simple docstring""" import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE_ : Union[str, Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ : Any = ['model.decoder.embed_positions.weights'] def _snake_case ( UpperCAmelCase_ : Optional[int] ): if "emb" in name: A__ = name.replace("""emb""" , """model.decoder.embed_tokens""" ) if "transformer" in name: A__ = name.replace("""transformer""" , """model.decoder""" ) if "cross_attention" in name: A__ = name.replace("""cross_attention""" , """encoder_attn""" ) if "linear1" in name: A__ = name.replace("""linear1""" , """fc1""" ) if "linear2" in name: A__ = name.replace("""linear2""" , """fc2""" ) if "norm1" in name: A__ = name.replace("""norm1""" , """self_attn_layer_norm""" ) if "norm_cross" in name: A__ = name.replace("""norm_cross""" , """encoder_attn_layer_norm""" ) if "norm2" in name: A__ = name.replace("""norm2""" , """final_layer_norm""" ) if "out_norm" in name: A__ = name.replace("""out_norm""" , """model.decoder.layer_norm""" ) if "linears" in name: A__ = name.replace("""linears""" , """lm_heads""" ) if "condition_provider.conditioners.description.output_proj" in name: A__ = name.replace("""condition_provider.conditioners.description.output_proj""" , """enc_to_dec_proj""" ) return name def _snake_case ( UpperCAmelCase_ : OrderedDict , UpperCAmelCase_ : int ): A__ = list(state_dict.keys() ) A__ = {} for key in keys: A__ = state_dict.pop(UpperCAmelCase_ ) A__ = rename_keys(UpperCAmelCase_ ) if "in_proj_weight" in key: # split fused qkv proj A__ = val[:hidden_size, :] A__ = val[hidden_size : 2 * hidden_size, :] A__ = val[-hidden_size:, :] elif "enc_to_dec_proj" in key: A__ = val else: A__ = val return state_dict, enc_dec_proj_state_dict def _snake_case ( UpperCAmelCase_ : str ): if checkpoint == "small": # default config values A__ = 1024 A__ = 24 A__ = 16 elif checkpoint == "medium": A__ = 1536 A__ = 48 A__ = 24 elif checkpoint == "large": A__ = 2048 A__ = 48 A__ = 32 else: raise ValueError(F"""Checkpoint should be one of `['small', 'medium', 'large']`, got {checkpoint}.""" ) A__ = MusicgenDecoderConfig( hidden_size=UpperCAmelCase_ , ffn_dim=hidden_size * 4 , num_hidden_layers=UpperCAmelCase_ , num_attention_heads=UpperCAmelCase_ , ) return config @torch.no_grad() def _snake_case ( UpperCAmelCase_ : int , UpperCAmelCase_ : List[Any]=None , UpperCAmelCase_ : Tuple=None , UpperCAmelCase_ : Any="cpu" ): A__ = MusicGen.get_pretrained(UpperCAmelCase_ , device=UpperCAmelCase_ ) A__ = decoder_config_from_checkpoint(UpperCAmelCase_ ) A__ = fairseq_model.lm.state_dict() A__ , A__ = rename_state_dict( UpperCAmelCase_ , hidden_size=decoder_config.hidden_size ) A__ = TaEncoderModel.from_pretrained("""t5-base""" ) A__ = EncodecModel.from_pretrained("""facebook/encodec_32khz""" ) A__ = MusicgenForCausalLM(UpperCAmelCase_ ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection A__ , A__ = decoder.load_state_dict(UpperCAmelCase_ , strict=UpperCAmelCase_ ) for key in missing_keys.copy(): if key.startswith(("""text_encoder""", """audio_encoder""") ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(UpperCAmelCase_ ) if len(UpperCAmelCase_ ) > 0: raise ValueError(F"""Missing key(s) in state_dict: {missing_keys}""" ) if len(UpperCAmelCase_ ) > 0: raise ValueError(F"""Unexpected key(s) in state_dict: {unexpected_keys}""" ) # init the composite model A__ = MusicgenForConditionalGeneration(text_encoder=UpperCAmelCase_ , audio_encoder=UpperCAmelCase_ , decoder=UpperCAmelCase_ ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(UpperCAmelCase_ ) # check we can do a forward pass A__ = torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) A__ = input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): A__ = model(input_ids=UpperCAmelCase_ , decoder_input_ids=UpperCAmelCase_ ).logits if logits.shape != (8, 1, 2048): raise ValueError("""Incorrect shape for logits""" ) # now construct the processor A__ = AutoTokenizer.from_pretrained("""t5-base""" ) A__ = AutoFeatureExtractor.from_pretrained("""facebook/encodec_32khz""" , padding_side="""left""" ) A__ = MusicgenProcessor(feature_extractor=UpperCAmelCase_ , tokenizer=UpperCAmelCase_ ) # set the appropriate bos/pad token ids A__ = 2048 A__ = 2048 # set other default generation config params A__ = int(30 * audio_encoder.config.frame_rate ) A__ = True A__ = 3.0 if pytorch_dump_folder is not None: Path(UpperCAmelCase_ ).mkdir(exist_ok=UpperCAmelCase_ ) logger.info(F"""Saving model {checkpoint} to {pytorch_dump_folder}""" ) model.save_pretrained(UpperCAmelCase_ ) processor.save_pretrained(UpperCAmelCase_ ) if repo_id: logger.info(F"""Pushing model {checkpoint} to {repo_id}""" ) model.push_to_hub(UpperCAmelCase_ ) processor.push_to_hub(UpperCAmelCase_ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint', default='small', type=str, help='Checkpoint size of the MusicGen model you\'d like to convert. Can be one of: `[\'small\', \'medium\', \'large\']`.', ) parser.add_argument( '--pytorch_dump_folder', required=True, default=None, type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) parser.add_argument( '--device', default='cpu', type=str, help='Torch device to run the conversion, either cpu or cuda.' ) SCREAMING_SNAKE_CASE_ : Tuple = parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
500
0
from .imports import is_tqdm_available if is_tqdm_available(): from tqdm.auto import tqdm as _tqdm from ..state import PartialState def __lowerCAmelCase ( A = True , *A , **A ): if not is_tqdm_available(): raise ImportError("Accelerate's `tqdm` module requires `tqdm` to be installed. Please run `pip install tqdm`." ) UpperCAmelCase_ = False if main_process_only: UpperCAmelCase_ = PartialState().local_process_index == 0 return _tqdm(*A , **A , disable=A )
162
def _UpperCAmelCase ( UpperCAmelCase : int = 600_851_475_143 ): """simple docstring""" try: __lowerCamelCase : Any = int(UpperCAmelCase ) except (TypeError, ValueError): raise TypeError("""Parameter n must be int or castable to int.""" ) if n <= 0: raise ValueError("""Parameter n must be greater than or equal to one.""" ) __lowerCamelCase : int = 1 __lowerCamelCase : str = 2 while i * i <= n: while n % i == 0: __lowerCamelCase : Union[str, Any] = i n //= i i += 1 if n > 1: __lowerCamelCase : Dict = n return int(UpperCAmelCase ) if __name__ == "__main__": print(F'''{solution() = }''')
519
0
import argparse import glob import logging import os import sys import time from collections import defaultdict from pathlib import Path from typing import Dict, List, Tuple import numpy as np import pytorch_lightning as pl import torch from callbacks import SeqaSeqLoggingCallback, get_checkpoint_callback, get_early_stopping_callback from torch import nn from torch.utils.data import DataLoader from transformers import MBartTokenizer, TaForConditionalGeneration from transformers.models.bart.modeling_bart import shift_tokens_right from utils import ( ROUGE_KEYS, LegacySeqaSeqDataset, SeqaSeqDataset, assert_all_frozen, calculate_bleu, calculate_rouge, check_output_dir, flatten_list, freeze_embeds, freeze_params, get_git_info, label_smoothed_nll_loss, lmap, pickle_save, save_git_info, save_json, use_task_specific_params, ) # need the parent dir module sys.path.insert(2, str(Path(__file__).resolve().parents[1])) from lightning_base import BaseTransformer, add_generic_args, generic_train # noqa _snake_case : List[str] = logging.getLogger(__name__) class _UpperCAmelCase ( lowercase__ ): UpperCamelCase = 'summarization' UpperCamelCase = ['loss'] UpperCamelCase = ROUGE_KEYS UpperCamelCase = 'rouge2' def __init__( self :Any , __UpperCamelCase :Optional[int] , **__UpperCamelCase :Dict ): if hparams.sortish_sampler and hparams.gpus > 1: A = False elif hparams.max_tokens_per_batch is not None: if hparams.gpus > 1: raise NotImplementedError("Dynamic Batch size does not work for multi-gpu training" ) if hparams.sortish_sampler: raise ValueError("--sortish_sampler and --max_tokens_per_batch may not be used simultaneously" ) super().__init__(__UpperCamelCase , num_labels=__UpperCamelCase , mode=self.mode , **__UpperCamelCase ) use_task_specific_params(self.model , "summarization" ) save_git_info(self.hparams.output_dir ) A = Path(self.output_dir ) / "metrics.json" A = Path(self.output_dir ) / "hparams.pkl" pickle_save(self.hparams , self.hparams_save_path ) A = 0 A = defaultdict(__UpperCamelCase ) A = self.config.model_type A = self.config.tgt_vocab_size if self.model_type == "fsmt" else self.config.vocab_size A = { "data_dir": self.hparams.data_dir, "max_source_length": self.hparams.max_source_length, "prefix": self.model.config.prefix or "", } A = { "train": self.hparams.n_train, "val": self.hparams.n_val, "test": self.hparams.n_test, } A = {k: v if v >= 0 else None for k, v in n_observations_per_split.items()} A = { "train": self.hparams.max_target_length, "val": self.hparams.val_max_target_length, "test": self.hparams.test_max_target_length, } assert self.target_lens["train"] <= self.target_lens["val"], f"target_lens: {self.target_lens}" assert self.target_lens["train"] <= self.target_lens["test"], f"target_lens: {self.target_lens}" if self.hparams.freeze_embeds: freeze_embeds(self.model ) if self.hparams.freeze_encoder: freeze_params(self.model.get_encoder() ) assert_all_frozen(self.model.get_encoder() ) A = get_git_info()["repo_sha"] A = hparams.num_workers A = None # default to config if self.model.config.decoder_start_token_id is None and isinstance(self.tokenizer , __UpperCamelCase ): A = self.tokenizer.lang_code_to_id[hparams.tgt_lang] A = self.decoder_start_token_id A = ( SeqaSeqDataset if hasattr(self.tokenizer , "prepare_seq2seq_batch" ) else LegacySeqaSeqDataset ) A = False A = self.model.config.num_beams if self.hparams.eval_beams is None else self.hparams.eval_beams if self.hparams.eval_max_gen_length is not None: A = self.hparams.eval_max_gen_length else: A = self.model.config.max_length A = self.default_val_metric if self.hparams.val_metric is None else self.hparams.val_metric def lowerCamelCase ( self :List[str] , __UpperCamelCase :List[str] ): A = { k: self.tokenizer.batch_decode(v.tolist() ) if "mask" not in k else v.shape for k, v in batch.items() } save_json(__UpperCamelCase , Path(self.output_dir ) / "text_batch.json" ) save_json({k: v.tolist() for k, v in batch.items()} , Path(self.output_dir ) / "tok_batch.json" ) A = True return readable_batch def lowerCamelCase ( self :Optional[int] , __UpperCamelCase :Dict , **__UpperCamelCase :Union[str, Any] ): return self.model(__UpperCamelCase , **__UpperCamelCase ) def lowerCamelCase ( self :Tuple , __UpperCamelCase :Dict ): A = self.tokenizer.batch_decode( __UpperCamelCase , skip_special_tokens=__UpperCamelCase , clean_up_tokenization_spaces=__UpperCamelCase ) return lmap(str.strip , __UpperCamelCase ) def lowerCamelCase ( self :Any , __UpperCamelCase :int ): A = self.tokenizer.pad_token_id A, A = batch["input_ids"], batch["attention_mask"] A = batch["labels"] if isinstance(self.model , __UpperCamelCase ): A = self.model._shift_right(__UpperCamelCase ) else: A = shift_tokens_right(__UpperCamelCase , __UpperCamelCase ) if not self.already_saved_batch: # This would be slightly better if it only happened on rank zero A = decoder_input_ids self.save_readable_batch(__UpperCamelCase ) A = self(__UpperCamelCase , attention_mask=__UpperCamelCase , decoder_input_ids=__UpperCamelCase , use_cache=__UpperCamelCase ) A = outputs["logits"] if self.hparams.label_smoothing == 0: # Same behavior as modeling_bart.py, besides ignoring pad_token_id A = nn.CrossEntropyLoss(ignore_index=__UpperCamelCase ) assert lm_logits.shape[-1] == self.vocab_size A = ce_loss_fct(lm_logits.view(-1 , lm_logits.shape[-1] ) , tgt_ids.view(-1 ) ) else: A = nn.functional.log_softmax(__UpperCamelCase , dim=-1 ) A, A = label_smoothed_nll_loss( __UpperCamelCase , __UpperCamelCase , self.hparams.label_smoothing , ignore_index=__UpperCamelCase ) return (loss,) @property def lowerCamelCase ( self :Tuple ): return self.tokenizer.pad_token_id def lowerCamelCase ( self :Any , __UpperCamelCase :List[str] , __UpperCamelCase :Optional[Any] ): A = self._step(__UpperCamelCase ) A = dict(zip(self.loss_names , __UpperCamelCase ) ) # tokens per batch A = batch["input_ids"].ne(self.pad ).sum() + batch["labels"].ne(self.pad ).sum() A = batch["input_ids"].shape[0] A = batch["input_ids"].eq(self.pad ).sum() A = batch["input_ids"].eq(self.pad ).float().mean() # TODO(SS): make a wandb summary metric for this return {"loss": loss_tensors[0], "log": logs} def lowerCamelCase ( self :Union[str, Any] , __UpperCamelCase :Union[str, Any] , __UpperCamelCase :Dict ): return self._generative_step(__UpperCamelCase ) def lowerCamelCase ( self :str , __UpperCamelCase :Union[str, Any] , __UpperCamelCase :Tuple="val" ): self.step_count += 1 A = {k: torch.stack([x[k] for x in outputs] ).mean() for k in self.loss_names} A = losses["loss"] A = { k: np.array([x[k] for x in outputs] ).mean() for k in self.metric_names + ["gen_time", "gen_len"] } A = ( generative_metrics[self.val_metric] if self.val_metric in generative_metrics else losses[self.val_metric] ) A = torch.tensor(__UpperCamelCase ).type_as(__UpperCamelCase ) generative_metrics.update({k: v.item() for k, v in losses.items()} ) losses.update(__UpperCamelCase ) A = {f"{prefix}_avg_{k}": x for k, x in losses.items()} A = self.step_count self.metrics[prefix].append(__UpperCamelCase ) # callback writes this to self.metrics_save_path A = flatten_list([x["preds"] for x in outputs] ) return { "log": all_metrics, "preds": preds, f"{prefix}_loss": loss, f"{prefix}_{self.val_metric}": metric_tensor, } def lowerCamelCase ( self :Optional[Any] , __UpperCamelCase :Optional[Any] , __UpperCamelCase :List[Any] ): return calculate_rouge(__UpperCamelCase , __UpperCamelCase ) def lowerCamelCase ( self :List[str] , __UpperCamelCase :List[Any] ): A = time.time() # parser.add_argument('--eval_max_gen_length', type=int, default=None, help='never generate more than n tokens') A = self.model.generate( batch["input_ids"] , attention_mask=batch["attention_mask"] , use_cache=__UpperCamelCase , decoder_start_token_id=self.decoder_start_token_id , num_beams=self.eval_beams , max_length=self.eval_max_length , ) A = (time.time() - ta) / batch["input_ids"].shape[0] A = self.ids_to_clean_text(__UpperCamelCase ) A = self.ids_to_clean_text(batch["labels"] ) A = self._step(__UpperCamelCase ) A = dict(zip(self.loss_names , __UpperCamelCase ) ) A = self.calc_generative_metrics(__UpperCamelCase , __UpperCamelCase ) A = np.mean(lmap(__UpperCamelCase , __UpperCamelCase ) ) base_metrics.update(gen_time=__UpperCamelCase , gen_len=__UpperCamelCase , preds=__UpperCamelCase , target=__UpperCamelCase , **__UpperCamelCase ) return base_metrics def lowerCamelCase ( self :Any , __UpperCamelCase :int , __UpperCamelCase :List[str] ): return self._generative_step(__UpperCamelCase ) def lowerCamelCase ( self :List[str] , __UpperCamelCase :List[Any] ): return self.validation_epoch_end(__UpperCamelCase , prefix="test" ) def lowerCamelCase ( self :Optional[int] , __UpperCamelCase :Optional[int] ): A = self.n_obs[type_path] A = self.target_lens[type_path] A = self.dataset_class( self.tokenizer , type_path=__UpperCamelCase , n_obs=__UpperCamelCase , max_target_length=__UpperCamelCase , **self.dataset_kwargs , ) return dataset def lowerCamelCase ( self :Any , __UpperCamelCase :Tuple , __UpperCamelCase :str , __UpperCamelCase :List[str] = False ): A = self.get_dataset(__UpperCamelCase ) if self.hparams.sortish_sampler and type_path != "test" and type_path != "val": A = dataset.make_sortish_sampler(__UpperCamelCase , distributed=self.hparams.gpus > 1 ) return DataLoader( __UpperCamelCase , batch_size=__UpperCamelCase , collate_fn=dataset.collate_fn , shuffle=__UpperCamelCase , num_workers=self.num_workers , sampler=__UpperCamelCase , ) elif self.hparams.max_tokens_per_batch is not None and type_path != "test" and type_path != "val": A = dataset.make_dynamic_sampler( self.hparams.max_tokens_per_batch , distributed=self.hparams.gpus > 1 ) return DataLoader( __UpperCamelCase , batch_sampler=__UpperCamelCase , collate_fn=dataset.collate_fn , num_workers=self.num_workers , ) else: return DataLoader( __UpperCamelCase , batch_size=__UpperCamelCase , collate_fn=dataset.collate_fn , shuffle=__UpperCamelCase , num_workers=self.num_workers , sampler=__UpperCamelCase , ) def lowerCamelCase ( self :Tuple ): A = self.get_dataloader("train" , batch_size=self.hparams.train_batch_size , shuffle=__UpperCamelCase ) return dataloader def lowerCamelCase ( self :str ): return self.get_dataloader("val" , batch_size=self.hparams.eval_batch_size ) def lowerCamelCase ( self :int ): return self.get_dataloader("test" , batch_size=self.hparams.eval_batch_size ) @staticmethod def lowerCamelCase ( __UpperCamelCase :int , __UpperCamelCase :List[str] ): BaseTransformer.add_model_specific_args(__UpperCamelCase , __UpperCamelCase ) add_generic_args(__UpperCamelCase , __UpperCamelCase ) parser.add_argument( "--max_source_length" , default=10_24 , type=__UpperCamelCase , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--max_target_length" , default=56 , type=__UpperCamelCase , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--val_max_target_length" , default=1_42 , type=__UpperCamelCase , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument( "--test_max_target_length" , default=1_42 , type=__UpperCamelCase , help=( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) , ) parser.add_argument("--freeze_encoder" , action="store_true" ) parser.add_argument("--freeze_embeds" , action="store_true" ) parser.add_argument("--sortish_sampler" , action="store_true" , default=__UpperCamelCase ) parser.add_argument("--overwrite_output_dir" , action="store_true" , default=__UpperCamelCase ) parser.add_argument("--max_tokens_per_batch" , type=__UpperCamelCase , default=__UpperCamelCase ) parser.add_argument("--logger_name" , type=__UpperCamelCase , choices=["default", "wandb", "wandb_shared"] , default="default" ) parser.add_argument("--n_train" , type=__UpperCamelCase , default=-1 , required=__UpperCamelCase , help="# examples. -1 means use all." ) parser.add_argument("--n_val" , type=__UpperCamelCase , default=5_00 , required=__UpperCamelCase , help="# examples. -1 means use all." ) parser.add_argument("--n_test" , type=__UpperCamelCase , default=-1 , required=__UpperCamelCase , help="# examples. -1 means use all." ) parser.add_argument( "--task" , type=__UpperCamelCase , default="summarization" , required=__UpperCamelCase , help="# examples. -1 means use all." ) parser.add_argument("--label_smoothing" , type=__UpperCamelCase , default=0.0 , required=__UpperCamelCase ) parser.add_argument("--src_lang" , type=__UpperCamelCase , default="" , required=__UpperCamelCase ) parser.add_argument("--tgt_lang" , type=__UpperCamelCase , default="" , required=__UpperCamelCase ) parser.add_argument("--eval_beams" , type=__UpperCamelCase , default=__UpperCamelCase , required=__UpperCamelCase ) parser.add_argument( "--val_metric" , type=__UpperCamelCase , default=__UpperCamelCase , required=__UpperCamelCase , choices=["bleu", "rouge2", "loss", None] ) parser.add_argument("--eval_max_gen_length" , type=__UpperCamelCase , default=__UpperCamelCase , help="never generate more than n tokens" ) parser.add_argument("--save_top_k" , type=__UpperCamelCase , default=1 , required=__UpperCamelCase , help="How many checkpoints to save" ) parser.add_argument( "--early_stopping_patience" , type=__UpperCamelCase , default=-1 , required=__UpperCamelCase , help=( "-1 means never early stop. early_stopping_patience is measured in validation checks, not epochs. So" " val_check_interval will effect it." ) , ) return parser class _UpperCAmelCase ( lowercase__ ): UpperCamelCase = 'translation' UpperCamelCase = ['loss'] UpperCamelCase = ['bleu'] UpperCamelCase = 'bleu' def __init__( self :Dict , __UpperCamelCase :Union[str, Any] , **__UpperCamelCase :int ): super().__init__(__UpperCamelCase , **__UpperCamelCase ) A = hparams.src_lang A = hparams.tgt_lang def lowerCamelCase ( self :str , __UpperCamelCase :List[Any] , __UpperCamelCase :List[Any] ): return calculate_bleu(__UpperCamelCase , __UpperCamelCase ) def A__ ( UpperCamelCase , UpperCamelCase=None ): Path(args.output_dir ).mkdir(exist_ok=__SCREAMING_SNAKE_CASE ) check_output_dir(__SCREAMING_SNAKE_CASE , expected_items=3 ) if model is None: if "summarization" in args.task: A = SummarizationModule(__SCREAMING_SNAKE_CASE ) else: A = TranslationModule(__SCREAMING_SNAKE_CASE ) A = Path(args.data_dir ).name if ( args.logger_name == "default" or args.fast_dev_run or str(args.output_dir ).startswith("/tmp" ) or str(args.output_dir ).startswith("/var" ) ): A = True # don't pollute wandb logs unnecessarily elif args.logger_name == "wandb": from pytorch_lightning.loggers import WandbLogger A = os.environ.get("WANDB_PROJECT" , __SCREAMING_SNAKE_CASE ) A = WandbLogger(name=model.output_dir.name , project=__SCREAMING_SNAKE_CASE ) elif args.logger_name == "wandb_shared": from pytorch_lightning.loggers import WandbLogger A = WandbLogger(name=model.output_dir.name , project=F"hf_{dataset}" ) if args.early_stopping_patience >= 0: A = get_early_stopping_callback(model.val_metric , args.early_stopping_patience ) else: A = False A = args.val_metric == "loss" A = generic_train( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , logging_callback=SeqaSeqLoggingCallback() , checkpoint_callback=get_checkpoint_callback( args.output_dir , model.val_metric , args.save_top_k , __SCREAMING_SNAKE_CASE ) , early_stopping_callback=__SCREAMING_SNAKE_CASE , logger=__SCREAMING_SNAKE_CASE , ) pickle_save(model.hparams , model.output_dir / "hparams.pkl" ) if not args.do_predict: return model A = "" A = sorted(glob.glob(os.path.join(args.output_dir , "*.ckpt" ) , recursive=__SCREAMING_SNAKE_CASE ) ) if checkpoints: A = checkpoints[-1] A = checkpoints[-1] trainer.logger.log_hyperparams(model.hparams ) # test() without a model tests using the best checkpoint automatically trainer.test() return model if __name__ == "__main__": _snake_case : Optional[Any] = argparse.ArgumentParser() _snake_case : List[str] = pl.Trainer.add_argparse_args(parser) _snake_case : int = SummarizationModule.add_model_specific_args(parser, os.getcwd()) _snake_case : Tuple = parser.parse_args() main(args)
701
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging _snake_case : Optional[Any] = logging.get_logger(__name__) _snake_case : Optional[Any] = { 'Visual-Attention-Network/van-base': ( 'https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json' ), } class _UpperCAmelCase ( lowercase_ ): UpperCamelCase = '''van''' def __init__( self :Optional[int] , __UpperCamelCase :Tuple=2_24 , __UpperCamelCase :Tuple=3 , __UpperCamelCase :int=[7, 3, 3, 3] , __UpperCamelCase :List[str]=[4, 2, 2, 2] , __UpperCamelCase :str=[64, 1_28, 3_20, 5_12] , __UpperCamelCase :Union[str, Any]=[3, 3, 12, 3] , __UpperCamelCase :Dict=[8, 8, 4, 4] , __UpperCamelCase :List[Any]="gelu" , __UpperCamelCase :str=0.02 , __UpperCamelCase :str=1e-6 , __UpperCamelCase :Tuple=1e-2 , __UpperCamelCase :Optional[Any]=0.0 , __UpperCamelCase :List[Any]=0.0 , **__UpperCamelCase :List[str] , ): super().__init__(**__UpperCamelCase ) A = image_size A = num_channels A = patch_sizes A = strides A = hidden_sizes A = depths A = mlp_ratios A = hidden_act A = initializer_range A = layer_norm_eps A = layer_scale_init_value A = drop_path_rate A = dropout_rate
524
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowerCAmelCase : Any = logging.get_logger(__name__) lowerCAmelCase : Union[str, Any] = { """microsoft/swin-tiny-patch4-window7-224""": ( """https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json""" ), # See all Swin models at https://huggingface.co/models?filter=swin } class a ( __lowercase ,__lowercase ): SCREAMING_SNAKE_CASE__ : Union[str, Any] = """swin""" SCREAMING_SNAKE_CASE__ : Optional[int] = { """num_attention_heads""": """num_heads""", """num_hidden_layers""": """num_layers""", } def __init__( self , _lowerCAmelCase=224 , _lowerCAmelCase=4 , _lowerCAmelCase=3 , _lowerCAmelCase=96 , _lowerCAmelCase=[2, 2, 6, 2] , _lowerCAmelCase=[3, 6, 12, 24] , _lowerCAmelCase=7 , _lowerCAmelCase=4.0 , _lowerCAmelCase=True , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.1 , _lowerCAmelCase="gelu" , _lowerCAmelCase=False , _lowerCAmelCase=0.02 , _lowerCAmelCase=1e-5 , _lowerCAmelCase=32 , _lowerCAmelCase=None , _lowerCAmelCase=None , **_lowerCAmelCase , ): """simple docstring""" super().__init__(**_lowercase ) __SCREAMING_SNAKE_CASE: Any = image_size __SCREAMING_SNAKE_CASE: Optional[int] = patch_size __SCREAMING_SNAKE_CASE: Union[str, Any] = num_channels __SCREAMING_SNAKE_CASE: Dict = embed_dim __SCREAMING_SNAKE_CASE: Tuple = depths __SCREAMING_SNAKE_CASE: Union[str, Any] = len(_lowercase ) __SCREAMING_SNAKE_CASE: Any = num_heads __SCREAMING_SNAKE_CASE: Tuple = window_size __SCREAMING_SNAKE_CASE: Dict = mlp_ratio __SCREAMING_SNAKE_CASE: str = qkv_bias __SCREAMING_SNAKE_CASE: List[Any] = hidden_dropout_prob __SCREAMING_SNAKE_CASE: Optional[Any] = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE: Any = drop_path_rate __SCREAMING_SNAKE_CASE: Tuple = hidden_act __SCREAMING_SNAKE_CASE: Any = use_absolute_embeddings __SCREAMING_SNAKE_CASE: Optional[int] = layer_norm_eps __SCREAMING_SNAKE_CASE: Dict = initializer_range __SCREAMING_SNAKE_CASE: Any = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __SCREAMING_SNAKE_CASE: List[Any] = int(embed_dim * 2 ** (len(_lowercase ) - 1) ) __SCREAMING_SNAKE_CASE: Dict = ['''stem'''] + [f"""stage{idx}""" for idx in range(1 , len(_lowercase ) + 1 )] __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE: Union[str, Any] = get_aligned_output_features_output_indices( out_features=_lowercase , out_indices=_lowercase , stage_names=self.stage_names ) class a ( __lowercase ): SCREAMING_SNAKE_CASE__ : Any = version.parse('''1.11''' ) @property def snake_case_ ( self ): """simple docstring""" return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def snake_case_ ( self ): """simple docstring""" return 1e-4
202
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinvaConfig, SwinvaForImageClassification def lowerCAmelCase( SCREAMING_SNAKE_CASE_ )-> Optional[int]: """simple docstring""" UpperCamelCase_ = SwinvaConfig() UpperCamelCase_ = swinva_name.split("_" ) UpperCamelCase_ = name_split[1] if "to" in name_split[3]: UpperCamelCase_ = int(name_split[3][-3:] ) else: UpperCamelCase_ = int(name_split[3] ) if "to" in name_split[2]: UpperCamelCase_ = int(name_split[2][-2:] ) else: UpperCamelCase_ = int(name_split[2][6:] ) if model_size == "tiny": UpperCamelCase_ = 9_6 UpperCamelCase_ = (2, 2, 6, 2) UpperCamelCase_ = (3, 6, 1_2, 2_4) elif model_size == "small": UpperCamelCase_ = 9_6 UpperCamelCase_ = (2, 2, 1_8, 2) UpperCamelCase_ = (3, 6, 1_2, 2_4) elif model_size == "base": UpperCamelCase_ = 1_2_8 UpperCamelCase_ = (2, 2, 1_8, 2) UpperCamelCase_ = (4, 8, 1_6, 3_2) else: UpperCamelCase_ = 1_9_2 UpperCamelCase_ = (2, 2, 1_8, 2) UpperCamelCase_ = (6, 1_2, 2_4, 4_8) if "to" in swinva_name: UpperCamelCase_ = (1_2, 1_2, 1_2, 6) if ("22k" in swinva_name) and ("to" not in swinva_name): UpperCamelCase_ = 2_1_8_4_1 UpperCamelCase_ = "huggingface/label-files" UpperCamelCase_ = "imagenet-22k-id2label.json" UpperCamelCase_ = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , repo_type="dataset" ) , "r" ) ) UpperCamelCase_ = {int(SCREAMING_SNAKE_CASE_ ): v for k, v in idalabel.items()} UpperCamelCase_ = idalabel UpperCamelCase_ = {v: k for k, v in idalabel.items()} else: UpperCamelCase_ = 1_0_0_0 UpperCamelCase_ = "huggingface/label-files" UpperCamelCase_ = "imagenet-1k-id2label.json" UpperCamelCase_ = json.load(open(hf_hub_download(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , repo_type="dataset" ) , "r" ) ) UpperCamelCase_ = {int(SCREAMING_SNAKE_CASE_ ): v for k, v in idalabel.items()} UpperCamelCase_ = idalabel UpperCamelCase_ = {v: k for k, v in idalabel.items()} UpperCamelCase_ = img_size UpperCamelCase_ = num_classes UpperCamelCase_ = embed_dim UpperCamelCase_ = depths UpperCamelCase_ = num_heads UpperCamelCase_ = window_size return config def lowerCAmelCase( SCREAMING_SNAKE_CASE_ )-> Optional[Any]: """simple docstring""" if "patch_embed.proj" in name: UpperCamelCase_ = name.replace("patch_embed.proj" , "embeddings.patch_embeddings.projection" ) if "patch_embed.norm" in name: UpperCamelCase_ = name.replace("patch_embed.norm" , "embeddings.norm" ) if "layers" in name: UpperCamelCase_ = "encoder." + name if "attn.proj" in name: UpperCamelCase_ = name.replace("attn.proj" , "attention.output.dense" ) if "attn" in name: UpperCamelCase_ = name.replace("attn" , "attention.self" ) if "norm1" in name: UpperCamelCase_ = name.replace("norm1" , "layernorm_before" ) if "norm2" in name: UpperCamelCase_ = name.replace("norm2" , "layernorm_after" ) if "mlp.fc1" in name: UpperCamelCase_ = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: UpperCamelCase_ = name.replace("mlp.fc2" , "output.dense" ) if "q_bias" in name: UpperCamelCase_ = name.replace("q_bias" , "query.bias" ) if "k_bias" in name: UpperCamelCase_ = name.replace("k_bias" , "key.bias" ) if "v_bias" in name: UpperCamelCase_ = name.replace("v_bias" , "value.bias" ) if "cpb_mlp" in name: UpperCamelCase_ = name.replace("cpb_mlp" , "continuous_position_bias_mlp" ) if name == "norm.weight": UpperCamelCase_ = "layernorm.weight" if name == "norm.bias": UpperCamelCase_ = "layernorm.bias" if "head" in name: UpperCamelCase_ = name.replace("head" , "classifier" ) else: UpperCamelCase_ = "swinv2." + name return name def lowerCAmelCase( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )-> Optional[int]: """simple docstring""" for key in orig_state_dict.copy().keys(): UpperCamelCase_ = orig_state_dict.pop(SCREAMING_SNAKE_CASE_ ) if "mask" in key: continue elif "qkv" in key: UpperCamelCase_ = key.split("." ) UpperCamelCase_ = int(key_split[1] ) UpperCamelCase_ = int(key_split[3] ) UpperCamelCase_ = model.swinva.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: UpperCamelCase_ = val[:dim, :] UpperCamelCase_ = val[dim : dim * 2, :] UpperCamelCase_ = val[-dim:, :] else: UpperCamelCase_ = val[:dim] UpperCamelCase_ = val[ dim : dim * 2 ] UpperCamelCase_ = val[-dim:] else: UpperCamelCase_ = val return orig_state_dict def lowerCAmelCase( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )-> Any: """simple docstring""" UpperCamelCase_ = timm.create_model(SCREAMING_SNAKE_CASE_ , pretrained=SCREAMING_SNAKE_CASE_ ) timm_model.eval() UpperCamelCase_ = get_swinva_config(SCREAMING_SNAKE_CASE_ ) UpperCamelCase_ = SwinvaForImageClassification(SCREAMING_SNAKE_CASE_ ) model.eval() UpperCamelCase_ = convert_state_dict(timm_model.state_dict() , SCREAMING_SNAKE_CASE_ ) model.load_state_dict(SCREAMING_SNAKE_CASE_ ) UpperCamelCase_ = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCamelCase_ = AutoImageProcessor.from_pretrained("microsoft/{}".format(swinva_name.replace("_" , "-" ) ) ) UpperCamelCase_ = Image.open(requests.get(SCREAMING_SNAKE_CASE_ , stream=SCREAMING_SNAKE_CASE_ ).raw ) UpperCamelCase_ = image_processor(images=SCREAMING_SNAKE_CASE_ , return_tensors="pt" ) UpperCamelCase_ = timm_model(inputs["pixel_values"] ) UpperCamelCase_ = model(**SCREAMING_SNAKE_CASE_ ).logits assert torch.allclose(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , atol=1E-3 ) print(f"Saving model {swinva_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(SCREAMING_SNAKE_CASE_ ) print(f"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(SCREAMING_SNAKE_CASE_ ) model.push_to_hub( repo_path_or_name=Path(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) , organization="nandwalritik" , commit_message="Add model" , ) if __name__ == "__main__": SCREAMING_SNAKE_CASE :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( """--swinv2_name""", default="""swinv2_tiny_patch4_window8_256""", type=str, help="""Name of the Swinv2 timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) SCREAMING_SNAKE_CASE :int = parser.parse_args() convert_swinva_checkpoint(args.swinva_name, args.pytorch_dump_folder_path)
628
0
import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def _lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' A_ = MobileBertConfig.from_json_file(SCREAMING_SNAKE_CASE ) print(f"Building PyTorch model from configuration: {config}" ) A_ = MobileBertForPreTraining(SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint A_ = load_tf_weights_in_mobilebert(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f"Save PyTorch model to {pytorch_dump_path}" ) torch.save(model.state_dict() , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--mobilebert_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained MobileBERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __lowercase = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.mobilebert_config_file, args.pytorch_dump_path)
563
def _lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' A_ = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def _lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' A_ = 0 while b > 0: if b & 1: A_ = ((res % c) + (a % c)) % c a += a b >>= 1 return res
563
1
'''simple docstring''' from __future__ import annotations from math import pow, sqrt def _snake_case ( _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : float ) -> dict[str, float]: """simple docstring""" if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError("""One and only one argument must be 0""" ) if resistance == 0: return {"resistance": sqrt(pow(_SCREAMING_SNAKE_CASE , 2 ) - pow(_SCREAMING_SNAKE_CASE , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(_SCREAMING_SNAKE_CASE , 2 ) - pow(_SCREAMING_SNAKE_CASE , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(_SCREAMING_SNAKE_CASE , 2 ) + pow(_SCREAMING_SNAKE_CASE , 2 ) )} else: raise ValueError("""Exactly one argument must be 0""" ) if __name__ == "__main__": import doctest doctest.testmod()
433
'''simple docstring''' import os import textwrap import pyarrow as pa import pytest from datasets import ClassLabel, Features, Image from datasets.packaged_modules.csv.csv import Csv from ..utils import require_pil @pytest.fixture def _snake_case ( _SCREAMING_SNAKE_CASE : List[str] ) -> str: """simple docstring""" lowerCAmelCase = tmp_path / """file.csv""" lowerCAmelCase = textwrap.dedent( """\ header1,header2 1,2 10,20 """ ) with open(_SCREAMING_SNAKE_CASE , """w""" ) as f: f.write(_SCREAMING_SNAKE_CASE ) return str(_SCREAMING_SNAKE_CASE ) @pytest.fixture def _snake_case ( _SCREAMING_SNAKE_CASE : Union[str, Any] ) -> Tuple: """simple docstring""" lowerCAmelCase = tmp_path / """malformed_file.csv""" lowerCAmelCase = textwrap.dedent( """\ header1,header2 1,2 10,20, """ ) with open(_SCREAMING_SNAKE_CASE , """w""" ) as f: f.write(_SCREAMING_SNAKE_CASE ) return str(_SCREAMING_SNAKE_CASE ) @pytest.fixture def _snake_case ( _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Union[str, Any] ) -> Optional[Any]: """simple docstring""" lowerCAmelCase = tmp_path / """csv_with_image.csv""" lowerCAmelCase = textwrap.dedent( f'\\n image\n {image_file}\n ' ) with open(_SCREAMING_SNAKE_CASE , """w""" ) as f: f.write(_SCREAMING_SNAKE_CASE ) return str(_SCREAMING_SNAKE_CASE ) @pytest.fixture def _snake_case ( _SCREAMING_SNAKE_CASE : int ) -> Any: """simple docstring""" lowerCAmelCase = tmp_path / """csv_with_label.csv""" lowerCAmelCase = textwrap.dedent( """\ label good bad good """ ) with open(_SCREAMING_SNAKE_CASE , """w""" ) as f: f.write(_SCREAMING_SNAKE_CASE ) return str(_SCREAMING_SNAKE_CASE ) @pytest.fixture def _snake_case ( _SCREAMING_SNAKE_CASE : Optional[int] ) -> Optional[Any]: """simple docstring""" lowerCAmelCase = tmp_path / """csv_with_int_list.csv""" lowerCAmelCase = textwrap.dedent( """\ int_list 1 2 3 4 5 6 7 8 9 """ ) with open(_SCREAMING_SNAKE_CASE , """w""" ) as f: f.write(_SCREAMING_SNAKE_CASE ) return str(_SCREAMING_SNAKE_CASE ) def _snake_case ( _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Any , _SCREAMING_SNAKE_CASE : List[str] ) -> List[Any]: """simple docstring""" lowerCAmelCase = Csv() lowerCAmelCase = csv._generate_tables([[csv_file, malformed_csv_file]] ) with pytest.raises(_SCREAMING_SNAKE_CASE , match="""Error tokenizing data""" ): for _ in generator: pass assert any( record.levelname == """ERROR""" and """Failed to read file""" in record.message and os.path.basename(_SCREAMING_SNAKE_CASE ) in record.message for record in caplog.records ) @require_pil def _snake_case ( _SCREAMING_SNAKE_CASE : Dict ) -> Tuple: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , encoding="""utf-8""" ) as f: lowerCAmelCase = f.read().splitlines()[1] lowerCAmelCase = Csv(encoding="""utf-8""" , features=Features({"""image""": Image()} ) ) lowerCAmelCase = csv._generate_tables([[csv_file_with_image]] ) lowerCAmelCase = pa.concat_tables([table for _, table in generator] ) assert pa_table.schema.field("""image""" ).type == Image()() lowerCAmelCase = pa_table.to_pydict()["""image"""] assert generated_content == [{"path": image_file, "bytes": None}] def _snake_case ( _SCREAMING_SNAKE_CASE : List[str] ) -> Union[str, Any]: """simple docstring""" with open(_SCREAMING_SNAKE_CASE , encoding="""utf-8""" ) as f: lowerCAmelCase = f.read().splitlines()[1:] lowerCAmelCase = Csv(encoding="""utf-8""" , features=Features({"""label""": ClassLabel(names=["""good""", """bad"""] )} ) ) lowerCAmelCase = csv._generate_tables([[csv_file_with_label]] ) lowerCAmelCase = pa.concat_tables([table for _, table in generator] ) assert pa_table.schema.field("""label""" ).type == ClassLabel(names=["""good""", """bad"""] )() lowerCAmelCase = pa_table.to_pydict()["""label"""] assert generated_content == [ClassLabel(names=["""good""", """bad"""] ).straint(_SCREAMING_SNAKE_CASE ) for label in labels] def _snake_case ( _SCREAMING_SNAKE_CASE : Optional[Any] ) -> Tuple: """simple docstring""" lowerCAmelCase = Csv(encoding="""utf-8""" , sep=""",""" , converters={"""int_list""": lambda _SCREAMING_SNAKE_CASE : [int(_SCREAMING_SNAKE_CASE ) for i in x.split()]} ) lowerCAmelCase = csv._generate_tables([[csv_file_with_int_list]] ) lowerCAmelCase = pa.concat_tables([table for _, table in generator] ) assert pa.types.is_list(pa_table.schema.field("""int_list""" ).type ) lowerCAmelCase = pa_table.to_pydict()["""int_list"""] assert generated_content == [[1, 2, 3], [4, 5, 6], [7, 8, 9]]
433
1
"""simple docstring""" import argparse import os import torch from transformers.utils import WEIGHTS_NAME lowerCAmelCase__ = ['''small''', '''medium''', '''large'''] lowerCAmelCase__ = '''lm_head.decoder.weight''' lowerCAmelCase__ = '''lm_head.weight''' def a__ ( SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : str ): '''simple docstring''' lowerCAmelCase : List[str] = torch.load(SCREAMING_SNAKE_CASE ) lowerCAmelCase : Optional[int] = d.pop(SCREAMING_SNAKE_CASE ) os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) torch.save(SCREAMING_SNAKE_CASE , os.path.join(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument('''--dialogpt_path''', default='''.''', type=str) lowerCAmelCase__ = parser.parse_args() for MODEL in DIALOGPT_MODELS: lowerCAmelCase__ = os.path.join(args.dialogpt_path, F"{MODEL}_ft.pkl") lowerCAmelCase__ = F"./DialoGPT-{MODEL}" convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
681
"""simple docstring""" def a__ ( SCREAMING_SNAKE_CASE : int = 1_0_0_0 ): '''simple docstring''' return sum(e for e in range(3 , SCREAMING_SNAKE_CASE ) if e % 3 == 0 or e % 5 == 0 ) if __name__ == "__main__": print(F"{solution() = }")
681
1
"""simple docstring""" import os from typing import Optional import fsspec from fsspec.archive import AbstractArchiveFileSystem from fsspec.utils import DEFAULT_BLOCK_SIZE class __lowercase ( __lowerCamelCase ): snake_case_ = """""" snake_case_ = ( None # protocol passed in prefix to the url. ex: "gzip", for gzip://file.txt::http://foo.bar/file.txt.gz ) snake_case_ = None # compression type in fsspec. ex: "gzip" snake_case_ = None # extension of the filename to strip. ex: "".gz" to get file.txt from file.txt.gz def __init__( self : Optional[int] ,A : str = "" ,A : Optional[str] = None ,A : Optional[dict] = None ,**A : Any ): '''simple docstring''' super().__init__(self ,**A ) # always open as "rb" since fsspec can then use the TextIOWrapper to make it work for "r" mode UpperCAmelCase__ : Optional[int] = fsspec.open( A ,mode="""rb""" ,protocol=A ,compression=self.compression ,client_kwargs={ """requote_redirect_url""": False, # see https://github.com/huggingface/datasets/pull/5459 """trust_env""": True, # Enable reading proxy env variables. **(target_options or {}).pop("""client_kwargs""" ,{} ), # To avoid issues if it was already passed. } ,**(target_options or {}) ,) UpperCAmelCase__ : int = os.path.basename(self.file.path.split("""::""" )[0] ) UpperCAmelCase__ : List[Any] = ( self.compressed_name[: self.compressed_name.rindex(""".""" )] if """.""" in self.compressed_name else self.compressed_name ) UpperCAmelCase__ : Optional[int] = None @classmethod def __lowercase ( cls : Optional[Any] ,A : Any ): '''simple docstring''' # compressed file paths are always relative to the archive root return super()._strip_protocol(A ).lstrip("""/""" ) def __lowercase ( self : str ): '''simple docstring''' if self.dir_cache is None: UpperCAmelCase__ : List[str] = {**self.file.fs.info(self.file.path ), """name""": self.uncompressed_name} UpperCAmelCase__ : List[str] = {f["""name"""]: f} def __lowercase ( self : Dict ,A : str ): '''simple docstring''' return self.file.open().read() def __lowercase ( self : Optional[int] ,A : str ,A : str = "rb" ,A : str=None ,A : List[Any]=True ,A : List[str]=None ,**A : Tuple ,): '''simple docstring''' UpperCAmelCase__ : Optional[int] = self._strip_protocol(A ) if mode != "rb": raise ValueError(f"Tried to read with mode {mode} on file {self.file.path} opened with mode 'rb'" ) return self.file.open() class __lowercase ( __lowerCamelCase ): snake_case_ = """bz2""" snake_case_ = """bz2""" snake_case_ = """.bz2""" class __lowercase ( __lowerCamelCase ): snake_case_ = """gzip""" snake_case_ = """gzip""" snake_case_ = """.gz""" class __lowercase ( __lowerCamelCase ): snake_case_ = """lz4""" snake_case_ = """lz4""" snake_case_ = """.lz4""" class __lowercase ( __lowerCamelCase ): snake_case_ = """xz""" snake_case_ = """xz""" snake_case_ = """.xz""" class __lowercase ( __lowerCamelCase ): snake_case_ = """zstd""" snake_case_ = """zstd""" snake_case_ = """.zst""" def __init__( self : Union[str, Any] ,A : str ,A : str = "rb" ,A : Optional[str] = None ,A : Optional[dict] = None ,A : int = DEFAULT_BLOCK_SIZE ,**A : List[str] ,): '''simple docstring''' super().__init__( fo=A ,mode=A ,target_protocol=A ,target_options=A ,block_size=A ,**A ,) # We need to wrap the zstd decompressor to avoid this error in fsspec==2021.7.0 and zstandard==0.15.2: # # File "/Users/user/.virtualenvs/hf-datasets/lib/python3.7/site-packages/fsspec/core.py", line 145, in open # out.close = close # AttributeError: 'zstd.ZstdDecompressionReader' object attribute 'close' is read-only # # see https://github.com/intake/filesystem_spec/issues/725 UpperCAmelCase__ : str = self.file.__enter__ class __lowercase : def __init__( self : Tuple ,A : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = file_ def __enter__( self : List[Any] ): '''simple docstring''' self._file.__enter__() return self def __exit__( self : List[Any] ,*A : Union[str, Any] ,**A : Tuple ): '''simple docstring''' self._file.__exit__(*A ,**A ) def __iter__( self : Tuple ): '''simple docstring''' return iter(self._file ) def __lowercase ( self : Optional[Any] ): '''simple docstring''' return next(self._file ) def __getattr__( self : Tuple ,A : Any ): '''simple docstring''' return getattr(self._file ,A ) def fixed_enter(*A : List[Any] ,**A : List[str] ): return WrappedFile(_enter(*A ,**A ) ) UpperCAmelCase__ : str = fixed_enter
65
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) a_ = { '''configuration_trocr''': ['''TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TrOCRConfig'''], '''processing_trocr''': ['''TrOCRProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ '''TROCR_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TrOCRForCausalLM''', '''TrOCRPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys a_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
480
0
'''simple docstring''' import argparse import requests import torch from PIL import Image from transformers import SwinConfig, SwinForMaskedImageModeling, ViTImageProcessor def _SCREAMING_SNAKE_CASE ( snake_case_ ): _lowercase = SwinConfig(image_size=192 ) if "base" in model_name: _lowercase = 6 _lowercase = 128 _lowercase = (2, 2, 18, 2) _lowercase = (4, 8, 16, 32) elif "large" in model_name: _lowercase = 12 _lowercase = 192 _lowercase = (2, 2, 18, 2) _lowercase = (6, 12, 24, 48) else: raise ValueError("""Model not supported, only supports base and large variants""" ) _lowercase = window_size _lowercase = embed_dim _lowercase = depths _lowercase = num_heads return config def _SCREAMING_SNAKE_CASE ( snake_case_ ): if "encoder.mask_token" in name: _lowercase = name.replace("""encoder.mask_token""" , """embeddings.mask_token""" ) if "encoder.patch_embed.proj" in name: _lowercase = name.replace("""encoder.patch_embed.proj""" , """embeddings.patch_embeddings.projection""" ) if "encoder.patch_embed.norm" in name: _lowercase = name.replace("""encoder.patch_embed.norm""" , """embeddings.norm""" ) if "attn.proj" in name: _lowercase = name.replace("""attn.proj""" , """attention.output.dense""" ) if "attn" in name: _lowercase = name.replace("""attn""" , """attention.self""" ) if "norm1" in name: _lowercase = name.replace("""norm1""" , """layernorm_before""" ) if "norm2" in name: _lowercase = name.replace("""norm2""" , """layernorm_after""" ) if "mlp.fc1" in name: _lowercase = name.replace("""mlp.fc1""" , """intermediate.dense""" ) if "mlp.fc2" in name: _lowercase = name.replace("""mlp.fc2""" , """output.dense""" ) if name == "encoder.norm.weight": _lowercase = """layernorm.weight""" if name == "encoder.norm.bias": _lowercase = """layernorm.bias""" if "decoder" in name: pass else: _lowercase = """swin.""" + name return name def _SCREAMING_SNAKE_CASE ( snake_case_ , snake_case_ ): for key in orig_state_dict.copy().keys(): _lowercase = orig_state_dict.pop(snake_case_ ) if "attn_mask" in key: pass elif "qkv" in key: _lowercase = key.split(""".""" ) _lowercase = int(key_split[2] ) _lowercase = int(key_split[4] ) _lowercase = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: _lowercase = val[:dim, :] _lowercase = val[ dim : dim * 2, : ] _lowercase = val[-dim:, :] else: _lowercase = val[ :dim ] _lowercase = val[ dim : dim * 2 ] _lowercase = val[ -dim: ] else: _lowercase = val return orig_state_dict def _SCREAMING_SNAKE_CASE ( snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _lowercase = torch.load(snake_case_ , map_location="""cpu""" )["""model"""] _lowercase = get_swin_config(snake_case_ ) _lowercase = SwinForMaskedImageModeling(snake_case_ ) model.eval() _lowercase = convert_state_dict(snake_case_ , snake_case_ ) model.load_state_dict(snake_case_ ) _lowercase = """http://images.cocodataset.org/val2017/000000039769.jpg""" _lowercase = ViTImageProcessor(size={"""height""": 192, """width""": 192} ) _lowercase = Image.open(requests.get(snake_case_ , stream=snake_case_ ).raw ) _lowercase = image_processor(images=snake_case_ , return_tensors="""pt""" ) with torch.no_grad(): _lowercase = model(**snake_case_ ).logits print(outputs.keys() ) print("""Looks ok!""" ) if pytorch_dump_folder_path is not None: print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(snake_case_ ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(snake_case_ ) if push_to_hub: print(F"""Pushing model and image processor for {model_name} to hub""" ) model.push_to_hub(F"""microsoft/{model_name}""" ) image_processor.push_to_hub(F"""microsoft/{model_name}""" ) if __name__ == "__main__": _lowerCamelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='swin-base-simmim-window6-192', type=str, choices=['swin-base-simmim-window6-192', 'swin-large-simmim-window12-192'], help='Name of the Swin SimMIM model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default='/Users/nielsrogge/Documents/SwinSimMIM/simmim_pretrain__swin_base__img192_window6__100ep.pth', type=str, help='Path to the original PyTorch checkpoint (.pth file).', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) _lowerCamelCase = parser.parse_args() convert_swin_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
572
'''simple docstring''' def _SCREAMING_SNAKE_CASE ( snake_case_ ): _lowercase = int(snake_case_ ) if n_element < 1: _lowercase = ValueError("""a should be a positive number""" ) raise my_error _lowercase = [1] _lowercase , _lowercase , _lowercase = (0, 0, 0) _lowercase = 1 while index < n_element: while hamming_list[i] * 2 <= hamming_list[-1]: i += 1 while hamming_list[j] * 3 <= hamming_list[-1]: j += 1 while hamming_list[k] * 5 <= hamming_list[-1]: k += 1 hamming_list.append( min(hamming_list[i] * 2 , hamming_list[j] * 3 , hamming_list[k] * 5 ) ) index += 1 return hamming_list if __name__ == "__main__": _lowerCamelCase = input('Enter the last number (nth term) of the Hamming Number Series: ') print('Formula of Hamming Number Series => 2^i * 3^j * 5^k') _lowerCamelCase = hamming(int(n)) print('-----------------------------------------------------') print(F"""The list with nth numbers is: {hamming_numbers}""") print('-----------------------------------------------------')
572
1
from math import factorial lowercase : dict[str, int] = {str(digit): factorial(digit) for digit in range(10)} def snake_case__ ( lowerCamelCase_ ): if not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise TypeError('''Parameter number must be int''' ) if number < 0: raise ValueError('''Parameter number must be greater than or equal to 0''' ) # Converts number in string to iterate on its digits and adds its factorial. return sum(DIGIT_FACTORIAL[digit] for digit in str(lowerCamelCase_ ) ) def snake_case__ ( lowerCamelCase_ = 60 , lowerCamelCase_ = 1000000 ): if not isinstance(lowerCamelCase_ , lowerCamelCase_ ) or not isinstance(lowerCamelCase_ , lowerCamelCase_ ): raise TypeError('''Parameters chain_length and number_limit must be int''' ) if chain_length <= 0 or number_limit <= 0: raise ValueError( '''Parameters chain_length and number_limit must be greater than 0''' ) # the counter for the chains with the exact desired length A : Optional[int] = 0 # the cached sizes of the previous chains A : dict[int, int] = {} for start_chain_element in range(1 , lowerCamelCase_ ): # The temporary set will contain the elements of the chain A : Any = set() A : Any = 0 # Stop computing the chain when you find a cached size, a repeating item or the # length is greater then the desired one. A : List[str] = start_chain_element while ( chain_element not in chain_sets_lengths and chain_element not in chain_set and chain_set_length <= chain_length ): chain_set.add(lowerCamelCase_ ) chain_set_length += 1 A : str = digit_factorial_sum(lowerCamelCase_ ) if chain_element in chain_sets_lengths: chain_set_length += chain_sets_lengths[chain_element] A : str = chain_set_length # If chain contains the exact amount of elements increase the counter if chain_set_length == chain_length: chains_counter += 1 return chains_counter if __name__ == "__main__": import doctest doctest.testmod() print(F"{solution()}")
542
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowercase : Tuple = logging.get_logger(__name__) class __lowercase ( _SCREAMING_SNAKE_CASE ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = ['''pixel_values'''] def __init__( self , __UpperCAmelCase = True , __UpperCAmelCase = None , __UpperCAmelCase = PIL.Image.BICUBIC , __UpperCAmelCase = True , __UpperCAmelCase = None , __UpperCAmelCase = 1 / 2_55 , __UpperCAmelCase = True , __UpperCAmelCase = True , __UpperCAmelCase = None , __UpperCAmelCase = None , **__UpperCAmelCase , ) -> None: super().__init__(**__UpperCAmelCase ) A : Any = size if size is not None else {'''height''': 2_56, '''width''': 2_56} A : Any = get_size_dict(__UpperCAmelCase ) A : List[Any] = crop_size if crop_size is not None else {'''height''': 2_24, '''width''': 2_24} A : List[Any] = get_size_dict(__UpperCAmelCase , param_name='''crop_size''' ) A : Dict = do_resize A : Tuple = size A : Union[str, Any] = resample A : Dict = do_center_crop A : int = crop_size A : Union[str, Any] = do_rescale A : str = rescale_factor A : Optional[Any] = do_normalize A : Dict = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN A : Dict = image_std if image_std is not None else IMAGENET_STANDARD_STD def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = PIL.Image.BICUBIC , __UpperCAmelCase = None , **__UpperCAmelCase , ) -> np.ndarray: A : int = get_size_dict(__UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}' ) return resize( __UpperCAmelCase , size=(size['''height'''], size['''width''']) , resample=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , **__UpperCAmelCase , ) -> np.ndarray: A : Optional[Any] = get_size_dict(__UpperCAmelCase ) if "height" not in size or "width" not in size: raise ValueError(f'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}' ) return center_crop(__UpperCAmelCase , size=(size['''height'''], size['''width''']) , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , **__UpperCAmelCase , ) -> List[str]: return rescale(__UpperCAmelCase , scale=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , **__UpperCAmelCase , ) -> np.ndarray: return normalize(__UpperCAmelCase , mean=__UpperCAmelCase , std=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase=None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = ChannelDimension.FIRST , **__UpperCAmelCase , ) -> PIL.Image.Image: A : Optional[int] = do_resize if do_resize is not None else self.do_resize A : Optional[Any] = resample if resample is not None else self.resample A : Tuple = do_center_crop if do_center_crop is not None else self.do_center_crop A : Optional[Any] = do_rescale if do_rescale is not None else self.do_rescale A : str = rescale_factor if rescale_factor is not None else self.rescale_factor A : Tuple = do_normalize if do_normalize is not None else self.do_normalize A : Optional[int] = image_mean if image_mean is not None else self.image_mean A : Dict = image_std if image_std is not None else self.image_std A : List[Any] = size if size is not None else self.size A : List[Any] = get_size_dict(__UpperCAmelCase ) A : Dict = crop_size if crop_size is not None else self.crop_size A : Tuple = get_size_dict(__UpperCAmelCase , param_name='''crop_size''' ) A : Any = make_list_of_images(__UpperCAmelCase ) if not valid_images(__UpperCAmelCase ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. A : List[Any] = [to_numpy_array(__UpperCAmelCase ) for image in images] if do_resize: A : List[str] = [self.resize(image=__UpperCAmelCase , size=__UpperCAmelCase , resample=__UpperCAmelCase ) for image in images] if do_center_crop: A : Dict = [self.center_crop(image=__UpperCAmelCase , size=__UpperCAmelCase ) for image in images] if do_rescale: A : Union[str, Any] = [self.rescale(image=__UpperCAmelCase , scale=__UpperCAmelCase ) for image in images] if do_normalize: A : Tuple = [self.normalize(image=__UpperCAmelCase , mean=__UpperCAmelCase , std=__UpperCAmelCase ) for image in images] A : Optional[Any] = [to_channel_dimension_format(__UpperCAmelCase , __UpperCAmelCase ) for image in images] A : List[Any] = {'''pixel_values''': images} return BatchFeature(data=__UpperCAmelCase , tensor_type=__UpperCAmelCase )
542
1
import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class a (unittest.TestCase ): """simple docstring""" def __init__( self : Any , lowerCamelCase : Union[str, Any] , lowerCamelCase : Dict=13 , lowerCamelCase : int=7 , lowerCamelCase : Optional[int]=True , lowerCamelCase : List[str]=True , lowerCamelCase : str=True , lowerCamelCase : List[str]=True , lowerCamelCase : Optional[int]=99 , lowerCamelCase : int=32 , lowerCamelCase : List[str]=5 , lowerCamelCase : Tuple=4 , lowerCamelCase : List[Any]=37 , lowerCamelCase : str="gelu" , lowerCamelCase : Union[str, Any]=0.1 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : List[Any]=512 , lowerCamelCase : Optional[Any]=16 , lowerCamelCase : Dict=2 , lowerCamelCase : int=0.02 , lowerCamelCase : Optional[int]=4 , ) -> Union[str, Any]: __snake_case : Union[str, Any] = parent __snake_case : str = batch_size __snake_case : int = seq_length __snake_case : List[str] = is_training __snake_case : Union[str, Any] = use_attention_mask __snake_case : Dict = use_token_type_ids __snake_case : Any = use_labels __snake_case : Tuple = vocab_size __snake_case : str = hidden_size __snake_case : int = num_hidden_layers __snake_case : Union[str, Any] = num_attention_heads __snake_case : Tuple = intermediate_size __snake_case : int = hidden_act __snake_case : int = hidden_dropout_prob __snake_case : List[str] = attention_probs_dropout_prob __snake_case : str = max_position_embeddings __snake_case : Union[str, Any] = type_vocab_size __snake_case : Tuple = type_sequence_label_size __snake_case : int = initializer_range __snake_case : Any = num_choices def __snake_case ( self : Optional[Any] ) -> Tuple: __snake_case : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case : Union[str, Any] = None if self.use_attention_mask: __snake_case : str = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case : Optional[int] = None if self.use_token_type_ids: __snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __snake_case : Optional[Any] = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowerCamelCase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def __snake_case ( self : Dict ) -> int: __snake_case : Optional[Any] = self.prepare_config_and_inputs() __snake_case , __snake_case , __snake_case , __snake_case : Optional[Any] = config_and_inputs __snake_case : int = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class a (_lowerCAmelCase , unittest.TestCase ): """simple docstring""" __UpperCAmelCase : Tuple = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def __snake_case ( self : Optional[int] ) -> List[str]: __snake_case : str = FlaxAlbertModelTester(self ) @slow def __snake_case ( self : Optional[int] ) -> Tuple: for model_class_name in self.all_model_classes: __snake_case : List[Any] = model_class_name.from_pretrained("albert-base-v2" ) __snake_case : str = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase ) @require_flax class a (unittest.TestCase ): """simple docstring""" @slow def __snake_case ( self : int ) -> Tuple: __snake_case : Union[str, Any] = FlaxAlbertModel.from_pretrained("albert-base-v2" ) __snake_case : Any = np.array([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __snake_case : int = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) __snake_case : Optional[Any] = model(lowerCamelCase , attention_mask=lowerCamelCase )[0] __snake_case : Optional[int] = (1, 11, 768) self.assertEqual(output.shape , lowerCamelCase ) __snake_case : Optional[int] = np.array( [[[-0.65_13, 1.50_35, -0.27_66], [-0.65_15, 1.50_46, -0.27_80], [-0.65_12, 1.50_49, -0.27_84]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCamelCase , atol=1E-4 ) )
203
import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel _snake_case : int = { "text_branch": "text_model", "audio_branch": "audio_model.audio_encoder", "attn": "attention.self", "self.proj": "output.dense", "attention.self_mask": "attn_mask", "mlp.fc1": "intermediate.dense", "mlp.fc2": "output.dense", "norm1": "layernorm_before", "norm2": "layernorm_after", "bn0": "batch_norm", } _snake_case : Dict = AutoFeatureExtractor.from_pretrained("laion/clap-htsat-unfused", truncation="rand_trunc") def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase=False ): __snake_case , __snake_case : Any = create_model( "HTSAT-tiny" , "roberta" , __lowerCamelCase , precision="fp32" , device="cuda:0" if torch.cuda.is_available() else "cpu" , enable_fusion=__lowerCamelCase , fusion_type="aff_2d" if enable_fusion else None , ) return model, model_cfg def lowerCAmelCase_ ( __lowerCamelCase ): __snake_case : int = {} __snake_case : List[Any] = R".*sequential.(\d+).*" __snake_case : Any = R".*_projection.(\d+).*" for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: __snake_case : Optional[int] = key.replace(__lowerCamelCase , __lowerCamelCase ) if re.match(__lowerCamelCase , __lowerCamelCase ): # replace sequential layers with list __snake_case : List[Any] = re.match(__lowerCamelCase , __lowerCamelCase ).group(1 ) __snake_case : str = key.replace(F'sequential.{sequential_layer}.' , F'layers.{int(__lowerCamelCase )//3}.linear.' ) elif re.match(__lowerCamelCase , __lowerCamelCase ): __snake_case : Any = int(re.match(__lowerCamelCase , __lowerCamelCase ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... __snake_case : List[str] = 1 if projecton_layer == 0 else 2 __snake_case : Tuple = key.replace(F'_projection.{projecton_layer}.' , F'_projection.linear{transformers_projection_layer}.' ) if "audio" and "qkv" in key: # split qkv into query key and value __snake_case : List[str] = value __snake_case : Optional[int] = mixed_qkv.size(0 ) // 3 __snake_case : List[str] = mixed_qkv[:qkv_dim] __snake_case : int = mixed_qkv[qkv_dim : qkv_dim * 2] __snake_case : str = mixed_qkv[qkv_dim * 2 :] __snake_case : int = query_layer __snake_case : Tuple = key_layer __snake_case : List[str] = value_layer else: __snake_case : Tuple = value return model_state_dict def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase=False ): __snake_case , __snake_case : List[str] = init_clap(__lowerCamelCase , enable_fusion=__lowerCamelCase ) clap_model.eval() __snake_case : Union[str, Any] = clap_model.state_dict() __snake_case : Any = rename_state_dict(__lowerCamelCase ) __snake_case : Dict = ClapConfig() __snake_case : Dict = enable_fusion __snake_case : Optional[Any] = ClapModel(__lowerCamelCase ) # ignore the spectrogram embedding layer model.load_state_dict(__lowerCamelCase , strict=__lowerCamelCase ) model.save_pretrained(__lowerCamelCase ) transformers_config.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _snake_case : Union[str, Any] = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument("--enable_fusion", action="store_true", help="Whether to enable fusion or not") _snake_case : Tuple = parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
203
1
'''simple docstring''' # limitations under the License. from typing import Optional, Tuple, Union import torch from diffusers import DiffusionPipeline, ImagePipelineOutput class lowerCAmelCase_ ( a_ ): def __init__( self : int, _snake_case : Dict, _snake_case : Union[str, Any] ): '''simple docstring''' super().__init__() self.register_modules(unet=_snake_case, scheduler=_snake_case ) @torch.no_grad() def __call__( self : str, _snake_case : int = 1, _snake_case : Optional[torch.Generator] = None, _snake_case : int = 50, _snake_case : Optional[str] = "pil", _snake_case : bool = True, **_snake_case : str, ): '''simple docstring''' snake_case : Any =torch.randn( (batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size), generator=_snake_case, ) snake_case : Dict =image.to(self.device ) # set step values self.scheduler.set_timesteps(_snake_case ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output snake_case : int =self.unet(_snake_case, _snake_case ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 snake_case : List[Any] =self.scheduler.step(_snake_case, _snake_case, _snake_case ).prev_sample snake_case : str =(image / 2 + 0.5).clamp(0, 1 ) snake_case : Dict =image.cpu().permute(0, 2, 3, 1 ).numpy() if output_type == "pil": snake_case : Any =self.numpy_to_pil(_snake_case ) if not return_dict: return (image,), "This is a local test" return ImagePipelineOutput(images=_snake_case ), "This is a local test"
349
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import BatchEncoding, MarianTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import is_sentencepiece_available, is_tf_available, is_torch_available if is_sentencepiece_available(): from transformers.models.marian.tokenization_marian import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin A : List[str] = get_tests_dir("""fixtures/test_sentencepiece.model""") A : int = {"""target_lang""": """fi""", """source_lang""": """en"""} A : Tuple = """>>zh<<""" A : Optional[int] = """Helsinki-NLP/""" if is_torch_available(): A : Dict = """pt""" elif is_tf_available(): A : Optional[int] = """tf""" else: A : List[str] = """jax""" @require_sentencepiece class lowerCAmelCase_ ( a_ , unittest.TestCase ): __UpperCAmelCase = MarianTokenizer __UpperCAmelCase = False __UpperCAmelCase = True def __snake_case ( self : List[str] ): '''simple docstring''' super().setUp() snake_case : Dict =['''</s>''', '''<unk>''', '''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est''', '''\u0120''', '''<pad>'''] snake_case : Optional[Any] =dict(zip(_snake_case, range(len(_snake_case ) ) ) ) snake_case : Dict =Path(self.tmpdirname ) save_json(_snake_case, save_dir / VOCAB_FILES_NAMES['''vocab'''] ) save_json(_snake_case, save_dir / VOCAB_FILES_NAMES['''tokenizer_config_file'''] ) if not (save_dir / VOCAB_FILES_NAMES["source_spm"]).exists(): copyfile(_snake_case, save_dir / VOCAB_FILES_NAMES['''source_spm'''] ) copyfile(_snake_case, save_dir / VOCAB_FILES_NAMES['''target_spm'''] ) snake_case : Any =MarianTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def __snake_case ( self : List[str], **_snake_case : Tuple ): '''simple docstring''' return MarianTokenizer.from_pretrained(self.tmpdirname, **_snake_case ) def __snake_case ( self : Any, _snake_case : Dict ): '''simple docstring''' return ( "This is a test", "This is a test", ) def __snake_case ( self : Union[str, Any] ): '''simple docstring''' snake_case : Optional[int] ='''</s>''' snake_case : int =0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(_snake_case ), _snake_case ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(_snake_case ), _snake_case ) def __snake_case ( self : Any ): '''simple docstring''' snake_case : Tuple =list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0], '''</s>''' ) self.assertEqual(vocab_keys[1], '''<unk>''' ) self.assertEqual(vocab_keys[-1], '''<pad>''' ) self.assertEqual(len(_snake_case ), 9 ) def __snake_case ( self : str ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size, 9 ) def __snake_case ( self : Tuple ): '''simple docstring''' snake_case : Tuple =MarianTokenizer.from_pretrained(f'''{ORG_NAME}opus-mt-en-de''' ) snake_case : List[str] =en_de_tokenizer(['''I am a small frog'''], return_tensors=_snake_case ) self.assertIsInstance(_snake_case, _snake_case ) snake_case : Any =[38, 121, 14, 697, 38_848, 0] self.assertListEqual(_snake_case, batch.input_ids[0] ) snake_case : List[Any] =tempfile.mkdtemp() en_de_tokenizer.save_pretrained(_snake_case ) snake_case : List[Any] =[x.name for x in Path(_snake_case ).glob('''*''' )] self.assertIn('''source.spm''', _snake_case ) MarianTokenizer.from_pretrained(_snake_case ) def __snake_case ( self : Optional[Any] ): '''simple docstring''' snake_case : Any =self.get_tokenizer() snake_case : int =tok( ['''I am a small frog''' * 1_000, '''I am a small frog'''], padding=_snake_case, truncation=_snake_case, return_tensors=_snake_case ) self.assertIsInstance(_snake_case, _snake_case ) self.assertEqual(batch.input_ids.shape, (2, 512) ) def __snake_case ( self : int ): '''simple docstring''' snake_case : List[str] =self.get_tokenizer() snake_case : int =tok(['''I am a tiny frog''', '''I am a small frog'''], padding=_snake_case, return_tensors=_snake_case ) self.assertIsInstance(_snake_case, _snake_case ) self.assertEqual(batch_smaller.input_ids.shape, (2, 10) ) @slow def __snake_case ( self : Optional[int] ): '''simple docstring''' snake_case : List[Any] ={'''input_ids''': [[43_495, 462, 20, 42_164, 1_369, 52, 464, 132, 1_703, 492, 13, 7_491, 38_999, 6, 8, 464, 132, 1_703, 492, 13, 4_669, 37_867, 13, 7_525, 27, 1_593, 988, 13, 33_972, 7_029, 6, 20, 8_251, 383, 2, 270, 5_866, 3_788, 2, 2_353, 8_251, 12_338, 2, 13_958, 387, 2, 3_629, 6_953, 188, 2_900, 2, 13_958, 8_011, 11_501, 23, 8_460, 4_073, 34_009, 20, 435, 11_439, 27, 8, 8_460, 4_073, 6_004, 20, 9_988, 375, 27, 33, 266, 1_945, 1_076, 1_350, 37_867, 3_288, 5, 577, 1_076, 4_374, 8, 5_082, 5, 26_453, 257, 556, 403, 2, 242, 132, 383, 316, 492, 8, 10_767, 6, 316, 304, 4_239, 3, 0], [148, 15_722, 19, 1_839, 12, 1_350, 13, 22_327, 5_082, 5_418, 47_567, 35_938, 59, 318, 19_552, 108, 2_183, 54, 14_976, 4_835, 32, 547, 1_114, 8, 315, 2_417, 5, 92, 19_088, 3, 0, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100], [36, 6_395, 12_570, 39_147, 11_597, 6, 266, 4, 45_405, 7_296, 3, 0, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100, 58_100]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=_snake_case, model_name='''Helsinki-NLP/opus-mt-en-de''', revision='''1a8c2263da11e68e50938f97e10cd57820bd504c''', decode_kwargs={'''use_source_tokenizer''': True}, ) def __snake_case ( self : Optional[int] ): '''simple docstring''' snake_case : Optional[Any] =MarianTokenizer.from_pretrained('''hf-internal-testing/test-marian-two-vocabs''' ) snake_case : List[str] ='''Tämä on testi''' snake_case : Optional[int] ='''This is a test''' snake_case : Optional[Any] =[76, 7, 2_047, 2] snake_case : int =[69, 12, 11, 940, 2] snake_case : Optional[int] =tokenizer(_snake_case ).input_ids self.assertListEqual(_snake_case, _snake_case ) snake_case : Optional[Any] =tokenizer(text_target=_snake_case ).input_ids self.assertListEqual(_snake_case, _snake_case ) snake_case : Optional[int] =tokenizer.decode(_snake_case, skip_special_tokens=_snake_case ) self.assertEqual(_snake_case, _snake_case )
349
1
import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def A__ ( __A , __A , __A ): '''simple docstring''' if gpta_config_file == "": _lowerCamelCase : Optional[Any] = GPTaConfig() else: _lowerCamelCase : str = GPTaConfig.from_json_file(UpperCAmelCase__ ) _lowerCamelCase : int = GPTaModel(UpperCAmelCase__ ) # Load weights from numpy load_tf_weights_in_gpta(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) # Save pytorch-model _lowerCamelCase : Dict = pytorch_dump_folder_path + """/""" + WEIGHTS_NAME _lowerCamelCase : Optional[Any] = pytorch_dump_folder_path + """/""" + CONFIG_NAME print(F"""Save PyTorch model to {pytorch_weights_dump_path}""" ) torch.save(model.state_dict() , UpperCAmelCase__ ) print(F"""Save configuration file to {pytorch_config_dump_path}""" ) with open(UpperCAmelCase__ , """w""" , encoding="""utf-8""" ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCAmelCase : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( "--gpt2_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--gpt2_config_file", default="", type=str, help=( "An optional config json file corresponding to the pre-trained OpenAI model. \n" "This specifies the model architecture." ), ) lowerCAmelCase : Optional[int] =parser.parse_args() convert_gpta_checkpoint_to_pytorch(args.gpta_checkpoint_path, args.gpta_config_file, args.pytorch_dump_folder_path)
709
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL lowerCAmelCase : Tuple =logging.get_logger(__name__) class __snake_case ( __lowerCAmelCase ): '''simple docstring''' _snake_case = ['pixel_values'] def __init__( self : Optional[Any] , _UpperCamelCase : bool = True , _UpperCamelCase : Dict[str, int] = None , _UpperCamelCase : PILImageResampling = PIL.Image.BICUBIC , _UpperCamelCase : bool = True , _UpperCamelCase : Dict[str, int] = None , _UpperCamelCase : Union[int, float] = 1 / 255 , _UpperCamelCase : bool = True , _UpperCamelCase : bool = True , _UpperCamelCase : Optional[Union[float, List[float]]] = None , _UpperCamelCase : Optional[Union[float, List[float]]] = None , **_UpperCamelCase : str , ) ->None: """simple docstring""" super().__init__(**_UpperCamelCase) _lowerCamelCase : Tuple = size if size is not None else {"""height""": 256, """width""": 256} _lowerCamelCase : Optional[Any] = get_size_dict(_UpperCamelCase) _lowerCamelCase : Any = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} _lowerCamelCase : Any = get_size_dict(_UpperCamelCase , param_name="""crop_size""") _lowerCamelCase : int = do_resize _lowerCamelCase : int = size _lowerCamelCase : Optional[int] = resample _lowerCamelCase : int = do_center_crop _lowerCamelCase : Optional[Any] = crop_size _lowerCamelCase : Union[str, Any] = do_rescale _lowerCamelCase : List[str] = rescale_factor _lowerCamelCase : List[Any] = do_normalize _lowerCamelCase : Dict = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN _lowerCamelCase : int = image_std if image_std is not None else IMAGENET_STANDARD_STD def _SCREAMING_SNAKE_CASE ( self : Any , _UpperCamelCase : np.ndarray , _UpperCamelCase : Dict[str, int] , _UpperCamelCase : PILImageResampling = PIL.Image.BICUBIC , _UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **_UpperCamelCase : Union[str, Any] , ) ->np.ndarray: """simple docstring""" _lowerCamelCase : Dict = get_size_dict(_UpperCamelCase) if "height" not in size or "width" not in size: raise ValueError(F"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""") return resize( _UpperCamelCase , size=(size["""height"""], size["""width"""]) , resample=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase) def _SCREAMING_SNAKE_CASE ( self : List[str] , _UpperCamelCase : np.ndarray , _UpperCamelCase : Dict[str, int] , _UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **_UpperCamelCase : List[str] , ) ->np.ndarray: """simple docstring""" _lowerCamelCase : Union[str, Any] = get_size_dict(_UpperCamelCase) if "height" not in size or "width" not in size: raise ValueError(F"""The size dictionary must have keys 'height' and 'width'. Got {size.keys()}""") return center_crop(_UpperCamelCase , size=(size["""height"""], size["""width"""]) , data_format=_UpperCamelCase , **_UpperCamelCase) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , _UpperCamelCase : np.ndarray , _UpperCamelCase : Union[int, float] , _UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **_UpperCamelCase : Union[str, Any] , ) ->str: """simple docstring""" return rescale(_UpperCamelCase , scale=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase) def _SCREAMING_SNAKE_CASE ( self : List[Any] , _UpperCamelCase : np.ndarray , _UpperCamelCase : Union[float, List[float]] , _UpperCamelCase : Union[float, List[float]] , _UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **_UpperCamelCase : Union[str, Any] , ) ->np.ndarray: """simple docstring""" return normalize(_UpperCamelCase , mean=_UpperCamelCase , std=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , _UpperCamelCase : ImageInput , _UpperCamelCase : bool = None , _UpperCamelCase : Dict[str, int] = None , _UpperCamelCase : Tuple=None , _UpperCamelCase : bool = None , _UpperCamelCase : Dict[str, int] = None , _UpperCamelCase : bool = None , _UpperCamelCase : float = None , _UpperCamelCase : bool = None , _UpperCamelCase : Optional[Union[float, List[float]]] = None , _UpperCamelCase : Optional[Union[float, List[float]]] = None , _UpperCamelCase : Optional[Union[str, TensorType]] = None , _UpperCamelCase : ChannelDimension = ChannelDimension.FIRST , **_UpperCamelCase : List[Any] , ) ->PIL.Image.Image: """simple docstring""" _lowerCamelCase : Any = do_resize if do_resize is not None else self.do_resize _lowerCamelCase : List[str] = resample if resample is not None else self.resample _lowerCamelCase : Dict = do_center_crop if do_center_crop is not None else self.do_center_crop _lowerCamelCase : Optional[Any] = do_rescale if do_rescale is not None else self.do_rescale _lowerCamelCase : Dict = rescale_factor if rescale_factor is not None else self.rescale_factor _lowerCamelCase : Dict = do_normalize if do_normalize is not None else self.do_normalize _lowerCamelCase : int = image_mean if image_mean is not None else self.image_mean _lowerCamelCase : Dict = image_std if image_std is not None else self.image_std _lowerCamelCase : Optional[Any] = size if size is not None else self.size _lowerCamelCase : Optional[int] = get_size_dict(_UpperCamelCase) _lowerCamelCase : List[str] = crop_size if crop_size is not None else self.crop_size _lowerCamelCase : Dict = get_size_dict(_UpperCamelCase , param_name="""crop_size""") _lowerCamelCase : int = make_list_of_images(_UpperCamelCase) if not valid_images(_UpperCamelCase): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""") if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""") if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""") if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""") if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""") # All transformations expect numpy arrays. _lowerCamelCase : Union[str, Any] = [to_numpy_array(_UpperCamelCase) for image in images] if do_resize: _lowerCamelCase : Any = [self.resize(image=_UpperCamelCase , size=_UpperCamelCase , resample=_UpperCamelCase) for image in images] if do_center_crop: _lowerCamelCase : str = [self.center_crop(image=_UpperCamelCase , size=_UpperCamelCase) for image in images] if do_rescale: _lowerCamelCase : Optional[int] = [self.rescale(image=_UpperCamelCase , scale=_UpperCamelCase) for image in images] if do_normalize: _lowerCamelCase : List[str] = [self.normalize(image=_UpperCamelCase , mean=_UpperCamelCase , std=_UpperCamelCase) for image in images] _lowerCamelCase : List[str] = [to_channel_dimension_format(_UpperCamelCase , _UpperCamelCase) for image in images] _lowerCamelCase : int = {"""pixel_values""": images} return BatchFeature(data=_UpperCamelCase , tensor_type=_UpperCamelCase)
15
0
import numpy as np class _A : """simple docstring""" def __init__( self : List[str] , __SCREAMING_SNAKE_CASE : Optional[Any]=None , __SCREAMING_SNAKE_CASE : Any=None , __SCREAMING_SNAKE_CASE : Any=None , __SCREAMING_SNAKE_CASE : Dict=None , __SCREAMING_SNAKE_CASE : Union[str, Any]=None ) -> List[Any]: self.set_matricies(red=lowerCamelCase_ , green=lowerCamelCase_ , blue=lowerCamelCase_ , red_edge=lowerCamelCase_ , nir=lowerCamelCase_ ) def _a ( self : List[Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=None , __SCREAMING_SNAKE_CASE : List[Any]=None , __SCREAMING_SNAKE_CASE : List[Any]=None , __SCREAMING_SNAKE_CASE : Tuple=None , __SCREAMING_SNAKE_CASE : Optional[int]=None ) -> List[str]: if red is not None: __UpperCAmelCase =red if green is not None: __UpperCAmelCase =green if blue is not None: __UpperCAmelCase =blue if red_edge is not None: __UpperCAmelCase =red_edge if nir is not None: __UpperCAmelCase =nir return True def _a ( self : str , __SCREAMING_SNAKE_CASE : int="" , __SCREAMING_SNAKE_CASE : Tuple=None , __SCREAMING_SNAKE_CASE : int=None , __SCREAMING_SNAKE_CASE : Any=None , __SCREAMING_SNAKE_CASE : Dict=None , __SCREAMING_SNAKE_CASE : List[str]=None ) -> Optional[Any]: self.set_matricies(red=lowerCamelCase_ , green=lowerCamelCase_ , blue=lowerCamelCase_ , red_edge=lowerCamelCase_ , nir=lowerCamelCase_ ) __UpperCAmelCase ={ """ARVI2""": self.arvaa, """CCCI""": self.ccci, """CVI""": self.cvi, """GLI""": self.gli, """NDVI""": self.ndvi, """BNDVI""": self.bndvi, """redEdgeNDVI""": self.red_edge_ndvi, """GNDVI""": self.gndvi, """GBNDVI""": self.gbndvi, """GRNDVI""": self.grndvi, """RBNDVI""": self.rbndvi, """PNDVI""": self.pndvi, """ATSAVI""": self.atsavi, """BWDRVI""": self.bwdrvi, """CIgreen""": self.ci_green, """CIrededge""": self.ci_rededge, """CI""": self.ci, """CTVI""": self.ctvi, """GDVI""": self.gdvi, """EVI""": self.evi, """GEMI""": self.gemi, """GOSAVI""": self.gosavi, """GSAVI""": self.gsavi, """Hue""": self.hue, """IVI""": self.ivi, """IPVI""": self.ipvi, """I""": self.i, """RVI""": self.rvi, """MRVI""": self.mrvi, """MSAVI""": self.m_savi, """NormG""": self.norm_g, """NormNIR""": self.norm_nir, """NormR""": self.norm_r, """NGRDI""": self.ngrdi, """RI""": self.ri, """S""": self.s, """IF""": self._if, """DVI""": self.dvi, """TVI""": self.tvi, """NDRE""": self.ndre, } try: return funcs[index]() except KeyError: print("""Index not in the list!""" ) return False def _a ( self : Dict ) -> List[Any]: return -0.18 + (1.17 * ((self.nir - self.red) / (self.nir + self.red))) def _a ( self : Optional[Any] ) -> int: return ((self.nir - self.redEdge) / (self.nir + self.redEdge)) / ( (self.nir - self.red) / (self.nir + self.red) ) def _a ( self : Union[str, Any] ) -> Union[str, Any]: return self.nir * (self.red / (self.green**2)) def _a ( self : List[Any] ) -> Optional[int]: return (2 * self.green - self.red - self.blue) / ( 2 * self.green + self.red + self.blue ) def _a ( self : Optional[Any] ) -> str: return (self.nir - self.red) / (self.nir + self.red) def _a ( self : Union[str, Any] ) -> Dict: return (self.nir - self.blue) / (self.nir + self.blue) def _a ( self : Optional[Any] ) -> int: return (self.redEdge - self.red) / (self.redEdge + self.red) def _a ( self : int ) -> int: return (self.nir - self.green) / (self.nir + self.green) def _a ( self : int ) -> List[str]: return (self.nir - (self.green + self.blue)) / ( self.nir + (self.green + self.blue) ) def _a ( self : str ) -> Dict: return (self.nir - (self.green + self.red)) / ( self.nir + (self.green + self.red) ) def _a ( self : Union[str, Any] ) -> str: return (self.nir - (self.blue + self.red)) / (self.nir + (self.blue + self.red)) def _a ( self : Optional[Any] ) -> Optional[int]: return (self.nir - (self.green + self.red + self.blue)) / ( self.nir + (self.green + self.red + self.blue) ) def _a ( self : Dict , __SCREAMING_SNAKE_CASE : List[str]=0.08 , __SCREAMING_SNAKE_CASE : Optional[int]=1.22 , __SCREAMING_SNAKE_CASE : Any=0.03 ) -> int: return a * ( (self.nir - a * self.red - b) / (a * self.nir + self.red - a * b + x * (1 + a**2)) ) def _a ( self : List[str] ) -> Union[str, Any]: return (0.1 * self.nir - self.blue) / (0.1 * self.nir + self.blue) def _a ( self : List[Any] ) -> Any: return (self.nir / self.green) - 1 def _a ( self : Optional[int] ) -> Optional[int]: return (self.nir / self.redEdge) - 1 def _a ( self : Optional[int] ) -> List[Any]: return (self.red - self.blue) / self.red def _a ( self : Any ) -> Optional[Any]: __UpperCAmelCase =self.ndvi() return ((ndvi + 0.5) / (abs(ndvi + 0.5 ))) * (abs(ndvi + 0.5 ) ** (1 / 2)) def _a ( self : Optional[Any] ) -> Tuple: return self.nir - self.green def _a ( self : List[Any] ) -> int: return 2.5 * ( (self.nir - self.red) / (self.nir + 6 * self.red - 7.5 * self.blue + 1) ) def _a ( self : Dict ) -> Tuple: __UpperCAmelCase =(2 * (self.nir**2 - self.red**2) + 1.5 * self.nir + 0.5 * self.red) / ( self.nir + self.red + 0.5 ) return n * (1 - 0.25 * n) - (self.red - 0.125) / (1 - self.red) def _a ( self : Any , __SCREAMING_SNAKE_CASE : Union[str, Any]=0.16 ) -> Optional[Any]: return (self.nir - self.green) / (self.nir + self.green + y) def _a ( self : int , __SCREAMING_SNAKE_CASE : Optional[Any]=0.5 ) -> Union[str, Any]: return ((self.nir - self.green) / (self.nir + self.green + n)) * (1 + n) def _a ( self : Tuple ) -> List[str]: return np.arctan( ((2 * self.red - self.green - self.blue) / 30.5) * (self.green - self.blue) ) def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : Any=None , __SCREAMING_SNAKE_CASE : Union[str, Any]=None ) -> Any: return (self.nir - b) / (a * self.red) def _a ( self : str ) -> Union[str, Any]: return (self.nir / ((self.nir + self.red) / 2)) * (self.ndvi() + 1) def _a ( self : str ) -> Union[str, Any]: return (self.red + self.green + self.blue) / 30.5 def _a ( self : Optional[int] ) -> Tuple: return self.nir / self.red def _a ( self : Tuple ) -> int: return (self.rvi() - 1) / (self.rvi() + 1) def _a ( self : Tuple ) -> List[str]: return ( (2 * self.nir + 1) - ((2 * self.nir + 1) ** 2 - 8 * (self.nir - self.red)) ** (1 / 2) ) / 2 def _a ( self : Union[str, Any] ) -> Optional[Any]: return self.green / (self.nir + self.red + self.green) def _a ( self : Tuple ) -> Union[str, Any]: return self.nir / (self.nir + self.red + self.green) def _a ( self : Any ) -> Optional[int]: return self.red / (self.nir + self.red + self.green) def _a ( self : List[str] ) -> List[Any]: return (self.green - self.red) / (self.green + self.red) def _a ( self : str ) -> Any: return (self.red - self.green) / (self.red + self.green) def _a ( self : Optional[int] ) -> str: __UpperCAmelCase =np.max([np.max(self.red ), np.max(self.green ), np.max(self.blue )] ) __UpperCAmelCase =np.min([np.min(self.red ), np.min(self.green ), np.min(self.blue )] ) return (max_value - min_value) / max_value def _a ( self : List[str] ) -> Union[str, Any]: return (2 * self.red - self.green - self.blue) / (self.green - self.blue) def _a ( self : Union[str, Any] ) -> int: return self.nir / self.red def _a ( self : int ) -> Optional[int]: return (self.ndvi() + 0.5) ** (1 / 2) def _a ( self : Union[str, Any] ) -> Dict: return (self.nir - self.redEdge) / (self.nir + self.redEdge)
68
'''simple docstring''' import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def _snake_case ( ) -> Union[str, Any]: raise RuntimeError('''CUDA out of memory.''' ) class a__ ( nn.Module ): '''simple docstring''' def __init__( self ) -> int: super().__init__() lowerCAmelCase__ = nn.Linear(3 , 4 ) lowerCAmelCase__ = nn.BatchNormad(4 ) lowerCAmelCase__ = nn.Linear(4 , 5 ) def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ ) -> Optional[Any]: return self.lineara(self.batchnorm(self.lineara(lowerCamelCase_ ) ) ) class a__ ( unittest.TestCase ): '''simple docstring''' def __SCREAMING_SNAKE_CASE ( self ) -> str: lowerCAmelCase__ = [] @find_executable_batch_size(starting_batch_size=1_28 ) def mock_training_loop_function(lowerCamelCase_ ): nonlocal batch_sizes batch_sizes.append(lowerCamelCase_ ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(lowerCamelCase_ , [1_28, 64, 32, 16, 8] ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: lowerCAmelCase__ = [] @find_executable_batch_size(starting_batch_size=1_28 ) def mock_training_loop_function(lowerCamelCase_ , lowerCamelCase_ ): nonlocal batch_sizes batch_sizes.append(lowerCamelCase_ ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga lowerCAmelCase__ , lowerCAmelCase__ = mock_training_loop_function('''hello''' ) self.assertListEqual(lowerCamelCase_ , [1_28, 64, 32, 16, 8] ) self.assertListEqual([bs, arga] , [8, '''hello'''] ) def __SCREAMING_SNAKE_CASE ( self ) -> Tuple: @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(lowerCamelCase_ ): pass with self.assertRaises(lowerCamelCase_ ) as cm: mock_training_loop_function() self.assertIn('''No executable batch size found, reached zero.''' , cm.exception.args[0] ) def __SCREAMING_SNAKE_CASE ( self ) -> Dict: @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(lowerCamelCase_ ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(lowerCamelCase_ ) as cm: mock_training_loop_function() self.assertIn('''No executable batch size found, reached zero.''' , cm.exception.args[0] ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: @find_executable_batch_size(starting_batch_size=1_28 ) def mock_training_loop_function(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(lowerCamelCase_ ) as cm: mock_training_loop_function(1_28 , '''hello''' , '''world''' ) self.assertIn('''Batch size was passed into `f`''' , cm.exception.args[0] ) self.assertIn('''`f(arg1=\'hello\', arg2=\'world\')''' , cm.exception.args[0] ) def __SCREAMING_SNAKE_CASE ( self ) -> Union[str, Any]: @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(lowerCamelCase_ ): raise ValueError('''Oops, we had an error!''' ) with self.assertRaises(lowerCamelCase_ ) as cm: mock_training_loop_function() self.assertIn('''Oops, we had an error!''' , cm.exception.args[0] ) @require_cuda def __SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: lowerCAmelCase__ = torch.cuda.memory_allocated() lowerCAmelCase__ = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , lowerCamelCase_ ) lowerCAmelCase__ = release_memory(lowerCamelCase_ ) self.assertEqual(torch.cuda.memory_allocated() , lowerCamelCase_ )
90
0
'''simple docstring''' import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler __A =16 __A =32 def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ = 1_6 , UpperCamelCase__ = "bert-base-cased" ): UpperCAmelCase__ : str = AutoTokenizer.from_pretrained(UpperCamelCase__ ) UpperCAmelCase__ : Any = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(UpperCamelCase__ ): # max_length=None => use the model max length (it's actually the default) UpperCAmelCase__ : Optional[int] = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=UpperCamelCase__ , max_length=UpperCamelCase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset UpperCAmelCase__ : Optional[Any] = datasets.map( UpperCamelCase__ , batched=UpperCamelCase__ , remove_columns=["""idx""", """sentence1""", """sentence2"""] , load_from_cache_file=UpperCamelCase__ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library UpperCAmelCase__ : Optional[Any] = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(UpperCamelCase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(UpperCamelCase__ , padding="""max_length""" , max_length=1_2_8 , return_tensors="""pt""" ) return tokenizer.pad(UpperCamelCase__ , padding="""longest""" , return_tensors="""pt""" ) # Instantiate dataloaders. UpperCAmelCase__ : Optional[int] = DataLoader( tokenized_datasets["""train"""] , shuffle=UpperCamelCase__ , collate_fn=UpperCamelCase__ , batch_size=UpperCamelCase__ ) UpperCAmelCase__ : Any = DataLoader( tokenized_datasets["""validation"""] , shuffle=UpperCamelCase__ , collate_fn=UpperCamelCase__ , batch_size=UpperCamelCase__ ) return train_dataloader, eval_dataloader def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): # Initialize accelerator UpperCAmelCase__ : Union[str, Any] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs UpperCAmelCase__ : List[Any] = config["""lr"""] UpperCAmelCase__ : Any = int(config["""num_epochs"""] ) UpperCAmelCase__ : int = int(config["""seed"""] ) UpperCAmelCase__ : int = int(config["""batch_size"""] ) UpperCAmelCase__ : Optional[int] = args.model_name_or_path set_seed(UpperCamelCase__ ) UpperCAmelCase__ , UpperCAmelCase__ : Any = get_dataloaders(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) UpperCAmelCase__ : int = AutoModelForSequenceClassification.from_pretrained(UpperCamelCase__ , return_dict=UpperCamelCase__ ) # Instantiate optimizer UpperCAmelCase__ : Union[str, Any] = ( AdamW if accelerator.state.deepspeed_plugin is None or """optimizer""" not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) UpperCAmelCase__ : Tuple = optimizer_cls(params=model.parameters() , lr=UpperCamelCase__ ) if accelerator.state.deepspeed_plugin is not None: UpperCAmelCase__ : List[Any] = accelerator.state.deepspeed_plugin.deepspeed_config[ """gradient_accumulation_steps""" ] else: UpperCAmelCase__ : int = 1 UpperCAmelCase__ : Union[str, Any] = (len(UpperCamelCase__ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): UpperCAmelCase__ : Optional[int] = get_linear_schedule_with_warmup( optimizer=UpperCamelCase__ , num_warmup_steps=0 , num_training_steps=UpperCamelCase__ , ) else: UpperCAmelCase__ : Optional[int] = DummyScheduler(UpperCamelCase__ , total_num_steps=UpperCamelCase__ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ : Tuple = accelerator.prepare( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # We need to keep track of how many total steps we have iterated over UpperCAmelCase__ : Union[str, Any] = 0 # We also need to keep track of the stating epoch so files are named properly UpperCAmelCase__ : Optional[Any] = 0 # Now we train the model UpperCAmelCase__ : int = evaluate.load("""glue""" , """mrpc""" ) UpperCAmelCase__ : Dict = 0 UpperCAmelCase__ : Tuple = {} for epoch in range(UpperCamelCase__ , UpperCamelCase__ ): model.train() for step, batch in enumerate(UpperCamelCase__ ): UpperCAmelCase__ : Dict = model(**UpperCamelCase__ ) UpperCAmelCase__ : int = outputs.loss UpperCAmelCase__ : int = loss / gradient_accumulation_steps accelerator.backward(UpperCamelCase__ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() UpperCAmelCase__ : Any = 0 for step, batch in enumerate(UpperCamelCase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): UpperCAmelCase__ : Dict = model(**UpperCamelCase__ ) UpperCAmelCase__ : Optional[int] = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times UpperCAmelCase__ , UpperCAmelCase__ : Tuple = accelerator.gather( (predictions, batch["""labels"""]) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(UpperCamelCase__ ) - 1: UpperCAmelCase__ : str = predictions[: len(eval_dataloader.dataset ) - samples_seen] UpperCAmelCase__ : Union[str, Any] = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=UpperCamelCase__ , references=UpperCamelCase__ , ) UpperCAmelCase__ : Optional[int] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f'''epoch {epoch}:''' , UpperCamelCase__ ) UpperCAmelCase__ : Optional[Any] = eval_metric["""accuracy"""] if best_performance < eval_metric["accuracy"]: UpperCAmelCase__ : List[str] = eval_metric["""accuracy"""] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f'''Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}''' accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , """all_results.json""" ) , """w""" ) as f: json.dump(UpperCamelCase__ , UpperCamelCase__ ) def _UpperCamelCase ( ): UpperCAmelCase__ : str = argparse.ArgumentParser(description="""Simple example of training script tracking peak GPU memory usage.""" ) parser.add_argument( """--model_name_or_path""" , type=UpperCamelCase__ , default="""bert-base-cased""" , help="""Path to pretrained model or model identifier from huggingface.co/models.""" , required=UpperCamelCase__ , ) parser.add_argument( """--output_dir""" , type=UpperCamelCase__ , default=""".""" , help="""Optional save directory where all checkpoint folders will be stored. Default is the current working directory.""" , ) parser.add_argument( """--performance_lower_bound""" , type=UpperCamelCase__ , default=UpperCamelCase__ , help="""Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.""" , ) parser.add_argument( """--num_epochs""" , type=UpperCamelCase__ , default=3 , help="""Number of train epochs.""" , ) UpperCAmelCase__ : str = parser.parse_args() UpperCAmelCase__ : Optional[Any] = {"""lr""": 2e-5, """num_epochs""": args.num_epochs, """seed""": 4_2, """batch_size""": 1_6} training_function(UpperCamelCase__ , UpperCamelCase__ ) if __name__ == "__main__": main()
113
'''simple docstring''' import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging __A =logging.get_logger(__name__) def _UpperCamelCase ( UpperCamelCase__ ): UpperCAmelCase__ : List[str] = R"""\w+[.]\d+""" UpperCAmelCase__ : List[Any] = re.findall(UpperCamelCase__ , UpperCamelCase__ ) for pat in pats: UpperCAmelCase__ : str = key.replace(UpperCamelCase__ , """_""".join(pat.split(""".""" ) ) ) return key def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): UpperCAmelCase__ : str = pt_tuple_key[:-1] + ("""scale""",) if ( any("""norm""" in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): UpperCAmelCase__ : Union[str, Any] = pt_tuple_key[:-1] + ("""scale""",) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: UpperCAmelCase__ : str = pt_tuple_key[:-1] + ("""scale""",) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: UpperCAmelCase__ : Union[str, Any] = pt_tuple_key[:-1] + ("""embedding""",) return renamed_pt_tuple_key, pt_tensor # conv layer UpperCAmelCase__ : Any = pt_tuple_key[:-1] + ("""kernel""",) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: UpperCAmelCase__ : Any = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer UpperCAmelCase__ : Optional[int] = pt_tuple_key[:-1] + ("""kernel""",) if pt_tuple_key[-1] == "weight": UpperCAmelCase__ : Any = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight UpperCAmelCase__ : Optional[Any] = pt_tuple_key[:-1] + ("""weight""",) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias UpperCAmelCase__ : Dict = pt_tuple_key[:-1] + ("""bias""",) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=4_2 ): # Step 1: Convert pytorch tensor to numpy UpperCAmelCase__ : str = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params UpperCAmelCase__ : str = flax_model.init_weights(PRNGKey(UpperCamelCase__ ) ) UpperCAmelCase__ : Union[str, Any] = flatten_dict(UpperCamelCase__ ) UpperCAmelCase__ : int = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): UpperCAmelCase__ : str = rename_key(UpperCamelCase__ ) UpperCAmelCase__ : Optional[Any] = tuple(renamed_pt_key.split(""".""" ) ) # Correctly rename weight parameters UpperCAmelCase__ , UpperCAmelCase__ : Tuple = rename_key_and_reshape_tensor(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f'''PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ''' f'''{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.''' ) # also add unexpected weight so that warning is thrown UpperCAmelCase__ : str = jnp.asarray(UpperCamelCase__ ) return unflatten_dict(UpperCamelCase__ )
113
1
'''simple docstring''' import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) lowerCAmelCase__ : Tuple = { """iou_prediction_head.layers.0""": """iou_prediction_head.proj_in""", """iou_prediction_head.layers.1""": """iou_prediction_head.layers.0""", """iou_prediction_head.layers.2""": """iou_prediction_head.proj_out""", """mask_decoder.output_upscaling.0""": """mask_decoder.upscale_conv1""", """mask_decoder.output_upscaling.1""": """mask_decoder.upscale_layer_norm""", """mask_decoder.output_upscaling.3""": """mask_decoder.upscale_conv2""", """mask_downscaling.0""": """mask_embed.conv1""", """mask_downscaling.1""": """mask_embed.layer_norm1""", """mask_downscaling.3""": """mask_embed.conv2""", """mask_downscaling.4""": """mask_embed.layer_norm2""", """mask_downscaling.6""": """mask_embed.conv3""", """point_embeddings""": """point_embed""", """pe_layer.positional_encoding_gaussian_matrix""": """shared_embedding.positional_embedding""", """image_encoder""": """vision_encoder""", """neck.0""": """neck.conv1""", """neck.1""": """neck.layer_norm1""", """neck.2""": """neck.conv2""", """neck.3""": """neck.layer_norm2""", """patch_embed.proj""": """patch_embed.projection""", """.norm""": """.layer_norm""", """blocks""": """layers""", } def _a ( __lowerCAmelCase : str ): """simple docstring""" snake_case__ : int = {} state_dict.pop('''pixel_mean''' , __lowerCAmelCase ) state_dict.pop('''pixel_std''' , __lowerCAmelCase ) snake_case__ : str = r'''.*.output_hypernetworks_mlps.(\d+).layers.(\d+).*''' for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: snake_case__ : Dict = key.replace(__lowerCAmelCase , __lowerCAmelCase ) if re.match(__lowerCAmelCase , __lowerCAmelCase ): snake_case__ : int = int(re.match(__lowerCAmelCase , __lowerCAmelCase ).group(2 ) ) if layer_nb == 0: snake_case__ : str = key.replace('''layers.0''' , '''proj_in''' ) elif layer_nb == 1: snake_case__ : Tuple = key.replace('''layers.1''' , '''layers.0''' ) elif layer_nb == 2: snake_case__ : Any = key.replace('''layers.2''' , '''proj_out''' ) snake_case__ : str = value snake_case__ : Dict = model_state_dict[ '''prompt_encoder.shared_embedding.positional_embedding''' ] return model_state_dict def _a ( __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Any , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Union[str, Any]="ybelkada/segment-anything" ): """simple docstring""" snake_case__ : int = hf_hub_download(__lowerCAmelCase , F"""checkpoints/{model_name}.pth""" ) if "sam_vit_b" in model_name: snake_case__ : Tuple = SamConfig() elif "sam_vit_l" in model_name: snake_case__ : Any = SamVisionConfig( hidden_size=10_24 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) snake_case__ : Optional[int] = SamConfig( vision_config=__lowerCAmelCase , ) elif "sam_vit_h" in model_name: snake_case__ : List[str] = SamVisionConfig( hidden_size=12_80 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) snake_case__ : List[str] = SamConfig( vision_config=__lowerCAmelCase , ) snake_case__ : str = torch.load(__lowerCAmelCase , map_location='''cpu''' ) snake_case__ : Dict = replace_keys(__lowerCAmelCase ) snake_case__ : Optional[Any] = SamImageProcessor() snake_case__ : Optional[Any] = SamProcessor(image_processor=__lowerCAmelCase ) snake_case__ : List[str] = SamModel(__lowerCAmelCase ) hf_model.load_state_dict(__lowerCAmelCase ) snake_case__ : Union[str, Any] = hf_model.to('''cuda''' ) snake_case__ : str = '''https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png''' snake_case__ : Dict = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ).convert('''RGB''' ) snake_case__ : Optional[int] = [[[4_00, 6_50]]] snake_case__ : str = [[1]] snake_case__ : Any = processor(images=np.array(__lowerCAmelCase ) , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): snake_case__ : Any = hf_model(**__lowerCAmelCase ) snake_case__ : Union[str, Any] = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579890251159668 snake_case__ : List[str] = processor( images=np.array(__lowerCAmelCase ) , input_points=__lowerCAmelCase , input_labels=__lowerCAmelCase , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): snake_case__ : List[str] = hf_model(**__lowerCAmelCase ) snake_case__ : List[Any] = output.iou_scores.squeeze() assert scores[-1].item() == 0.9712603092193604 snake_case__ : List[str] = ((75, 2_75, 17_25, 8_50),) snake_case__ : Dict = processor(images=np.array(__lowerCAmelCase ) , input_boxes=__lowerCAmelCase , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): snake_case__ : Any = hf_model(**__lowerCAmelCase ) snake_case__ : Optional[int] = output.iou_scores.squeeze() assert scores[-1].item() == 0.8686015605926514 # Test with 2 points and 1 image. snake_case__ : Any = [[[4_00, 6_50], [8_00, 6_50]]] snake_case__ : Union[str, Any] = [[1, 1]] snake_case__ : List[str] = processor( images=np.array(__lowerCAmelCase ) , input_points=__lowerCAmelCase , input_labels=__lowerCAmelCase , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): snake_case__ : List[Any] = hf_model(**__lowerCAmelCase ) snake_case__ : Any = output.iou_scores.squeeze() assert scores[-1].item() == 0.9936047792434692 if __name__ == "__main__": lowerCAmelCase__ : Tuple = argparse.ArgumentParser() lowerCAmelCase__ : Tuple = ["""sam_vit_b_01ec64""", """sam_vit_h_4b8939""", """sam_vit_l_0b3195"""] parser.add_argument( """--model_name""", default="""sam_vit_h_4b8939""", choices=choices, type=str, help="""Path to hf config.json of model to convert""", ) parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to push the model and processor to the hub after converting""", ) parser.add_argument( """--model_hub_id""", default="""ybelkada/segment-anything""", choices=choices, type=str, help="""Path to hf config.json of model to convert""", ) lowerCAmelCase__ : Optional[Any] = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
347
'''simple docstring''' import inspect import unittest from transformers import ViTHybridConfig from transformers.testing_utils import require_accelerate, require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel from transformers.models.vit_hybrid.modeling_vit_hybrid import VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class a : """simple docstring""" def __init__( self : Optional[int] , snake_case_ : str , snake_case_ : List[Any]=1_3 , snake_case_ : Any=6_4 , snake_case_ : Optional[int]=2 , snake_case_ : int=3 , snake_case_ : str=True , snake_case_ : Dict=True , snake_case_ : Optional[Any]=3_2 , snake_case_ : Optional[Any]=5 , snake_case_ : List[Any]=4 , snake_case_ : List[Any]=3_7 , snake_case_ : Dict="gelu" , snake_case_ : Union[str, Any]=0.1 , snake_case_ : int=0.1 , snake_case_ : int=1_0 , snake_case_ : int=0.0_2 , snake_case_ : List[Any]=[1, 1_6, 4, 4] , snake_case_ : Any=None , ): '''simple docstring''' snake_case__ : Dict = parent snake_case__ : Optional[int] = batch_size snake_case__ : Union[str, Any] = image_size snake_case__ : str = patch_size snake_case__ : Optional[Any] = num_channels snake_case__ : List[str] = is_training snake_case__ : List[Any] = use_labels snake_case__ : List[str] = hidden_size snake_case__ : List[Any] = num_hidden_layers snake_case__ : str = num_attention_heads snake_case__ : Dict = intermediate_size snake_case__ : Tuple = hidden_act snake_case__ : str = hidden_dropout_prob snake_case__ : List[Any] = attention_probs_dropout_prob snake_case__ : List[str] = type_sequence_label_size snake_case__ : Any = initializer_range snake_case__ : Optional[Any] = scope snake_case__ : Optional[int] = backbone_featmap_shape # in ViT hybrid, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) # the number of patches is based on the feature map of the backbone, which by default uses an output stride # of 32, which means that the feature map has a spatial resolution of 1/32 of the input image size snake_case__ : Dict = (self.image_size // 3_2) ** 2 snake_case__ : List[str] = num_patches + 1 def __magic_name__ ( self : List[str] ): '''simple docstring''' snake_case__ : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case__ : int = None if self.use_labels: snake_case__ : List[str] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case__ : Any = self.get_config() return config, pixel_values, labels def __magic_name__ ( self : str ): '''simple docstring''' snake_case__ : str = { '''global_padding''': '''same''', '''layer_type''': '''bottleneck''', '''depths''': [3, 4, 9], '''out_features''': ['''stage1''', '''stage2''', '''stage3'''], '''embedding_dynamic_padding''': True, '''hidden_sizes''': [4, 8, 1_6, 3_2], '''num_groups''': 2, } return ViTHybridConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case_ , initializer_range=self.initializer_range , backbone_featmap_shape=self.backbone_featmap_shape , backbone_config=snake_case_ , ) def __magic_name__ ( self : Optional[Any] , snake_case_ : str , snake_case_ : Any , snake_case_ : List[str] ): '''simple docstring''' snake_case__ : Dict = ViTHybridModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() snake_case__ : List[str] = model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __magic_name__ ( self : str , snake_case_ : Any , snake_case_ : Tuple , snake_case_ : Any ): '''simple docstring''' snake_case__ : Optional[int] = self.type_sequence_label_size snake_case__ : int = ViTHybridForImageClassification(snake_case_ ) model.to(snake_case_ ) model.eval() snake_case__ : Optional[int] = model(snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __magic_name__ ( self : Union[str, Any] ): '''simple docstring''' snake_case__ : List[Any] = self.prepare_config_and_inputs() snake_case__ , snake_case__ , snake_case__ : str = config_and_inputs snake_case__ : str = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" __UpperCAmelCase = (ViTHybridModel, ViTHybridForImageClassification) if is_torch_available() else () __UpperCAmelCase = ( {"""feature-extraction""": ViTHybridModel, """image-classification""": ViTHybridForImageClassification} if is_torch_available() else {} ) __UpperCAmelCase = False __UpperCAmelCase = False __UpperCAmelCase = False def __magic_name__ ( self : Tuple ): '''simple docstring''' snake_case__ : Tuple = ViTHybridModelTester(self ) snake_case__ : int = ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ , hidden_size=3_7 ) def __magic_name__ ( self : Union[str, Any] ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def __magic_name__ ( self : List[str] ): '''simple docstring''' pass def __magic_name__ ( self : str ): '''simple docstring''' snake_case__ , snake_case__ : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case__ : Any = model_class(snake_case_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case__ : Tuple = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case_ , nn.Linear ) ) def __magic_name__ ( self : Union[str, Any] ): '''simple docstring''' snake_case__ , snake_case__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case__ : Tuple = model_class(snake_case_ ) snake_case__ : Optional[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case__ : str = [*signature.parameters.keys()] snake_case__ : List[Any] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , snake_case_ ) def __magic_name__ ( self : int ): '''simple docstring''' snake_case__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def __magic_name__ ( self : Dict ): '''simple docstring''' snake_case__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case_ ) def __magic_name__ ( self : List[Any] ): '''simple docstring''' snake_case__ , snake_case__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() snake_case__ : Tuple = _config_zero_init(snake_case_ ) for model_class in self.all_model_classes: snake_case__ : Optional[Any] = model_class(config=snake_case_ ) # Skip the check for the backbone for name, module in model.named_modules(): if module.__class__.__name__ == "ViTHybridPatchEmbeddings": snake_case__ : Optional[Any] = [F"""{name}.{key}""" for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @slow def __magic_name__ ( self : List[Any] ): '''simple docstring''' for model_name in VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case__ : Any = ViTHybridModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def _a ( ): """simple docstring""" snake_case__ : Optional[Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class a ( unittest.TestCase ): """simple docstring""" @cached_property def __magic_name__ ( self : List[Any] ): '''simple docstring''' return ( ViTHybridImageProcessor.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def __magic_name__ ( self : str ): '''simple docstring''' snake_case__ : Tuple = ViTHybridForImageClassification.from_pretrained(VIT_HYBRID_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( snake_case_ ) snake_case__ : Dict = self.default_image_processor snake_case__ : Optional[int] = prepare_img() snake_case__ : Optional[Any] = image_processor(images=snake_case_ , return_tensors='''pt''' ).to(snake_case_ ) # forward pass with torch.no_grad(): snake_case__ : Dict = model(**snake_case_ ) # verify the logits snake_case__ : Any = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , snake_case_ ) snake_case__ : Union[str, Any] = torch.tensor([-1.9_0_9_0, -0.4_9_9_3, -0.2_3_8_9] ).to(snake_case_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case_ , atol=1e-4 ) ) @slow @require_accelerate def __magic_name__ ( self : Tuple ): '''simple docstring''' snake_case__ : Tuple = ViTHybridImageProcessor.from_pretrained('''google/vit-hybrid-base-bit-384''' ) snake_case__ : Optional[int] = ViTHybridForImageClassification.from_pretrained('''google/vit-hybrid-base-bit-384''' , device_map='''auto''' ) snake_case__ : int = prepare_img() snake_case__ : int = image_processor(images=snake_case_ , return_tensors='''pt''' ) snake_case__ : int = model(**snake_case_ ) snake_case__ : List[str] = outputs.logits # model predicts one of the 1000 ImageNet classes snake_case__ : Union[str, Any] = logits.argmax(-1 ).item() self.assertTrue(model.config.idalabel[predicted_class_idx] , '''tabby, tabby cat''' )
347
1
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class __magic_name__ (unittest.TestCase ): '''simple docstring''' def __init__( self:Optional[Any] , _a:List[Any] , _a:Any=7 , _a:str=3 , _a:Tuple=10 , _a:str=18 , _a:List[str]=30 , _a:Tuple=4_00 , _a:str=True , _a:List[str]=None , _a:List[str]=True , _a:Optional[Any]=[0.5, 0.5, 0.5] , _a:List[str]=[0.5, 0.5, 0.5] , _a:int=None , ): snake_case__ = size if size is not None else {'''shortest_edge''': 18} snake_case__ = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} snake_case__ = parent snake_case__ = batch_size snake_case__ = num_channels snake_case__ = num_frames snake_case__ = image_size snake_case__ = min_resolution snake_case__ = max_resolution snake_case__ = do_resize snake_case__ = size snake_case__ = do_normalize snake_case__ = image_mean snake_case__ = image_std snake_case__ = crop_size def SCREAMING_SNAKE_CASE__ ( self:Dict ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class __magic_name__ (snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : Union[str, Any] = VivitImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE__ ( self:int ): snake_case__ = VivitImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): snake_case__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_a , '''image_mean''' ) ) self.assertTrue(hasattr(_a , '''image_std''' ) ) self.assertTrue(hasattr(_a , '''do_normalize''' ) ) self.assertTrue(hasattr(_a , '''do_resize''' ) ) self.assertTrue(hasattr(_a , '''do_center_crop''' ) ) self.assertTrue(hasattr(_a , '''size''' ) ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): snake_case__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18} ) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18} ) snake_case__ = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42} ) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84} ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): # Initialize image_processing snake_case__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos snake_case__ = prepare_video_inputs(self.image_processor_tester , equal_resolution=_a ) for video in video_inputs: self.assertIsInstance(_a , _a ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input snake_case__ = image_processing(video_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched snake_case__ = image_processing(_a , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): # Initialize image_processing snake_case__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case__ = prepare_video_inputs(self.image_processor_tester , equal_resolution=_a , numpify=_a ) for video in video_inputs: self.assertIsInstance(_a , _a ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input snake_case__ = image_processing(video_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched snake_case__ = image_processing(_a , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): # Initialize image_processing snake_case__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case__ = prepare_video_inputs(self.image_processor_tester , equal_resolution=_a , torchify=_a ) for video in video_inputs: self.assertIsInstance(_a , _a ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input snake_case__ = image_processing(video_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched snake_case__ = image_processing(_a , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
708
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class __magic_name__ (unittest.TestCase ): '''simple docstring''' def __init__( self:Optional[Any] , _a:List[Any] , _a:Any=7 , _a:str=3 , _a:Tuple=10 , _a:str=18 , _a:List[str]=30 , _a:Tuple=4_00 , _a:str=True , _a:List[str]=None , _a:List[str]=True , _a:Optional[Any]=[0.5, 0.5, 0.5] , _a:List[str]=[0.5, 0.5, 0.5] , _a:int=None , ): snake_case__ = size if size is not None else {'''shortest_edge''': 18} snake_case__ = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} snake_case__ = parent snake_case__ = batch_size snake_case__ = num_channels snake_case__ = num_frames snake_case__ = image_size snake_case__ = min_resolution snake_case__ = max_resolution snake_case__ = do_resize snake_case__ = size snake_case__ = do_normalize snake_case__ = image_mean snake_case__ = image_std snake_case__ = crop_size def SCREAMING_SNAKE_CASE__ ( self:Dict ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class __magic_name__ (snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : Union[str, Any] = VivitImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE__ ( self:int ): snake_case__ = VivitImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): snake_case__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_a , '''image_mean''' ) ) self.assertTrue(hasattr(_a , '''image_std''' ) ) self.assertTrue(hasattr(_a , '''do_normalize''' ) ) self.assertTrue(hasattr(_a , '''do_resize''' ) ) self.assertTrue(hasattr(_a , '''do_center_crop''' ) ) self.assertTrue(hasattr(_a , '''size''' ) ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): snake_case__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18} ) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18} ) snake_case__ = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42} ) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84} ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): # Initialize image_processing snake_case__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos snake_case__ = prepare_video_inputs(self.image_processor_tester , equal_resolution=_a ) for video in video_inputs: self.assertIsInstance(_a , _a ) self.assertIsInstance(video[0] , Image.Image ) # Test not batched input snake_case__ = image_processing(video_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched snake_case__ = image_processing(_a , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): # Initialize image_processing snake_case__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case__ = prepare_video_inputs(self.image_processor_tester , equal_resolution=_a , numpify=_a ) for video in video_inputs: self.assertIsInstance(_a , _a ) self.assertIsInstance(video[0] , np.ndarray ) # Test not batched input snake_case__ = image_processing(video_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched snake_case__ = image_processing(_a , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): # Initialize image_processing snake_case__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case__ = prepare_video_inputs(self.image_processor_tester , equal_resolution=_a , torchify=_a ) for video in video_inputs: self.assertIsInstance(_a , _a ) self.assertIsInstance(video[0] , torch.Tensor ) # Test not batched input snake_case__ = image_processing(video_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape , ( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched snake_case__ = image_processing(_a , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
208
0
# flake8: noqa # Lint as: python3 from typing import Dict, List, Optional, Type from .. import config from ..utils import logging from .formatting import ( ArrowFormatter, CustomFormatter, Formatter, PandasFormatter, PythonFormatter, TensorFormatter, format_table, query_table, ) from .np_formatter import NumpyFormatter _UpperCAmelCase = logging.get_logger(__name__) _UpperCAmelCase = {} _UpperCAmelCase = {} _UpperCAmelCase = {} def UpperCamelCase ( __lowercase : Optional[Any] ,__lowercase : List[str] ,__lowercase : List[Any] = None ,): '''simple docstring''' A_ : List[str] = aliases if aliases is not None else [] if format_type in _FORMAT_TYPES: logger.warning( f'''Overwriting format type \'{format_type}\' ({_FORMAT_TYPES[format_type].__name__} -> {formatter_cls.__name__})''' ) A_ : Tuple = formatter_cls for alias in set(aliases + [format_type] ): if alias in _FORMAT_TYPES_ALIASES: logger.warning( f'''Overwriting format type alias \'{alias}\' ({_FORMAT_TYPES_ALIASES[alias]} -> {format_type})''' ) A_ : Dict = format_type def UpperCamelCase ( __lowercase : int ,__lowercase : Optional[int] ,__lowercase : str = None ): '''simple docstring''' A_ : Union[str, Any] = aliases if aliases is not None else [] for alias in set(aliases + [format_type] ): A_ : str = unavailable_error # Here we define all the available formatting functions that can be used by `Dataset.set_format` _register_formatter(PythonFormatter, None, aliases=["""python"""]) _register_formatter(ArrowFormatter, """arrow""", aliases=["""pa""", """pyarrow"""]) _register_formatter(NumpyFormatter, """numpy""", aliases=["""np"""]) _register_formatter(PandasFormatter, """pandas""", aliases=["""pd"""]) _register_formatter(CustomFormatter, """custom""") if config.TORCH_AVAILABLE: from .torch_formatter import TorchFormatter _register_formatter(TorchFormatter, """torch""", aliases=["""pt""", """pytorch"""]) else: _UpperCAmelCase = ValueError("""PyTorch needs to be installed to be able to return PyTorch tensors.""") _register_unavailable_formatter(_torch_error, """torch""", aliases=["""pt""", """pytorch"""]) if config.TF_AVAILABLE: from .tf_formatter import TFFormatter _register_formatter(TFFormatter, """tensorflow""", aliases=["""tf"""]) else: _UpperCAmelCase = ValueError("""Tensorflow needs to be installed to be able to return Tensorflow tensors.""") _register_unavailable_formatter(_tf_error, """tensorflow""", aliases=["""tf"""]) if config.JAX_AVAILABLE: from .jax_formatter import JaxFormatter _register_formatter(JaxFormatter, """jax""", aliases=[]) else: _UpperCAmelCase = ValueError("""JAX needs to be installed to be able to return JAX arrays.""") _register_unavailable_formatter(_jax_error, """jax""", aliases=[]) def UpperCamelCase ( __lowercase : Dict ): '''simple docstring''' if format_type in _FORMAT_TYPES_ALIASES: return _FORMAT_TYPES_ALIASES[format_type] else: return format_type def UpperCamelCase ( __lowercase : str ,**__lowercase : str ): '''simple docstring''' A_ : Tuple = get_format_type_from_alias(_SCREAMING_SNAKE_CASE ) if format_type in _FORMAT_TYPES: return _FORMAT_TYPES[format_type](**_SCREAMING_SNAKE_CASE ) if format_type in _FORMAT_TYPES_ALIASES_UNAVAILABLE: raise _FORMAT_TYPES_ALIASES_UNAVAILABLE[format_type] else: raise ValueError( f'''Return type should be None or selected in {list(type for type in _FORMAT_TYPES.keys() if type != None )}, but got \'{format_type}\'''' )
558
import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments @require_torch class UpperCamelCase__ (unittest.TestCase ): '''simple docstring''' def _lowercase ( self , UpperCamelCase__ ) -> Tuple: for model_result in results.values(): for batch_size, sequence_length in zip(model_result["bs"] , model_result["ss"] ): lowerCamelCase : Union[str, Any] = model_result["result"][batch_size][sequence_length] self.assertIsNotNone(UpperCamelCase__ ) def _lowercase ( self ) -> Tuple: lowerCamelCase : Tuple = "sshleifer/tiny-gpt2" lowerCamelCase : Tuple = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) lowerCamelCase : Dict = PyTorchBenchmark(UpperCamelCase__ ) lowerCamelCase : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self ) -> int: lowerCamelCase : List[str] = "sgugger/tiny-distilbert-classification" lowerCamelCase : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , only_pretrain_model=UpperCamelCase__ , ) lowerCamelCase : Tuple = PyTorchBenchmark(UpperCamelCase__ ) lowerCamelCase : List[str] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self ) -> int: lowerCamelCase : Dict = "sshleifer/tiny-gpt2" lowerCamelCase : Dict = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , torchscript=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) lowerCamelCase : Union[str, Any] = PyTorchBenchmark(UpperCamelCase__ ) lowerCamelCase : int = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) @unittest.skipIf(torch_device == "cpu" , "Cant do half precision" ) def _lowercase ( self ) -> Optional[Any]: lowerCamelCase : Any = "sshleifer/tiny-gpt2" lowerCamelCase : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , fpaa=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) lowerCamelCase : Optional[int] = PyTorchBenchmark(UpperCamelCase__ ) lowerCamelCase : Tuple = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self ) -> Optional[int]: lowerCamelCase : int = "sshleifer/tiny-gpt2" lowerCamelCase : List[str] = AutoConfig.from_pretrained(UpperCamelCase__ ) # set architectures equal to `None` lowerCamelCase : Union[str, Any] = None lowerCamelCase : str = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) lowerCamelCase : List[Any] = PyTorchBenchmark(UpperCamelCase__ , configs=[config] ) lowerCamelCase : List[Any] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self ) -> str: lowerCamelCase : Optional[int] = "sshleifer/tiny-gpt2" lowerCamelCase : Dict = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) lowerCamelCase : int = PyTorchBenchmark(UpperCamelCase__ ) lowerCamelCase : str = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) @unittest.skipIf(torch_device == "cpu" , "Can't do half precision" ) def _lowercase ( self ) -> Dict: lowerCamelCase : List[str] = "sshleifer/tiny-gpt2" lowerCamelCase : Optional[Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , fpaa=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) lowerCamelCase : Dict = PyTorchBenchmark(UpperCamelCase__ ) lowerCamelCase : int = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _lowercase ( self ) -> Union[str, Any]: lowerCamelCase : Union[str, Any] = "sshleifer/tiny-gpt2" lowerCamelCase : Union[str, Any] = AutoConfig.from_pretrained(UpperCamelCase__ ) lowerCamelCase : Tuple = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) lowerCamelCase : Dict = PyTorchBenchmark(UpperCamelCase__ , configs=[config] ) lowerCamelCase : List[str] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self ) -> int: lowerCamelCase : Tuple = "sshleifer/tinier_bart" lowerCamelCase : Optional[Any] = AutoConfig.from_pretrained(UpperCamelCase__ ) lowerCamelCase : Tuple = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) lowerCamelCase : Any = PyTorchBenchmark(UpperCamelCase__ , configs=[config] ) lowerCamelCase : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_inference_result ) self.check_results_dict_not_empty(results.memory_inference_result ) def _lowercase ( self ) -> Any: lowerCamelCase : List[Any] = "sshleifer/tiny-gpt2" lowerCamelCase : Optional[Any] = AutoConfig.from_pretrained(UpperCamelCase__ ) lowerCamelCase : List[str] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) lowerCamelCase : int = PyTorchBenchmark(UpperCamelCase__ , configs=[config] ) lowerCamelCase : Optional[int] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _lowercase ( self ) -> Dict: lowerCamelCase : List[str] = "sshleifer/tinier_bart" lowerCamelCase : Optional[Any] = AutoConfig.from_pretrained(UpperCamelCase__ ) lowerCamelCase : Union[str, Any] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , multi_process=UpperCamelCase__ , ) lowerCamelCase : Tuple = PyTorchBenchmark(UpperCamelCase__ , configs=[config] ) lowerCamelCase : List[str] = benchmark.run() self.check_results_dict_not_empty(results.time_train_result ) self.check_results_dict_not_empty(results.memory_train_result ) def _lowercase ( self ) -> Union[str, Any]: lowerCamelCase : Optional[Any] = "sshleifer/tiny-gpt2" with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase : Optional[int] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , save_to_csv=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , inference_time_csv_file=os.path.join(UpperCamelCase__ , "inf_time.csv" ) , train_memory_csv_file=os.path.join(UpperCamelCase__ , "train_mem.csv" ) , inference_memory_csv_file=os.path.join(UpperCamelCase__ , "inf_mem.csv" ) , train_time_csv_file=os.path.join(UpperCamelCase__ , "train_time.csv" ) , env_info_csv_file=os.path.join(UpperCamelCase__ , "env.csv" ) , multi_process=UpperCamelCase__ , ) lowerCamelCase : str = PyTorchBenchmark(UpperCamelCase__ ) benchmark.run() self.assertTrue(Path(os.path.join(UpperCamelCase__ , "inf_time.csv" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , "train_time.csv" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , "inf_mem.csv" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , "train_mem.csv" ) ).exists() ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , "env.csv" ) ).exists() ) def _lowercase ( self ) -> Dict: lowerCamelCase : Union[str, Any] = "sshleifer/tiny-gpt2" def _check_summary_is_not_empty(UpperCamelCase__ ): self.assertTrue(hasattr(UpperCamelCase__ , "sequential" ) ) self.assertTrue(hasattr(UpperCamelCase__ , "cumulative" ) ) self.assertTrue(hasattr(UpperCamelCase__ , "current" ) ) self.assertTrue(hasattr(UpperCamelCase__ , "total" ) ) with tempfile.TemporaryDirectory() as tmp_dir: lowerCamelCase : List[str] = PyTorchBenchmarkArguments( models=[MODEL_ID] , training=UpperCamelCase__ , inference=UpperCamelCase__ , sequence_lengths=[8] , batch_sizes=[1] , log_filename=os.path.join(UpperCamelCase__ , "log.txt" ) , log_print=UpperCamelCase__ , trace_memory_line_by_line=UpperCamelCase__ , multi_process=UpperCamelCase__ , ) lowerCamelCase : Any = PyTorchBenchmark(UpperCamelCase__ ) lowerCamelCase : Dict = benchmark.run() _check_summary_is_not_empty(result.inference_summary ) _check_summary_is_not_empty(result.train_summary ) self.assertTrue(Path(os.path.join(UpperCamelCase__ , "log.txt" ) ).exists() )
311
0
"""simple docstring""" import warnings warnings.warn( "memory_utils has been reorganized to utils.memory. Import `find_executable_batchsize` from the main `__init__`: " "`from accelerate import find_executable_batch_size` to avoid this warning.", FutureWarning, )
625
"""simple docstring""" from __future__ import annotations import unittest from transformers import BlenderbotSmallConfig, BlenderbotSmallTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel @require_tf class __SCREAMING_SNAKE_CASE : '''simple docstring''' _a = BlenderbotSmallConfig _a = {} _a = 'gelu' def __init__( self : Union[str, Any], lowerCamelCase : List[str], lowerCamelCase : Dict=13, lowerCamelCase : Optional[Any]=7, lowerCamelCase : Optional[int]=True, lowerCamelCase : int=False, lowerCamelCase : Union[str, Any]=99, lowerCamelCase : str=32, lowerCamelCase : List[Any]=2, lowerCamelCase : Optional[int]=4, lowerCamelCase : Union[str, Any]=37, lowerCamelCase : str=0.1, lowerCamelCase : Optional[int]=0.1, lowerCamelCase : Optional[Any]=20, lowerCamelCase : int=2, lowerCamelCase : Any=1, lowerCamelCase : Optional[Any]=0, )-> List[str]: lowerCamelCase__ : Any =parent lowerCamelCase__ : Dict =batch_size lowerCamelCase__ : Optional[int] =seq_length lowerCamelCase__ : Tuple =is_training lowerCamelCase__ : Dict =use_labels lowerCamelCase__ : List[Any] =vocab_size lowerCamelCase__ : str =hidden_size lowerCamelCase__ : str =num_hidden_layers lowerCamelCase__ : Union[str, Any] =num_attention_heads lowerCamelCase__ : Any =intermediate_size lowerCamelCase__ : Dict =hidden_dropout_prob lowerCamelCase__ : List[Any] =attention_probs_dropout_prob lowerCamelCase__ : str =max_position_embeddings lowerCamelCase__ : Optional[int] =eos_token_id lowerCamelCase__ : str =pad_token_id lowerCamelCase__ : Union[str, Any] =bos_token_id def snake_case ( self : Any )-> Any: lowerCamelCase__ : Any =ids_tensor([self.batch_size, self.seq_length - 1], self.vocab_size ) lowerCamelCase__ : Tuple =tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ), 1 ) lowerCamelCase__ : Any =tf.concat([input_ids, eos_tensor], axis=1 ) lowerCamelCase__ : Optional[int] =ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) lowerCamelCase__ : int =self.config_cls( vocab_size=self.vocab_size, d_model=self.hidden_size, encoder_layers=self.num_hidden_layers, decoder_layers=self.num_hidden_layers, encoder_attention_heads=self.num_attention_heads, decoder_attention_heads=self.num_attention_heads, encoder_ffn_dim=self.intermediate_size, decoder_ffn_dim=self.intermediate_size, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, eos_token_ids=[2], bos_token_id=self.bos_token_id, pad_token_id=self.pad_token_id, decoder_start_token_id=self.pad_token_id, **self.config_updates, ) lowerCamelCase__ : Optional[int] =prepare_blenderbot_small_inputs_dict(lowerCamelCase, lowerCamelCase, lowerCamelCase ) return config, inputs_dict def snake_case ( self : Any, lowerCamelCase : str, lowerCamelCase : Any )-> Optional[Any]: lowerCamelCase__ : Union[str, Any] =TFBlenderbotSmallModel(config=lowerCamelCase ).get_decoder() lowerCamelCase__ : List[Any] =inputs_dict['''input_ids'''] lowerCamelCase__ : Optional[int] =input_ids[:1, :] lowerCamelCase__ : str =inputs_dict['''attention_mask'''][:1, :] lowerCamelCase__ : Union[str, Any] =inputs_dict['''head_mask'''] lowerCamelCase__ : Optional[Any] =1 # first forward pass lowerCamelCase__ : Dict =model(lowerCamelCase, attention_mask=lowerCamelCase, head_mask=lowerCamelCase, use_cache=lowerCamelCase ) lowerCamelCase__ , lowerCamelCase__ : List[str] =outputs.to_tuple() # create hypothetical next token and extent to next_input_ids lowerCamelCase__ : Union[str, Any] =ids_tensor((self.batch_size, 3), config.vocab_size ) lowerCamelCase__ : Tuple =tf.cast(ids_tensor((self.batch_size, 3), 2 ), tf.inta ) # append to next input_ids and lowerCamelCase__ : List[str] =tf.concat([input_ids, next_tokens], axis=-1 ) lowerCamelCase__ : str =tf.concat([attention_mask, next_attn_mask], axis=-1 ) lowerCamelCase__ : Optional[int] =model(lowerCamelCase, attention_mask=lowerCamelCase )[0] lowerCamelCase__ : Union[str, Any] =model(lowerCamelCase, attention_mask=lowerCamelCase, past_key_values=lowerCamelCase )[0] self.parent.assertEqual(next_tokens.shape[1], output_from_past.shape[1] ) # select random slice lowerCamelCase__ : Tuple =int(ids_tensor((1,), output_from_past.shape[-1] ) ) lowerCamelCase__ : int =output_from_no_past[:, -3:, random_slice_idx] lowerCamelCase__ : List[str] =output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowerCamelCase, lowerCamelCase, rtol=1E-3 ) def snake_case__ ( __lowerCamelCase : Tuple , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Union[str, Any] , __lowerCamelCase : Any=None , __lowerCamelCase : List[str]=None , __lowerCamelCase : Tuple=None , __lowerCamelCase : Dict=None , __lowerCamelCase : Optional[int]=None , ): """simple docstring""" if attention_mask is None: lowerCamelCase__ : List[str] =tf.cast(tf.math.not_equal(__lowerCamelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: lowerCamelCase__ : str =tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: lowerCamelCase__ : int =tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: lowerCamelCase__ : int =tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: lowerCamelCase__ : List[str] =tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): '''simple docstring''' _a = ( (TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel) if is_tf_available() else () ) _a = (TFBlenderbotSmallForConditionalGeneration,) if is_tf_available() else () _a = ( { 'conversational': TFBlenderbotSmallForConditionalGeneration, 'feature-extraction': TFBlenderbotSmallModel, 'summarization': TFBlenderbotSmallForConditionalGeneration, 'text2text-generation': TFBlenderbotSmallForConditionalGeneration, 'translation': TFBlenderbotSmallForConditionalGeneration, } if is_tf_available() else {} ) _a = True _a = False _a = False def snake_case ( self : Any )-> str: lowerCamelCase__ : Tuple =TFBlenderbotSmallModelTester(self ) lowerCamelCase__ : Union[str, Any] =ConfigTester(self, config_class=lowerCamelCase ) def snake_case ( self : Any )-> Optional[int]: self.config_tester.run_common_tests() def snake_case ( self : int )-> str: lowerCamelCase__ : List[Any] =self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCamelCase ) @require_tokenizers @require_tf class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' _a = [ 'Social anxiety\nWow, I am never shy. Do you have anxiety?\nYes. I end up sweating and blushing and feel like ' ' i\'m going to throw up.\nand why is that?' ] _a = 'facebook/blenderbot_small-90M' @cached_property def snake_case ( self : Any )-> List[Any]: # use "old" tokenizer here because of bug when downloading new tokenizer return BlenderbotSmallTokenizer.from_pretrained('''facebook/blenderbot-90M''' ) @cached_property def snake_case ( self : int )-> List[Any]: lowerCamelCase__ : str =TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def snake_case ( self : Tuple )-> int: lowerCamelCase__ : Dict =self.tokenizer(self.src_text, return_tensors='''tf''' ) lowerCamelCase__ : Any =self.model.generate( model_inputs.input_ids, attention_mask=model_inputs.attention_mask, num_beams=2, use_cache=lowerCamelCase, ) lowerCamelCase__ : Any =self.tokenizer.batch_decode(generated_ids.numpy(), skip_special_tokens=lowerCamelCase )[0] assert generated_words in ( "i don't know. i just feel like i'm going to throw up. it's not fun.", "i'm not sure. i just feel like i've been feeling like i have to be in a certain place", "i'm not sure. i just feel like i've been in a bad situation.", )
625
1
from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import ( BackboneOutput, BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import ( add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging, replace_return_docstrings, ) from ...utils.backbone_utils import BackboneMixin from .configuration_resnet import ResNetConfig __lowerCAmelCase : List[str] = logging.get_logger(__name__) # General docstring __lowerCAmelCase : Union[str, Any] = 'ResNetConfig' # Base docstring __lowerCAmelCase : List[Any] = 'microsoft/resnet-50' __lowerCAmelCase : int = [1, 2048, 7, 7] # Image classification docstring __lowerCAmelCase : Optional[int] = 'microsoft/resnet-50' __lowerCAmelCase : str = 'tiger cat' __lowerCAmelCase : Union[str, Any] = [ 'microsoft/resnet-50', # See all resnet models at https://huggingface.co/models?filter=resnet ] class UpperCAmelCase_ ( nn.Module ): '''simple docstring''' def __init__( self : Union[str, Any] , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int = 3 , UpperCamelCase__ : int = 1 , UpperCamelCase__ : str = "relu" ) -> List[str]: """simple docstring""" super().__init__() __magic_name__ = nn.Convad( UpperCamelCase__ , UpperCamelCase__ , kernel_size=UpperCamelCase__ , stride=UpperCamelCase__ , padding=kernel_size // 2 , bias=UpperCamelCase__ ) __magic_name__ = nn.BatchNormad(UpperCamelCase__ ) __magic_name__ = ACTaFN[activation] if activation is not None else nn.Identity() def _lowercase ( self : List[Any] , UpperCamelCase__ : Tensor ) -> Tensor: """simple docstring""" __magic_name__ = self.convolution(UpperCamelCase__ ) __magic_name__ = self.normalization(UpperCamelCase__ ) __magic_name__ = self.activation(UpperCamelCase__ ) return hidden_state class UpperCAmelCase_ ( nn.Module ): '''simple docstring''' def __init__( self : Tuple , UpperCamelCase__ : ResNetConfig ) -> str: """simple docstring""" super().__init__() __magic_name__ = ResNetConvLayer( config.num_channels , config.embedding_size , kernel_size=7 , stride=2 , activation=config.hidden_act ) __magic_name__ = nn.MaxPoolad(kernel_size=3 , stride=2 , padding=1 ) __magic_name__ = config.num_channels def _lowercase ( self : Optional[int] , UpperCamelCase__ : Tensor ) -> Tensor: """simple docstring""" __magic_name__ = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( """Make sure that the channel dimension of the pixel values match with the one set in the configuration.""" ) __magic_name__ = self.embedder(UpperCamelCase__ ) __magic_name__ = self.pooler(UpperCamelCase__ ) return embedding class UpperCAmelCase_ ( nn.Module ): '''simple docstring''' def __init__( self : List[str] , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int = 2 ) -> Any: """simple docstring""" super().__init__() __magic_name__ = nn.Convad(UpperCamelCase__ , UpperCamelCase__ , kernel_size=1 , stride=UpperCamelCase__ , bias=UpperCamelCase__ ) __magic_name__ = nn.BatchNormad(UpperCamelCase__ ) def _lowercase ( self : List[str] , UpperCamelCase__ : Tensor ) -> Tensor: """simple docstring""" __magic_name__ = self.convolution(UpperCamelCase__ ) __magic_name__ = self.normalization(UpperCamelCase__ ) return hidden_state class UpperCAmelCase_ ( nn.Module ): '''simple docstring''' def __init__( self : Union[str, Any] , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int = 1 , UpperCamelCase__ : str = "relu" ) -> Optional[int]: """simple docstring""" super().__init__() __magic_name__ = in_channels != out_channels or stride != 1 __magic_name__ = ( ResNetShortCut(UpperCamelCase__ , UpperCamelCase__ , stride=UpperCamelCase__ ) if should_apply_shortcut else nn.Identity() ) __magic_name__ = nn.Sequential( ResNetConvLayer(UpperCamelCase__ , UpperCamelCase__ , stride=UpperCamelCase__ ) , ResNetConvLayer(UpperCamelCase__ , UpperCamelCase__ , activation=UpperCamelCase__ ) , ) __magic_name__ = ACTaFN[activation] def _lowercase ( self : int , UpperCamelCase__ : List[str] ) -> Optional[Any]: """simple docstring""" __magic_name__ = hidden_state __magic_name__ = self.layer(UpperCamelCase__ ) __magic_name__ = self.shortcut(UpperCamelCase__ ) hidden_state += residual __magic_name__ = self.activation(UpperCamelCase__ ) return hidden_state class UpperCAmelCase_ ( nn.Module ): '''simple docstring''' def __init__( self : List[str] , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int = 1 , UpperCamelCase__ : str = "relu" , UpperCamelCase__ : int = 4 ) -> List[str]: """simple docstring""" super().__init__() __magic_name__ = in_channels != out_channels or stride != 1 __magic_name__ = out_channels // reduction __magic_name__ = ( ResNetShortCut(UpperCamelCase__ , UpperCamelCase__ , stride=UpperCamelCase__ ) if should_apply_shortcut else nn.Identity() ) __magic_name__ = nn.Sequential( ResNetConvLayer(UpperCamelCase__ , UpperCamelCase__ , kernel_size=1 ) , ResNetConvLayer(UpperCamelCase__ , UpperCamelCase__ , stride=UpperCamelCase__ ) , ResNetConvLayer(UpperCamelCase__ , UpperCamelCase__ , kernel_size=1 , activation=UpperCamelCase__ ) , ) __magic_name__ = ACTaFN[activation] def _lowercase ( self : Union[str, Any] , UpperCamelCase__ : Tuple ) -> Tuple: """simple docstring""" __magic_name__ = hidden_state __magic_name__ = self.layer(UpperCamelCase__ ) __magic_name__ = self.shortcut(UpperCamelCase__ ) hidden_state += residual __magic_name__ = self.activation(UpperCamelCase__ ) return hidden_state class UpperCAmelCase_ ( nn.Module ): '''simple docstring''' def __init__( self : Any , UpperCamelCase__ : ResNetConfig , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int = 2 , UpperCamelCase__ : int = 2 , ) -> Union[str, Any]: """simple docstring""" super().__init__() __magic_name__ = ResNetBottleNeckLayer if config.layer_type == """bottleneck""" else ResNetBasicLayer __magic_name__ = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer(UpperCamelCase__ , UpperCamelCase__ , stride=UpperCamelCase__ , activation=config.hidden_act ) , *[layer(UpperCamelCase__ , UpperCamelCase__ , activation=config.hidden_act ) for _ in range(depth - 1 )] , ) def _lowercase ( self : int , UpperCamelCase__ : Tensor ) -> Tensor: """simple docstring""" __magic_name__ = input for layer in self.layers: __magic_name__ = layer(UpperCamelCase__ ) return hidden_state class UpperCAmelCase_ ( nn.Module ): '''simple docstring''' def __init__( self : str , UpperCamelCase__ : ResNetConfig ) -> Any: """simple docstring""" super().__init__() __magic_name__ = nn.ModuleList([] ) # based on `downsample_in_first_stage` the first layer of the first stage may or may not downsample the input self.stages.append( ResNetStage( UpperCamelCase__ , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) __magic_name__ = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(UpperCamelCase__ , config.depths[1:] ): self.stages.append(ResNetStage(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , depth=UpperCamelCase__ ) ) def _lowercase ( self : List[Any] , UpperCamelCase__ : Tensor , UpperCamelCase__ : bool = False , UpperCamelCase__ : bool = True ) -> BaseModelOutputWithNoAttention: """simple docstring""" __magic_name__ = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: __magic_name__ = hidden_states + (hidden_state,) __magic_name__ = stage_module(UpperCamelCase__ ) if output_hidden_states: __magic_name__ = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention( last_hidden_state=UpperCamelCase__ , hidden_states=UpperCamelCase__ , ) class UpperCAmelCase_ ( _A ): '''simple docstring''' a__ = ResNetConfig a__ = """resnet""" a__ = """pixel_values""" a__ = True def _lowercase ( self : Tuple , UpperCamelCase__ : List[Any] ) -> List[str]: """simple docstring""" if isinstance(UpperCamelCase__ , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode="""fan_out""" , nonlinearity="""relu""" ) elif isinstance(UpperCamelCase__ , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def _lowercase ( self : Optional[Any] , UpperCamelCase__ : Tuple , UpperCamelCase__ : Union[str, Any]=False ) -> Tuple: """simple docstring""" if isinstance(UpperCamelCase__ , UpperCamelCase__ ): __magic_name__ = value __lowerCAmelCase : int = R'\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it\n as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`ResNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n' __lowerCAmelCase : Dict = R'\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConvNextImageProcessor.__call__`] for details.\n\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( """The bare ResNet model outputting raw features without any specific head on top.""" , _A , ) class UpperCAmelCase_ ( _A ): '''simple docstring''' def __init__( self : Optional[int] , UpperCamelCase__ : List[str] ) -> int: """simple docstring""" super().__init__(UpperCamelCase__ ) __magic_name__ = config __magic_name__ = ResNetEmbeddings(UpperCamelCase__ ) __magic_name__ = ResNetEncoder(UpperCamelCase__ ) __magic_name__ = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCamelCase__ ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=UpperCamelCase__ , config_class=_CONFIG_FOR_DOC , modality="""vision""" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def _lowercase ( self : int , UpperCamelCase__ : Tensor , UpperCamelCase__ : Optional[bool] = None , UpperCamelCase__ : Optional[bool] = None ) -> BaseModelOutputWithPoolingAndNoAttention: """simple docstring""" __magic_name__ = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __magic_name__ = return_dict if return_dict is not None else self.config.use_return_dict __magic_name__ = self.embedder(UpperCamelCase__ ) __magic_name__ = self.encoder( UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , return_dict=UpperCamelCase__ ) __magic_name__ = encoder_outputs[0] __magic_name__ = self.pooler(UpperCamelCase__ ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=UpperCamelCase__ , pooler_output=UpperCamelCase__ , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( """ ResNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. """ , _A , ) class UpperCAmelCase_ ( _A ): '''simple docstring''' def __init__( self : Tuple , UpperCamelCase__ : Optional[int] ) -> Optional[int]: """simple docstring""" super().__init__(UpperCamelCase__ ) __magic_name__ = config.num_labels __magic_name__ = ResNetModel(UpperCamelCase__ ) # classification head __magic_name__ = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCamelCase__ ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=UpperCamelCase__ , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def _lowercase ( self : List[Any] , UpperCamelCase__ : Optional[torch.FloatTensor] = None , UpperCamelCase__ : Optional[torch.LongTensor] = None , UpperCamelCase__ : Optional[bool] = None , UpperCamelCase__ : Optional[bool] = None , ) -> ImageClassifierOutputWithNoAttention: """simple docstring""" __magic_name__ = return_dict if return_dict is not None else self.config.use_return_dict __magic_name__ = self.resnet(UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , return_dict=UpperCamelCase__ ) __magic_name__ = outputs.pooler_output if return_dict else outputs[1] __magic_name__ = self.classifier(UpperCamelCase__ ) __magic_name__ = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: __magic_name__ = """regression""" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): __magic_name__ = """single_label_classification""" else: __magic_name__ = """multi_label_classification""" if self.config.problem_type == "regression": __magic_name__ = MSELoss() if self.num_labels == 1: __magic_name__ = loss_fct(logits.squeeze() , labels.squeeze() ) else: __magic_name__ = loss_fct(UpperCamelCase__ , UpperCamelCase__ ) elif self.config.problem_type == "single_label_classification": __magic_name__ = CrossEntropyLoss() __magic_name__ = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": __magic_name__ = BCEWithLogitsLoss() __magic_name__ = loss_fct(UpperCamelCase__ , UpperCamelCase__ ) if not return_dict: __magic_name__ = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=UpperCamelCase__ , logits=UpperCamelCase__ , hidden_states=outputs.hidden_states ) @add_start_docstrings( """ ResNet backbone, to be used with frameworks like DETR and MaskFormer. """ , _A , ) class UpperCAmelCase_ ( _A , _A ): '''simple docstring''' def __init__( self : List[str] , UpperCamelCase__ : Optional[Any] ) -> str: """simple docstring""" super().__init__(UpperCamelCase__ ) super()._init_backbone(UpperCamelCase__ ) __magic_name__ = [config.embedding_size] + config.hidden_sizes __magic_name__ = ResNetEmbeddings(UpperCamelCase__ ) __magic_name__ = ResNetEncoder(UpperCamelCase__ ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UpperCamelCase__ ) @replace_return_docstrings(output_type=UpperCamelCase__ , config_class=_CONFIG_FOR_DOC ) def _lowercase ( self : Optional[int] , UpperCamelCase__ : Tensor , UpperCamelCase__ : Optional[bool] = None , UpperCamelCase__ : Optional[bool] = None ) -> BackboneOutput: """simple docstring""" __magic_name__ = return_dict if return_dict is not None else self.config.use_return_dict __magic_name__ = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __magic_name__ = self.embedder(UpperCamelCase__ ) __magic_name__ = self.encoder(UpperCamelCase__ , output_hidden_states=UpperCamelCase__ , return_dict=UpperCamelCase__ ) __magic_name__ = outputs.hidden_states __magic_name__ = () for idx, stage in enumerate(self.stage_names ): if stage in self.out_features: feature_maps += (hidden_states[idx],) if not return_dict: __magic_name__ = (feature_maps,) if output_hidden_states: output += (outputs.hidden_states,) return output return BackboneOutput( feature_maps=UpperCamelCase__ , hidden_states=outputs.hidden_states if output_hidden_states else None , attentions=UpperCamelCase__ , )
529
from __future__ import annotations def a__ ( A_, A_, A_, A_ ): '''simple docstring''' if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): __magic_name__ , __magic_name__ = array[indexa], array[indexa] def a__ ( A_, A_, A_, A_ ): '''simple docstring''' if length > 1: __magic_name__ = int(length / 2 ) for i in range(A_, low + middle ): comp_and_swap(A_, A_, i + middle, A_ ) bitonic_merge(A_, A_, A_, A_ ) bitonic_merge(A_, low + middle, A_, A_ ) def a__ ( A_, A_, A_, A_ ): '''simple docstring''' if length > 1: __magic_name__ = int(length / 2 ) bitonic_sort(A_, A_, A_, 1 ) bitonic_sort(A_, low + middle, A_, 0 ) bitonic_merge(A_, A_, A_, A_ ) if __name__ == "__main__": __lowerCAmelCase : Optional[Any] = input('Enter numbers separated by a comma:\n').strip() __lowerCAmelCase : List[Any] = [int(item.strip()) for item in user_input.split(',')] bitonic_sort(unsorted, 0, len(unsorted), 1) print('\nSorted array in ascending order is: ', end='') print(*unsorted, sep=', ') bitonic_merge(unsorted, 0, len(unsorted), 0) print('Sorted array in descending order is: ', end='') print(*unsorted, sep=', ')
529
1
'''simple docstring''' import os import unittest from transformers import FunnelTokenizer, FunnelTokenizerFast from transformers.models.funnel.tokenization_funnel import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __a ( _snake_case, unittest.TestCase ): __UpperCamelCase : Union[str, Any] = FunnelTokenizer __UpperCamelCase : Dict = FunnelTokenizerFast __UpperCamelCase : List[str] = True __UpperCamelCase : Optional[Any] = True def UpperCAmelCase__ ( self : List[str] ): '''simple docstring''' super().setUp() __SCREAMING_SNAKE_CASE = [ """<unk>""", """<cls>""", """<sep>""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] __SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file ,"""w""" ,encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) def UpperCAmelCase__ ( self : List[Any] ,**lowerCamelCase : str ): '''simple docstring''' return FunnelTokenizer.from_pretrained(self.tmpdirname ,**lowerCamelCase ) def UpperCAmelCase__ ( self : List[str] ,**lowerCamelCase : Union[str, Any] ): '''simple docstring''' return FunnelTokenizerFast.from_pretrained(self.tmpdirname ,**lowerCamelCase ) def UpperCAmelCase__ ( self : Optional[Any] ,lowerCamelCase : int ): '''simple docstring''' __SCREAMING_SNAKE_CASE = """UNwant\u00E9d,running""" __SCREAMING_SNAKE_CASE = """unwanted, running""" return input_text, output_text def UpperCAmelCase__ ( self : Tuple ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.tokenizer_class(self.vocab_file ) __SCREAMING_SNAKE_CASE = tokenizer.tokenize("""UNwant\u00E9d,running""" ) self.assertListEqual(lowerCamelCase ,["""un""", """##want""", """##ed""", """,""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase ) ,[7, 4, 5, 10, 8, 9] ) def UpperCAmelCase__ ( self : Any ): '''simple docstring''' __SCREAMING_SNAKE_CASE = self.get_tokenizers(do_lower_case=lowerCamelCase ) for tokenizer in tokenizers: __SCREAMING_SNAKE_CASE = tokenizer("""UNwant\u00E9d,running""" ) __SCREAMING_SNAKE_CASE = len(inputs["""input_ids"""] ) - 1 self.assertListEqual(inputs["""token_type_ids"""] ,[2] + [0] * sentence_len ) __SCREAMING_SNAKE_CASE = tokenizer("""UNwant\u00E9d,running""" ,"""UNwant\u00E9d,running""" ) self.assertListEqual(inputs["""token_type_ids"""] ,[2] + [0] * sentence_len + [1] * sentence_len )
13
'''simple docstring''' def __magic_name__ ( __UpperCAmelCase = 1 , __UpperCAmelCase = 1000 ) -> int: '''simple docstring''' __SCREAMING_SNAKE_CASE = 1 __SCREAMING_SNAKE_CASE = 0 for divide_by_number in range(__UpperCAmelCase , digit + 1 ): __SCREAMING_SNAKE_CASE = [] __SCREAMING_SNAKE_CASE = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(__UpperCAmelCase ): __SCREAMING_SNAKE_CASE = len(__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = divide_by_number else: has_been_divided.append(__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
13
1
'''simple docstring''' # Imports import numpy as np class UpperCAmelCase_ : """simple docstring""" def __init__( self , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None ) -> str: self.set_matricies(red=SCREAMING_SNAKE_CASE_ , green=SCREAMING_SNAKE_CASE_ , blue=SCREAMING_SNAKE_CASE_ , red_edge=SCREAMING_SNAKE_CASE_ , nir=SCREAMING_SNAKE_CASE_ ) def lowercase_ ( self , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None ) -> Any: if red is not None: __lowerCamelCase : Dict = red if green is not None: __lowerCamelCase : List[Any] = green if blue is not None: __lowerCamelCase : int = blue if red_edge is not None: __lowerCamelCase : Optional[int] = red_edge if nir is not None: __lowerCamelCase : Any = nir return True def lowercase_ ( self , SCREAMING_SNAKE_CASE_="" , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None ) -> Optional[Any]: self.set_matricies(red=SCREAMING_SNAKE_CASE_ , green=SCREAMING_SNAKE_CASE_ , blue=SCREAMING_SNAKE_CASE_ , red_edge=SCREAMING_SNAKE_CASE_ , nir=SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : List[Any] = { 'ARVI2': self.arvaa, 'CCCI': self.ccci, 'CVI': self.cvi, 'GLI': self.gli, 'NDVI': self.ndvi, 'BNDVI': self.bndvi, 'redEdgeNDVI': self.red_edge_ndvi, 'GNDVI': self.gndvi, 'GBNDVI': self.gbndvi, 'GRNDVI': self.grndvi, 'RBNDVI': self.rbndvi, 'PNDVI': self.pndvi, 'ATSAVI': self.atsavi, 'BWDRVI': self.bwdrvi, 'CIgreen': self.ci_green, 'CIrededge': self.ci_rededge, 'CI': self.ci, 'CTVI': self.ctvi, 'GDVI': self.gdvi, 'EVI': self.evi, 'GEMI': self.gemi, 'GOSAVI': self.gosavi, 'GSAVI': self.gsavi, 'Hue': self.hue, 'IVI': self.ivi, 'IPVI': self.ipvi, 'I': self.i, 'RVI': self.rvi, 'MRVI': self.mrvi, 'MSAVI': self.m_savi, 'NormG': self.norm_g, 'NormNIR': self.norm_nir, 'NormR': self.norm_r, 'NGRDI': self.ngrdi, 'RI': self.ri, 'S': self.s, 'IF': self._if, 'DVI': self.dvi, 'TVI': self.tvi, 'NDRE': self.ndre, } try: return funcs[index]() except KeyError: print('Index not in the list!' ) return False def lowercase_ ( self ) -> Optional[Any]: return -0.1_8 + (1.1_7 * ((self.nir - self.red) / (self.nir + self.red))) def lowercase_ ( self ) -> Dict: return ((self.nir - self.redEdge) / (self.nir + self.redEdge)) / ( (self.nir - self.red) / (self.nir + self.red) ) def lowercase_ ( self ) -> Any: return self.nir * (self.red / (self.green**2)) def lowercase_ ( self ) -> Optional[Any]: return (2 * self.green - self.red - self.blue) / ( 2 * self.green + self.red + self.blue ) def lowercase_ ( self ) -> Union[str, Any]: return (self.nir - self.red) / (self.nir + self.red) def lowercase_ ( self ) -> Optional[int]: return (self.nir - self.blue) / (self.nir + self.blue) def lowercase_ ( self ) -> Optional[int]: return (self.redEdge - self.red) / (self.redEdge + self.red) def lowercase_ ( self ) -> Any: return (self.nir - self.green) / (self.nir + self.green) def lowercase_ ( self ) -> int: return (self.nir - (self.green + self.blue)) / ( self.nir + (self.green + self.blue) ) def lowercase_ ( self ) -> List[Any]: return (self.nir - (self.green + self.red)) / ( self.nir + (self.green + self.red) ) def lowercase_ ( self ) -> List[Any]: return (self.nir - (self.blue + self.red)) / (self.nir + (self.blue + self.red)) def lowercase_ ( self ) -> str: return (self.nir - (self.green + self.red + self.blue)) / ( self.nir + (self.green + self.red + self.blue) ) def lowercase_ ( self , SCREAMING_SNAKE_CASE_=0.0_8 , SCREAMING_SNAKE_CASE_=1.2_2 , SCREAMING_SNAKE_CASE_=0.0_3 ) -> str: return a * ( (self.nir - a * self.red - b) / (a * self.nir + self.red - a * b + x * (1 + a**2)) ) def lowercase_ ( self ) -> Any: return (0.1 * self.nir - self.blue) / (0.1 * self.nir + self.blue) def lowercase_ ( self ) -> Optional[Any]: return (self.nir / self.green) - 1 def lowercase_ ( self ) -> Tuple: return (self.nir / self.redEdge) - 1 def lowercase_ ( self ) -> List[Any]: return (self.red - self.blue) / self.red def lowercase_ ( self ) -> Optional[Any]: __lowerCamelCase : Any = self.ndvi() return ((ndvi + 0.5) / (abs(ndvi + 0.5 ))) * (abs(ndvi + 0.5 ) ** (1 / 2)) def lowercase_ ( self ) -> List[str]: return self.nir - self.green def lowercase_ ( self ) -> Tuple: return 2.5 * ( (self.nir - self.red) / (self.nir + 6 * self.red - 7.5 * self.blue + 1) ) def lowercase_ ( self ) -> Union[str, Any]: __lowerCamelCase : str = (2 * (self.nir**2 - self.red**2) + 1.5 * self.nir + 0.5 * self.red) / ( self.nir + self.red + 0.5 ) return n * (1 - 0.2_5 * n) - (self.red - 0.1_2_5) / (1 - self.red) def lowercase_ ( self , SCREAMING_SNAKE_CASE_=0.1_6 ) -> Tuple: return (self.nir - self.green) / (self.nir + self.green + y) def lowercase_ ( self , SCREAMING_SNAKE_CASE_=0.5 ) -> Optional[int]: return ((self.nir - self.green) / (self.nir + self.green + n)) * (1 + n) def lowercase_ ( self ) -> str: return np.arctan( ((2 * self.red - self.green - self.blue) / 3_0.5) * (self.green - self.blue) ) def lowercase_ ( self , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None ) -> int: return (self.nir - b) / (a * self.red) def lowercase_ ( self ) -> Union[str, Any]: return (self.nir / ((self.nir + self.red) / 2)) * (self.ndvi() + 1) def lowercase_ ( self ) -> int: return (self.red + self.green + self.blue) / 3_0.5 def lowercase_ ( self ) -> Optional[int]: return self.nir / self.red def lowercase_ ( self ) -> Dict: return (self.rvi() - 1) / (self.rvi() + 1) def lowercase_ ( self ) -> Dict: return ( (2 * self.nir + 1) - ((2 * self.nir + 1) ** 2 - 8 * (self.nir - self.red)) ** (1 / 2) ) / 2 def lowercase_ ( self ) -> Optional[int]: return self.green / (self.nir + self.red + self.green) def lowercase_ ( self ) -> List[str]: return self.nir / (self.nir + self.red + self.green) def lowercase_ ( self ) -> List[str]: return self.red / (self.nir + self.red + self.green) def lowercase_ ( self ) -> List[Any]: return (self.green - self.red) / (self.green + self.red) def lowercase_ ( self ) -> List[str]: return (self.red - self.green) / (self.red + self.green) def lowercase_ ( self ) -> Any: __lowerCamelCase : Union[str, Any] = np.max([np.max(self.red ), np.max(self.green ), np.max(self.blue )] ) __lowerCamelCase : Optional[Any] = np.min([np.min(self.red ), np.min(self.green ), np.min(self.blue )] ) return (max_value - min_value) / max_value def lowercase_ ( self ) -> Optional[Any]: return (2 * self.red - self.green - self.blue) / (self.green - self.blue) def lowercase_ ( self ) -> Any: return self.nir / self.red def lowercase_ ( self ) -> str: return (self.ndvi() + 0.5) ** (1 / 2) def lowercase_ ( self ) -> Optional[int]: return (self.nir - self.redEdge) / (self.nir + self.redEdge)
13
__UpperCamelCase : str = {str(digit): digit**5 for digit in range(10)} def a_ ( _A ) -> int: """simple docstring""" return sum(DIGITS_FIFTH_POWER[digit] for digit in str(_A ) ) def a_ ( ) -> int: """simple docstring""" return sum( number for number in range(1000 , 1000000 ) if number == digits_fifth_powers_sum(_A ) ) if __name__ == "__main__": print(solution())
328
0
import torch from diffusers import DiffusionPipeline class _SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ ): def __init__( self : List[Any] , UpperCamelCase : Dict , UpperCamelCase : Union[str, Any] )->Optional[int]: super().__init__() self.register_modules(unet=snake_case__ , scheduler=snake_case__ ) def __call__( self : Any )->Dict: __SCREAMING_SNAKE_CASE : Optional[Any] = torch.randn( (1, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) , ) __SCREAMING_SNAKE_CASE : List[str] = 1 __SCREAMING_SNAKE_CASE : int = self.unet(snake_case__ , snake_case__ ).sample __SCREAMING_SNAKE_CASE : Optional[Any] = self.scheduler.step(snake_case__ , snake_case__ , snake_case__ ).prev_sample __SCREAMING_SNAKE_CASE : List[Any] = scheduler_output - scheduler_output + torch.ones_like(snake_case__ ) return result
719
import io import math from typing import Dict, Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import convert_to_rgb, normalize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, get_image_size, infer_channel_dimension_format, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_vision_available, logging from ...utils.import_utils import requires_backends if is_vision_available(): import textwrap from PIL import Image, ImageDraw, ImageFont if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: _lowerCamelCase = False _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = """ybelkada/fonts""" def _lowerCAmelCase ( ): """simple docstring""" if is_torch_available() and not is_torch_greater_or_equal_than_1_11: raise ImportError( F"""You are using torch=={torch.__version__}, but torch>=1.11.0 is required to use """ "Pix2StructImageProcessor. Please upgrade torch." ) def _lowerCAmelCase ( __lowerCamelCase : List[str] , __lowerCamelCase : int , __lowerCamelCase : Dict ): """simple docstring""" requires_backends(__lowerCamelCase , ["torch"] ) _check_torch_version() __SCREAMING_SNAKE_CASE : List[Any] = image_tensor.unsqueeze(0 ) __SCREAMING_SNAKE_CASE : Optional[Any] = torch.nn.functional.unfold(__lowerCamelCase , (patch_height, patch_width) , stride=(patch_height, patch_width) ) __SCREAMING_SNAKE_CASE : List[Any] = patches.reshape(image_tensor.size(0 ) , image_tensor.size(1 ) , __lowerCamelCase , __lowerCamelCase , -1 ) __SCREAMING_SNAKE_CASE : Any = patches.permute(0 , 4 , 2 , 3 , 1 ).reshape( image_tensor.size(2 ) // patch_height , image_tensor.size(3 ) // patch_width , image_tensor.size(1 ) * patch_height * patch_width , ) return patches.unsqueeze(0 ) def _lowerCAmelCase ( __lowerCamelCase : str , __lowerCamelCase : int = 36 , __lowerCamelCase : str = "black" , __lowerCamelCase : str = "white" , __lowerCamelCase : int = 5 , __lowerCamelCase : int = 5 , __lowerCamelCase : int = 5 , __lowerCamelCase : int = 5 , __lowerCamelCase : Optional[bytes] = None , __lowerCamelCase : Optional[str] = None , ): """simple docstring""" requires_backends(__lowerCamelCase , "vision" ) # Add new lines so that each line is no more than 80 characters. __SCREAMING_SNAKE_CASE : Union[str, Any] = textwrap.TextWrapper(width=80 ) __SCREAMING_SNAKE_CASE : Optional[Any] = wrapper.wrap(text=__lowerCamelCase ) __SCREAMING_SNAKE_CASE : str = "\n".join(__lowerCamelCase ) if font_bytes is not None and font_path is None: __SCREAMING_SNAKE_CASE : List[Any] = io.BytesIO(__lowerCamelCase ) elif font_path is not None: __SCREAMING_SNAKE_CASE : Union[str, Any] = font_path else: __SCREAMING_SNAKE_CASE : Tuple = hf_hub_download(__lowerCamelCase , "Arial.TTF" ) __SCREAMING_SNAKE_CASE : List[Any] = ImageFont.truetype(__lowerCamelCase , encoding="UTF-8" , size=__lowerCamelCase ) # Use a temporary canvas to determine the width and height in pixels when # rendering the text. __SCREAMING_SNAKE_CASE : str = ImageDraw.Draw(Image.new("RGB" , (1, 1) , __lowerCamelCase ) ) __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : int = temp_draw.textbbox((0, 0) , __lowerCamelCase , __lowerCamelCase ) # Create the actual image with a bit of padding around the text. __SCREAMING_SNAKE_CASE : Union[str, Any] = text_width + left_padding + right_padding __SCREAMING_SNAKE_CASE : Tuple = text_height + top_padding + bottom_padding __SCREAMING_SNAKE_CASE : Tuple = Image.new("RGB" , (image_width, image_height) , __lowerCamelCase ) __SCREAMING_SNAKE_CASE : List[Any] = ImageDraw.Draw(__lowerCamelCase ) draw.text(xy=(left_padding, top_padding) , text=__lowerCamelCase , fill=__lowerCamelCase , font=__lowerCamelCase ) return image def _lowerCAmelCase ( __lowerCamelCase : np.ndarray , __lowerCamelCase : str , **__lowerCamelCase : Union[str, Any] ): """simple docstring""" requires_backends(__lowerCamelCase , "vision" ) # Convert to PIL image if necessary __SCREAMING_SNAKE_CASE : int = to_pil_image(__lowerCamelCase ) __SCREAMING_SNAKE_CASE : int = render_text(__lowerCamelCase , **__lowerCamelCase ) __SCREAMING_SNAKE_CASE : Tuple = max(header_image.width , image.width ) __SCREAMING_SNAKE_CASE : Tuple = int(image.height * (new_width / image.width) ) __SCREAMING_SNAKE_CASE : Any = int(header_image.height * (new_width / header_image.width) ) __SCREAMING_SNAKE_CASE : Tuple = Image.new("RGB" , (new_width, new_height + new_header_height) , "white" ) new_image.paste(header_image.resize((new_width, new_header_height) ) , (0, 0) ) new_image.paste(image.resize((new_width, new_height) ) , (0, new_header_height) ) # Convert back to the original framework if necessary __SCREAMING_SNAKE_CASE : Dict = to_numpy_array(__lowerCamelCase ) if infer_channel_dimension_format(__lowerCamelCase ) == ChannelDimension.LAST: __SCREAMING_SNAKE_CASE : Any = to_channel_dimension_format(__lowerCamelCase , ChannelDimension.LAST ) return new_image class _SCREAMING_SNAKE_CASE (UpperCamelCase ): lowerCAmelCase = ["""flattened_patches"""] def __init__( self : Tuple , UpperCamelCase : bool = True , UpperCamelCase : bool = True , UpperCamelCase : Dict[str, int] = None , UpperCamelCase : int = 2_0_4_8 , UpperCamelCase : bool = False , **UpperCamelCase : Optional[int] , )->None: super().__init__(**UpperCamelCase ) __SCREAMING_SNAKE_CASE : Tuple = patch_size if patch_size is not None else {"height": 1_6, "width": 1_6} __SCREAMING_SNAKE_CASE : List[Any] = do_normalize __SCREAMING_SNAKE_CASE : List[Any] = do_convert_rgb __SCREAMING_SNAKE_CASE : List[str] = max_patches __SCREAMING_SNAKE_CASE : List[Any] = is_vqa def __snake_case ( self : Tuple , UpperCamelCase : np.ndarray , UpperCamelCase : int , UpperCamelCase : dict , **UpperCamelCase : Tuple )->np.ndarray: requires_backends(self.extract_flattened_patches , "torch" ) _check_torch_version() # convert to torch __SCREAMING_SNAKE_CASE : Optional[Any] = to_channel_dimension_format(UpperCamelCase , ChannelDimension.FIRST ) __SCREAMING_SNAKE_CASE : int = torch.from_numpy(UpperCamelCase ) __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : List[str] = patch_size["height"], patch_size["width"] __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE : Optional[Any] = get_image_size(UpperCamelCase ) # maximize scale s.t. __SCREAMING_SNAKE_CASE : List[str] = math.sqrt(max_patches * (patch_height / image_height) * (patch_width / image_width) ) __SCREAMING_SNAKE_CASE : List[str] = max(min(math.floor(scale * image_height / patch_height ) , UpperCamelCase ) , 1 ) __SCREAMING_SNAKE_CASE : Tuple = max(min(math.floor(scale * image_width / patch_width ) , UpperCamelCase ) , 1 ) __SCREAMING_SNAKE_CASE : List[str] = max(num_feasible_rows * patch_height , 1 ) __SCREAMING_SNAKE_CASE : int = max(num_feasible_cols * patch_width , 1 ) __SCREAMING_SNAKE_CASE : Any = torch.nn.functional.interpolate( image.unsqueeze(0 ) , size=(resized_height, resized_width) , mode="bilinear" , align_corners=UpperCamelCase , antialias=UpperCamelCase , ).squeeze(0 ) # [1, rows, columns, patch_height * patch_width * image_channels] __SCREAMING_SNAKE_CASE : List[str] = torch_extract_patches(UpperCamelCase , UpperCamelCase , UpperCamelCase ) __SCREAMING_SNAKE_CASE : Optional[Any] = patches.shape __SCREAMING_SNAKE_CASE : int = patches_shape[1] __SCREAMING_SNAKE_CASE : List[str] = patches_shape[2] __SCREAMING_SNAKE_CASE : List[str] = patches_shape[3] # [rows * columns, patch_height * patch_width * image_channels] __SCREAMING_SNAKE_CASE : Union[str, Any] = patches.reshape([rows * columns, depth] ) # [rows * columns, 1] __SCREAMING_SNAKE_CASE : Any = torch.arange(UpperCamelCase ).reshape([rows, 1] ).repeat(1 , UpperCamelCase ).reshape([rows * columns, 1] ) __SCREAMING_SNAKE_CASE : Optional[int] = torch.arange(UpperCamelCase ).reshape([1, columns] ).repeat(UpperCamelCase , 1 ).reshape([rows * columns, 1] ) # Offset by 1 so the ids do not contain zeros, which represent padding. row_ids += 1 col_ids += 1 # Prepare additional patch features. # [rows * columns, 1] __SCREAMING_SNAKE_CASE : str = row_ids.to(torch.floataa ) __SCREAMING_SNAKE_CASE : int = col_ids.to(torch.floataa ) # [rows * columns, 2 + patch_height * patch_width * image_channels] __SCREAMING_SNAKE_CASE : Optional[int] = torch.cat([row_ids, col_ids, patches] , -1 ) # [max_patches, 2 + patch_height * patch_width * image_channels] __SCREAMING_SNAKE_CASE : Any = torch.nn.functional.pad(UpperCamelCase , [0, 0, 0, max_patches - (rows * columns)] ).float() __SCREAMING_SNAKE_CASE : str = to_numpy_array(UpperCamelCase ) return result def __snake_case ( self : Optional[Any] , UpperCamelCase : np.ndarray , UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **UpperCamelCase : Optional[int] )->np.ndarray: if image.dtype == np.uinta: __SCREAMING_SNAKE_CASE : Optional[int] = image.astype(np.floataa ) # take mean across the whole `image` __SCREAMING_SNAKE_CASE : int = np.mean(UpperCamelCase ) __SCREAMING_SNAKE_CASE : List[str] = np.std(UpperCamelCase ) __SCREAMING_SNAKE_CASE : Optional[int] = max(UpperCamelCase , 1.0 / math.sqrt(np.prod(image.shape ) ) ) return normalize(UpperCamelCase , mean=UpperCamelCase , std=UpperCamelCase , **UpperCamelCase ) def __snake_case ( self : Union[str, Any] , UpperCamelCase : ImageInput , UpperCamelCase : Optional[str] = None , UpperCamelCase : bool = None , UpperCamelCase : Optional[bool] = None , UpperCamelCase : Optional[int] = None , UpperCamelCase : Optional[Dict[str, int]] = None , UpperCamelCase : Optional[Union[str, TensorType]] = None , UpperCamelCase : ChannelDimension = ChannelDimension.FIRST , **UpperCamelCase : Dict , )->ImageInput: __SCREAMING_SNAKE_CASE : int = do_normalize if do_normalize is not None else self.do_normalize __SCREAMING_SNAKE_CASE : Optional[Any] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb __SCREAMING_SNAKE_CASE : List[Any] = patch_size if patch_size is not None else self.patch_size __SCREAMING_SNAKE_CASE : List[str] = max_patches if max_patches is not None else self.max_patches __SCREAMING_SNAKE_CASE : List[Any] = self.is_vqa if kwargs.get("data_format" , UpperCamelCase ) is not None: raise ValueError("data_format is not an accepted input as the outputs are " ) __SCREAMING_SNAKE_CASE : Union[str, Any] = make_list_of_images(UpperCamelCase ) if not valid_images(UpperCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) # PIL RGBA images are converted to RGB if do_convert_rgb: __SCREAMING_SNAKE_CASE : Optional[int] = [convert_to_rgb(UpperCamelCase ) for image in images] # All transformations expect numpy arrays. __SCREAMING_SNAKE_CASE : str = [to_numpy_array(UpperCamelCase ) for image in images] if is_vqa: if header_text is None: raise ValueError("A header text must be provided for VQA models." ) __SCREAMING_SNAKE_CASE : Any = kwargs.pop("font_bytes" , UpperCamelCase ) __SCREAMING_SNAKE_CASE : Any = kwargs.pop("font_path" , UpperCamelCase ) if isinstance(UpperCamelCase , UpperCamelCase ): __SCREAMING_SNAKE_CASE : Dict = [header_text] * len(UpperCamelCase ) __SCREAMING_SNAKE_CASE : Optional[int] = [ render_header(UpperCamelCase , header_text[i] , font_bytes=UpperCamelCase , font_path=UpperCamelCase ) for i, image in enumerate(UpperCamelCase ) ] if do_normalize: __SCREAMING_SNAKE_CASE : List[str] = [self.normalize(image=UpperCamelCase ) for image in images] # convert to torch tensor and permute __SCREAMING_SNAKE_CASE : str = [ self.extract_flattened_patches(image=UpperCamelCase , max_patches=UpperCamelCase , patch_size=UpperCamelCase ) for image in images ] # create attention mask in numpy __SCREAMING_SNAKE_CASE : List[Any] = [(image.sum(axis=-1 ) != 0).astype(np.floataa ) for image in images] __SCREAMING_SNAKE_CASE : Dict = BatchFeature( data={"flattened_patches": images, "attention_mask": attention_masks} , tensor_type=UpperCamelCase ) return encoded_outputs
447
0
'''simple docstring''' import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse('''9.1.0'''): A_ : Optional[int] ={ '''linear''': PIL.Image.Resampling.BILINEAR, '''bilinear''': PIL.Image.Resampling.BILINEAR, '''bicubic''': PIL.Image.Resampling.BICUBIC, '''lanczos''': PIL.Image.Resampling.LANCZOS, '''nearest''': PIL.Image.Resampling.NEAREST, } else: A_ : Union[str, Any] ={ '''linear''': PIL.Image.LINEAR, '''bilinear''': PIL.Image.BILINEAR, '''bicubic''': PIL.Image.BICUBIC, '''lanczos''': PIL.Image.LANCZOS, '''nearest''': PIL.Image.NEAREST, } def snake_case_ ( __snake_case : Optional[Any]) -> List[str]: lowerCAmelCase_ = (images / 2 + 0.5).clamp(0 , 1) lowerCAmelCase_ = images.cpu().permute(0 , 2 , 3 , 1).float().numpy() lowerCAmelCase_ = numpy_to_pil(SCREAMING_SNAKE_CASE_) return images def snake_case_ ( __snake_case : str) -> Tuple: if images.ndim == 3: lowerCAmelCase_ = images[None, ...] lowerCAmelCase_ = (images * 255).round().astype('''uint8''') if images.shape[-1] == 1: # special case for grayscale (single channel) images lowerCAmelCase_ = [Image.fromarray(image.squeeze() , mode='''L''') for image in images] else: lowerCAmelCase_ = [Image.fromarray(SCREAMING_SNAKE_CASE_) for image in images] return pil_images
274
from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder snake_case_ = datasets.utils.logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ (folder_based_builder.FolderBasedBuilderConfig ): __lowerCamelCase : bool = None __lowerCamelCase : bool = None class SCREAMING_SNAKE_CASE__ (folder_based_builder.FolderBasedBuilder ): __lowerCamelCase : Any = datasets.Audio() __lowerCamelCase : List[Any] = """audio""" __lowerCamelCase : List[Any] = AudioFolderConfig __lowerCamelCase : List[str] # definition at the bottom of the script __lowerCamelCase : List[str] = AudioClassification(audio_column="""audio""" , label_column="""label""" ) snake_case_ = [ '''.aiff''', '''.au''', '''.avr''', '''.caf''', '''.flac''', '''.htk''', '''.svx''', '''.mat4''', '''.mat5''', '''.mpc2k''', '''.ogg''', '''.paf''', '''.pvf''', '''.raw''', '''.rf64''', '''.sd2''', '''.sds''', '''.ircam''', '''.voc''', '''.w64''', '''.wav''', '''.nist''', '''.wavex''', '''.wve''', '''.xi''', '''.mp3''', '''.opus''', ] snake_case_ = AUDIO_EXTENSIONS
164
0
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging A_ : List[str] = logging.get_logger(__name__) A_ : Union[str, Any] = { """google/pix2struct-textcaps-base""": ( """https://huggingface.co/google/pix2struct-textcaps-base/resolve/main/config.json""" ), } class _lowercase ( SCREAMING_SNAKE_CASE__ ): _UpperCAmelCase = '''pix2struct_text_model''' _UpperCAmelCase = ['''past_key_values'''] _UpperCAmelCase = { '''hidden_size''': '''hidden_size''', '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self : Union[str, Any] , __lowerCAmelCase : Tuple=5_0244 , __lowerCAmelCase : str=768 , __lowerCAmelCase : List[str]=64 , __lowerCAmelCase : Optional[int]=2048 , __lowerCAmelCase : Dict=12 , __lowerCAmelCase : Dict=12 , __lowerCAmelCase : Union[str, Any]=32 , __lowerCAmelCase : Any=128 , __lowerCAmelCase : Union[str, Any]=0.1 , __lowerCAmelCase : Dict=1E-6 , __lowerCAmelCase : Optional[Any]=1.0 , __lowerCAmelCase : Union[str, Any]="gelu_new" , __lowerCAmelCase : Optional[Any]=0 , __lowerCAmelCase : Optional[int]=False , __lowerCAmelCase : Any=0 , __lowerCAmelCase : Dict=1 , __lowerCAmelCase : Optional[int]=False , __lowerCAmelCase : str=True , **__lowerCAmelCase : Union[str, Any] , ) -> Tuple: """simple docstring""" a = vocab_size a = hidden_size a = d_kv a = d_ff a = num_layers a = num_heads a = relative_attention_num_buckets a = relative_attention_max_distance a = dropout_rate a = layer_norm_epsilon a = initializer_factor a = use_cache a = eos_token_id a = decoder_start_token_id # for backwards compatibility a = dense_act_fn super().__init__( pad_token_id=snake_case__ , eos_token_id=snake_case__ , decoder_start_token_id=snake_case__ , tie_word_embeddings=snake_case__ , is_decoder=snake_case__ , **snake_case__ , ) @classmethod def A ( cls : Union[str, Any] , __lowerCAmelCase : Optional[int] , **__lowerCAmelCase : int ) -> List[Any]: """simple docstring""" cls._set_token_in_kwargs(snake_case__ ) a = cls.get_config_dict(snake_case__ , **snake_case__ ) # get the text config dict if we are loading from Pix2StructConfig if config_dict.get("model_type" ) == "pix2struct": a = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(snake_case__ , **snake_case__ ) class _lowercase ( SCREAMING_SNAKE_CASE__ ): _UpperCAmelCase = '''pix2struct_vision_model''' def __init__( self : str , __lowerCAmelCase : List[str]=768 , __lowerCAmelCase : Optional[int]=768 , __lowerCAmelCase : Dict=2048 , __lowerCAmelCase : str=64 , __lowerCAmelCase : List[str]=12 , __lowerCAmelCase : Optional[int]=12 , __lowerCAmelCase : int="gelu_new" , __lowerCAmelCase : Optional[int]=1E-6 , __lowerCAmelCase : Tuple=0.0 , __lowerCAmelCase : List[Any]=0.0 , __lowerCAmelCase : Tuple=1E-10 , __lowerCAmelCase : Any=1.0 , __lowerCAmelCase : Tuple=4096 , __lowerCAmelCase : Optional[int]=32 , __lowerCAmelCase : Union[str, Any]=128 , **__lowerCAmelCase : Optional[int] , ) -> List[Any]: """simple docstring""" super().__init__(**snake_case__ ) a = hidden_size a = patch_embed_hidden_size a = d_ff a = dropout_rate a = num_hidden_layers a = num_attention_heads a = initializer_range a = initializer_factor a = attention_dropout a = layer_norm_eps a = dense_act_fn a = seq_len a = relative_attention_num_buckets a = relative_attention_max_distance a = d_kv @classmethod def A ( cls : int , __lowerCAmelCase : Dict , **__lowerCAmelCase : Union[str, Any] ) -> Optional[int]: """simple docstring""" cls._set_token_in_kwargs(snake_case__ ) a = cls.get_config_dict(snake_case__ , **snake_case__ ) # get the vision config dict if we are loading from Pix2StructConfig if config_dict.get("model_type" ) == "pix2struct": a = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(snake_case__ , **snake_case__ ) class _lowercase ( SCREAMING_SNAKE_CASE__ ): _UpperCAmelCase = '''pix2struct''' _UpperCAmelCase = True def __init__( self : Dict , __lowerCAmelCase : str=None , __lowerCAmelCase : str=None , __lowerCAmelCase : str=1.0 , __lowerCAmelCase : Dict=0.0_2 , __lowerCAmelCase : int=False , __lowerCAmelCase : List[str]=False , __lowerCAmelCase : str=True , **__lowerCAmelCase : Tuple , ) -> int: """simple docstring""" super().__init__(tie_word_embeddings=snake_case__ , is_encoder_decoder=snake_case__ , **snake_case__ ) if text_config is None: a = {} logger.info("text_config is None. Initializing the Pix2StructTextConfig with default values." ) if vision_config is None: a = {} logger.info("vision_config is None. Initializing the Pix2StructVisionConfig with default values." ) a = PixaStructTextConfig(**snake_case__ ) a = PixaStructVisionConfig(**snake_case__ ) a = self.text_config.decoder_start_token_id a = self.text_config.pad_token_id a = self.text_config.eos_token_id a = initializer_factor a = initializer_range a = self.initializer_range a = self.initializer_range a = is_vqa @classmethod def A ( cls : Tuple , __lowerCAmelCase : List[str] , __lowerCAmelCase : List[Any] , **__lowerCAmelCase : List[Any] ) -> Union[str, Any]: """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **snake_case__ ) def A ( self : Any ) -> str: """simple docstring""" a = copy.deepcopy(self.__dict__ ) a = self.text_config.to_dict() a = self.vision_config.to_dict() a = self.__class__.model_type return output
718
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging A_ : List[str] = logging.get_logger(__name__) class _lowercase ( UpperCAmelCase__ ): _UpperCAmelCase = ['''audio_values''', '''audio_mask'''] def __init__( self : List[Any] , __lowerCAmelCase : Dict=2048 , __lowerCAmelCase : List[Any]=1 , __lowerCAmelCase : Dict=[16, 16] , __lowerCAmelCase : str=128 , __lowerCAmelCase : Optional[int]=4_4100 , __lowerCAmelCase : int=86 , __lowerCAmelCase : Optional[Any]=2048 , __lowerCAmelCase : str=0.0 , **__lowerCAmelCase : Optional[int] , ) -> Union[str, Any]: """simple docstring""" super().__init__( feature_size=__lowerCAmelCase , sampling_rate=__lowerCAmelCase , padding_value=__lowerCAmelCase , **__lowerCAmelCase , ) a = spectrogram_length a = num_channels a = patch_size a = feature_size // self.patch_size[1] a = n_fft a = sampling_rate // hop_length_to_sampling_rate a = sampling_rate a = padding_value a = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=__lowerCAmelCase , min_frequency=0.0 , max_frequency=2_2_0_5_0.0 , sampling_rate=__lowerCAmelCase , norm="slaney" , mel_scale="slaney" , ).T def A ( self : List[str] , __lowerCAmelCase : np.array ) -> np.ndarray: """simple docstring""" a = spectrogram( __lowerCAmelCase , window_function(self.n_fft , "hann" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel="dB" , db_range=8_0.0 , ) a = log_spec[:, :-1] a = log_spec - 2_0.0 a = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : Union[str, Any] , __lowerCAmelCase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __lowerCAmelCase : Optional[Union[str, TensorType]] = None , __lowerCAmelCase : Optional[bool] = True , __lowerCAmelCase : Optional[int] = None , __lowerCAmelCase : bool = False , __lowerCAmelCase : bool = False , **__lowerCAmelCase : Optional[int] , ) -> BatchFeature: """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( "This feature extractor is set to support sampling rate" f""" of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled""" f""" with {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) a = isinstance(__lowerCAmelCase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"""Only mono-channel audio is supported for input to {self}""" ) a = is_batched_numpy or ( isinstance(__lowerCAmelCase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: a = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(__lowerCAmelCase , np.ndarray ): a = np.asarray(__lowerCAmelCase , dtype=np.floataa ) elif isinstance(__lowerCAmelCase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): a = raw_speech.astype(np.floataa ) # always return batch if not is_batched: a = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis a = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , __lowerCAmelCase ): a = [np.asarray(__lowerCAmelCase , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask a = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: a = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] a = np.array(__lowerCAmelCase ).astype(np.floataa ) # convert into correct format for padding a = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch a = np.ones([len(__lowerCAmelCase ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) a = padded_audio_features * self.padding_value for i in range(len(__lowerCAmelCase ) ): a = audio_features[i] a = feature # return as BatchFeature if return_attention_mask: a = {"audio_values": padded_audio_features, "audio_mask": audio_mask} else: a = {"audio_values": padded_audio_features} a = BatchFeature(data=__lowerCAmelCase , tensor_type=__lowerCAmelCase ) return encoded_inputs
32
0
import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging __magic_name__ =logging.get_logger(__name__) __magic_name__ =r''' Args: input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`): Indices of input sequence tokens in the vocabulary. Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and [`PreTrainedTokenizer.__call__`] for details. [What are input IDs?](../glossary#input-ids) scores (`torch.FloatTensor` of shape `(batch_size, config.vocab_size)`): Prediction scores of a language modeling head. These can be scores for each vocabulary token before SoftMax or scores for each vocabulary token after SoftMax. kwargs (`Dict[str, Any]`, *optional*): Additional stopping criteria specific kwargs. Return: `bool`. `False` indicates we should continue, `True` indicates we should stop. ''' class _A ( __UpperCamelCase ): @add_start_docstrings(SCREAMING_SNAKE_CASE_ ) def __call__(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) -> bool: '''simple docstring''' raise NotImplementedError('''StoppingCriteria needs to be subclassed''' ) class _A ( __UpperCamelCase ): def __init__(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None ) -> Optional[int]: '''simple docstring''' UpperCamelCase__ = max_length UpperCamelCase__ = max_position_embeddings @add_start_docstrings(SCREAMING_SNAKE_CASE_ ) def __call__(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) -> bool: '''simple docstring''' UpperCamelCase__ = input_ids.shape[-1] UpperCamelCase__ = cur_len >= self.max_length if self.max_position_embeddings is not None and not is_done and cur_len >= self.max_position_embeddings: logger.warning_once( '''This is a friendly reminder - the current text generation call will exceed the model\'s predefined ''' F"maximum length ({self.max_position_embeddings}). Depending on the model, you may observe " '''exceptions, performance degradation, or nothing at all.''' ) return is_done class _A ( __UpperCamelCase ): def __init__(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> str: '''simple docstring''' warnings.warn( '''The class `MaxNewTokensCriteria` is deprecated. ''' F"Please use `MaxLengthCriteria(max_length={start_length + max_new_tokens})` " '''with `max_length = start_length + max_new_tokens` instead.''' , SCREAMING_SNAKE_CASE_ , ) UpperCamelCase__ = start_length UpperCamelCase__ = max_new_tokens UpperCamelCase__ = start_length + max_new_tokens @add_start_docstrings(SCREAMING_SNAKE_CASE_ ) def __call__(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) -> bool: '''simple docstring''' return input_ids.shape[-1] >= self.max_length class _A ( __UpperCamelCase ): def __init__(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None ) -> Optional[int]: '''simple docstring''' UpperCamelCase__ = max_time UpperCamelCase__ = time.time() if initial_timestamp is None else initial_timestamp @add_start_docstrings(SCREAMING_SNAKE_CASE_ ) def __call__(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) -> bool: '''simple docstring''' return time.time() - self.initial_timestamp > self.max_time class _A ( __UpperCamelCase ): @add_start_docstrings(SCREAMING_SNAKE_CASE_ ) def __call__(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) -> bool: '''simple docstring''' return any(criteria(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for criteria in self ) @property def _a (self ) -> Optional[int]: '''simple docstring''' for stopping_criterium in self: if isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): return stopping_criterium.max_length elif isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): return stopping_criterium.max_length return None def __UpperCamelCase ( A , A ): UpperCamelCase__ = stopping_criteria.max_length UpperCamelCase__ = deepcopy(A ) if stopping_max_length is not None and stopping_max_length != max_length: warnings.warn('''You set different `max_length` for stopping criteria and `max_length` parameter''' , A ) elif stopping_max_length is None: new_stopping_criteria.append(MaxLengthCriteria(max_length=A ) ) return new_stopping_criteria
415
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DetaImageProcessor class _A ( unittest.TestCase ): def __init__(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=7 , SCREAMING_SNAKE_CASE_=3 , SCREAMING_SNAKE_CASE_=30 , SCREAMING_SNAKE_CASE_=400 , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=[0.5, 0.5, 0.5] , SCREAMING_SNAKE_CASE_=[0.5, 0.5, 0.5] , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=1 / 255 , SCREAMING_SNAKE_CASE_=True , ) -> List[Any]: '''simple docstring''' UpperCamelCase__ = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 1333} UpperCamelCase__ = parent UpperCamelCase__ = batch_size UpperCamelCase__ = num_channels UpperCamelCase__ = min_resolution UpperCamelCase__ = max_resolution UpperCamelCase__ = do_resize UpperCamelCase__ = size UpperCamelCase__ = do_normalize UpperCamelCase__ = image_mean UpperCamelCase__ = image_std UpperCamelCase__ = do_rescale UpperCamelCase__ = rescale_factor UpperCamelCase__ = do_pad def _a (self ) -> List[str]: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def _a (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False ) -> str: '''simple docstring''' if not batched: UpperCamelCase__ = image_inputs[0] if isinstance(SCREAMING_SNAKE_CASE_ , Image.Image ): UpperCamelCase__ , UpperCamelCase__ = image.size else: UpperCamelCase__ , UpperCamelCase__ = image.shape[1], image.shape[2] if w < h: UpperCamelCase__ = int(self.size['''shortest_edge'''] * h / w ) UpperCamelCase__ = self.size['''shortest_edge'''] elif w > h: UpperCamelCase__ = self.size['''shortest_edge'''] UpperCamelCase__ = int(self.size['''shortest_edge'''] * w / h ) else: UpperCamelCase__ = self.size['''shortest_edge'''] UpperCamelCase__ = self.size['''shortest_edge'''] else: UpperCamelCase__ = [] for image in image_inputs: UpperCamelCase__ , UpperCamelCase__ = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) UpperCamelCase__ = max(SCREAMING_SNAKE_CASE_ , key=lambda SCREAMING_SNAKE_CASE_ : item[0] )[0] UpperCamelCase__ = max(SCREAMING_SNAKE_CASE_ , key=lambda SCREAMING_SNAKE_CASE_ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class _A ( __UpperCamelCase , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Optional[int] =DetaImageProcessor if is_vision_available() else None def _a (self ) -> Tuple: '''simple docstring''' UpperCamelCase__ = DetaImageProcessingTester(self ) @property def _a (self ) -> List[str]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def _a (self ) -> Dict: '''simple docstring''' UpperCamelCase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''image_mean''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''image_std''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''do_normalize''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''do_resize''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''do_rescale''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''do_pad''' ) ) self.assertTrue(hasattr(SCREAMING_SNAKE_CASE_ , '''size''' ) ) def _a (self ) -> Dict: '''simple docstring''' UpperCamelCase__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 1333} ) self.assertEqual(image_processor.do_pad , SCREAMING_SNAKE_CASE_ ) def _a (self ) -> List[Any]: '''simple docstring''' pass def _a (self ) -> List[str]: '''simple docstring''' UpperCamelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCamelCase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE_ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE_ , Image.Image ) # Test not batched input UpperCamelCase__ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values UpperCamelCase__ , UpperCamelCase__ = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase__ , UpperCamelCase__ = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = image_processing(SCREAMING_SNAKE_CASE_ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _a (self ) -> int: '''simple docstring''' UpperCamelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCamelCase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE_ , numpify=SCREAMING_SNAKE_CASE_ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE_ , np.ndarray ) # Test not batched input UpperCamelCase__ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values UpperCamelCase__ , UpperCamelCase__ = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase__ = image_processing(SCREAMING_SNAKE_CASE_ , return_tensors='''pt''' ).pixel_values UpperCamelCase__ , UpperCamelCase__ = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _a (self ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCamelCase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=SCREAMING_SNAKE_CASE_ , torchify=SCREAMING_SNAKE_CASE_ ) for image in image_inputs: self.assertIsInstance(SCREAMING_SNAKE_CASE_ , torch.Tensor ) # Test not batched input UpperCamelCase__ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values UpperCamelCase__ , UpperCamelCase__ = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched UpperCamelCase__ = image_processing(SCREAMING_SNAKE_CASE_ , return_tensors='''pt''' ).pixel_values UpperCamelCase__ , UpperCamelCase__ = self.image_processor_tester.get_expected_values(SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def _a (self ) -> Optional[int]: '''simple docstring''' UpperCamelCase__ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: UpperCamelCase__ = json.loads(f.read() ) UpperCamelCase__ = {'''image_id''': 3_9769, '''annotations''': target} # encode them UpperCamelCase__ = DetaImageProcessor() UpperCamelCase__ = image_processing(images=SCREAMING_SNAKE_CASE_ , annotations=SCREAMING_SNAKE_CASE_ , return_tensors='''pt''' ) # verify pixel values UpperCamelCase__ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['''pixel_values'''].shape , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , SCREAMING_SNAKE_CASE_ , atol=1E-4 ) ) # verify area UpperCamelCase__ = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , SCREAMING_SNAKE_CASE_ ) ) # verify boxes UpperCamelCase__ = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , SCREAMING_SNAKE_CASE_ , atol=1E-3 ) ) # verify image_id UpperCamelCase__ = torch.tensor([3_9769] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , SCREAMING_SNAKE_CASE_ ) ) # verify is_crowd UpperCamelCase__ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , SCREAMING_SNAKE_CASE_ ) ) # verify class_labels UpperCamelCase__ = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , SCREAMING_SNAKE_CASE_ ) ) # verify orig_size UpperCamelCase__ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , SCREAMING_SNAKE_CASE_ ) ) # verify size UpperCamelCase__ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , SCREAMING_SNAKE_CASE_ ) ) @slow def _a (self ) -> Optional[Any]: '''simple docstring''' UpperCamelCase__ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: UpperCamelCase__ = json.loads(f.read() ) UpperCamelCase__ = {'''file_name''': '''000000039769.png''', '''image_id''': 3_9769, '''segments_info''': target} UpperCamelCase__ = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them UpperCamelCase__ = DetaImageProcessor(format='''coco_panoptic''' ) UpperCamelCase__ = image_processing(images=SCREAMING_SNAKE_CASE_ , annotations=SCREAMING_SNAKE_CASE_ , masks_path=SCREAMING_SNAKE_CASE_ , return_tensors='''pt''' ) # verify pixel values UpperCamelCase__ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['''pixel_values'''].shape , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , SCREAMING_SNAKE_CASE_ , atol=1E-4 ) ) # verify area UpperCamelCase__ = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , SCREAMING_SNAKE_CASE_ ) ) # verify boxes UpperCamelCase__ = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , SCREAMING_SNAKE_CASE_ , atol=1E-3 ) ) # verify image_id UpperCamelCase__ = torch.tensor([3_9769] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , SCREAMING_SNAKE_CASE_ ) ) # verify is_crowd UpperCamelCase__ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , SCREAMING_SNAKE_CASE_ ) ) # verify class_labels UpperCamelCase__ = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , SCREAMING_SNAKE_CASE_ ) ) # verify masks UpperCamelCase__ = 82_2873 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , SCREAMING_SNAKE_CASE_ ) # verify orig_size UpperCamelCase__ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , SCREAMING_SNAKE_CASE_ ) ) # verify size UpperCamelCase__ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , SCREAMING_SNAKE_CASE_ ) )
415
1
'''simple docstring''' def snake_case ( a_ : int = 4_000_000 ) -> int: """simple docstring""" UpperCamelCase_ : Any = [] UpperCamelCase_ , UpperCamelCase_ : Union[str, Any] = 0, 1 while b <= n: if b % 2 == 0: even_fibs.append(a_ ) UpperCamelCase_ , UpperCamelCase_ : List[Any] = b, a + b return sum(a_ ) if __name__ == "__main__": print(f"{solution() = }")
543
'''simple docstring''' import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class A ( unittest.TestCase ): """simple docstring""" def _UpperCAmelCase ( self ): UpperCamelCase_ : str = """| <pad> <unk> <s> </s> a b c d e f g h i j k""".split() UpperCamelCase_ : int = dict(zip(__lowerCAmelCase , range(len(__lowerCAmelCase ) ) ) ) UpperCamelCase_ : Tuple = { """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>""", } UpperCamelCase_ : Optional[Any] = { """feature_size""": 1, """padding_value""": 0.0, """sampling_rate""": 1_60_00, """return_attention_mask""": False, """do_normalize""": True, } UpperCamelCase_ : int = tempfile.mkdtemp() UpperCamelCase_ : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) UpperCamelCase_ : List[Any] = os.path.join(self.tmpdirname , __lowerCAmelCase ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCAmelCase ) + """\n""" ) with open(self.feature_extraction_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(__lowerCAmelCase ) + """\n""" ) # load decoder from hub UpperCamelCase_ : Union[str, Any] = """hf-internal-testing/ngram-beam-search-decoder""" def _UpperCAmelCase ( self , **__lowerCAmelCase ): UpperCamelCase_ : str = self.add_kwargs_tokens_map.copy() kwargs.update(__lowerCAmelCase ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **__lowerCAmelCase ) def _UpperCAmelCase ( self , **__lowerCAmelCase ): return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **__lowerCAmelCase ) def _UpperCAmelCase ( self , **__lowerCAmelCase ): return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **__lowerCAmelCase ) def _UpperCAmelCase ( self ): shutil.rmtree(self.tmpdirname ) def _UpperCAmelCase ( self ): UpperCamelCase_ : Union[str, Any] = self.get_tokenizer() UpperCamelCase_ : str = self.get_feature_extractor() UpperCamelCase_ : Tuple = self.get_decoder() UpperCamelCase_ : int = WavaVecaProcessorWithLM(tokenizer=__lowerCAmelCase , feature_extractor=__lowerCAmelCase , decoder=__lowerCAmelCase ) processor.save_pretrained(self.tmpdirname ) UpperCamelCase_ : Dict = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , __lowerCAmelCase ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , __lowerCAmelCase ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , __lowerCAmelCase ) def _UpperCAmelCase ( self ): UpperCamelCase_ : Union[str, Any] = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match UpperCamelCase_ : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def _UpperCAmelCase ( self ): UpperCamelCase_ : int = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(["""xx"""] ) with self.assertRaisesRegex(__lowerCAmelCase , """include""" ): WavaVecaProcessorWithLM( tokenizer=__lowerCAmelCase , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def _UpperCAmelCase ( self ): UpperCamelCase_ : Tuple = self.get_feature_extractor() UpperCamelCase_ : Tuple = self.get_tokenizer() UpperCamelCase_ : Any = self.get_decoder() UpperCamelCase_ : List[Any] = WavaVecaProcessorWithLM(tokenizer=__lowerCAmelCase , feature_extractor=__lowerCAmelCase , decoder=__lowerCAmelCase ) UpperCamelCase_ : List[Any] = floats_list((3, 10_00) ) UpperCamelCase_ : Tuple = feature_extractor(__lowerCAmelCase , return_tensors="""np""" ) UpperCamelCase_ : str = processor(__lowerCAmelCase , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _UpperCAmelCase ( self ): UpperCamelCase_ : Dict = self.get_feature_extractor() UpperCamelCase_ : List[Any] = self.get_tokenizer() UpperCamelCase_ : List[Any] = self.get_decoder() UpperCamelCase_ : Optional[Any] = WavaVecaProcessorWithLM(tokenizer=__lowerCAmelCase , feature_extractor=__lowerCAmelCase , decoder=__lowerCAmelCase ) UpperCamelCase_ : List[str] = """This is a test string""" UpperCamelCase_ : Optional[Any] = processor(text=__lowerCAmelCase ) UpperCamelCase_ : int = tokenizer(__lowerCAmelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _UpperCAmelCase ( self , __lowerCAmelCase=(2, 10, 16) , __lowerCAmelCase=77 ): np.random.seed(__lowerCAmelCase ) return np.random.rand(*__lowerCAmelCase ) def _UpperCAmelCase ( self ): UpperCamelCase_ : int = self.get_feature_extractor() UpperCamelCase_ : Tuple = self.get_tokenizer() UpperCamelCase_ : Optional[int] = self.get_decoder() UpperCamelCase_ : int = WavaVecaProcessorWithLM(tokenizer=__lowerCAmelCase , feature_extractor=__lowerCAmelCase , decoder=__lowerCAmelCase ) UpperCamelCase_ : List[str] = self._get_dummy_logits(shape=(10, 16) , seed=13 ) UpperCamelCase_ : Any = processor.decode(__lowerCAmelCase ) UpperCamelCase_ : Any = decoder.decode_beams(__lowerCAmelCase )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual("""</s> <s> </s>""" , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ["""fork"""], ["""spawn"""]] ) def _UpperCAmelCase ( self , __lowerCAmelCase ): UpperCamelCase_ : Union[str, Any] = self.get_feature_extractor() UpperCamelCase_ : str = self.get_tokenizer() UpperCamelCase_ : List[Any] = self.get_decoder() UpperCamelCase_ : Optional[Any] = WavaVecaProcessorWithLM(tokenizer=__lowerCAmelCase , feature_extractor=__lowerCAmelCase , decoder=__lowerCAmelCase ) UpperCamelCase_ : Union[str, Any] = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: UpperCamelCase_ : List[Any] = processor.batch_decode(__lowerCAmelCase ) else: with get_context(__lowerCAmelCase ).Pool() as pool: UpperCamelCase_ : Any = processor.batch_decode(__lowerCAmelCase , __lowerCAmelCase ) UpperCamelCase_ : Tuple = list(__lowerCAmelCase ) with get_context("""fork""" ).Pool() as p: UpperCamelCase_ : Optional[int] = decoder.decode_beams_batch(__lowerCAmelCase , __lowerCAmelCase ) UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ : Union[str, Any] = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(__lowerCAmelCase , decoded_processor.text ) self.assertListEqual(["""<s> <s> </s>""", """<s> <s> <s>"""] , decoded_processor.text ) self.assertListEqual(__lowerCAmelCase , decoded_processor.logit_score ) self.assertListEqual(__lowerCAmelCase , decoded_processor.lm_score ) def _UpperCAmelCase ( self ): UpperCamelCase_ : Dict = self.get_feature_extractor() UpperCamelCase_ : Tuple = self.get_tokenizer() UpperCamelCase_ : Tuple = self.get_decoder() UpperCamelCase_ : Optional[Any] = WavaVecaProcessorWithLM(tokenizer=__lowerCAmelCase , feature_extractor=__lowerCAmelCase , decoder=__lowerCAmelCase ) UpperCamelCase_ : List[str] = self._get_dummy_logits() UpperCamelCase_ : Dict = 15 UpperCamelCase_ : str = -20.0 UpperCamelCase_ : Dict = -4.0 UpperCamelCase_ : Union[str, Any] = processor.batch_decode( __lowerCAmelCase , beam_width=__lowerCAmelCase , beam_prune_logp=__lowerCAmelCase , token_min_logp=__lowerCAmelCase , ) UpperCamelCase_ : Any = decoded_processor_out.text UpperCamelCase_ : Tuple = list(__lowerCAmelCase ) with get_context("""fork""" ).Pool() as pool: UpperCamelCase_ : str = decoder.decode_beams_batch( __lowerCAmelCase , __lowerCAmelCase , beam_width=__lowerCAmelCase , beam_prune_logp=__lowerCAmelCase , token_min_logp=__lowerCAmelCase , ) UpperCamelCase_ : str = [d[0][0] for d in decoded_decoder_out] UpperCamelCase_ : List[str] = [d[0][2] for d in decoded_decoder_out] UpperCamelCase_ : Union[str, Any] = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase ) self.assertListEqual(["""</s> <s> <s>""", """<s> <s> <s>"""] , __lowerCAmelCase ) self.assertTrue(np.array_equal(__lowerCAmelCase , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.0_54, -18.4_47] , __lowerCAmelCase , atol=1E-3 ) ) self.assertTrue(np.array_equal(__lowerCAmelCase , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.5_54, -13.94_74] , __lowerCAmelCase , atol=1E-3 ) ) def _UpperCAmelCase ( self ): UpperCamelCase_ : Optional[int] = self.get_feature_extractor() UpperCamelCase_ : Union[str, Any] = self.get_tokenizer() UpperCamelCase_ : int = self.get_decoder() UpperCamelCase_ : Dict = WavaVecaProcessorWithLM(tokenizer=__lowerCAmelCase , feature_extractor=__lowerCAmelCase , decoder=__lowerCAmelCase ) UpperCamelCase_ : str = self._get_dummy_logits() UpperCamelCase_ : Optional[int] = 2.0 UpperCamelCase_ : List[str] = 5.0 UpperCamelCase_ : Optional[Any] = -20.0 UpperCamelCase_ : Optional[Any] = True UpperCamelCase_ : Union[str, Any] = processor.batch_decode( __lowerCAmelCase , alpha=__lowerCAmelCase , beta=__lowerCAmelCase , unk_score_offset=__lowerCAmelCase , lm_score_boundary=__lowerCAmelCase , ) UpperCamelCase_ : List[str] = decoded_processor_out.text UpperCamelCase_ : List[str] = list(__lowerCAmelCase ) decoder.reset_params( alpha=__lowerCAmelCase , beta=__lowerCAmelCase , unk_score_offset=__lowerCAmelCase , lm_score_boundary=__lowerCAmelCase , ) with get_context("""fork""" ).Pool() as pool: UpperCamelCase_ : int = decoder.decode_beams_batch( __lowerCAmelCase , __lowerCAmelCase , ) UpperCamelCase_ : Any = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase ) self.assertListEqual(["""<s> </s> <s> </s> </s>""", """</s> </s> <s> </s> </s>"""] , __lowerCAmelCase ) UpperCamelCase_ : Union[str, Any] = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , __lowerCAmelCase ) def _UpperCAmelCase ( self ): UpperCamelCase_ : Dict = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCamelCase_ : Optional[int] = processor.decoder.model_container[processor.decoder._model_key] UpperCamelCase_ : int = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCamelCase_ : Any = os.listdir(__lowerCAmelCase ) UpperCamelCase_ : Optional[Any] = ["""alphabet.json""", """language_model"""] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase ) def _UpperCAmelCase ( self ): UpperCamelCase_ : Union[str, Any] = snapshot_download("""hf-internal-testing/processor_with_lm""" ) UpperCamelCase_ : Union[str, Any] = WavaVecaProcessorWithLM.from_pretrained(__lowerCAmelCase ) UpperCamelCase_ : Union[str, Any] = processor.decoder.model_container[processor.decoder._model_key] UpperCamelCase_ : Tuple = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() UpperCamelCase_ : Union[str, Any] = os.listdir(__lowerCAmelCase ) UpperCamelCase_ : str = os.listdir(__lowerCAmelCase ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase ) def _UpperCAmelCase ( self ): UpperCamelCase_ : Any = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCamelCase_ : Tuple = AutoProcessor.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCamelCase_ : Dict = floats_list((3, 10_00) ) UpperCamelCase_ : List[Any] = processor_wavaveca(__lowerCAmelCase , return_tensors="""np""" ) UpperCamelCase_ : Tuple = processor_auto(__lowerCAmelCase , return_tensors="""np""" ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1E-2 ) UpperCamelCase_ : Optional[int] = self._get_dummy_logits() UpperCamelCase_ : Dict = processor_wavaveca.batch_decode(__lowerCAmelCase ) UpperCamelCase_ : Any = processor_auto.batch_decode(__lowerCAmelCase ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def _UpperCAmelCase ( self ): UpperCamelCase_ : Tuple = self.get_feature_extractor() UpperCamelCase_ : int = self.get_tokenizer() UpperCamelCase_ : List[Any] = self.get_decoder() UpperCamelCase_ : List[str] = WavaVecaProcessorWithLM(tokenizer=__lowerCAmelCase , feature_extractor=__lowerCAmelCase , decoder=__lowerCAmelCase ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg="""`processor` and `feature_extractor` model input names do not match""" , ) @staticmethod def _UpperCAmelCase ( __lowerCAmelCase , __lowerCAmelCase ): UpperCamelCase_ : Union[str, Any] = [d[key] for d in offsets] return retrieved_list def _UpperCAmelCase ( self ): UpperCamelCase_ : Dict = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCamelCase_ : int = self._get_dummy_logits()[0] UpperCamelCase_ : List[Any] = processor.decode(__lowerCAmelCase , output_word_offsets=__lowerCAmelCase ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(__lowerCAmelCase , __lowerCAmelCase ) ) self.assertEqual(""" """.join(self.get_from_offsets(outputs["""word_offsets"""] , """word""" ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] , """word""" ) , ["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] , """start_offset""" ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] , """end_offset""" ) , [1, 3, 5] ) def _UpperCAmelCase ( self ): UpperCamelCase_ : Any = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) UpperCamelCase_ : Union[str, Any] = self._get_dummy_logits() UpperCamelCase_ : Dict = processor.batch_decode(__lowerCAmelCase , output_word_offsets=__lowerCAmelCase ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(__lowerCAmelCase , __lowerCAmelCase ) ) self.assertListEqual( [""" """.join(self.get_from_offsets(__lowerCAmelCase , """word""" ) ) for o in outputs["""word_offsets"""]] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] , """word""" ) , ["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] , """start_offset""" ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] , """end_offset""" ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def _UpperCAmelCase ( self ): import torch UpperCamelCase_ : str = load_dataset("""common_voice""" , """en""" , split="""train""" , streaming=__lowerCAmelCase ) UpperCamelCase_ : Union[str, Any] = ds.cast_column("""audio""" , datasets.Audio(sampling_rate=1_60_00 ) ) UpperCamelCase_ : Union[str, Any] = iter(__lowerCAmelCase ) UpperCamelCase_ : int = next(__lowerCAmelCase ) UpperCamelCase_ : List[str] = AutoProcessor.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) UpperCamelCase_ : Tuple = WavaVecaForCTC.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train UpperCamelCase_ : Dict = processor(sample["""audio"""]["""array"""] , return_tensors="""pt""" ).input_values with torch.no_grad(): UpperCamelCase_ : List[Any] = model(__lowerCAmelCase ).logits.cpu().numpy() UpperCamelCase_ : Tuple = processor.decode(logits[0] , output_word_offsets=__lowerCAmelCase ) UpperCamelCase_ : List[Any] = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate UpperCamelCase_ : Dict = [ { """start_time""": d["""start_offset"""] * time_offset, """end_time""": d["""end_offset"""] * time_offset, """word""": d["""word"""], } for d in output["""word_offsets"""] ] UpperCamelCase_ : Optional[Any] = """WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL""" # output words self.assertEqual(""" """.join(self.get_from_offsets(__lowerCAmelCase , """word""" ) ) , __lowerCAmelCase ) self.assertEqual(""" """.join(self.get_from_offsets(__lowerCAmelCase , """word""" ) ) , output.text ) # output times UpperCamelCase_ : str = torch.tensor(self.get_from_offsets(__lowerCAmelCase , """start_time""" ) ) UpperCamelCase_ : Union[str, Any] = torch.tensor(self.get_from_offsets(__lowerCAmelCase , """end_time""" ) ) # fmt: off UpperCamelCase_ : Union[str, Any] = torch.tensor([1.41_99, 1.65_99, 2.25_99, 3.0, 3.24, 3.59_99, 3.79_99, 4.09_99, 4.26, 4.94, 5.28, 5.65_99, 5.78, 5.94, 6.32, 6.53_99, 6.65_99] ) UpperCamelCase_ : Union[str, Any] = torch.tensor([1.53_99, 1.89_99, 2.9, 3.16, 3.53_99, 3.72, 4.01_99, 4.17_99, 4.76, 5.15_99, 5.55_99, 5.69_99, 5.86, 6.19_99, 6.38, 6.61_99, 6.94] ) # fmt: on self.assertTrue(torch.allclose(__lowerCAmelCase , __lowerCAmelCase , atol=0.01 ) ) self.assertTrue(torch.allclose(__lowerCAmelCase , __lowerCAmelCase , atol=0.01 ) )
543
1
'''simple docstring''' import warnings from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging lowerCAmelCase_ : Dict = logging.get_logger(__name__) class UpperCamelCase__ ( __lowerCAmelCase ): lowerCAmelCase__ : Union[str, Any] = ["input_values", "attention_mask"] def __init__( self : Union[str, Any] , lowerCamelCase : int = 1 , lowerCamelCase : int = 1_6_0_0_0 , lowerCamelCase : float = 0.0 , lowerCamelCase : bool = False , lowerCamelCase : int = 8_0 , lowerCamelCase : int = 1_6 , lowerCamelCase : int = 6_4 , lowerCamelCase : str = "hann_window" , lowerCamelCase : float = 1.0 , lowerCamelCase : float = 8_0 , lowerCamelCase : float = 7_6_0_0 , lowerCamelCase : float = 1e-10 , lowerCamelCase : int = 2 , lowerCamelCase : bool = True , **lowerCamelCase : Optional[Any] , ): '''simple docstring''' super().__init__(feature_size=lowerCamelCase , sampling_rate=lowerCamelCase , padding_value=lowerCamelCase , **lowerCamelCase ) a__ = do_normalize a__ = return_attention_mask a__ = num_mel_bins a__ = hop_length a__ = win_length a__ = win_function a__ = frame_signal_scale a__ = fmin a__ = fmax a__ = mel_floor a__ = reduction_factor a__ = win_length * sampling_rate // 1_0_0_0 a__ = hop_length * sampling_rate // 1_0_0_0 a__ = optimal_fft_length(self.sample_size ) a__ = (self.n_fft // 2) + 1 a__ = window_function(window_length=self.sample_size , name=self.win_function , periodic=lowerCamelCase ) a__ = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.num_mel_bins , min_frequency=self.fmin , max_frequency=self.fmax , sampling_rate=self.sampling_rate , norm="slaney" , mel_scale="slaney" , ) if frame_signal_scale != 1.0: warnings.warn( "The argument `frame_signal_scale` is deprecated and will be removed in version 4.30.0 of Transformers" , lowerCamelCase , ) if reduction_factor != 2.0: warnings.warn( "The argument `reduction_factor` is deprecated and will be removed in version 4.30.0 of Transformers" , lowerCamelCase , ) @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def __a ( lowerCamelCase : List[np.ndarray] , lowerCamelCase : List[np.ndarray] , lowerCamelCase : float = 0.0 ): '''simple docstring''' if attention_mask is not None: a__ = np.array(lowerCamelCase , np.intaa ) a__ = [] for vector, length in zip(lowerCamelCase , attention_mask.sum(-1 ) ): a__ = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1e-7 ) if length < normed_slice.shape[0]: a__ = padding_value normed_input_values.append(lowerCamelCase ) else: a__ = [(x - x.mean()) / np.sqrt(x.var() + 1e-7 ) for x in input_values] return normed_input_values def __a ( self : Any , lowerCamelCase : np.ndarray , ): '''simple docstring''' a__ = spectrogram( lowerCamelCase , window=self.window , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , mel_filters=self.mel_filters , mel_floor=self.mel_floor , log_mel="log10" , ) return log_mel_spec.T def __call__( self : Any , lowerCamelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , lowerCamelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , lowerCamelCase : Union[bool, str, PaddingStrategy] = False , lowerCamelCase : Optional[int] = None , lowerCamelCase : bool = False , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , lowerCamelCase : Optional[int] = None , **lowerCamelCase : Optional[int] , ): '''simple docstring''' if audio is None and audio_target is None: raise ValueError("You must provide either `audio` or `audio_target` values." ) if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' F''' {self.sampling_rate}. Please make sure that the provided audio input was sampled with''' F''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) if audio is not None: a__ = self._process_audio( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase , ) else: a__ = None if audio_target is not None: a__ = self._process_audio( lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , **lowerCamelCase , ) if inputs is None: return inputs_target else: a__ = inputs_target["input_values"] a__ = inputs_target.get("attention_mask" ) if decoder_attention_mask is not None: a__ = decoder_attention_mask return inputs def __a ( self : int , lowerCamelCase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , lowerCamelCase : bool = False , lowerCamelCase : Union[bool, str, PaddingStrategy] = False , lowerCamelCase : Optional[int] = None , lowerCamelCase : bool = False , lowerCamelCase : Optional[int] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[Union[str, TensorType]] = None , **lowerCamelCase : int , ): '''simple docstring''' a__ = isinstance(lowerCamelCase , np.ndarray ) and len(speech.shape ) > 1 if is_batched_numpy and len(speech.shape ) > 2: raise ValueError(F'''Only mono-channel audio is supported for input to {self}''' ) a__ = is_batched_numpy or ( isinstance(lowerCamelCase , (list, tuple) ) and (isinstance(speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: a__ = [np.asarray(lowerCamelCase , dtype=np.floataa ) for speech in speech] elif not is_batched and not isinstance(lowerCamelCase , np.ndarray ): a__ = np.asarray(lowerCamelCase , dtype=np.floataa ) elif isinstance(lowerCamelCase , np.ndarray ) and speech.dtype is np.dtype(np.floataa ): a__ = speech.astype(np.floataa ) # always return batch if not is_batched: a__ = [speech] # needed to make pad() work on spectrogram inputs a__ = self.feature_size # convert into correct format for padding if is_target: a__ = [self._extract_mel_features(lowerCamelCase ) for waveform in speech] a__ = BatchFeature({"input_values": features} ) a__ = self.num_mel_bins else: a__ = BatchFeature({"input_values": speech} ) a__ = self.pad( lowerCamelCase , padding=lowerCamelCase , max_length=lowerCamelCase , truncation=lowerCamelCase , pad_to_multiple_of=lowerCamelCase , return_attention_mask=lowerCamelCase , **lowerCamelCase , ) a__ = feature_size_hack # convert input values to correct format a__ = padded_inputs["input_values"] if not isinstance(input_values[0] , np.ndarray ): a__ = [np.asarray(lowerCamelCase , dtype=np.floataa ) for array in input_values] elif ( not isinstance(lowerCamelCase , np.ndarray ) and isinstance(input_values[0] , np.ndarray ) and input_values[0].dtype is np.dtype(np.floataa ) ): a__ = [array.astype(np.floataa ) for array in input_values] elif isinstance(lowerCamelCase , np.ndarray ) and input_values.dtype is np.dtype(np.floataa ): a__ = input_values.astype(np.floataa ) # convert attention_mask to correct format a__ = padded_inputs.get("attention_mask" ) if attention_mask is not None: a__ = [np.asarray(lowerCamelCase , dtype=np.intaa ) for array in attention_mask] # zero-mean and unit-variance normalization if not is_target and self.do_normalize: a__ = ( attention_mask if self._get_padding_strategies(lowerCamelCase , max_length=lowerCamelCase ) is not PaddingStrategy.DO_NOT_PAD else None ) a__ = self.zero_mean_unit_var_norm( padded_inputs["input_values"] , attention_mask=lowerCamelCase , padding_value=self.padding_value ) if return_tensors is not None: a__ = padded_inputs.convert_to_tensors(lowerCamelCase ) return padded_inputs def __a ( self : Dict ): '''simple docstring''' a__ = super().to_dict() # Don't serialize these as they are derived from the other properties. a__ = ["window", "mel_filters", "sample_size", "sample_stride", "n_fft", "n_freqs"] for name in names: if name in output: del output[name] return output
489
'''simple docstring''' import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def _lowerCamelCase (__lowerCamelCase : Dict ) -> Tuple: a__ = [ "encoder.version", "decoder.version", "model.encoder.version", "model.decoder.version", "decoder.output_projection.weight", "_float_tensor", "encoder.embed_positions._float_tensor", "decoder.embed_positions._float_tensor", ] for k in ignore_keys: state_dict.pop(__lowerCamelCase , __lowerCamelCase ) def _lowerCamelCase (__lowerCamelCase : List[Any] ) -> Dict: a__ , a__ = emb.weight.shape a__ = nn.Linear(__lowerCamelCase , __lowerCamelCase , bias=__lowerCamelCase ) a__ = emb.weight.data return lin_layer def _lowerCamelCase (__lowerCamelCase : Union[str, Any] ) -> Tuple: a__ = torch.load(__lowerCamelCase , map_location="cpu" ) a__ = mam_aaa["args"] or mam_aaa["cfg"]["model"] a__ = mam_aaa["model"] remove_ignore_keys_(__lowerCamelCase ) a__ = state_dict["encoder.embed_tokens.weight"].shape[0] a__ = MaMaaaConfig( vocab_size=__lowerCamelCase , max_position_embeddings=1024 , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , encoder_layerdrop=args.encoder_layerdrop , decoder_layerdrop=args.decoder_layerdrop , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function="relu" , ) a__ = state_dict["decoder.embed_tokens.weight"] a__ = MaMaaaForConditionalGeneration(__lowerCamelCase ) model.model.load_state_dict(__lowerCamelCase , strict=__lowerCamelCase ) a__ = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": lowerCAmelCase_ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument("fairseq_path", type=str, help="path to a model.pt on local filesystem.") parser.add_argument("pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") lowerCAmelCase_ : Dict = parser.parse_args() lowerCAmelCase_ : Union[str, Any] = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
489
1
'''simple docstring''' import argparse import shutil import time from json import JSONDecodeError from logging import getLogger from pathlib import Path from typing import Dict, List import torch from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import ( SeqaSeqDataset, calculate_bleu, calculate_rouge, chunks, lmap, load_json, parse_numeric_n_bool_cl_kwargs, save_json, use_task_specific_params, write_txt_file, ) lowerCamelCase_ : Optional[Any] = getLogger(__name__) def __magic_name__( _A , _A , _A , _A = 8 , _A = 1024 , _A="val" , _A=None , _A=False , _A="summarization" , _A=None , _A=1 , _A = None , _A="" , **_A , ): '''simple docstring''' UpperCamelCase__ = str(_A ) assert local_rank is not None torch.distributed.init_process_group(backend="""nccl""" , rank=_A ) UpperCamelCase__ = Path(_A ) UpperCamelCase__ = save_dir.joinpath(f"rank_{local_rank}_output.json" ) torch.cuda.set_device(_A ) UpperCamelCase__ = AutoModelForSeqaSeqLM.from_pretrained(_A ).cuda() if fpaa: UpperCamelCase__ = model.half() # determine if we need to increase num_beams use_task_specific_params(_A , _A ) # update config with task specific params UpperCamelCase__ = generate_kwargs.pop("""num_beams""" , model.config.num_beams ) # AttributeError risk? if num_return_sequences > num_beams: UpperCamelCase__ = num_return_sequences UpperCamelCase__ = AutoTokenizer.from_pretrained(_A ) logger.info(f"Inferred tokenizer type: {tokenizer.__class__}" ) # if this is wrong, check config.model_type. if max_source_length is None: UpperCamelCase__ = tokenizer.model_max_length if prefix is None: UpperCamelCase__ = prefix or getattr(model.config , """prefix""" , """""" ) or """""" UpperCamelCase__ = SeqaSeqDataset( _A , _A , _A , max_target_length=1024 , type_path=_A , n_obs=_A , prefix=_A , **_A , ) # I set shuffle=True for a more accurate progress bar. # If all the longest samples are first, the prog bar estimate is too high at the beginning. UpperCamelCase__ = ds.make_sortish_sampler(_A , distributed=_A , add_extra_examples=_A , shuffle=_A ) UpperCamelCase__ = DataLoader(_A , sampler=_A , batch_size=_A , collate_fn=ds.collate_fn ) UpperCamelCase__ = [] for batch in tqdm(_A ): UpperCamelCase__ = model.generate( input_ids=batch["""input_ids"""].to(model.device ) , attention_mask=batch["""attention_mask"""].to(model.device ) , num_return_sequences=_A , num_beams=_A , **_A , ) UpperCamelCase__ = tokenizer.batch_decode(_A , skip_special_tokens=_A , clean_up_tokenization_spaces=_A ) UpperCamelCase__ = batch["""ids"""] if num_return_sequences > 1: UpperCamelCase__ = chunks(_A , _A ) # batch size chunks, each of size num_return_seq for i, pred in enumerate(_A ): results.append({"""pred""": pred, """id""": ids[i].item()} ) save_json(_A , _A ) return results, sampler.num_replicas def __magic_name__( ): '''simple docstring''' UpperCamelCase__ = argparse.ArgumentParser( epilog="""Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate""" ) parser.add_argument("""--data_dir""" , type=_A , help="""like cnn_dm/test.source""" ) parser.add_argument( """--model_name""" , type=_A , help="""like facebook/bart-large-cnn,t5-base, etc.""" , default="""sshleifer/distilbart-xsum-12-3""" , ) parser.add_argument("""--save_dir""" , type=_A , help="""where to save""" , default="""tmp_gen""" ) parser.add_argument("""--max_source_length""" , type=_A , default=_A ) parser.add_argument( """--type_path""" , type=_A , default="""test""" , help="""which subset to evaluate typically train/val/test""" ) parser.add_argument("""--task""" , type=_A , default="""summarization""" , help="""used for task_specific_params + metrics""" ) parser.add_argument("""--bs""" , type=_A , default=8 , required=_A , help="""batch size""" ) parser.add_argument( """--local_rank""" , type=_A , default=-1 , required=_A , help="""should be passed by distributed.launch""" ) parser.add_argument( """--n_obs""" , type=_A , default=_A , required=_A , help="""How many observations. Defaults to all.""" ) parser.add_argument( """--num_return_sequences""" , type=_A , default=1 , required=_A , help="""How many sequences to return""" ) parser.add_argument( """--sync_timeout""" , type=_A , default=600 , required=_A , help="""How long should master process wait for other processes to finish.""" , ) parser.add_argument("""--src_lang""" , type=_A , default=_A , required=_A ) parser.add_argument("""--tgt_lang""" , type=_A , default=_A , required=_A ) parser.add_argument( """--prefix""" , type=_A , required=_A , default=_A , help="""will be added to the begininng of src examples""" ) parser.add_argument("""--fp16""" , action="""store_true""" ) parser.add_argument("""--debug""" , action="""store_true""" ) UpperCamelCase__ = time.time() UpperCamelCase__ , UpperCamelCase__ = parser.parse_known_args() UpperCamelCase__ = parse_numeric_n_bool_cl_kwargs(_A ) if generate_kwargs and args.local_rank <= 0: print(f"parsed the following generate kwargs: {generate_kwargs}" ) UpperCamelCase__ = Path(args.save_dir + """_tmp""" ) Path(_A ).mkdir(exist_ok=_A ) # this handles locking. UpperCamelCase__ = list(json_save_dir.glob("""rank_*.json""" ) ) if intermediate_files: raise ValueError(f"Found files at {json_save_dir} please move or remove them." ) # In theory, a node could finish and save before another node hits this. If this happens, we can address later. UpperCamelCase__ = {} if args.src_lang is not None: UpperCamelCase__ = args.src_lang if args.tgt_lang is not None: UpperCamelCase__ = args.tgt_lang Path(args.save_dir ).mkdir(exist_ok=_A ) UpperCamelCase__ , UpperCamelCase__ = eval_data_dir( args.data_dir , _A , args.model_name , type_path=args.type_path , bs=args.bs , fpaa=args.fpaa , task=args.task , local_rank=args.local_rank , n_obs=args.n_obs , max_source_length=args.max_source_length , num_return_sequences=args.num_return_sequences , prefix=args.prefix , dataset_kwargs=_A , **_A , ) if args.local_rank <= 0: UpperCamelCase__ = Path(args.save_dir ) save_dir.mkdir(exist_ok=_A ) UpperCamelCase__ = gather_results_from_each_node(_A , _A , args.sync_timeout ) UpperCamelCase__ = combine_partial_results(_A ) if args.num_return_sequences > 1: UpperCamelCase__ = save_dir.joinpath("""pseudolabel_results.json""" ) print(f"Saving aggregated results at {save_path}, intermediate in {json_save_dir}/" ) save_json(_A , _A ) return UpperCamelCase__ = Path(args.data_dir ).joinpath(args.type_path + """.target""" ) with open(_A ) as f: UpperCamelCase__ = [x.rstrip() for x in f.readlines()][: len(_A )] # Calculate metrics, save metrics, and save _generations.txt UpperCamelCase__ = """translation""" in args.task UpperCamelCase__ = calculate_bleu if calc_bleu else calculate_rouge UpperCamelCase__ = """bleu""" if calc_bleu else """rouge""" UpperCamelCase__ = score_fn(_A , _A ) UpperCamelCase__ = len(_A ) UpperCamelCase__ = time.time() - start_time UpperCamelCase__ = round(runtime / metrics["""n_obs"""] , 4 ) UpperCamelCase__ = num_replicas # TODO(@stas00): add whatever metadata to metrics UpperCamelCase__ = save_dir.joinpath(f"{args.type_path}_{metric_name}.json" ) save_json(_A , _A , indent=_A ) print(_A ) write_txt_file(_A , save_dir.joinpath(f"{args.type_path}_generations.txt" ) ) if args.debug: write_txt_file(_A , save_dir.joinpath(f"{args.type_path}.target" ) ) else: shutil.rmtree(_A ) def __magic_name__( _A ): '''simple docstring''' UpperCamelCase__ = [] for partial_result in partial_results: records.extend(_A ) UpperCamelCase__ = sorted(_A , key=lambda _A : x["id"] ) UpperCamelCase__ = [x["""pred"""] for x in records] return preds def __magic_name__( _A , _A , _A ): '''simple docstring''' UpperCamelCase__ = time.time() logger.info("""waiting for all nodes to finish""" ) UpperCamelCase__ = None while (time.time() - start_wait) < timeout: UpperCamelCase__ = list(save_dir.glob("""rank_*.json""" ) ) if len(_A ) < num_replicas: continue try: # make sure all json files are fully saved UpperCamelCase__ = lmap(_A , _A ) return json_data except JSONDecodeError: continue else: raise TimeoutError("""Rank 0 gave up on waiting for other processes""" ) # Unreachable if __name__ == "__main__": # Usage for MT: run_generate()
265
'''simple docstring''' from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor lowerCamelCase_ : str = transforms.Compose( [ transforms.Resize((256, 256)), transforms.ToTensor(), transforms.Normalize([0.5], [0.5]), ] ) def __magic_name__( _A ): '''simple docstring''' if isinstance(_A , torch.Tensor ): return image elif isinstance(_A , PIL.Image.Image ): UpperCamelCase__ = [image] UpperCamelCase__ = [trans(img.convert("""RGB""" ) ) for img in image] UpperCamelCase__ = torch.stack(_A ) return image class _SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : Tuple , lowercase : Dict , lowercase : Optional[Any] ) -> Any: '''simple docstring''' super().__init__() # make sure scheduler can always be converted to DDIM UpperCamelCase__ = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=lowercase , scheduler=lowercase ) def A ( self : Tuple , lowercase : Tuple ) -> Any: '''simple docstring''' if strength < 0 or strength > 1: raise ValueError(f"The value of strength should in [0.0, 1.0] but is {strength}" ) def A ( self : str , lowercase : List[Any] , lowercase : Dict , lowercase : List[str] ) -> str: '''simple docstring''' UpperCamelCase__ = min(int(num_inference_steps * strength ) , lowercase ) UpperCamelCase__ = max(num_inference_steps - init_timestep , 0 ) UpperCamelCase__ = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def A ( self : Union[str, Any] , lowercase : Optional[Any] , lowercase : Any , lowercase : Union[str, Any] , lowercase : Dict , lowercase : Tuple , lowercase : int=None ) -> int: '''simple docstring''' if not isinstance(lowercase , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( f"`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(lowercase )}" ) UpperCamelCase__ = image.to(device=lowercase , dtype=lowercase ) if isinstance(lowercase , lowercase ) and len(lowercase ) != batch_size: raise ValueError( f"You have passed a list of generators of length {len(lowercase )}, but requested an effective batch" f" size of {batch_size}. Make sure the batch size matches the length of the generators." ) UpperCamelCase__ = init_latents.shape UpperCamelCase__ = randn_tensor(lowercase , generator=lowercase , device=lowercase , dtype=lowercase ) # get latents print("""add noise to latents at timestep""" , lowercase ) UpperCamelCase__ = self.scheduler.add_noise(lowercase , lowercase , lowercase ) UpperCamelCase__ = init_latents return latents @torch.no_grad() def __call__( self : Any , lowercase : Union[torch.FloatTensor, PIL.Image.Image] = None , lowercase : float = 0.8 , lowercase : int = 1 , lowercase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowercase : float = 0.0 , lowercase : int = 5_0 , lowercase : Optional[bool] = None , lowercase : Optional[str] = "pil" , lowercase : bool = True , ) -> Union[ImagePipelineOutput, Tuple]: '''simple docstring''' self.check_inputs(lowercase ) # 2. Preprocess image UpperCamelCase__ = preprocess(lowercase ) # 3. set timesteps self.scheduler.set_timesteps(lowercase , device=self.device ) UpperCamelCase__ , UpperCamelCase__ = self.get_timesteps(lowercase , lowercase , self.device ) UpperCamelCase__ = timesteps[:1].repeat(lowercase ) # 4. Prepare latent variables UpperCamelCase__ = self.prepare_latents(lowercase , lowercase , lowercase , self.unet.dtype , self.device , lowercase ) UpperCamelCase__ = latents # 5. Denoising loop for t in self.progress_bar(lowercase ): # 1. predict noise model_output UpperCamelCase__ = self.unet(lowercase , lowercase ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 UpperCamelCase__ = self.scheduler.step( lowercase , lowercase , lowercase , eta=lowercase , use_clipped_model_output=lowercase , generator=lowercase , ).prev_sample UpperCamelCase__ = (image / 2 + 0.5).clamp(0 , 1 ) UpperCamelCase__ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCamelCase__ = self.numpy_to_pil(lowercase ) if not return_dict: return (image, latent_timestep.item()) return ImagePipelineOutput(images=lowercase )
265
1
"""simple docstring""" import numpy # List of input, output pairs __SCREAMING_SNAKE_CASE = ( ((5, 2, 3), 15), ((6, 5, 9), 25), ((11, 12, 13), 41), ((1, 1, 1), 8), ((11, 12, 13), 41), ) __SCREAMING_SNAKE_CASE = (((5_15, 22, 13), 5_55), ((61, 35, 49), 1_50)) __SCREAMING_SNAKE_CASE = [2, 4, 1, 5] __SCREAMING_SNAKE_CASE = len(train_data) __SCREAMING_SNAKE_CASE = 0.009 def __a ( a, a="train" ): """simple docstring""" return calculate_hypothesis_value(a, a ) - output( a, a ) def __a ( a ): """simple docstring""" _a = 0 for i in range(len(a ) - 1 ): hyp_val += data_input_tuple[i] * parameter_vector[i + 1] hyp_val += parameter_vector[0] return hyp_val def __a ( a, a ): """simple docstring""" if data_set == "train": return train_data[example_no][1] elif data_set == "test": return test_data[example_no][1] return None def __a ( a, a ): """simple docstring""" if data_set == "train": return _hypothesis_value(train_data[example_no][0] ) elif data_set == "test": return _hypothesis_value(test_data[example_no][0] ) return None def __a ( a, a=m ): """simple docstring""" _a = 0 for i in range(a ): if index == -1: summation_value += _error(a ) else: summation_value += _error(a ) * train_data[i][0][index] return summation_value def __a ( a ): """simple docstring""" _a = summation_of_cost_derivative(a, a ) / m return cost_derivative_value def __a ( ): """simple docstring""" global parameter_vector # Tune these values to set a tolerance value for predicted output _a = 0.00_0002 _a = 0 _a = 0 while True: j += 1 _a = [0, 0, 0, 0] for i in range(0, len(a ) ): _a = get_cost_derivative(i - 1 ) _a = ( parameter_vector[i] - LEARNING_RATE * cost_derivative ) if numpy.allclose( a, a, atol=a, rtol=a, ): break _a = temp_parameter_vector print(("Number of iterations:", j) ) def __a ( ): """simple docstring""" for i in range(len(a ) ): print(("Actual output value:", output(a, "test" )) ) print(("Hypothesis output:", calculate_hypothesis_value(a, "test" )) ) if __name__ == "__main__": run_gradient_descent() print("""\nTesting gradient descent for a linear hypothesis function.\n""") test_gradient_descent()
388
"""simple docstring""" import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoFeatureExtractor, WavaVecaFeatureExtractor from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 __SCREAMING_SNAKE_CASE = get_tests_dir("""fixtures""") class __snake_case ( unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE_ ( self :Tuple ): # A mock response for an HTTP head request to emulate server down _a = mock.Mock() _a = 500 _a = {} _a = HTTPError _a = {} # Download this model to make sure it's in the cache. _a = WavaVecaFeatureExtractor.from_pretrained("hf-internal-testing/tiny-random-wav2vec2" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("requests.Session.request" , return_value=UpperCamelCase__ ) as mock_head: _a = WavaVecaFeatureExtractor.from_pretrained("hf-internal-testing/tiny-random-wav2vec2" ) # This check we did call the fake head request mock_head.assert_called() def SCREAMING_SNAKE_CASE_ ( self :str ): # This test is for deprecated behavior and can be removed in v5 _a = WavaVecaFeatureExtractor.from_pretrained( "https://huggingface.co/hf-internal-testing/tiny-random-wav2vec2/resolve/main/preprocessor_config.json" ) @is_staging_test class __snake_case ( unittest.TestCase ): """simple docstring""" @classmethod def SCREAMING_SNAKE_CASE_ ( cls :Any ): _a = TOKEN HfFolder.save_token(UpperCamelCase__ ) @classmethod def SCREAMING_SNAKE_CASE_ ( cls :Any ): try: delete_repo(token=cls._token , repo_id="test-feature-extractor" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-feature-extractor-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-feature-extractor" ) except HTTPError: pass def SCREAMING_SNAKE_CASE_ ( self :Dict ): _a = WavaVecaFeatureExtractor.from_pretrained(UpperCamelCase__ ) feature_extractor.push_to_hub("test-feature-extractor" , use_auth_token=self._token ) _a = WavaVecaFeatureExtractor.from_pretrained(f'{USER}/test-feature-extractor' ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(UpperCamelCase__ , getattr(UpperCamelCase__ , UpperCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="test-feature-extractor" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained( UpperCamelCase__ , repo_id="test-feature-extractor" , push_to_hub=UpperCamelCase__ , use_auth_token=self._token ) _a = WavaVecaFeatureExtractor.from_pretrained(f'{USER}/test-feature-extractor' ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(UpperCamelCase__ , getattr(UpperCamelCase__ , UpperCamelCase__ ) ) def SCREAMING_SNAKE_CASE_ ( self :str ): _a = WavaVecaFeatureExtractor.from_pretrained(UpperCamelCase__ ) feature_extractor.push_to_hub("valid_org/test-feature-extractor" , use_auth_token=self._token ) _a = WavaVecaFeatureExtractor.from_pretrained("valid_org/test-feature-extractor" ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(UpperCamelCase__ , getattr(UpperCamelCase__ , UpperCamelCase__ ) ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-feature-extractor" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained( UpperCamelCase__ , repo_id="valid_org/test-feature-extractor-org" , push_to_hub=UpperCamelCase__ , use_auth_token=self._token ) _a = WavaVecaFeatureExtractor.from_pretrained("valid_org/test-feature-extractor-org" ) for k, v in feature_extractor.__dict__.items(): self.assertEqual(UpperCamelCase__ , getattr(UpperCamelCase__ , UpperCamelCase__ ) ) def SCREAMING_SNAKE_CASE_ ( self :Dict ): CustomFeatureExtractor.register_for_auto_class() _a = CustomFeatureExtractor.from_pretrained(UpperCamelCase__ ) feature_extractor.push_to_hub("test-dynamic-feature-extractor" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual( feature_extractor.auto_map , {"AutoFeatureExtractor": "custom_feature_extraction.CustomFeatureExtractor"} , ) _a = AutoFeatureExtractor.from_pretrained( f'{USER}/test-dynamic-feature-extractor' , trust_remote_code=UpperCamelCase__ ) # Can't make an isinstance check because the new_feature_extractor is from the CustomFeatureExtractor class of a dynamic module self.assertEqual(new_feature_extractor.__class__.__name__ , "CustomFeatureExtractor" )
388
1
'''simple docstring''' import copy import tempfile import unittest from transformers import MaMaaaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.utils import cached_property from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaTokenizer from transformers.models.mam_aaa.modeling_mam_aaa import MaMaaaDecoder, MaMaaaEncoder def __a(SCREAMING_SNAKE_CASE_ : Optional[Any] , SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : int , SCREAMING_SNAKE_CASE_ : Dict=None , SCREAMING_SNAKE_CASE_ : str=None , SCREAMING_SNAKE_CASE_ : Optional[int]=None , SCREAMING_SNAKE_CASE_ : Optional[Any]=None , SCREAMING_SNAKE_CASE_ : List[str]=None , ): '''simple docstring''' if attention_mask is None: _lowerCAmelCase = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: _lowerCAmelCase = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: _lowerCAmelCase = torch.ones(config.encoder_layers , config.encoder_attention_heads , device=SCREAMING_SNAKE_CASE_ ) if decoder_head_mask is None: _lowerCAmelCase = torch.ones(config.decoder_layers , config.decoder_attention_heads , device=SCREAMING_SNAKE_CASE_ ) if cross_attn_head_mask is None: _lowerCAmelCase = torch.ones(config.decoder_layers , config.decoder_attention_heads , device=SCREAMING_SNAKE_CASE_ ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } class lowerCAmelCase_ : def __init__( self , _lowerCAmelCase , _lowerCAmelCase=13 , _lowerCAmelCase=7 , _lowerCAmelCase=True , _lowerCAmelCase=False , _lowerCAmelCase=99 , _lowerCAmelCase=16 , _lowerCAmelCase=2 , _lowerCAmelCase=4 , _lowerCAmelCase=4 , _lowerCAmelCase="relu" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.0 , _lowerCAmelCase=0.0 , _lowerCAmelCase=20 , _lowerCAmelCase=2 , _lowerCAmelCase=1 , _lowerCAmelCase=0 , ) -> Tuple: _lowerCAmelCase = parent _lowerCAmelCase = batch_size _lowerCAmelCase = seq_length _lowerCAmelCase = is_training _lowerCAmelCase = use_labels _lowerCAmelCase = vocab_size _lowerCAmelCase = hidden_size _lowerCAmelCase = num_hidden_layers _lowerCAmelCase = num_attention_heads _lowerCAmelCase = intermediate_size _lowerCAmelCase = hidden_act _lowerCAmelCase = hidden_dropout_prob _lowerCAmelCase = attention_probs_dropout_prob _lowerCAmelCase = encoder_layerdrop _lowerCAmelCase = decoder_layerdrop _lowerCAmelCase = max_position_embeddings _lowerCAmelCase = eos_token_id _lowerCAmelCase = pad_token_id _lowerCAmelCase = bos_token_id def _snake_case ( self ) -> Optional[int]: _lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCAmelCase = self.eos_token_id # Eos Token _lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for M2M100 the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input _lowerCAmelCase = input_ids.clamp(self.pad_token_id + 1 ) _lowerCAmelCase = decoder_input_ids.clamp(self.pad_token_id + 1 ) _lowerCAmelCase = self.get_config() _lowerCAmelCase = prepare_mam_aaa_inputs_dict(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return config, inputs_dict def _snake_case ( self ) -> List[Any]: return MaMaaaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , encoder_layerdrop=self.encoder_layerdrop , decoder_layerdrop=self.decoder_layerdrop , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , ) def _snake_case ( self ) -> Optional[Any]: _lowerCAmelCase , _lowerCAmelCase = self.prepare_config_and_inputs() return config, inputs_dict def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase ) -> int: _lowerCAmelCase = MaMaaaModel(config=_lowerCAmelCase ).get_decoder().to(_lowerCAmelCase ).eval() _lowerCAmelCase = inputs_dict["input_ids"] _lowerCAmelCase = inputs_dict["attention_mask"] _lowerCAmelCase = inputs_dict["head_mask"] # first forward pass _lowerCAmelCase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , head_mask=_lowerCAmelCase , use_cache=_lowerCAmelCase ) _lowerCAmelCase , _lowerCAmelCase = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids _lowerCAmelCase = ids_tensor((self.batch_size, 3) , config.vocab_size ) _lowerCAmelCase = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and _lowerCAmelCase = torch.cat([input_ids, next_tokens] , dim=-1 ) _lowerCAmelCase = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) _lowerCAmelCase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase )["last_hidden_state"] _lowerCAmelCase = model(_lowerCAmelCase , attention_mask=_lowerCAmelCase , past_key_values=_lowerCAmelCase )[ "last_hidden_state" ] # select random slice _lowerCAmelCase = ids_tensor((1,) , output_from_past.shape[-1] ).item() _lowerCAmelCase = output_from_no_past[:, -3:, random_slice_idx].detach() _lowerCAmelCase = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(_lowerCAmelCase , _lowerCAmelCase , atol=1E-2 ) ) def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase ) -> int: _lowerCAmelCase = MaMaaaModel(config=_lowerCAmelCase ).to(_lowerCAmelCase ).eval() _lowerCAmelCase = model(**_lowerCAmelCase ) _lowerCAmelCase = outputs.encoder_last_hidden_state _lowerCAmelCase = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: _lowerCAmelCase = model.get_encoder() encoder.save_pretrained(_lowerCAmelCase ) _lowerCAmelCase = MaMaaaEncoder.from_pretrained(_lowerCAmelCase ).to(_lowerCAmelCase ) _lowerCAmelCase = encoder(inputs_dict["input_ids"] , attention_mask=inputs_dict["attention_mask"] )[ 0 ] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1E-3 ) with tempfile.TemporaryDirectory() as tmpdirname: _lowerCAmelCase = model.get_decoder() decoder.save_pretrained(_lowerCAmelCase ) _lowerCAmelCase = MaMaaaDecoder.from_pretrained(_lowerCAmelCase ).to(_lowerCAmelCase ) _lowerCAmelCase = decoder( input_ids=inputs_dict["decoder_input_ids"] , attention_mask=inputs_dict["decoder_attention_mask"] , encoder_hidden_states=_lowerCAmelCase , encoder_attention_mask=inputs_dict["attention_mask"] , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1E-3 ) @require_torch class lowerCAmelCase_ ( __magic_name__ ,__magic_name__ ,__magic_name__ ,unittest.TestCase ): __lowerCamelCase : Tuple = ( ( MaMaaaModel, MaMaaaForConditionalGeneration, ) if is_torch_available() else () ) __lowerCamelCase : int = (MaMaaaForConditionalGeneration,) if is_torch_available() else () __lowerCamelCase : Union[str, Any] = ( { "conversational": MaMaaaForConditionalGeneration, "feature-extraction": MaMaaaModel, "summarization": MaMaaaForConditionalGeneration, "text2text-generation": MaMaaaForConditionalGeneration, "translation": MaMaaaForConditionalGeneration, } if is_torch_available() else {} ) __lowerCamelCase : Union[str, Any] = True __lowerCamelCase : List[str] = True __lowerCamelCase : Any = False __lowerCamelCase : Union[str, Any] = False def _snake_case ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Tuple: if pipeline_test_casse_name == "TranslationPipelineTests": # Get `ValueError: Translation requires a `src_lang` and a `tgt_lang` for this model`. # `M2M100Config` was never used in pipeline tests: cannot create a simple tokenizer. return True return False def _snake_case ( self ) -> str: _lowerCAmelCase = MaMaaaModelTester(self ) _lowerCAmelCase = ConfigTester(self , config_class=_lowerCAmelCase ) def _snake_case ( self ) -> Optional[Any]: self.config_tester.run_common_tests() def _snake_case ( self ) -> Any: _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: _lowerCAmelCase = model_class(_lowerCAmelCase ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_lowerCAmelCase ) _lowerCAmelCase , _lowerCAmelCase = model_class.from_pretrained(_lowerCAmelCase , output_loading_info=_lowerCAmelCase ) self.assertEqual(info["missing_keys"] , [] ) def _snake_case ( self ) -> List[str]: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(*_lowerCAmelCase ) def _snake_case ( self ) -> str: _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*_lowerCAmelCase ) def _snake_case ( self ) -> int: _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in (MaMaaaModel, MaMaaaForConditionalGeneration): _lowerCAmelCase = model_class(_lowerCAmelCase ) model.to(_lowerCAmelCase ) model.eval() _lowerCAmelCase = copy.deepcopy(self._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase ) ) if not self.is_encoder_decoder: _lowerCAmelCase = inputs["input_ids"] del inputs["input_ids"] else: _lowerCAmelCase = inputs["input_ids"] _lowerCAmelCase = inputs.get("decoder_input_ids" , _lowerCAmelCase ) del inputs["input_ids"] inputs.pop("decoder_input_ids" , _lowerCAmelCase ) _lowerCAmelCase = model.get_input_embeddings() if not self.is_encoder_decoder: _lowerCAmelCase = wte(_lowerCAmelCase ) else: _lowerCAmelCase = wte(_lowerCAmelCase ) _lowerCAmelCase = wte(_lowerCAmelCase ) with torch.no_grad(): model(**_lowerCAmelCase )[0] def _snake_case ( self ) -> Union[str, Any]: _lowerCAmelCase , _lowerCAmelCase = self.model_tester.prepare_config_and_inputs() _lowerCAmelCase = input_dict["input_ids"] _lowerCAmelCase = input_ids.ne(1 ).to(_lowerCAmelCase ) _lowerCAmelCase = MaMaaaForConditionalGeneration(_lowerCAmelCase ).eval().to(_lowerCAmelCase ) if torch_device == "cuda": model.half() model.generate(_lowerCAmelCase , attention_mask=_lowerCAmelCase ) model.generate(num_beams=4 , do_sample=_lowerCAmelCase , early_stopping=_lowerCAmelCase , num_return_sequences=3 ) def __a(SCREAMING_SNAKE_CASE_ : Any ): '''simple docstring''' return torch.tensor(SCREAMING_SNAKE_CASE_ , dtype=torch.long , device=SCREAMING_SNAKE_CASE_ ) _SCREAMING_SNAKE_CASE = 1e-4 @require_torch @require_sentencepiece @require_tokenizers @slow class lowerCAmelCase_ ( unittest.TestCase ): @cached_property def _snake_case ( self ) -> Dict: return MaMaaaTokenizer.from_pretrained("facebook/m2m100_418M" ) def _snake_case ( self ) -> int: _lowerCAmelCase = MaMaaaModel.from_pretrained("facebook/m2m100_418M" ).to(_lowerCAmelCase ) _lowerCAmelCase = _long_tensor([[128028, 98, 12, 30527, 2732, 159, 7755, 61904, 39144, 38, 2]] ) _lowerCAmelCase = _long_tensor([[2, 128028, 98, 12, 30527, 2732, 159, 7755, 61904, 39144, 38]] ) _lowerCAmelCase = prepare_mam_aaa_inputs_dict(model.config , _lowerCAmelCase , _lowerCAmelCase ) with torch.no_grad(): _lowerCAmelCase = model(**_lowerCAmelCase )[0] _lowerCAmelCase = torch.Size((1, 11, 1024) ) self.assertEqual(output.shape , _lowerCAmelCase ) # change to expected output here _lowerCAmelCase = torch.tensor( [[-0.7780, -0.1676, 0.1038], [-6.7556, -1.3992, 0.0567], [-7.5383, -0.5920, -0.2779]] , device=_lowerCAmelCase ) self.assertTrue(torch.allclose(output[:, :3, :3] , _lowerCAmelCase , atol=_lowerCAmelCase ) ) def _snake_case ( self ) -> Dict: _lowerCAmelCase = MaMaaaForConditionalGeneration.from_pretrained("facebook/m2m100_418M" ).to(_lowerCAmelCase ) # change to intended input _lowerCAmelCase = _long_tensor([[128028, 98, 12, 30527, 2732, 159, 7755, 61904, 39144, 38, 2]] ) _lowerCAmelCase = _long_tensor([[2, 128028, 98, 12, 30527, 2732, 159, 7755, 61904, 39144, 38]] ) _lowerCAmelCase = prepare_mam_aaa_inputs_dict(model.config , _lowerCAmelCase , _lowerCAmelCase ) with torch.no_grad(): _lowerCAmelCase = model(**_lowerCAmelCase )[0] _lowerCAmelCase = torch.Size((1, 11, model.config.vocab_size) ) self.assertEqual(output.shape , _lowerCAmelCase ) # change to expected output here _lowerCAmelCase = torch.tensor( [[-1.0448, -1.0411, 3.7992], [-3.2191, -3.2386, -1.3451], [-3.6210, -3.5993, 0.4925]] , device=_lowerCAmelCase ) self.assertTrue(torch.allclose(output[:, :3, :3] , _lowerCAmelCase , atol=_lowerCAmelCase ) ) def _snake_case ( self ) -> Tuple: _lowerCAmelCase = MaMaaaForConditionalGeneration.from_pretrained("facebook/m2m100_418M" ).to(_lowerCAmelCase ) _lowerCAmelCase = MaMaaaTokenizer.from_pretrained("facebook/m2m100_418M" , src_lang="fr" , tgt_lang="en" ) _lowerCAmelCase = [ "L'affaire NSA souligne l'absence totale de débat sur le renseignement", "Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.", "Lorsque François Hollande téléphone à Barack Obama ou quand le ministre des affaires étrangères Laurent" " Fabius convoque l'ambassadeur des Etats-Unis, ils réagissent à une vraie découverte, qui est celle de" " l'ampleur de la surveillance américaine sur l'ensemble des communications en France.", ] # The below article tests that we don't add any hypotheses outside of the top n_beams _lowerCAmelCase = tokenizer(_lowerCAmelCase , padding=_lowerCAmelCase , return_tensors="pt" ) _lowerCAmelCase = model.generate( input_ids=dct["input_ids"].to(_lowerCAmelCase ) , attention_mask=dct["attention_mask"].to(_lowerCAmelCase ) , num_beams=5 , forced_bos_token_id=tokenizer.get_lang_id("en" ) , ) _lowerCAmelCase = [ "The NSA case highlights the total absence of intelligence debate", "I think there are two levels of response from the French government.", "When François Hollande calls Barack Obama or when Foreign Minister Laurent Fabius calls the U.S." " Ambassador, they respond to a real discovery, which is that of the scale of U.S. surveillance on all" " communications in France.", ] _lowerCAmelCase = tokenizer.batch_decode( hypotheses_batch.tolist() , clean_up_tokenization_spaces=_lowerCAmelCase , skip_special_tokens=_lowerCAmelCase ) assert generated == expected_en
489
'''simple docstring''' def __a(SCREAMING_SNAKE_CASE_ : int ): '''simple docstring''' if not isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): raise TypeError("Input value must be an 'int' type" ) _lowerCAmelCase = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
489
1
'''simple docstring''' import argparse import torch from huggingface_hub import hf_hub_download from transformers import AutoTokenizer, RobertaPreLayerNormConfig, RobertaPreLayerNormForMaskedLM from transformers.utils import logging logging.set_verbosity_info() UpperCAmelCase__ : str = logging.get_logger(__name__) def A ( UpperCamelCase_ : str , UpperCamelCase_ : str ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase__ = RobertaPreLayerNormConfig.from_pretrained( UpperCamelCase_ , architectures=["RobertaPreLayerNormForMaskedLM"] ) # convert state_dict lowerCAmelCase__ = torch.load(hf_hub_download(repo_id=UpperCamelCase_ , filename="pytorch_model.bin" ) ) lowerCAmelCase__ = {} for tensor_key, tensor_value in original_state_dict.items(): # The transformer implementation gives the model a unique name, rather than overwiriting 'roberta' if tensor_key.startswith("roberta." ): lowerCAmelCase__ = "roberta_prelayernorm." + tensor_key[len("roberta." ) :] # The original implementation contains weights which are not used, remove them from the state_dict if tensor_key.endswith(".self.LayerNorm.weight" ) or tensor_key.endswith(".self.LayerNorm.bias" ): continue lowerCAmelCase__ = tensor_value lowerCAmelCase__ = RobertaPreLayerNormForMaskedLM.from_pretrained( pretrained_model_name_or_path=UpperCamelCase_ , config=UpperCamelCase_ , state_dict=UpperCamelCase_ ) model.save_pretrained(UpperCamelCase_ ) # convert tokenizer lowerCAmelCase__ = AutoTokenizer.from_pretrained(UpperCamelCase_ ) tokenizer.save_pretrained(UpperCamelCase_ ) if __name__ == "__main__": UpperCAmelCase__ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint-repo", default=None, type=str, required=True, help="Path the official PyTorch dump, e.g. 'andreasmadsen/efficient_mlm_m0.40'.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) UpperCAmelCase__ : Tuple = parser.parse_args() convert_roberta_prelayernorm_checkpoint_to_pytorch(args.checkpoint_repo, args.pytorch_dump_folder_path)
48
'''simple docstring''' import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class __magic_name__ ( __SCREAMING_SNAKE_CASE ): UpperCamelCase__ = ['image_processor', 'tokenizer'] UpperCamelCase__ = 'BlipImageProcessor' UpperCamelCase__ = 'AutoTokenizer' def __init__( self , snake_case_ , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) # add QFormer tokenizer lowercase =qformer_tokenizer def __call__( self , snake_case_ = None , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): if images is None and text is None: raise ValueError('''You have to specify at least images or text.''' ) lowercase =BatchFeature() if text is not None: lowercase =self.tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) encoding.update(snake_case_ ) lowercase =self.qformer_tokenizer( text=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_token_type_ids=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) lowercase =qformer_text_encoding.pop('''input_ids''' ) lowercase =qformer_text_encoding.pop('''attention_mask''' ) if images is not None: lowercase =self.image_processor(snake_case_ , return_tensors=snake_case_ ) encoding.update(snake_case_ ) return encoding def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def _A( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _A( self ): lowercase =self.tokenizer.model_input_names lowercase =self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _A( self , snake_case_ , **snake_case_ ): if os.path.isfile(snake_case_ ): raise ValueError(f'Provided path ({save_directory}) should be a directory, not a file' ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) lowercase =os.path.join(snake_case_ , '''qformer_tokenizer''' ) self.qformer_tokenizer.save_pretrained(snake_case_ ) return super().save_pretrained(snake_case_ , **snake_case_ ) @classmethod def _A( cls , snake_case_ , **snake_case_ ): lowercase =AutoTokenizer.from_pretrained(snake_case_ , subfolder='''qformer_tokenizer''' ) lowercase =cls._get_arguments_from_pretrained(snake_case_ , **snake_case_ ) args.append(snake_case_ ) return cls(*snake_case_ )
72
0
import unicodedata from dataclasses import dataclass from typing import Optional, Union import numpy as np from transformers.data.data_collator import DataCollatorMixin from transformers.file_utils import PaddingStrategy from transformers.tokenization_utils_base import PreTrainedTokenizerBase def UpperCamelCase__ ( lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ): if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ): lowercase = np.full((len(lowerCAmelCase__ ), sequence_length, 2) ,lowerCAmelCase__ ) else: lowercase = np.full((len(lowerCAmelCase__ ), sequence_length) ,lowerCAmelCase__ ) for i, tensor in enumerate(lowerCAmelCase__ ): if padding_side == "right": if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ): lowercase = tensor[:sequence_length] else: lowercase = tensor[:sequence_length] else: if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ): lowercase = tensor[:sequence_length] else: lowercase = tensor[:sequence_length] return out_tensor.tolist() def UpperCamelCase__ ( lowerCAmelCase__ ): lowercase = ord(lowerCAmelCase__ ) if (cp >= 33 and cp <= 47) or (cp >= 58 and cp <= 64) or (cp >= 91 and cp <= 96) or (cp >= 123 and cp <= 126): return True lowercase = unicodedata.category(lowerCAmelCase__ ) if cat.startswith("""P""" ): return True return False @dataclass class A_ ( __a ): _A :PreTrainedTokenizerBase _A :Union[bool, str, PaddingStrategy] = True _A :Optional[int] = None _A :Optional[int] = None _A :int = -100 _A :str = "pt" def SCREAMING_SNAKE_CASE__ ( self : Tuple , snake_case__ : List[str] ): import torch lowercase = """label""" if """label""" in features[0].keys() else """labels""" lowercase = [feature[label_name] for feature in features] if label_name in features[0].keys() else None lowercase = self.tokenizer.pad( snake_case__ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors="""pt""" if labels is None else None , ) if labels is None: return batch lowercase = torch.tensor(batch["""entity_ids"""] ).shape[1] lowercase = self.tokenizer.padding_side if padding_side == "right": lowercase = [ list(snake_case__ ) + [self.label_pad_token_id] * (sequence_length - len(snake_case__ )) for label in labels ] else: lowercase = [ [self.label_pad_token_id] * (sequence_length - len(snake_case__ )) + list(snake_case__ ) for label in labels ] lowercase = [feature["""ner_tags"""] for feature in features] lowercase = padding_tensor(snake_case__ , -1 , snake_case__ , snake_case__ ) lowercase = [feature["""original_entity_spans"""] for feature in features] lowercase = padding_tensor(snake_case__ , (-1, -1) , snake_case__ , snake_case__ ) lowercase = {k: torch.tensor(snake_case__ , dtype=torch.intaa ) for k, v in batch.items()} return batch
72
import math from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : str =logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : str ={ '''facebook/data2vec-base-960h''': '''https://huggingface.co/facebook/data2vec-audio-base-960h/resolve/main/config.json''', # See all Data2VecAudio models at https://huggingface.co/models?filter=data2vec-audio } class A_ ( __a ): _A :Tuple = '''data2vec-audio''' def __init__( self : Optional[Any] , snake_case__ : List[Any]=32 , snake_case__ : List[Any]=7_68 , snake_case__ : int=12 , snake_case__ : Dict=12 , snake_case__ : List[str]=30_72 , snake_case__ : List[str]="gelu" , snake_case__ : Optional[int]=0.1 , snake_case__ : List[Any]=0.1 , snake_case__ : int=0.1 , snake_case__ : Tuple=0.0 , snake_case__ : Tuple=0.1 , snake_case__ : Any=0.1 , snake_case__ : Dict=0.02 , snake_case__ : List[str]=1E-5 , snake_case__ : Optional[Any]="gelu" , snake_case__ : Union[str, Any]=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , snake_case__ : List[str]=(5, 2, 2, 2, 2, 2, 2) , snake_case__ : str=(10, 3, 3, 3, 3, 2, 2) , snake_case__ : Any=False , snake_case__ : List[str]=16 , snake_case__ : Any=19 , snake_case__ : Optional[Any]=5 , snake_case__ : str=0.05 , snake_case__ : Tuple=10 , snake_case__ : Optional[Any]=2 , snake_case__ : Dict=0.0 , snake_case__ : int=10 , snake_case__ : Any=0 , snake_case__ : int="sum" , snake_case__ : str=False , snake_case__ : str=False , snake_case__ : Optional[int]=2_56 , snake_case__ : List[str]=(5_12, 5_12, 5_12, 5_12, 15_00) , snake_case__ : List[str]=(5, 3, 3, 1, 1) , snake_case__ : int=(1, 2, 3, 1, 1) , snake_case__ : Optional[Any]=5_12 , snake_case__ : Dict=0 , snake_case__ : Optional[Any]=1 , snake_case__ : Tuple=2 , snake_case__ : Tuple=False , snake_case__ : List[str]=3 , snake_case__ : List[str]=2 , snake_case__ : Tuple=3 , snake_case__ : List[str]=None , **snake_case__ : str , ): super().__init__(**snake_case__ , pad_token_id=snake_case__ , bos_token_id=snake_case__ , eos_token_id=snake_case__ ) lowercase = hidden_size lowercase = feat_extract_activation lowercase = list(snake_case__ ) lowercase = list(snake_case__ ) lowercase = list(snake_case__ ) lowercase = conv_bias lowercase = num_conv_pos_embeddings lowercase = num_conv_pos_embedding_groups lowercase = conv_pos_kernel_size lowercase = len(self.conv_dim ) lowercase = num_hidden_layers lowercase = intermediate_size lowercase = hidden_act lowercase = num_attention_heads lowercase = hidden_dropout lowercase = attention_dropout lowercase = activation_dropout lowercase = feat_proj_dropout lowercase = final_dropout lowercase = layerdrop lowercase = layer_norm_eps lowercase = initializer_range lowercase = vocab_size lowercase = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" F""" {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,""" F""" `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowercase = mask_time_prob lowercase = mask_time_length lowercase = mask_time_min_masks lowercase = mask_feature_prob lowercase = mask_feature_length lowercase = mask_feature_min_masks # ctc loss lowercase = ctc_loss_reduction lowercase = ctc_zero_infinity # adapter lowercase = add_adapter lowercase = adapter_kernel_size lowercase = adapter_stride lowercase = num_adapter_layers lowercase = output_hidden_size or hidden_size # SequenceClassification-specific parameter. Feel free to ignore for other classes. lowercase = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. lowercase = list(snake_case__ ) lowercase = list(snake_case__ ) lowercase = list(snake_case__ ) lowercase = xvector_output_dim @property def SCREAMING_SNAKE_CASE__ ( self : Dict ): return math.prod(self.conv_stride )
72
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCamelCase : Any = { 'configuration_altclip': [ 'ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'AltCLIPConfig', 'AltCLIPTextConfig', 'AltCLIPVisionConfig', ], 'processing_altclip': ['AltCLIPProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase : List[str] = [ 'ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST', 'AltCLIPPreTrainedModel', 'AltCLIPModel', 'AltCLIPTextModel', 'AltCLIPVisionModel', ] if TYPE_CHECKING: from .configuration_altclip import ( ALTCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, AltCLIPConfig, AltCLIPTextConfig, AltCLIPVisionConfig, ) from .processing_altclip import AltCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_altclip import ( ALTCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, AltCLIPModel, AltCLIPPreTrainedModel, AltCLIPTextModel, AltCLIPVisionModel, ) else: import sys UpperCamelCase : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
50
'''simple docstring''' from PIL import Image def A__ ( __lowerCAmelCase : Image , __lowerCAmelCase : float ): def brightness(__lowerCAmelCase : int ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError("""level must be between -255.0 (black) and 255.0 (white)""" ) return img.point(__lowerCAmelCase ) if __name__ == "__main__": # Load image with Image.open('image_data/lena.jpg') as img: # Change brightness to 100 UpperCamelCase : Union[str, Any] = change_brightness(img, 1_00) brigt_img.save('image_data/lena_brightness.png', format='png')
50
1
'''simple docstring''' from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging _lowerCamelCase : str = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" _SCREAMING_SNAKE_CASE = ["""audio_values""", """audio_mask"""] def __init__( self : Dict , UpperCamelCase__ : Optional[Any]=2_0_4_8 , UpperCamelCase__ : Dict=1 , UpperCamelCase__ : List[str]=[1_6, 1_6] , UpperCamelCase__ : Union[str, Any]=1_2_8 , UpperCamelCase__ : int=4_4_1_0_0 , UpperCamelCase__ : int=8_6 , UpperCamelCase__ : Optional[Any]=2_0_4_8 , UpperCamelCase__ : int=0.0 , **UpperCamelCase__ : Dict , ): """simple docstring""" super().__init__( feature_size=UpperCamelCase__ , sampling_rate=UpperCamelCase__ , padding_value=UpperCamelCase__ , **UpperCamelCase__ , ) UpperCamelCase = spectrogram_length UpperCamelCase = num_channels UpperCamelCase = patch_size UpperCamelCase = feature_size // self.patch_size[1] UpperCamelCase = n_fft UpperCamelCase = sampling_rate // hop_length_to_sampling_rate UpperCamelCase = sampling_rate UpperCamelCase = padding_value UpperCamelCase = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=UpperCamelCase__ , min_frequency=0.0 , max_frequency=2_2_0_5_0.0 , sampling_rate=UpperCamelCase__ , norm='slaney' , mel_scale='slaney' , ).T def A ( self : Tuple , UpperCamelCase__ : np.array ): """simple docstring""" UpperCamelCase = spectrogram( UpperCamelCase__ , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='dB' , db_range=8_0.0 , ) UpperCamelCase = log_spec[:, :-1] UpperCamelCase = log_spec - 2_0.0 UpperCamelCase = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : List[Any] , UpperCamelCase__ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , UpperCamelCase__ : Optional[Union[str, TensorType]] = None , UpperCamelCase__ : Optional[bool] = True , UpperCamelCase__ : Optional[int] = None , UpperCamelCase__ : bool = False , UpperCamelCase__ : bool = False , **UpperCamelCase__ : Tuple , ): """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( 'This feature extractor is set to support sampling rate' f""" of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled""" f""" with {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) UpperCamelCase = isinstance(UpperCamelCase__ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"""Only mono-channel audio is supported for input to {self}""" ) UpperCamelCase = is_batched_numpy or ( isinstance(UpperCamelCase__ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCamelCase = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(UpperCamelCase__ , np.ndarray ): UpperCamelCase = np.asarray(UpperCamelCase__ , dtype=np.floataa ) elif isinstance(UpperCamelCase__ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCamelCase = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCamelCase = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis UpperCamelCase = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , UpperCamelCase__ ): UpperCamelCase = [np.asarray(UpperCamelCase__ , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask UpperCamelCase = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: UpperCamelCase = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] UpperCamelCase = np.array(UpperCamelCase__ ).astype(np.floataa ) # convert into correct format for padding UpperCamelCase = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch UpperCamelCase = np.ones([len(UpperCamelCase__ ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) UpperCamelCase = padded_audio_features * self.padding_value for i in range(len(UpperCamelCase__ ) ): UpperCamelCase = audio_features[i] UpperCamelCase = feature # return as BatchFeature if return_attention_mask: UpperCamelCase = {'audio_values': padded_audio_features, 'audio_mask': audio_mask} else: UpperCamelCase = {'audio_values': padded_audio_features} UpperCamelCase = BatchFeature(data=UpperCamelCase__ , tensor_type=UpperCamelCase__ ) return encoded_inputs
324
'''simple docstring''' import ast import os import re import shutil import tempfile import unittest from unittest import mock import torch from accelerate.test_utils.examples import compare_against_test from accelerate.test_utils.testing import TempDirTestCase, require_trackers, run_command, slow from accelerate.utils import write_basic_config # DataLoaders built from `test_samples/MRPC` for quick testing # Should mock `{script_name}.get_dataloaders` via: # @mock.patch("{script_name}.get_dataloaders", mocked_dataloaders) _lowerCamelCase : List[str] = [ "cross_validation.py", "gradient_accumulation.py", "local_sgd.py", "multi_process_metrics.py", "memory.py", "automatic_gradient_accumulation.py", "fsdp_with_peak_mem_tracking.py", "deepspeed_with_config_support.py", "megatron_lm_gpt_pretraining.py", ] class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def A ( self : Tuple , UpperCamelCase__ : str , UpperCamelCase__ : bool , UpperCamelCase__ : str = None , UpperCamelCase__ : list = None ): """simple docstring""" UpperCamelCase = None UpperCamelCase = os.path.abspath(os.path.join('examples' , 'by_feature' ) ) UpperCamelCase = os.path.abspath('examples' ) for item in os.listdir(UpperCamelCase__ ): if item not in EXCLUDE_EXAMPLES: UpperCamelCase = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) if os.path.isfile(UpperCamelCase__ ) and ".py" in item_path: with self.subTest( tested_script=UpperCamelCase__ , feature_script=UpperCamelCase__ , tested_section='main()' if parser_only else 'training_function()' , ): UpperCamelCase = compare_against_test( os.path.join(UpperCamelCase__ , UpperCamelCase__ ) , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) UpperCamelCase = '\n'.join(UpperCamelCase__ ) if special_strings is not None: for string in special_strings: UpperCamelCase = diff.replace(UpperCamelCase__ , '' ) self.assertEqual(UpperCamelCase__ , '' ) def A ( self : Optional[Any] ): """simple docstring""" self.one_complete_example('complete_nlp_example.py' , UpperCamelCase__ ) self.one_complete_example('complete_nlp_example.py' , UpperCamelCase__ ) def A ( self : Dict ): """simple docstring""" UpperCamelCase = os.path.abspath(os.path.join('examples' , 'cv_example.py' ) ) UpperCamelCase = [ ' ' * 1_6 + '{\n\n', ' ' * 2_0 + '"accuracy": eval_metric["accuracy"],\n\n', ' ' * 2_0 + '"f1": eval_metric["f1"],\n\n', ' ' * 2_0 + '"train_loss": total_loss.item() / len(train_dataloader),\n\n', ' ' * 2_0 + '"epoch": epoch,\n\n', ' ' * 1_6 + '},\n\n', ' ' * 1_6 + 'step=epoch,\n', ' ' * 1_2, ' ' * 8 + 'for step, batch in enumerate(active_dataloader):\n', ] self.one_complete_example('complete_cv_example.py' , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) self.one_complete_example('complete_cv_example.py' , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) @mock.patch.dict(os.environ , {"""TESTING_MOCKED_DATALOADERS""": """1"""} ) class SCREAMING_SNAKE_CASE ( _a ): """simple docstring""" _SCREAMING_SNAKE_CASE = False @classmethod def A ( cls : Union[str, Any] ): """simple docstring""" super().setUpClass() UpperCamelCase = tempfile.mkdtemp() UpperCamelCase = os.path.join(cls._tmpdir , 'default_config.yml' ) write_basic_config(save_location=cls.configPath ) UpperCamelCase = ['accelerate', 'launch', '--config_file', cls.configPath] @classmethod def A ( cls : List[Any] ): """simple docstring""" super().tearDownClass() shutil.rmtree(cls._tmpdir ) def A ( self : int ): """simple docstring""" UpperCamelCase = f""" examples/by_feature/checkpointing.py --checkpointing_steps epoch --output_dir {self.tmpdir} """.split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , 'epoch_0' ) ) ) def A ( self : Optional[Any] ): """simple docstring""" UpperCamelCase = f""" examples/by_feature/checkpointing.py --checkpointing_steps 1 --output_dir {self.tmpdir} """.split() UpperCamelCase = run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(self.tmpdir , 'step_2' ) ) ) def A ( self : Optional[Any] ): """simple docstring""" UpperCamelCase = f""" examples/by_feature/checkpointing.py --resume_from_checkpoint {os.path.join(self.tmpdir , 'epoch_0' )} """.split() UpperCamelCase = run_command(self._launch_args + testargs , return_stdout=UpperCamelCase__ ) self.assertNotIn('epoch 0:' , UpperCamelCase__ ) self.assertIn('epoch 1:' , UpperCamelCase__ ) def A ( self : Optional[Any] ): """simple docstring""" UpperCamelCase = f""" examples/by_feature/checkpointing.py --resume_from_checkpoint {os.path.join(self.tmpdir , 'step_2' )} """.split() UpperCamelCase = run_command(self._launch_args + testargs , return_stdout=UpperCamelCase__ ) if torch.cuda.is_available(): UpperCamelCase = torch.cuda.device_count() else: UpperCamelCase = 1 if num_processes > 1: self.assertNotIn('epoch 0:' , UpperCamelCase__ ) self.assertIn('epoch 1:' , UpperCamelCase__ ) else: self.assertIn('epoch 0:' , UpperCamelCase__ ) self.assertIn('epoch 1:' , UpperCamelCase__ ) @slow def A ( self : Optional[Any] ): """simple docstring""" UpperCamelCase = '\n examples/by_feature/cross_validation.py\n --num_folds 2\n '.split() with mock.patch.dict(os.environ , {'TESTING_MOCKED_DATALOADERS': '0'} ): UpperCamelCase = run_command(self._launch_args + testargs , return_stdout=UpperCamelCase__ ) UpperCamelCase = re.findall('({.+})' , UpperCamelCase__ ) UpperCamelCase = [r for r in results if 'accuracy' in r][-1] UpperCamelCase = ast.literal_eval(UpperCamelCase__ ) self.assertGreaterEqual(results['accuracy'] , 0.7_5 ) def A ( self : Optional[int] ): """simple docstring""" UpperCamelCase = ['examples/by_feature/multi_process_metrics.py'] run_command(self._launch_args + testargs ) @require_trackers @mock.patch.dict(os.environ , {'WANDB_MODE': 'offline'} ) def A ( self : Dict ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdir: UpperCamelCase = f""" examples/by_feature/tracking.py --with_tracking --project_dir {tmpdir} """.split() run_command(self._launch_args + testargs ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , 'tracking' ) ) ) def A ( self : Any ): """simple docstring""" UpperCamelCase = ['examples/by_feature/gradient_accumulation.py'] run_command(self._launch_args + testargs ) def A ( self : List[Any] ): """simple docstring""" UpperCamelCase = ['examples/by_feature/local_sgd.py'] run_command(self._launch_args + testargs )
324
1
import json import logging import os import sys from time import time from unittest.mock import patch from transformers.testing_utils import TestCasePlus, require_torch_tpu logging.basicConfig(level=logging.DEBUG) UpperCamelCase = logging.getLogger() def lowerCamelCase_ ( _lowercase ) -> str: __A : Union[str, Any] = {} __A : List[Any] = os.path.join(_lowercase , "all_results.json" ) if os.path.exists(_lowercase ): with open(_lowercase , "r" ) as f: __A : Any = json.load(_lowercase ) else: raise ValueError(F"can't find {path}" ) return results UpperCamelCase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) @require_torch_tpu class _a ( lowerCAmelCase__ ): '''simple docstring''' def __UpperCAmelCase( self ): import xla_spawn __A : Tuple = self.get_auto_remove_tmp_dir() __A : Optional[Any] = F"\n ./examples/pytorch/text-classification/run_glue.py\n --num_cores=8\n ./examples/pytorch/text-classification/run_glue.py\n --model_name_or_path distilbert-base-uncased\n --output_dir {tmp_dir}\n --overwrite_output_dir\n --train_file ./tests/fixtures/tests_samples/MRPC/train.csv\n --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv\n --do_train\n --do_eval\n --debug tpu_metrics_debug\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --learning_rate=1e-4\n --max_steps=10\n --warmup_steps=2\n --seed=42\n --max_seq_length=128\n ".split() with patch.object(__UpperCAmelCase , "argv" , __UpperCAmelCase ): __A : str = time() xla_spawn.main() __A : List[str] = time() __A : Optional[Any] = get_results(__UpperCAmelCase ) self.assertGreaterEqual(result["eval_accuracy"] , 0.75 ) # Assert that the script takes less than 500 seconds to make sure it doesn't hang. self.assertLess(end - start , 500 ) def __UpperCAmelCase( self ): import xla_spawn __A : List[str] = "\n ./tests/test_trainer_tpu.py\n --num_cores=8\n ./tests/test_trainer_tpu.py\n ".split() with patch.object(__UpperCAmelCase , "argv" , __UpperCAmelCase ): xla_spawn.main()
520
from math import isqrt, loga def lowerCamelCase_ ( _lowercase ) -> list[int]: __A : Dict = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , _lowercase , _lowercase ): __A : List[str] = False return [i for i in range(2 , _lowercase ) if is_prime[i]] def lowerCamelCase_ ( _lowercase = 800_800 , _lowercase = 800_800 ) -> int: __A : List[str] = degree * loga(_lowercase ) __A : List[Any] = int(_lowercase ) __A : Optional[Any] = calculate_prime_numbers(_lowercase ) __A : Optional[int] = 0 __A : Optional[int] = 0 __A : List[Any] = len(_lowercase ) - 1 while left < right: while ( prime_numbers[right] * loga(prime_numbers[left] ) + prime_numbers[left] * loga(prime_numbers[right] ) > upper_bound ): right -= 1 hybrid_integers_count += right - left left += 1 return hybrid_integers_count if __name__ == "__main__": print(F'''{solution() = }''')
520
1
import torch from diffusers import EulerDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class a_( _UpperCAmelCase ): """simple docstring""" __snake_case : Optional[int] =(EulerDiscreteScheduler,) __snake_case : str =10 def __UpperCamelCase ( self : Tuple , **lowerCAmelCase__ : Tuple) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE = { 'num_train_timesteps': 1_1_0_0, 'beta_start': 0.00_01, 'beta_end': 0.02, 'beta_schedule': 'linear', } config.update(**A_) return config def __UpperCamelCase ( self : List[Any]) -> Dict: """simple docstring""" for timesteps in [1_0, 5_0, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=A_) def __UpperCamelCase ( self : Union[str, Any]) -> Optional[Any]: """simple docstring""" for beta_start, beta_end in zip([0.0_00_01, 0.00_01, 0.0_01] , [0.00_02, 0.0_02, 0.02]): self.check_over_configs(beta_start=A_ , beta_end=A_) def __UpperCamelCase ( self : List[Any]) -> Tuple: """simple docstring""" for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=A_) def __UpperCamelCase ( self : int) -> Tuple: """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A_) def __UpperCamelCase ( self : List[str]) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE = self.scheduler_classes[0] SCREAMING_SNAKE_CASE = self.get_scheduler_config() SCREAMING_SNAKE_CASE = scheduler_class(**A_) scheduler.set_timesteps(self.num_inference_steps) SCREAMING_SNAKE_CASE = torch.manual_seed(0) SCREAMING_SNAKE_CASE = self.dummy_model() SCREAMING_SNAKE_CASE = self.dummy_sample_deter * scheduler.init_noise_sigma SCREAMING_SNAKE_CASE = sample.to(A_) for i, t in enumerate(scheduler.timesteps): SCREAMING_SNAKE_CASE = scheduler.scale_model_input(A_ , A_) SCREAMING_SNAKE_CASE = model(A_ , A_) SCREAMING_SNAKE_CASE = scheduler.step(A_ , A_ , A_ , generator=A_) SCREAMING_SNAKE_CASE = output.prev_sample SCREAMING_SNAKE_CASE = torch.sum(torch.abs(A_)) SCREAMING_SNAKE_CASE = torch.mean(torch.abs(A_)) assert abs(result_sum.item() - 10.08_07) < 1e-2 assert abs(result_mean.item() - 0.01_31) < 1e-3 def __UpperCamelCase ( self : List[str]) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE = self.scheduler_classes[0] SCREAMING_SNAKE_CASE = self.get_scheduler_config(prediction_type='v_prediction') SCREAMING_SNAKE_CASE = scheduler_class(**A_) scheduler.set_timesteps(self.num_inference_steps) SCREAMING_SNAKE_CASE = torch.manual_seed(0) SCREAMING_SNAKE_CASE = self.dummy_model() SCREAMING_SNAKE_CASE = self.dummy_sample_deter * scheduler.init_noise_sigma SCREAMING_SNAKE_CASE = sample.to(A_) for i, t in enumerate(scheduler.timesteps): SCREAMING_SNAKE_CASE = scheduler.scale_model_input(A_ , A_) SCREAMING_SNAKE_CASE = model(A_ , A_) SCREAMING_SNAKE_CASE = scheduler.step(A_ , A_ , A_ , generator=A_) SCREAMING_SNAKE_CASE = output.prev_sample SCREAMING_SNAKE_CASE = torch.sum(torch.abs(A_)) SCREAMING_SNAKE_CASE = torch.mean(torch.abs(A_)) assert abs(result_sum.item() - 0.00_02) < 1e-2 assert abs(result_mean.item() - 2.2_676e-06) < 1e-3 def __UpperCamelCase ( self : str) -> int: """simple docstring""" SCREAMING_SNAKE_CASE = self.scheduler_classes[0] SCREAMING_SNAKE_CASE = self.get_scheduler_config() SCREAMING_SNAKE_CASE = scheduler_class(**A_) scheduler.set_timesteps(self.num_inference_steps , device=A_) SCREAMING_SNAKE_CASE = torch.manual_seed(0) SCREAMING_SNAKE_CASE = self.dummy_model() SCREAMING_SNAKE_CASE = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() SCREAMING_SNAKE_CASE = sample.to(A_) for t in scheduler.timesteps: SCREAMING_SNAKE_CASE = scheduler.scale_model_input(A_ , A_) SCREAMING_SNAKE_CASE = model(A_ , A_) SCREAMING_SNAKE_CASE = scheduler.step(A_ , A_ , A_ , generator=A_) SCREAMING_SNAKE_CASE = output.prev_sample SCREAMING_SNAKE_CASE = torch.sum(torch.abs(A_)) SCREAMING_SNAKE_CASE = torch.mean(torch.abs(A_)) assert abs(result_sum.item() - 10.08_07) < 1e-2 assert abs(result_mean.item() - 0.01_31) < 1e-3 def __UpperCamelCase ( self : Optional[int]) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE = self.scheduler_classes[0] SCREAMING_SNAKE_CASE = self.get_scheduler_config() SCREAMING_SNAKE_CASE = scheduler_class(**A_ , use_karras_sigmas=A_) scheduler.set_timesteps(self.num_inference_steps , device=A_) SCREAMING_SNAKE_CASE = torch.manual_seed(0) SCREAMING_SNAKE_CASE = self.dummy_model() SCREAMING_SNAKE_CASE = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() SCREAMING_SNAKE_CASE = sample.to(A_) for t in scheduler.timesteps: SCREAMING_SNAKE_CASE = scheduler.scale_model_input(A_ , A_) SCREAMING_SNAKE_CASE = model(A_ , A_) SCREAMING_SNAKE_CASE = scheduler.step(A_ , A_ , A_ , generator=A_) SCREAMING_SNAKE_CASE = output.prev_sample SCREAMING_SNAKE_CASE = torch.sum(torch.abs(A_)) SCREAMING_SNAKE_CASE = torch.mean(torch.abs(A_)) assert abs(result_sum.item() - 1_2_4.5_2_2_9_9_4_9_9_5_1_1_7_1_9) < 1e-2 assert abs(result_mean.item() - 0.1_62_13_93_26_33_39_99_63) < 1e-3
705
import html from ...feature_extraction_utils import BatchFeature, FeatureExtractionMixin from ...utils import is_bsa_available, logging, requires_backends if is_bsa_available(): import bsa from bsa import BeautifulSoup __UpperCAmelCase = logging.get_logger(__name__) class a_( lowercase__ ): """simple docstring""" def __init__( self : Any , **lowerCAmelCase__ : Tuple) -> str: """simple docstring""" requires_backends(self , ['bs4']) super().__init__(**lowerCAmelCase__) def __UpperCamelCase ( self : str , lowerCAmelCase__ : Tuple) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = element if element.name else element.parent for parent in child.parents: # type: bs4.element.Tag SCREAMING_SNAKE_CASE = parent.find_all(child.name , recursive=lowerCAmelCase__) xpath_tags.append(child.name) xpath_subscripts.append( 0 if 1 == len(lowerCAmelCase__) else next(i for i, s in enumerate(lowerCAmelCase__ , 1) if s is child)) SCREAMING_SNAKE_CASE = parent xpath_tags.reverse() xpath_subscripts.reverse() return xpath_tags, xpath_subscripts def __UpperCamelCase ( self : Union[str, Any] , lowerCAmelCase__ : Union[str, Any]) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE = BeautifulSoup(lowerCAmelCase__ , 'html.parser') SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = [] for element in html_code.descendants: if type(lowerCAmelCase__) == bsa.element.NavigableString: if type(element.parent) != bsa.element.Tag: continue SCREAMING_SNAKE_CASE = html.unescape(lowerCAmelCase__).strip() if not text_in_this_tag: continue all_doc_strings.append(lowerCAmelCase__) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self.xpath_soup(lowerCAmelCase__) stringaxtag_seq.append(lowerCAmelCase__) stringaxsubs_seq.append(lowerCAmelCase__) if len(lowerCAmelCase__) != len(lowerCAmelCase__): raise ValueError('Number of doc strings and xtags does not correspond') if len(lowerCAmelCase__) != len(lowerCAmelCase__): raise ValueError('Number of doc strings and xsubs does not correspond') return all_doc_strings, stringaxtag_seq, stringaxsubs_seq def __UpperCamelCase ( self : Union[str, Any] , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Dict) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE = '' for tagname, subs in zip(lowerCAmelCase__ , lowerCAmelCase__): xpath += f'''/{tagname}''' if subs != 0: xpath += f'''[{subs}]''' return xpath def __call__( self : List[Any] , lowerCAmelCase__ : List[Any]) -> BatchFeature: """simple docstring""" SCREAMING_SNAKE_CASE = False # Check that strings has a valid type if isinstance(lowerCAmelCase__ , lowerCAmelCase__): SCREAMING_SNAKE_CASE = True elif isinstance(lowerCAmelCase__ , (list, tuple)): if len(lowerCAmelCase__) == 0 or isinstance(html_strings[0] , lowerCAmelCase__): SCREAMING_SNAKE_CASE = True if not valid_strings: raise ValueError( 'HTML strings must of type `str`, `List[str]` (batch of examples), ' f'''but is of type {type(lowerCAmelCase__)}.''') SCREAMING_SNAKE_CASE = bool(isinstance(lowerCAmelCase__ , (list, tuple)) and (isinstance(html_strings[0] , lowerCAmelCase__))) if not is_batched: SCREAMING_SNAKE_CASE = [html_strings] # Get nodes + xpaths SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = [] for html_string in html_strings: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self.get_three_from_single(lowerCAmelCase__) nodes.append(lowerCAmelCase__) SCREAMING_SNAKE_CASE = [] for node, tag_list, sub_list in zip(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__): SCREAMING_SNAKE_CASE = self.construct_xpath(lowerCAmelCase__ , lowerCAmelCase__) xpath_strings.append(lowerCAmelCase__) xpaths.append(lowerCAmelCase__) # return as Dict SCREAMING_SNAKE_CASE = {'nodes': nodes, 'xpaths': xpaths} SCREAMING_SNAKE_CASE = BatchFeature(data=lowerCAmelCase__ , tensor_type=lowerCAmelCase__) return encoded_inputs
259
0
from queue import PriorityQueue from typing import Any import numpy as np def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ): for nxt, d in graph[v]: if nxt in visited_forward: continue snake_case_ = cst_fwd.get(SCREAMING_SNAKE_CASE__ , np.inf ) snake_case_ = cst_fwd[v] + d if new_cost_f < old_cost_f: queue.put((new_cost_f, nxt) ) snake_case_ = new_cost_f snake_case_ = v if nxt in visited_backward: if cst_fwd[v] + d + cst_bwd[nxt] < shortest_distance: snake_case_ = cst_fwd[v] + d + cst_bwd[nxt] return shortest_distance def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): snake_case_ = -1 snake_case_ = set() snake_case_ = set() snake_case_ = {source: 0} snake_case_ = {destination: 0} snake_case_ = {source: None} snake_case_ = {destination: None} snake_case_ = PriorityQueue() snake_case_ = PriorityQueue() snake_case_ = np.inf queue_forward.put((0, source) ) queue_backward.put((0, destination) ) if source == destination: return 0 while not queue_forward.empty() and not queue_backward.empty(): snake_case_, snake_case_ = queue_forward.get() visited_forward.add(SCREAMING_SNAKE_CASE__ ) snake_case_, snake_case_ = queue_backward.get() visited_backward.add(SCREAMING_SNAKE_CASE__ ) snake_case_ = pass_and_relaxation( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ) snake_case_ = pass_and_relaxation( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ) if cst_fwd[v_fwd] + cst_bwd[v_bwd] >= shortest_distance: break if shortest_distance != np.inf: snake_case_ = shortest_distance return shortest_path_distance lowerCAmelCase_ = { '''B''': [['''C''', 1]], '''C''': [['''D''', 1]], '''D''': [['''F''', 1]], '''E''': [['''B''', 1], ['''G''', 2]], '''F''': [], '''G''': [['''F''', 1]], } lowerCAmelCase_ = { '''B''': [['''E''', 1]], '''C''': [['''B''', 1]], '''D''': [['''C''', 1]], '''F''': [['''D''', 1], ['''G''', 1]], '''E''': [[None, np.inf]], '''G''': [['''E''', 2]], } if __name__ == "__main__": import doctest doctest.testmod()
39
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase__ : str = { 'configuration_git': ['GIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GitConfig', 'GitVisionConfig'], 'processing_git': ['GitProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ : str = [ 'GIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'GitForCausalLM', 'GitModel', 'GitPreTrainedModel', 'GitVisionModel', ] if TYPE_CHECKING: from .configuration_git import GIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GitConfig, GitVisionConfig from .processing_git import GitProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_git import ( GIT_PRETRAINED_MODEL_ARCHIVE_LIST, GitForCausalLM, GitModel, GitPreTrainedModel, GitVisionModel, ) else: import sys UpperCamelCase__ : Tuple = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
614
0
class SCREAMING_SNAKE_CASE__ : '''simple docstring''' def __init__( self : List[str] , lowerCamelCase : Any , lowerCamelCase : Dict , lowerCamelCase : List[str] ) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = name _UpperCAmelCase = value _UpperCAmelCase = weight def __repr__( self : Optional[Any] ) -> Tuple: """simple docstring""" return f"""{self.__class__.__name__}({self.name}, {self.value}, {self.weight})""" def lowerCamelCase ( self : str ) -> Optional[int]: """simple docstring""" return self.value def lowerCamelCase ( self : str ) -> Tuple: """simple docstring""" return self.name def lowerCamelCase ( self : str ) -> Tuple: """simple docstring""" return self.weight def lowerCamelCase ( self : str ) -> Union[str, Any]: """simple docstring""" return self.value / self.weight def _SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , __snake_case ) -> List[Any]: _UpperCAmelCase = [] for i in range(len(__snake_case ) ): menu.append(Things(name[i] , value[i] , weight[i] ) ) return menu def _SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , __snake_case ) -> Tuple: _UpperCAmelCase = sorted(__snake_case , key=__snake_case , reverse=__snake_case ) _UpperCAmelCase = [] _UpperCAmelCase , _UpperCAmelCase = 0.0, 0.0 for i in range(len(__snake_case ) ): if (total_cost + items_copy[i].get_weight()) <= max_cost: result.append(items_copy[i] ) total_cost += items_copy[i].get_weight() total_value += items_copy[i].get_value() return (result, total_value) def _SCREAMING_SNAKE_CASE ( ) -> Optional[int]: pass if __name__ == "__main__": import doctest doctest.testmod()
402
from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig __a: Any = logging.get_logger(__name__) # General docstring __a: List[str] = '''RegNetConfig''' # Base docstring __a: int = '''facebook/regnet-y-040''' __a: Union[str, Any] = [1, 1088, 7, 7] # Image classification docstring __a: int = '''facebook/regnet-y-040''' __a: Any = '''tabby, tabby cat''' __a: List[str] = [ '''facebook/regnet-y-040''', # See all regnet models at https://huggingface.co/models?filter=regnet ] class SCREAMING_SNAKE_CASE__ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self : Optional[Any] , lowerCamelCase : int , lowerCamelCase : int = 3 , lowerCamelCase : int = 1 , lowerCamelCase : int = 1 , lowerCamelCase : Optional[str] = "relu" , **lowerCamelCase : Union[str, Any] , ) -> int: """simple docstring""" super().__init__(**lowerCamelCase ) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb _UpperCAmelCase = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2 ) _UpperCAmelCase = tf.keras.layers.ConvaD( filters=lowerCamelCase , kernel_size=lowerCamelCase , strides=lowerCamelCase , padding="""VALID""" , groups=lowerCamelCase , use_bias=lowerCamelCase , name="""convolution""" , ) _UpperCAmelCase = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="""normalization""" ) _UpperCAmelCase = ACTaFN[activation] if activation is not None else tf.identity def lowerCamelCase ( self : Any , lowerCamelCase : Any ) -> Any: """simple docstring""" _UpperCAmelCase = self.convolution(self.padding(lowerCamelCase ) ) _UpperCAmelCase = self.normalization(lowerCamelCase ) _UpperCAmelCase = self.activation(lowerCamelCase ) return hidden_state class SCREAMING_SNAKE_CASE__ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self : Optional[Any] , lowerCamelCase : RegNetConfig , **lowerCamelCase : str ) -> str: """simple docstring""" super().__init__(**lowerCamelCase ) _UpperCAmelCase = config.num_channels _UpperCAmelCase = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name="""embedder""" , ) def lowerCamelCase ( self : str , lowerCamelCase : Dict ) -> Optional[Any]: """simple docstring""" _UpperCAmelCase = shape_list(lowerCamelCase )[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( """Make sure that the channel dimension of the pixel values match with the one set in the configuration.""" ) # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) _UpperCAmelCase = tf.transpose(lowerCamelCase , perm=(0, 2, 3, 1) ) _UpperCAmelCase = self.embedder(lowerCamelCase ) return hidden_state class SCREAMING_SNAKE_CASE__ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self : Optional[int] , lowerCamelCase : int , lowerCamelCase : int = 2 , **lowerCamelCase : Any ) -> str: """simple docstring""" super().__init__(**lowerCamelCase ) _UpperCAmelCase = tf.keras.layers.ConvaD( filters=lowerCamelCase , kernel_size=1 , strides=lowerCamelCase , use_bias=lowerCamelCase , name="""convolution""" ) _UpperCAmelCase = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name="""normalization""" ) def lowerCamelCase ( self : int , lowerCamelCase : tf.Tensor , lowerCamelCase : bool = False ) -> tf.Tensor: """simple docstring""" return self.normalization(self.convolution(lowerCamelCase ) , training=lowerCamelCase ) class SCREAMING_SNAKE_CASE__ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self : List[Any] , lowerCamelCase : int , lowerCamelCase : int , **lowerCamelCase : Tuple ) -> List[Any]: """simple docstring""" super().__init__(**lowerCamelCase ) _UpperCAmelCase = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase , name="""pooler""" ) _UpperCAmelCase = [ tf.keras.layers.ConvaD(filters=lowerCamelCase , kernel_size=1 , activation="""relu""" , name="""attention.0""" ), tf.keras.layers.ConvaD(filters=lowerCamelCase , kernel_size=1 , activation="""sigmoid""" , name="""attention.2""" ), ] def lowerCamelCase ( self : str , lowerCamelCase : int ) -> Optional[Any]: """simple docstring""" # [batch_size, h, w, num_channels] -> [batch_size, 1, 1, num_channels] _UpperCAmelCase = self.pooler(lowerCamelCase ) for layer_module in self.attention: _UpperCAmelCase = layer_module(lowerCamelCase ) _UpperCAmelCase = hidden_state * pooled return hidden_state class SCREAMING_SNAKE_CASE__ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self : Dict , lowerCamelCase : RegNetConfig , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int = 1 , **lowerCamelCase : Dict ) -> Optional[int]: """simple docstring""" super().__init__(**lowerCamelCase ) _UpperCAmelCase = in_channels != out_channels or stride != 1 _UpperCAmelCase = max(1 , out_channels // config.groups_width ) _UpperCAmelCase = ( TFRegNetShortCut(lowerCamelCase , stride=lowerCamelCase , name="""shortcut""" ) if should_apply_shortcut else tf.keras.layers.Activation("""linear""" , name="""shortcut""" ) ) # `self.layers` instead of `self.layer` because that is a reserved argument. _UpperCAmelCase = [ TFRegNetConvLayer(lowerCamelCase , kernel_size=1 , activation=config.hidden_act , name="""layer.0""" ), TFRegNetConvLayer( lowerCamelCase , stride=lowerCamelCase , groups=lowerCamelCase , activation=config.hidden_act , name="""layer.1""" ), TFRegNetConvLayer(lowerCamelCase , kernel_size=1 , activation=lowerCamelCase , name="""layer.2""" ), ] _UpperCAmelCase = ACTaFN[config.hidden_act] def lowerCamelCase ( self : Dict , lowerCamelCase : str ) -> List[str]: """simple docstring""" _UpperCAmelCase = hidden_state for layer_module in self.layers: _UpperCAmelCase = layer_module(lowerCamelCase ) _UpperCAmelCase = self.shortcut(lowerCamelCase ) hidden_state += residual _UpperCAmelCase = self.activation(lowerCamelCase ) return hidden_state class SCREAMING_SNAKE_CASE__ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self : List[str] , lowerCamelCase : RegNetConfig , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int = 1 , **lowerCamelCase : List[str] ) -> Optional[int]: """simple docstring""" super().__init__(**lowerCamelCase ) _UpperCAmelCase = in_channels != out_channels or stride != 1 _UpperCAmelCase = max(1 , out_channels // config.groups_width ) _UpperCAmelCase = ( TFRegNetShortCut(lowerCamelCase , stride=lowerCamelCase , name="""shortcut""" ) if should_apply_shortcut else tf.keras.layers.Activation("""linear""" , name="""shortcut""" ) ) _UpperCAmelCase = [ TFRegNetConvLayer(lowerCamelCase , kernel_size=1 , activation=config.hidden_act , name="""layer.0""" ), TFRegNetConvLayer( lowerCamelCase , stride=lowerCamelCase , groups=lowerCamelCase , activation=config.hidden_act , name="""layer.1""" ), TFRegNetSELayer(lowerCamelCase , reduced_channels=int(round(in_channels / 4 ) ) , name="""layer.2""" ), TFRegNetConvLayer(lowerCamelCase , kernel_size=1 , activation=lowerCamelCase , name="""layer.3""" ), ] _UpperCAmelCase = ACTaFN[config.hidden_act] def lowerCamelCase ( self : Dict , lowerCamelCase : Tuple ) -> Dict: """simple docstring""" _UpperCAmelCase = hidden_state for layer_module in self.layers: _UpperCAmelCase = layer_module(lowerCamelCase ) _UpperCAmelCase = self.shortcut(lowerCamelCase ) hidden_state += residual _UpperCAmelCase = self.activation(lowerCamelCase ) return hidden_state class SCREAMING_SNAKE_CASE__ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self : Tuple , lowerCamelCase : RegNetConfig , lowerCamelCase : int , lowerCamelCase : int , lowerCamelCase : int = 2 , lowerCamelCase : int = 2 , **lowerCamelCase : Dict ) -> Dict: """simple docstring""" super().__init__(**lowerCamelCase ) _UpperCAmelCase = TFRegNetXLayer if config.layer_type == """x""" else TFRegNetYLayer _UpperCAmelCase = [ # downsampling is done in the first layer with stride of 2 layer(lowerCamelCase , lowerCamelCase , lowerCamelCase , stride=lowerCamelCase , name="""layers.0""" ), *[layer(lowerCamelCase , lowerCamelCase , lowerCamelCase , name=f"""layers.{i+1}""" ) for i in range(depth - 1 )], ] def lowerCamelCase ( self : List[str] , lowerCamelCase : Tuple ) -> int: """simple docstring""" for layer_module in self.layers: _UpperCAmelCase = layer_module(lowerCamelCase ) return hidden_state class SCREAMING_SNAKE_CASE__ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self : List[str] , lowerCamelCase : RegNetConfig , **lowerCamelCase : Tuple ) -> Any: """simple docstring""" super().__init__(**lowerCamelCase ) _UpperCAmelCase = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( lowerCamelCase , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name="""stages.0""" , ) ) _UpperCAmelCase = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for i, ((in_channels, out_channels), depth) in enumerate(zip(lowerCamelCase , config.depths[1:] ) ): self.stages.append(TFRegNetStage(lowerCamelCase , lowerCamelCase , lowerCamelCase , depth=lowerCamelCase , name=f"""stages.{i+1}""" ) ) def lowerCamelCase ( self : Optional[int] , lowerCamelCase : tf.Tensor , lowerCamelCase : bool = False , lowerCamelCase : bool = True ) -> TFBaseModelOutputWithNoAttention: """simple docstring""" _UpperCAmelCase = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: _UpperCAmelCase = hidden_states + (hidden_state,) _UpperCAmelCase = stage_module(lowerCamelCase ) if output_hidden_states: _UpperCAmelCase = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return TFBaseModelOutputWithNoAttention(last_hidden_state=lowerCamelCase , hidden_states=lowerCamelCase ) @keras_serializable class SCREAMING_SNAKE_CASE__ ( tf.keras.layers.Layer ): '''simple docstring''' _lowerCamelCase = RegNetConfig def __init__( self : int , lowerCamelCase : Dict , **lowerCamelCase : Dict ) -> Union[str, Any]: """simple docstring""" super().__init__(**lowerCamelCase ) _UpperCAmelCase = config _UpperCAmelCase = TFRegNetEmbeddings(lowerCamelCase , name="""embedder""" ) _UpperCAmelCase = TFRegNetEncoder(lowerCamelCase , name="""encoder""" ) _UpperCAmelCase = tf.keras.layers.GlobalAveragePoolingaD(keepdims=lowerCamelCase , name="""pooler""" ) @unpack_inputs def lowerCamelCase ( self : int , lowerCamelCase : tf.Tensor , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : bool = False , ) -> TFBaseModelOutputWithPoolingAndNoAttention: """simple docstring""" _UpperCAmelCase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase = self.embedder(lowerCamelCase , training=lowerCamelCase ) _UpperCAmelCase = self.encoder( lowerCamelCase , output_hidden_states=lowerCamelCase , return_dict=lowerCamelCase , training=lowerCamelCase ) _UpperCAmelCase = encoder_outputs[0] _UpperCAmelCase = self.pooler(lowerCamelCase ) # Change to NCHW output format have uniformity in the modules _UpperCAmelCase = tf.transpose(lowerCamelCase , perm=(0, 3, 1, 2) ) _UpperCAmelCase = tf.transpose(lowerCamelCase , perm=(0, 3, 1, 2) ) # Change the other hidden state outputs to NCHW as well if output_hidden_states: _UpperCAmelCase = tuple([tf.transpose(lowerCamelCase , perm=(0, 3, 1, 2) ) for h in encoder_outputs[1]] ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=lowerCamelCase , pooler_output=lowerCamelCase , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' _lowerCamelCase = RegNetConfig _lowerCamelCase = '''regnet''' _lowerCamelCase = '''pixel_values''' @property def lowerCamelCase ( self : Optional[int] ) -> Optional[Any]: """simple docstring""" return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 224, 224) , dtype=tf.floataa )} __a: Dict = R''' Parameters: This model is a Tensorflow [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and behavior. config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights. ''' __a: Union[str, Any] = R''' Args: pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConveNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( '''The bare RegNet model outputting raw features without any specific head on top.''' , UpperCAmelCase , ) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ): '''simple docstring''' def __init__( self : str , lowerCamelCase : RegNetConfig , *lowerCamelCase : int , **lowerCamelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" super().__init__(lowerCamelCase , *lowerCamelCase , **lowerCamelCase ) _UpperCAmelCase = TFRegNetMainLayer(lowerCamelCase , name="""regnet""" ) @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC , modality="""vision""" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowerCamelCase ( self : List[str] , lowerCamelCase : tf.Tensor , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Optional[bool] = None , lowerCamelCase : Tuple=False , ) -> Union[TFBaseModelOutputWithPoolingAndNoAttention, Tuple[tf.Tensor]]: """simple docstring""" _UpperCAmelCase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase = self.regnet( pixel_values=lowerCamelCase , output_hidden_states=lowerCamelCase , return_dict=lowerCamelCase , training=lowerCamelCase , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( ''' RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. ''' , UpperCAmelCase , ) class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase , UpperCAmelCase ): '''simple docstring''' def __init__( self : Tuple , lowerCamelCase : RegNetConfig , *lowerCamelCase : Tuple , **lowerCamelCase : List[Any] ) -> str: """simple docstring""" super().__init__(lowerCamelCase , *lowerCamelCase , **lowerCamelCase ) _UpperCAmelCase = config.num_labels _UpperCAmelCase = TFRegNetMainLayer(lowerCamelCase , name="""regnet""" ) # classification head _UpperCAmelCase = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name="""classifier.1""" ) if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(lowerCamelCase ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=lowerCamelCase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowerCamelCase ( self : List[Any] , lowerCamelCase : tf.Tensor = None , lowerCamelCase : tf.Tensor = None , lowerCamelCase : bool = None , lowerCamelCase : bool = None , lowerCamelCase : Optional[Any]=False , ) -> Union[TFSequenceClassifierOutput, Tuple[tf.Tensor]]: """simple docstring""" _UpperCAmelCase = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) _UpperCAmelCase = return_dict if return_dict is not None else self.config.use_return_dict _UpperCAmelCase = self.regnet( lowerCamelCase , output_hidden_states=lowerCamelCase , return_dict=lowerCamelCase , training=lowerCamelCase ) _UpperCAmelCase = outputs.pooler_output if return_dict else outputs[1] _UpperCAmelCase = self.classifier[0](lowerCamelCase ) _UpperCAmelCase = self.classifier[1](lowerCamelCase ) _UpperCAmelCase = None if labels is None else self.hf_compute_loss(labels=lowerCamelCase , logits=lowerCamelCase ) if not return_dict: _UpperCAmelCase = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=lowerCamelCase , logits=lowerCamelCase , hidden_states=outputs.hidden_states )
402
1
'''simple docstring''' from sklearn.metrics import fa_score import datasets __a: Any = """\nThe F1 score is the harmonic mean of the precision and recall. It can be computed with the equation:\nF1 = 2 * (precision * recall) / (precision + recall)\n""" __a: List[str] = """\nArgs:\n predictions (`list` of `int`): Predicted labels.\n references (`list` of `int`): Ground truth labels.\n labels (`list` of `int`): The set of labels to include when `average` is not set to `'binary'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None.\n pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1.\n average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `'binary'`.\n\n - 'binary': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary.\n - 'micro': Calculate metrics globally by counting the total true positives, false negatives and false positives.\n - 'macro': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.\n - 'weighted': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `'macro'` to account for label imbalance. This option can result in an F-score that is not between precision and recall.\n - 'samples': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification).\n sample_weight (`list` of `float`): Sample weights Defaults to None.\n\nReturns:\n f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better.\n\nExamples:\n\n Example 1-A simple binary example\n >>> f1_metric = datasets.load_metric(\"f1\")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0])\n >>> print(results)\n {'f1': 0.5}\n\n Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`.\n >>> f1_metric = datasets.load_metric(\"f1\")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0)\n >>> print(round(results['f1'], 2))\n 0.67\n\n Example 3-The same simple binary example as in Example 1, but with `sample_weight` included.\n >>> f1_metric = datasets.load_metric(\"f1\")\n >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3])\n >>> print(round(results['f1'], 2))\n 0.35\n\n Example 4-A multiclass example, with different values for the `average` input.\n >>> predictions = [0, 2, 1, 0, 0, 1]\n >>> references = [0, 1, 2, 0, 1, 2]\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=\"macro\")\n >>> print(round(results['f1'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=\"micro\")\n >>> print(round(results['f1'], 2))\n 0.33\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=\"weighted\")\n >>> print(round(results['f1'], 2))\n 0.27\n >>> results = f1_metric.compute(predictions=predictions, references=references, average=None)\n >>> print(results)\n {'f1': array([0.8, 0. , 0. ])}\n""" __a: List[str] = """\n@article{scikit-learn,\n title={Scikit-learn: Machine Learning in {P}ython},\n author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V.\n and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P.\n and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and\n Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.},\n journal={Journal of Machine Learning Research},\n volume={12},\n pages={2825--2830},\n year={2011}\n}\n""" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase ( datasets.Metric ): '''simple docstring''' def _lowerCAmelCase( self ) -> str: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''int32''' ) ), '''references''': datasets.Sequence(datasets.Value('''int32''' ) ), } if self.config_name == '''multilabel''' else { '''predictions''': datasets.Value('''int32''' ), '''references''': datasets.Value('''int32''' ), } ) , reference_urls=['''https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html'''] , ) def _lowerCAmelCase( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None , __lowerCAmelCase=1 , __lowerCAmelCase="binary" , __lowerCAmelCase=None ) -> Optional[int]: lowercase__ : int = fa_score( a_ , a_ , labels=a_ , pos_label=a_ , average=a_ , sample_weight=a_ ) return {"f1": float(a_ ) if score.size == 1 else score}
152
"""simple docstring""" def A_ ( lowercase ) -> int: """simple docstring""" return 1 if digit in (0, 1) else (digit * factorial(digit - 1 )) def A_ ( lowercase ) -> bool: """simple docstring""" UpperCAmelCase_ : str = 0 UpperCAmelCase_ : Union[str, Any] = number while duplicate > 0: UpperCAmelCase_ ,UpperCAmelCase_ : Tuple = divmod(lowercase , 10 ) fact_sum += factorial(lowercase ) return fact_sum == number if __name__ == "__main__": print("Program to check whether a number is a Krisnamurthy Number or not.") lowercase_ = int(input("Enter number: ").strip()) print( f"""{number} is {'' if krishnamurthy(number) else 'not '}a Krishnamurthy Number.""" )
470
0
'''simple docstring''' import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata def __a ( A__ , A__=False ) -> str: try: lowerCAmelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. lowerCAmelCase = default else: # KEY is set, convert it to True or False. try: lowerCAmelCase = strtobool(A__ ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f"If set, {key} must be yes or no." ) return _value lowercase : Any = parse_flag_from_env('RUN_SLOW', default=False) lowercase : Dict = parse_flag_from_env('RUN_REMOTE', default=False) lowercase : Union[str, Any] = parse_flag_from_env('RUN_LOCAL', default=True) lowercase : int = parse_flag_from_env('RUN_PACKAGED', default=True) # Compression lowercase : List[str] = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='test requires lz4') lowercase : Tuple = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='test requires py7zr') lowercase : str = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='test requires zstandard') # Audio lowercase : int = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('soundfile') is None or version.parse(importlib_metadata.version('soundfile')) < version.parse('0.12.0'), reason='test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ', ) # Beam lowercase : Tuple = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('0.3.2'), reason='test requires apache-beam and a compatible dill version', ) # Dill-cloudpickle compatibility lowercase : List[str] = pytest.mark.skipif( config.DILL_VERSION <= version.parse('0.3.2'), reason='test requires dill>0.3.2 for cloudpickle compatibility', ) # Windows lowercase : List[str] = pytest.mark.skipif( sys.platform == 'win32', reason='test should not be run on Windows', ) def __a ( A__ ) -> Union[str, Any]: try: import faiss # noqa except ImportError: lowerCAmelCase = unittest.skip("test requires faiss" )(A__ ) return test_case def __a ( A__ ) -> Optional[int]: try: import regex # noqa except ImportError: lowerCAmelCase = unittest.skip("test requires regex" )(A__ ) return test_case def __a ( A__ ) -> Union[str, Any]: try: import elasticsearch # noqa except ImportError: lowerCAmelCase = unittest.skip("test requires elasticsearch" )(A__ ) return test_case def __a ( A__ ) -> str: try: import sqlalchemy # noqa except ImportError: lowerCAmelCase = unittest.skip("test requires sqlalchemy" )(A__ ) return test_case def __a ( A__ ) -> Tuple: if not config.TORCH_AVAILABLE: lowerCAmelCase = unittest.skip("test requires PyTorch" )(A__ ) return test_case def __a ( A__ ) -> Tuple: if not config.TF_AVAILABLE: lowerCAmelCase = unittest.skip("test requires TensorFlow" )(A__ ) return test_case def __a ( A__ ) -> List[str]: if not config.JAX_AVAILABLE: lowerCAmelCase = unittest.skip("test requires JAX" )(A__ ) return test_case def __a ( A__ ) -> Tuple: if not config.PIL_AVAILABLE: lowerCAmelCase = unittest.skip("test requires Pillow" )(A__ ) return test_case def __a ( A__ ) -> List[str]: try: import transformers # noqa F401 except ImportError: return unittest.skip("test requires transformers" )(A__ ) else: return test_case def __a ( A__ ) -> List[str]: try: import tiktoken # noqa F401 except ImportError: return unittest.skip("test requires tiktoken" )(A__ ) else: return test_case def __a ( A__ ) -> Dict: try: import spacy # noqa F401 except ImportError: return unittest.skip("test requires spacy" )(A__ ) else: return test_case def __a ( A__ ) -> Dict: def _require_spacy_model(A__ ): try: import spacy # noqa F401 spacy.load(A__ ) except ImportError: return unittest.skip("test requires spacy" )(A__ ) except OSError: return unittest.skip("test requires spacy model '{}'".format(A__ ) )(A__ ) else: return test_case return _require_spacy_model def __a ( A__ ) -> List[Any]: try: import pyspark # noqa F401 except ImportError: return unittest.skip("test requires pyspark" )(A__ ) else: return test_case def __a ( A__ ) -> str: try: import joblibspark # noqa F401 except ImportError: return unittest.skip("test requires joblibspark" )(A__ ) else: return test_case def __a ( A__ ) -> Union[str, Any]: if not _run_slow_tests or _run_slow_tests == 0: lowerCAmelCase = unittest.skip("test is slow" )(A__ ) return test_case def __a ( A__ ) -> int: if not _run_local_tests or _run_local_tests == 0: lowerCAmelCase = unittest.skip("test is local" )(A__ ) return test_case def __a ( A__ ) -> List[Any]: if not _run_packaged_tests or _run_packaged_tests == 0: lowerCAmelCase = unittest.skip("test is packaged" )(A__ ) return test_case def __a ( A__ ) -> int: if not _run_remote_tests or _run_remote_tests == 0: lowerCAmelCase = unittest.skip("test requires remote" )(A__ ) return test_case def __a ( *A__ ) -> Tuple: def decorate(cls ): for name, fn in cls.__dict__.items(): if callable(A__ ) and name.startswith("test" ): for decorator in decorators: lowerCAmelCase = decorator(A__ ) setattr(cls , A__ , A__ ) return cls return decorate class _lowerCAmelCase ( UpperCamelCase_ ): """simple docstring""" pass class _lowerCAmelCase ( UpperCamelCase_ ): """simple docstring""" lowerCAmelCase = 0 lowerCAmelCase = 1 lowerCAmelCase = 2 @contextmanager def __a ( A__=OfflineSimulationMode.CONNECTION_FAILS , A__=1e-16 ) -> Tuple: lowerCAmelCase = requests.Session().request def timeout_request(A__ , A__ , A__ , **A__ ): # Change the url to an invalid url so that the connection hangs lowerCAmelCase = "https://10.255.255.1" if kwargs.get("timeout" ) is None: raise RequestWouldHangIndefinitelyError( f"Tried a call to {url} in offline mode with no timeout set. Please set a timeout." ) lowerCAmelCase = timeout try: return online_request(A__ , A__ , **A__ ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier lowerCAmelCase = url lowerCAmelCase = e.args[0] lowerCAmelCase = (max_retry_error.args[0].replace("10.255.255.1" , f"OfflineMock[{url}]" ),) lowerCAmelCase = (max_retry_error,) raise def raise_connection_error(A__ , A__ , **A__ ): raise requests.ConnectionError("Offline mode is enabled." , request=A__ ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch("requests.Session.send" , A__ ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch("requests.Session.request" , A__ ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch("datasets.config.HF_DATASETS_OFFLINE" , A__ ): yield else: raise ValueError("Please use a value from the OfflineSimulationMode enum." ) @contextmanager def __a ( *A__ , **A__ ) -> Optional[Any]: lowerCAmelCase = str(Path().resolve() ) with tempfile.TemporaryDirectory(*A__ , **A__ ) as tmp_dir: try: os.chdir(A__ ) yield finally: os.chdir(A__ ) @contextmanager def __a ( ) -> List[Any]: import gc gc.collect() lowerCAmelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def __a ( ) -> Dict: import gc gc.collect() lowerCAmelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def __a ( A__ , A__ ) -> Optional[int]: return deepcopy(A__ ).integers(0 , 100 , 10 ).tolist() == deepcopy(A__ ).integers(0 , 100 , 10 ).tolist() def __a ( A__ ) -> Optional[int]: import decorator from requests.exceptions import HTTPError def _wrapper(A__ , *A__ , **A__ ): try: return func(*A__ , **A__ ) except HTTPError as err: if str(A__ ).startswith("500" ) or str(A__ ).startswith("502" ): pytest.xfail(str(A__ ) ) raise err return decorator.decorator(_wrapper , A__ ) class _lowerCAmelCase : """simple docstring""" def __init__( self : int , SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : List[Any] ) -> Optional[int]: """simple docstring""" lowerCAmelCase = returncode lowerCAmelCase = stdout lowerCAmelCase = stderr async def __a ( A__ , A__ ) -> Union[str, Any]: while True: lowerCAmelCase = await stream.readline() if line: callback(A__ ) else: break async def __a ( A__ , A__=None , A__=None , A__=None , A__=False , A__=False ) -> _RunOutput: if echo: print("\nRunning: " , " ".join(A__ ) ) lowerCAmelCase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=A__ , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=A__ , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) lowerCAmelCase = [] lowerCAmelCase = [] def tee(A__ , A__ , A__ , A__="" ): lowerCAmelCase = line.decode("utf-8" ).rstrip() sink.append(A__ ) if not quiet: print(A__ , A__ , file=A__ ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout , lambda A__ : tee(A__ , A__ , sys.stdout , label="stdout:" ) ), _read_stream(p.stderr , lambda A__ : tee(A__ , A__ , sys.stderr , label="stderr:" ) ), ] , timeout=A__ , ) return _RunOutput(await p.wait() , A__ , A__ ) def __a ( A__ , A__=None , A__=None , A__=180 , A__=False , A__=True ) -> _RunOutput: lowerCAmelCase = asyncio.get_event_loop() lowerCAmelCase = loop.run_until_complete( _stream_subprocess(A__ , env=A__ , stdin=A__ , timeout=A__ , quiet=A__ , echo=A__ ) ) lowerCAmelCase = " ".join(A__ ) if result.returncode > 0: lowerCAmelCase = "\n".join(result.stderr ) raise RuntimeError( f"'{cmd_str}' failed with returncode {result.returncode}\n\n" f"The combined stderr from workers follows:\n{stderr}" ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f"'{cmd_str}' produced no output." ) return result def __a ( ) -> str: lowerCAmelCase = os.environ.get("PYTEST_XDIST_WORKER" , "gw0" ) lowerCAmelCase = re.sub(r"^gw" , "" , A__ , 0 , re.M ) return int(A__ ) def __a ( ) -> int: lowerCAmelCase = 2_9500 lowerCAmelCase = pytest_xdist_worker_id() return port + uniq_delta
159
'''simple docstring''' import functools import logging import os import sys import threading from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional import huggingface_hub.utils as hf_hub_utils from tqdm import auto as tqdm_lib lowercase : Any = threading.Lock() lowercase : Optional[logging.Handler] = None lowercase : List[Any] = { 'debug': logging.DEBUG, 'info': logging.INFO, 'warning': logging.WARNING, 'error': logging.ERROR, 'critical': logging.CRITICAL, } lowercase : List[Any] = logging.WARNING lowercase : int = True def __a ( ) -> Tuple: lowerCAmelCase = os.getenv("TRANSFORMERS_VERBOSITY" , A__ ) if env_level_str: if env_level_str in log_levels: return log_levels[env_level_str] else: logging.getLogger().warning( f"Unknown option TRANSFORMERS_VERBOSITY={env_level_str}, " f"has to be one of: { ', '.join(log_levels.keys() ) }" ) return _default_log_level def __a ( ) -> str: return __name__.split("." )[0] def __a ( ) -> logging.Logger: return logging.getLogger(_get_library_name() ) def __a ( ) -> None: global _default_handler with _lock: if _default_handler: # This library has already configured the library root logger. return lowerCAmelCase = logging.StreamHandler() # Set sys.stderr as stream. lowerCAmelCase = sys.stderr.flush # Apply our default configuration to the library root logger. lowerCAmelCase = _get_library_root_logger() library_root_logger.addHandler(_default_handler ) library_root_logger.setLevel(_get_default_logging_level() ) lowerCAmelCase = False def __a ( ) -> None: global _default_handler with _lock: if not _default_handler: return lowerCAmelCase = _get_library_root_logger() library_root_logger.removeHandler(_default_handler ) library_root_logger.setLevel(logging.NOTSET ) lowerCAmelCase = None def __a ( ) -> List[Any]: return log_levels def __a ( A__ = None ) -> logging.Logger: if name is None: lowerCAmelCase = _get_library_name() _configure_library_root_logger() return logging.getLogger(A__ ) def __a ( ) -> int: _configure_library_root_logger() return _get_library_root_logger().getEffectiveLevel() def __a ( A__ ) -> None: _configure_library_root_logger() _get_library_root_logger().setLevel(A__ ) def __a ( ) -> List[Any]: return set_verbosity(A__ ) def __a ( ) -> str: return set_verbosity(A__ ) def __a ( ) -> List[str]: return set_verbosity(A__ ) def __a ( ) -> Tuple: return set_verbosity(A__ ) def __a ( ) -> None: _configure_library_root_logger() assert _default_handler is not None _get_library_root_logger().removeHandler(_default_handler ) def __a ( ) -> None: _configure_library_root_logger() assert _default_handler is not None _get_library_root_logger().addHandler(_default_handler ) def __a ( A__ ) -> None: _configure_library_root_logger() assert handler is not None _get_library_root_logger().addHandler(A__ ) def __a ( A__ ) -> None: _configure_library_root_logger() assert handler is not None and handler not in _get_library_root_logger().handlers _get_library_root_logger().removeHandler(A__ ) def __a ( ) -> None: _configure_library_root_logger() lowerCAmelCase = False def __a ( ) -> None: _configure_library_root_logger() lowerCAmelCase = True def __a ( ) -> None: lowerCAmelCase = _get_library_root_logger().handlers for handler in handlers: lowerCAmelCase = logging.Formatter("[%(levelname)s|%(filename)s:%(lineno)s] %(asctime)s >> %(message)s" ) handler.setFormatter(A__ ) def __a ( ) -> None: lowerCAmelCase = _get_library_root_logger().handlers for handler in handlers: handler.setFormatter(A__ ) def __a ( self , *A__ , **A__ ) -> List[Any]: lowerCAmelCase = os.getenv("TRANSFORMERS_NO_ADVISORY_WARNINGS" , A__ ) if no_advisory_warnings: return self.warning(*A__ , **A__ ) lowercase : int = warning_advice @functools.lru_cache(A__ ) def __a ( self , *A__ , **A__ ) -> Tuple: self.warning(*A__ , **A__ ) lowercase : int = warning_once class _lowerCAmelCase : """simple docstring""" def __init__( self : Dict , *SCREAMING_SNAKE_CASE : Any , **SCREAMING_SNAKE_CASE : Optional[int] ) -> Tuple: # pylint: disable=unused-argument """simple docstring""" lowerCAmelCase = args[0] if args else None def __iter__( self : Optional[int] ) -> int: """simple docstring""" return iter(self._iterator ) def __getattr__( self : str , SCREAMING_SNAKE_CASE : Dict ) -> Any: """simple docstring""" def empty_fn(*SCREAMING_SNAKE_CASE : List[str] , **SCREAMING_SNAKE_CASE : List[Any] ): # pylint: disable=unused-argument return return empty_fn def __enter__( self : Optional[int] ) -> int: """simple docstring""" return self def __exit__( self : Optional[int] , SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : List[str] , SCREAMING_SNAKE_CASE : str ) -> str: """simple docstring""" return class _lowerCAmelCase : """simple docstring""" def __call__( self : Optional[Any] , *SCREAMING_SNAKE_CASE : Optional[int] , **SCREAMING_SNAKE_CASE : Any ) -> str: """simple docstring""" if _tqdm_active: return tqdm_lib.tqdm(*SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) else: return EmptyTqdm(*SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) def __A ( self : Union[str, Any] , *SCREAMING_SNAKE_CASE : int , **SCREAMING_SNAKE_CASE : Any ) -> Union[str, Any]: """simple docstring""" lowerCAmelCase = None if _tqdm_active: return tqdm_lib.tqdm.set_lock(*SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) def __A ( self : str ) -> Optional[Any]: """simple docstring""" if _tqdm_active: return tqdm_lib.tqdm.get_lock() lowercase : Union[str, Any] = _tqdm_cls() def __a ( ) -> bool: global _tqdm_active return bool(_tqdm_active ) def __a ( ) -> Optional[Any]: global _tqdm_active lowerCAmelCase = True hf_hub_utils.enable_progress_bars() def __a ( ) -> Optional[Any]: global _tqdm_active lowerCAmelCase = False hf_hub_utils.disable_progress_bars()
159
1
'''simple docstring''' import argparse import pathlib import fairseq import torch from fairseq.models.roberta import RobertaModel as FairseqRobertaModel from fairseq.modules import TransformerSentenceEncoderLayer from packaging import version from transformers import XLMRobertaConfig, XLMRobertaXLForMaskedLM, XLMRobertaXLForSequenceClassification from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.models.roberta.modeling_roberta import RobertaAttention from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse("""1.0.0a"""): raise Exception("""requires fairseq >= 1.0.0a""") logging.set_verbosity_info() __magic_name__ : Dict = logging.get_logger(__name__) __magic_name__ : Optional[Any] = "Hello world! cécé herlolip" def snake_case_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' _snake_case = FairseqRobertaModel.from_pretrained(SCREAMING_SNAKE_CASE__ ) roberta.eval() # disable dropout _snake_case = roberta.model.encoder.sentence_encoder _snake_case = XLMRobertaConfig( vocab_size=roberta_sent_encoder.embed_tokens.num_embeddings , hidden_size=roberta.cfg.model.encoder_embed_dim , num_hidden_layers=roberta.cfg.model.encoder_layers , num_attention_heads=roberta.cfg.model.encoder_attention_heads , intermediate_size=roberta.cfg.model.encoder_ffn_embed_dim , max_position_embeddings=5_14 , type_vocab_size=1 , layer_norm_eps=1E-5 , ) if classification_head: _snake_case = roberta.model.classification_heads["mnli"].out_proj.weight.shape[0] print("Our RoBERTa config:" , SCREAMING_SNAKE_CASE__ ) _snake_case = XLMRobertaXLForSequenceClassification(SCREAMING_SNAKE_CASE__ ) if classification_head else XLMRobertaXLForMaskedLM(SCREAMING_SNAKE_CASE__ ) model.eval() # Now let's copy all the weights. # Embeddings _snake_case = roberta_sent_encoder.embed_tokens.weight _snake_case = roberta_sent_encoder.embed_positions.weight _snake_case = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c RoBERTa doesn't use them. _snake_case = roberta_sent_encoder.layer_norm.weight _snake_case = roberta_sent_encoder.layer_norm.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer _snake_case = model.roberta.encoder.layer[i] _snake_case = roberta_sent_encoder.layers[i] _snake_case = layer.attention _snake_case = roberta_layer.self_attn_layer_norm.weight _snake_case = roberta_layer.self_attn_layer_norm.bias # self attention _snake_case = layer.attention.self assert ( roberta_layer.self_attn.k_proj.weight.data.shape == roberta_layer.self_attn.q_proj.weight.data.shape == roberta_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ) _snake_case = roberta_layer.self_attn.q_proj.weight _snake_case = roberta_layer.self_attn.q_proj.bias _snake_case = roberta_layer.self_attn.k_proj.weight _snake_case = roberta_layer.self_attn.k_proj.bias _snake_case = roberta_layer.self_attn.v_proj.weight _snake_case = roberta_layer.self_attn.v_proj.bias # self-attention output _snake_case = layer.attention.output assert self_output.dense.weight.shape == roberta_layer.self_attn.out_proj.weight.shape _snake_case = roberta_layer.self_attn.out_proj.weight _snake_case = roberta_layer.self_attn.out_proj.bias # this one is final layer norm _snake_case = roberta_layer.final_layer_norm.weight _snake_case = roberta_layer.final_layer_norm.bias # intermediate _snake_case = layer.intermediate assert intermediate.dense.weight.shape == roberta_layer.fca.weight.shape _snake_case = roberta_layer.fca.weight _snake_case = roberta_layer.fca.bias # output _snake_case = layer.output assert bert_output.dense.weight.shape == roberta_layer.fca.weight.shape _snake_case = roberta_layer.fca.weight _snake_case = roberta_layer.fca.bias # end of layer if classification_head: _snake_case = roberta.model.classification_heads["mnli"].dense.weight _snake_case = roberta.model.classification_heads["mnli"].dense.bias _snake_case = roberta.model.classification_heads["mnli"].out_proj.weight _snake_case = roberta.model.classification_heads["mnli"].out_proj.bias else: # LM Head _snake_case = roberta.model.encoder.lm_head.dense.weight _snake_case = roberta.model.encoder.lm_head.dense.bias _snake_case = roberta.model.encoder.lm_head.layer_norm.weight _snake_case = roberta.model.encoder.lm_head.layer_norm.bias _snake_case = roberta.model.encoder.lm_head.weight _snake_case = roberta.model.encoder.lm_head.bias # Let's check that we get the same results. _snake_case = roberta.encode(SCREAMING_SNAKE_CASE__ ).unsqueeze(0 ) # batch of size 1 _snake_case = model(SCREAMING_SNAKE_CASE__ )[0] if classification_head: _snake_case = roberta.model.classification_heads["mnli"](roberta.extract_features(SCREAMING_SNAKE_CASE__ ) ) else: _snake_case = roberta.model(SCREAMING_SNAKE_CASE__ )[0] print(our_output.shape , their_output.shape ) _snake_case = torch.max(torch.abs(our_output - their_output ) ).item() print(f'''max_absolute_diff = {max_absolute_diff}''' ) # ~ 1e-7 _snake_case = torch.allclose(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , atol=1E-3 ) print("Do both models output the same tensors?" , "🔥" if success else "💩" ) if not success: raise Exception("Something went wRoNg" ) pathlib.Path(SCREAMING_SNAKE_CASE__ ).mkdir(parents=SCREAMING_SNAKE_CASE__ , exist_ok=SCREAMING_SNAKE_CASE__ ) print(f'''Saving model to {pytorch_dump_folder_path}''' ) model.save_pretrained(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": __magic_name__ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--roberta_checkpoint_path""", default=None, type=str, required=True, help="""Path the official PyTorch dump.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--classification_head""", action="""store_true""", help="""Whether to convert a final classification head.""" ) __magic_name__ : Dict = parser.parse_args() convert_xlm_roberta_xl_checkpoint_to_pytorch( args.roberta_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
672
import inspect import tempfile import unittest from huggingface_hub import hf_hub_download from transformers import is_torch_available from transformers.testing_utils import is_flaky, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin SCREAMING_SNAKE_CASE__ : Optional[Any] = 1E-4 if is_torch_available(): import torch from transformers import AutoformerConfig, AutoformerForPrediction, AutoformerModel from transformers.models.autoformer.modeling_autoformer import AutoformerDecoder, AutoformerEncoder @require_torch class a_ : def __init__( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=16 , SCREAMING_SNAKE_CASE=13 , SCREAMING_SNAKE_CASE=7 , SCREAMING_SNAKE_CASE=14 , SCREAMING_SNAKE_CASE=10 , SCREAMING_SNAKE_CASE=19 , SCREAMING_SNAKE_CASE=5 , SCREAMING_SNAKE_CASE=4 , SCREAMING_SNAKE_CASE=True , SCREAMING_SNAKE_CASE=16 , SCREAMING_SNAKE_CASE=2 , SCREAMING_SNAKE_CASE=4 , SCREAMING_SNAKE_CASE=4 , SCREAMING_SNAKE_CASE="gelu" , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=0.1 , SCREAMING_SNAKE_CASE=[1, 2, 3, 4, 5] , SCREAMING_SNAKE_CASE=25 , SCREAMING_SNAKE_CASE=5 , ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE_ = d_model SCREAMING_SNAKE_CASE_ = parent SCREAMING_SNAKE_CASE_ = batch_size SCREAMING_SNAKE_CASE_ = prediction_length SCREAMING_SNAKE_CASE_ = context_length SCREAMING_SNAKE_CASE_ = cardinality SCREAMING_SNAKE_CASE_ = num_time_features SCREAMING_SNAKE_CASE_ = lags_sequence SCREAMING_SNAKE_CASE_ = embedding_dimension SCREAMING_SNAKE_CASE_ = is_training SCREAMING_SNAKE_CASE_ = hidden_size SCREAMING_SNAKE_CASE_ = num_hidden_layers SCREAMING_SNAKE_CASE_ = num_attention_heads SCREAMING_SNAKE_CASE_ = intermediate_size SCREAMING_SNAKE_CASE_ = hidden_act SCREAMING_SNAKE_CASE_ = hidden_dropout_prob SCREAMING_SNAKE_CASE_ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ = context_length SCREAMING_SNAKE_CASE_ = prediction_length + label_length SCREAMING_SNAKE_CASE_ = label_length SCREAMING_SNAKE_CASE_ = moving_average SCREAMING_SNAKE_CASE_ = autocorrelation_factor def A_( self ) -> List[str]: """simple docstring""" return AutoformerConfig( d_model=self.d_model , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , prediction_length=self.prediction_length , context_length=self.context_length , label_length=self.label_length , lags_sequence=self.lags_sequence , num_time_features=self.num_time_features , num_static_categorical_features=1 , cardinality=[self.cardinality] , embedding_dimension=[self.embedding_dimension] , moving_average=self.moving_average , ) def A_( self , SCREAMING_SNAKE_CASE ) -> Optional[int]: """simple docstring""" SCREAMING_SNAKE_CASE_ = config.context_length + max(config.lags_sequence ) SCREAMING_SNAKE_CASE_ = ids_tensor([self.batch_size, 1] , config.cardinality[0] ) SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, _past_length, config.num_time_features] ) SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, _past_length] ) SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, _past_length] ) > 0.5 # decoder inputs SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, config.prediction_length, config.num_time_features] ) SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, config.prediction_length] ) SCREAMING_SNAKE_CASE_ = { 'past_values': past_values, 'static_categorical_features': static_categorical_features, 'past_time_features': past_time_features, 'past_observed_mask': past_observed_mask, 'future_time_features': future_time_features, 'future_values': future_values, } return inputs_dict def A_( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE_ = self.get_config() SCREAMING_SNAKE_CASE_ = self.prepare_autoformer_inputs_dict(SCREAMING_SNAKE_CASE ) return config, inputs_dict def A_( self ) -> Union[str, Any]: """simple docstring""" SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.prepare_config_and_inputs() return config, inputs_dict def A_( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE_ = AutoformerModel(config=SCREAMING_SNAKE_CASE ).to(SCREAMING_SNAKE_CASE ).eval() SCREAMING_SNAKE_CASE_ = model(**SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = outputs.encoder_last_hidden_state SCREAMING_SNAKE_CASE_ = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: SCREAMING_SNAKE_CASE_ = model.get_encoder() encoder.save_pretrained(SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = AutoformerEncoder.from_pretrained(SCREAMING_SNAKE_CASE ).to(SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = model.create_network_inputs(**SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = model.decomposition_layer(transformer_inputs[:, : config.context_length, ...] ) SCREAMING_SNAKE_CASE_ = torch.cat( (transformer_inputs[:, : config.context_length, ...], feature[:, : config.context_length, ...]) , dim=-1 , ) SCREAMING_SNAKE_CASE_ = encoder(inputs_embeds=SCREAMING_SNAKE_CASE )[0] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1e-3 ) SCREAMING_SNAKE_CASE_ = ( torch.mean(transformer_inputs[:, : config.context_length, ...] , dim=1 ) .unsqueeze(1 ) .repeat(1 , config.prediction_length , 1 ) ) SCREAMING_SNAKE_CASE_ = torch.zeros( [transformer_inputs.shape[0], config.prediction_length, transformer_inputs.shape[2]] , device=enc_input.device , ) SCREAMING_SNAKE_CASE_ = torch.cat( ( torch.cat((seasonal_input[:, -config.label_length :, ...], zeros) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) SCREAMING_SNAKE_CASE_ = torch.cat( ( torch.cat((trend_input[:, -config.label_length :, ...], mean) , dim=1 ), feature[:, config.context_length - config.label_length :, ...], ) , dim=-1 , ) with tempfile.TemporaryDirectory() as tmpdirname: SCREAMING_SNAKE_CASE_ = model.get_decoder() decoder.save_pretrained(SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = AutoformerDecoder.from_pretrained(SCREAMING_SNAKE_CASE ).to(SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = decoder( trend=SCREAMING_SNAKE_CASE , inputs_embeds=SCREAMING_SNAKE_CASE , encoder_hidden_states=SCREAMING_SNAKE_CASE , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1e-3 ) @require_torch class a_ ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): A = (AutoformerModel, AutoformerForPrediction) if is_torch_available() else () A = (AutoformerForPrediction,) if is_torch_available() else () A = {'''feature-extraction''': AutoformerModel} if is_torch_available() else {} A = False A = False A = False A = False A = False A = False def A_( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE_ = AutoformerModelTester(self ) SCREAMING_SNAKE_CASE_ = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE , has_text_modality=SCREAMING_SNAKE_CASE ) def A_( self ) -> int: """simple docstring""" self.config_tester.run_common_tests() def A_( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = model_class(SCREAMING_SNAKE_CASE ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = model_class.from_pretrained(SCREAMING_SNAKE_CASE , output_loading_info=SCREAMING_SNAKE_CASE ) self.assertEqual(info['missing_keys'] , [] ) def A_( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*SCREAMING_SNAKE_CASE ) @unittest.skip(reason='Model has no tokens embeddings' ) def A_( self ) -> Union[str, Any]: """simple docstring""" pass def A_( self ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE_ = inspect.signature(getattr(SCREAMING_SNAKE_CASE , 'forward' ) ) # The main input is the name of the argument after `self` SCREAMING_SNAKE_CASE_ = list(model_signature.parameters.keys() )[1] self.assertEqual(AutoformerModel.main_input_name , SCREAMING_SNAKE_CASE ) def A_( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = model_class(SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE_ = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE_ = [ 'past_values', 'past_time_features', 'past_observed_mask', 'static_categorical_features', 'static_real_features', 'future_values', 'future_time_features', ] if model.__class__.__name__ in ["AutoformerForPrediction"]: expected_arg_names.append('future_observed_mask' ) expected_arg_names.extend( [ 'decoder_attention_mask', 'head_mask', 'decoder_head_mask', 'cross_attn_head_mask', 'encoder_outputs', 'past_key_values', 'output_hidden_states', 'output_attentions', 'use_cache', 'return_dict', ] ) self.assertListEqual(arg_names[: len(SCREAMING_SNAKE_CASE )] , SCREAMING_SNAKE_CASE ) def A_( self ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = getattr(self.model_tester , 'seq_length' , SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = getattr(self.model_tester , 'decoder_seq_length' , SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = getattr(self.model_tester , 'encoder_seq_length' , SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = getattr(self.model_tester , 'd_model' , SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = getattr(self.model_tester , 'num_attention_heads' , SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = d_model // num_attention_heads for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = False SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = model_class(SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) SCREAMING_SNAKE_CASE_ = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(SCREAMING_SNAKE_CASE ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = model_class(SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) SCREAMING_SNAKE_CASE_ = outputs.encoder_attentions self.assertEqual(len(SCREAMING_SNAKE_CASE ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) SCREAMING_SNAKE_CASE_ = len(SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = 7 if "last_hidden_state" in outputs: correct_outlen += 1 if "trend" in outputs: correct_outlen += 1 if "past_key_values" in outputs: correct_outlen += 1 # past_key_values have been returned if "loss" in outputs: correct_outlen += 1 if "params" in outputs: correct_outlen += 1 self.assertEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # decoder attentions SCREAMING_SNAKE_CASE_ = outputs.decoder_attentions self.assertIsInstance(SCREAMING_SNAKE_CASE , (list, tuple) ) self.assertEqual(len(SCREAMING_SNAKE_CASE ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # cross attentions SCREAMING_SNAKE_CASE_ = outputs.cross_attentions self.assertIsInstance(SCREAMING_SNAKE_CASE , (list, tuple) ) self.assertEqual(len(SCREAMING_SNAKE_CASE ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(cross_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, decoder_seq_length, dim] , ) # Check attention is always last and order is fine SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = True SCREAMING_SNAKE_CASE_ = model_class(SCREAMING_SNAKE_CASE ) model.to(SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model(**self._prepare_for_class(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) self.assertEqual(out_len + 2 , len(SCREAMING_SNAKE_CASE ) ) SCREAMING_SNAKE_CASE_ = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(SCREAMING_SNAKE_CASE ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, dim] , ) @is_flaky() def A_( self ) -> Any: """simple docstring""" super().test_retain_grad_hidden_states_attentions() def lowercase ( SCREAMING_SNAKE_CASE="train-batch.pt" ) -> str: '''simple docstring''' SCREAMING_SNAKE_CASE_ = hf_hub_download(repo_id='hf-internal-testing/tourism-monthly-batch' , filename=SCREAMING_SNAKE_CASE , repo_type='dataset' ) SCREAMING_SNAKE_CASE_ = torch.load(SCREAMING_SNAKE_CASE , map_location=SCREAMING_SNAKE_CASE ) return batch @require_torch @slow class a_ ( unittest.TestCase ): def A_( self ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE_ = AutoformerModel.from_pretrained('huggingface/autoformer-tourism-monthly' ).to(SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = prepare_batch() with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model( past_values=batch['past_values'] , past_time_features=batch['past_time_features'] , past_observed_mask=batch['past_observed_mask'] , static_categorical_features=batch['static_categorical_features'] , future_values=batch['future_values'] , future_time_features=batch['future_time_features'] , )[0] SCREAMING_SNAKE_CASE_ = torch.Size( (64, model.config.prediction_length + model.config.label_length, model.config.feature_size) ) self.assertEqual(output.shape , SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = torch.tensor( [[0.3_5_9_3, -1.3_3_9_8, 0.6_3_3_0], [0.2_2_7_9, 1.5_3_9_6, -0.1_7_9_2], [0.0_4_5_0, 1.3_2_2_5, -0.2_3_3_5]] , device=SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(output[0, :3, :3] , SCREAMING_SNAKE_CASE , atol=SCREAMING_SNAKE_CASE ) ) def A_( self ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE_ = AutoformerForPrediction.from_pretrained('huggingface/autoformer-tourism-monthly' ).to(SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = prepare_batch('val-batch.pt' ) with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model( past_values=batch['past_values'] , past_time_features=batch['past_time_features'] , past_observed_mask=batch['past_observed_mask'] , static_categorical_features=batch['static_categorical_features'] , ).encoder_last_hidden_state SCREAMING_SNAKE_CASE_ = torch.Size((64, model.config.context_length, model.config.d_model) ) self.assertEqual(output.shape , SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = torch.tensor( [[-0.0_7_3_4, -0.9_0_3_6, 0.8_3_5_8], [4.7_1_8_6, 2.4_1_1_3, 1.9_5_8_1], [1.7_9_5_3, 2.3_5_5_8, 1.2_9_7_0]] , device=SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(output[0, :3, :3] , SCREAMING_SNAKE_CASE , atol=SCREAMING_SNAKE_CASE ) ) def A_( self ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE_ = AutoformerForPrediction.from_pretrained('huggingface/autoformer-tourism-monthly' ).to(SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = prepare_batch('val-batch.pt' ) with torch.no_grad(): SCREAMING_SNAKE_CASE_ = model.generate( static_categorical_features=batch['static_categorical_features'] , past_time_features=batch['past_time_features'] , past_values=batch['past_values'] , future_time_features=batch['future_time_features'] , past_observed_mask=batch['past_observed_mask'] , ) SCREAMING_SNAKE_CASE_ = torch.Size((64, model.config.num_parallel_samples, model.config.prediction_length) ) self.assertEqual(outputs.sequences.shape , SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = torch.tensor([3_1_3_0.6_7_6_3, 4_0_5_6.5_2_9_3, 7_0_5_3.0_7_8_6] , device=SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ = outputs.sequences.mean(dim=1 ) self.assertTrue(torch.allclose(mean_prediction[0, -3:] , SCREAMING_SNAKE_CASE , rtol=1e-1 ) )
205
0
'''simple docstring''' from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration __UpperCAmelCase = '''facebook/wmt19-en-de''' __UpperCAmelCase = FSMTTokenizer.from_pretrained(mname) # get the correct vocab sizes, etc. from the master model __UpperCAmelCase = FSMTConfig.from_pretrained(mname) config.update( dict( d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) ) __UpperCAmelCase = FSMTForConditionalGeneration(config) print(f"""num of params {tiny_model.num_parameters()}""") # Test __UpperCAmelCase = tokenizer(['''Making tiny model'''], return_tensors='''pt''') __UpperCAmelCase = tiny_model(**batch) print('''test output:''', len(outputs.logits[0])) # Save __UpperCAmelCase = '''tiny-wmt19-en-de''' tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(f"""Generated {mname_tiny}""") # Upload # transformers-cli upload tiny-wmt19-en-de
718
'''simple docstring''' import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device __UpperCAmelCase = False class a__ ( unittest.TestCase ): '''simple docstring''' pass @slow @require_torch_gpu class a__ ( unittest.TestCase ): '''simple docstring''' def __SCREAMING_SNAKE_CASE ( self ) -> str: lowerCAmelCase__ = VersatileDiffusionImageVariationPipeline.from_pretrained('''shi-labs/versatile-diffusion''' ) pipe.to(lowerCamelCase_ ) pipe.set_progress_bar_config(disable=lowerCamelCase_ ) lowerCAmelCase__ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) lowerCAmelCase__ = torch.manual_seed(0 ) lowerCAmelCase__ = pipe( image=lowerCamelCase_ , generator=lowerCamelCase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' , ).images lowerCAmelCase__ = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) lowerCAmelCase__ = np.array([0.0_441, 0.0_469, 0.0_507, 0.0_575, 0.0_632, 0.0_650, 0.0_865, 0.0_909, 0.0_945] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
98
0
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin if is_torch_available(): __magic_name__ ='''pt''' elif is_tf_available(): __magic_name__ ='''tf''' else: __magic_name__ ='''jax''' class _A ( __UpperCamelCase , unittest.TestCase ): SCREAMING_SNAKE_CASE_ : Tuple =PerceiverTokenizer SCREAMING_SNAKE_CASE_ : int =False def _a (self ) -> List[str]: '''simple docstring''' super().setUp() UpperCamelCase__ = PerceiverTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def _a (self ) -> List[str]: '''simple docstring''' return PerceiverTokenizer.from_pretrained('''deepmind/language-perceiver''' ) def _a (self , **SCREAMING_SNAKE_CASE_ ) -> PerceiverTokenizer: '''simple docstring''' return self.tokenizer_class.from_pretrained(self.tmpdirname , **SCREAMING_SNAKE_CASE_ ) def _a (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=False , SCREAMING_SNAKE_CASE_=20 , SCREAMING_SNAKE_CASE_=5 ) -> Tuple[str, list]: '''simple docstring''' UpperCamelCase__ = [] for i in range(len(SCREAMING_SNAKE_CASE_ ) ): try: UpperCamelCase__ = tokenizer.decode([i] , clean_up_tokenization_spaces=SCREAMING_SNAKE_CASE_ ) except UnicodeDecodeError: pass toks.append((i, tok) ) UpperCamelCase__ = list(filter(lambda SCREAMING_SNAKE_CASE_ : re.match(r'''^[ a-zA-Z]+$''' , t[1] ) , SCREAMING_SNAKE_CASE_ ) ) UpperCamelCase__ = list(filter(lambda SCREAMING_SNAKE_CASE_ : [t[0]] == tokenizer.encode(t[1] , add_special_tokens=SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ ) ) if max_length is not None and len(SCREAMING_SNAKE_CASE_ ) > max_length: UpperCamelCase__ = toks[:max_length] if min_length is not None and len(SCREAMING_SNAKE_CASE_ ) < min_length and len(SCREAMING_SNAKE_CASE_ ) > 0: while len(SCREAMING_SNAKE_CASE_ ) < min_length: UpperCamelCase__ = toks + toks # toks_str = [t[1] for t in toks] UpperCamelCase__ = [t[0] for t in toks] # Ensure consistency UpperCamelCase__ = tokenizer.decode(SCREAMING_SNAKE_CASE_ , clean_up_tokenization_spaces=SCREAMING_SNAKE_CASE_ ) if " " not in output_txt and len(SCREAMING_SNAKE_CASE_ ) > 1: UpperCamelCase__ = ( tokenizer.decode([toks_ids[0]] , clean_up_tokenization_spaces=SCREAMING_SNAKE_CASE_ ) + ''' ''' + tokenizer.decode(toks_ids[1:] , clean_up_tokenization_spaces=SCREAMING_SNAKE_CASE_ ) ) if with_prefix_space: UpperCamelCase__ = ''' ''' + output_txt UpperCamelCase__ = tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) return output_txt, output_ids def _a (self ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ = self.perceiver_tokenizer UpperCamelCase__ = '''Unicode €.''' UpperCamelCase__ = tokenizer(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = [4, 91, 116, 111, 105, 117, 106, 107, 38, 232, 136, 178, 52, 5] self.assertEqual(encoded['''input_ids'''] , SCREAMING_SNAKE_CASE_ ) # decoding UpperCamelCase__ = tokenizer.decode(SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , '''[CLS]Unicode €.[SEP]''' ) UpperCamelCase__ = tokenizer('''e è é ê ë''' ) UpperCamelCase__ = [4, 107, 38, 201, 174, 38, 201, 175, 38, 201, 176, 38, 201, 177, 5] self.assertEqual(encoded['''input_ids'''] , SCREAMING_SNAKE_CASE_ ) # decoding UpperCamelCase__ = tokenizer.decode(SCREAMING_SNAKE_CASE_ ) self.assertEqual(SCREAMING_SNAKE_CASE_ , '''[CLS]e è é ê ë[SEP]''' ) # encode/decode, but with `encode` instead of `__call__` self.assertEqual(tokenizer.decode(tokenizer.encode('''e è é ê ë''' ) ) , '''[CLS]e è é ê ë[SEP]''' ) def _a (self ) -> Optional[int]: '''simple docstring''' UpperCamelCase__ = self.perceiver_tokenizer UpperCamelCase__ = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] # fmt: off UpperCamelCase__ = [4, 71, 38, 114, 117, 116, 109, 38, 118, 103, 120, 103, 109, 120, 103, 118, 110, 38, 108, 117, 120, 38, 121, 123, 115, 115, 103, 120, 111, 128, 103, 122, 111, 117, 116, 52, 5, 0] # fmt: on UpperCamelCase__ = tokenizer(SCREAMING_SNAKE_CASE_ , padding=SCREAMING_SNAKE_CASE_ , return_tensors=SCREAMING_SNAKE_CASE_ ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if FRAMEWORK != "jax": UpperCamelCase__ = list(batch.input_ids.numpy()[0] ) else: UpperCamelCase__ = list(batch.input_ids.tolist()[0] ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertEqual((2, 38) , batch.input_ids.shape ) self.assertEqual((2, 38) , batch.attention_mask.shape ) def _a (self ) -> int: '''simple docstring''' UpperCamelCase__ = self.perceiver_tokenizer UpperCamelCase__ = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] UpperCamelCase__ = tokenizer(SCREAMING_SNAKE_CASE_ , padding=SCREAMING_SNAKE_CASE_ , return_tensors=SCREAMING_SNAKE_CASE_ ) # check if input_ids are returned and no decoder_input_ids self.assertIn('''input_ids''' , SCREAMING_SNAKE_CASE_ ) self.assertIn('''attention_mask''' , SCREAMING_SNAKE_CASE_ ) self.assertNotIn('''decoder_input_ids''' , SCREAMING_SNAKE_CASE_ ) self.assertNotIn('''decoder_attention_mask''' , SCREAMING_SNAKE_CASE_ ) def _a (self ) -> List[Any]: '''simple docstring''' UpperCamelCase__ = self.perceiver_tokenizer UpperCamelCase__ = [ '''Summary of the text.''', '''Another summary.''', ] UpperCamelCase__ = tokenizer( text_target=SCREAMING_SNAKE_CASE_ , max_length=32 , padding='''max_length''' , truncation=SCREAMING_SNAKE_CASE_ , return_tensors=SCREAMING_SNAKE_CASE_ ) self.assertEqual(32 , targets['''input_ids'''].shape[1] ) def _a (self ) -> List[Any]: '''simple docstring''' UpperCamelCase__ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test UpperCamelCase__ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): # Isolate this from the other tests because we save additional tokens/etc UpperCamelCase__ = tempfile.mkdtemp() UpperCamelCase__ = ''' He is very happy, UNwant\u00E9d,running''' UpperCamelCase__ = tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) tokenizer.save_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.__class__.from_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = after_tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) shutil.rmtree(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): # Isolate this from the other tests because we save additional tokens/etc UpperCamelCase__ = tempfile.mkdtemp() UpperCamelCase__ = ''' He is very happy, UNwant\u00E9d,running''' tokenizer.add_tokens(['''bim''', '''bambam'''] ) UpperCamelCase__ = tokenizer.additional_special_tokens additional_special_tokens.append('''new_additional_special_token''' ) tokenizer.add_special_tokens({'''additional_special_tokens''': additional_special_tokens} ) UpperCamelCase__ = tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) tokenizer.save_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = tokenizer.__class__.from_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = after_tokenizer.encode(SCREAMING_SNAKE_CASE_ , add_special_tokens=SCREAMING_SNAKE_CASE_ ) self.assertListEqual(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertIn('''new_additional_special_token''' , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) UpperCamelCase__ = tokenizer.__class__.from_pretrained(SCREAMING_SNAKE_CASE_ , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(SCREAMING_SNAKE_CASE_ ) def _a (self ) -> str: '''simple docstring''' UpperCamelCase__ = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(SCREAMING_SNAKE_CASE_ ) with open(os.path.join(SCREAMING_SNAKE_CASE_ , '''special_tokens_map.json''' ) , encoding='''utf-8''' ) as json_file: UpperCamelCase__ = json.load(SCREAMING_SNAKE_CASE_ ) with open(os.path.join(SCREAMING_SNAKE_CASE_ , '''tokenizer_config.json''' ) , encoding='''utf-8''' ) as json_file: UpperCamelCase__ = json.load(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = [F"<extra_id_{i}>" for i in range(125 )] UpperCamelCase__ = added_tokens_extra_ids + [ '''an_additional_special_token''' ] UpperCamelCase__ = added_tokens_extra_ids + [ '''an_additional_special_token''' ] with open(os.path.join(SCREAMING_SNAKE_CASE_ , '''special_tokens_map.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) with open(os.path.join(SCREAMING_SNAKE_CASE_ , '''tokenizer_config.json''' ) , '''w''' , encoding='''utf-8''' ) as outfile: json.dump(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files UpperCamelCase__ = tokenizer_class.from_pretrained( SCREAMING_SNAKE_CASE_ , ) self.assertIn( '''an_additional_special_token''' , tokenizer_without_change_in_init.additional_special_tokens ) self.assertEqual( ['''an_additional_special_token'''] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids(['''an_additional_special_token'''] ) ) , ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained UpperCamelCase__ = added_tokens_extra_ids + [AddedToken('''a_new_additional_special_token''' , lstrip=SCREAMING_SNAKE_CASE_ )] UpperCamelCase__ = tokenizer_class.from_pretrained( SCREAMING_SNAKE_CASE_ , additional_special_tokens=SCREAMING_SNAKE_CASE_ , ) self.assertIn('''a_new_additional_special_token''' , tokenizer.additional_special_tokens ) self.assertEqual( ['''a_new_additional_special_token'''] , tokenizer.convert_ids_to_tokens( tokenizer.convert_tokens_to_ids(['''a_new_additional_special_token'''] ) ) , ) def _a (self ) -> int: '''simple docstring''' UpperCamelCase__ = self.perceiver_tokenizer self.assertEqual(tokenizer.decode([178] ) , '''�''' ) def _a (self ) -> List[str]: '''simple docstring''' pass def _a (self ) -> str: '''simple docstring''' pass def _a (self ) -> List[str]: '''simple docstring''' pass def _a (self ) -> List[Any]: '''simple docstring''' pass def _a (self ) -> Optional[int]: '''simple docstring''' UpperCamelCase__ = self.get_tokenizers(fast=SCREAMING_SNAKE_CASE_ , do_lower_case=SCREAMING_SNAKE_CASE_ ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): UpperCamelCase__ = ['''[CLS]''', '''t''', '''h''', '''i''', '''s''', ''' ''', '''i''', '''s''', ''' ''', '''a''', ''' ''', '''t''', '''e''', '''s''', '''t''', '''[SEP]'''] UpperCamelCase__ = tokenizer.convert_tokens_to_string(SCREAMING_SNAKE_CASE_ ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ )
415
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __magic_name__ =logging.get_logger(__name__) class _A ( __UpperCamelCase ): SCREAMING_SNAKE_CASE_ : Any =["pixel_values"] def __init__(self , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = PIL.Image.BICUBIC , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = 1 / 255 , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = True , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> None: '''simple docstring''' super().__init__(**SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = size if size is not None else {'''height''': 256, '''width''': 256} UpperCamelCase__ = get_size_dict(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = crop_size if crop_size is not None else {'''height''': 224, '''width''': 224} UpperCamelCase__ = get_size_dict(SCREAMING_SNAKE_CASE_ , param_name='''crop_size''' ) UpperCamelCase__ = do_resize UpperCamelCase__ = size UpperCamelCase__ = resample UpperCamelCase__ = do_center_crop UpperCamelCase__ = crop_size UpperCamelCase__ = do_rescale UpperCamelCase__ = rescale_factor UpperCamelCase__ = do_normalize UpperCamelCase__ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCamelCase__ = image_std if image_std is not None else IMAGENET_STANDARD_STD def _a (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = PIL.Image.BICUBIC , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> np.ndarray: '''simple docstring''' UpperCamelCase__ = get_size_dict(SCREAMING_SNAKE_CASE_ ) if "height" not in size or "width" not in size: raise ValueError(F"The size dictionary must have keys 'height' and 'width'. Got {size.keys()}" ) return resize( SCREAMING_SNAKE_CASE_ , size=(size['''height'''], size['''width''']) , resample=SCREAMING_SNAKE_CASE_ , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def _a (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> np.ndarray: '''simple docstring''' UpperCamelCase__ = get_size_dict(SCREAMING_SNAKE_CASE_ ) if "height" not in size or "width" not in size: raise ValueError(F"The size dictionary must have keys 'height' and 'width'. Got {size.keys()}" ) return center_crop(SCREAMING_SNAKE_CASE_ , size=(size['''height'''], size['''width''']) , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def _a (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> List[str]: '''simple docstring''' return rescale(SCREAMING_SNAKE_CASE_ , scale=SCREAMING_SNAKE_CASE_ , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def _a (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , **SCREAMING_SNAKE_CASE_ , ) -> np.ndarray: '''simple docstring''' return normalize(SCREAMING_SNAKE_CASE_ , mean=SCREAMING_SNAKE_CASE_ , std=SCREAMING_SNAKE_CASE_ , data_format=SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ) def _a (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = None , SCREAMING_SNAKE_CASE_ = ChannelDimension.FIRST , **SCREAMING_SNAKE_CASE_ , ) -> PIL.Image.Image: '''simple docstring''' UpperCamelCase__ = do_resize if do_resize is not None else self.do_resize UpperCamelCase__ = resample if resample is not None else self.resample UpperCamelCase__ = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCamelCase__ = do_rescale if do_rescale is not None else self.do_rescale UpperCamelCase__ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCamelCase__ = do_normalize if do_normalize is not None else self.do_normalize UpperCamelCase__ = image_mean if image_mean is not None else self.image_mean UpperCamelCase__ = image_std if image_std is not None else self.image_std UpperCamelCase__ = size if size is not None else self.size UpperCamelCase__ = get_size_dict(SCREAMING_SNAKE_CASE_ ) UpperCamelCase__ = crop_size if crop_size is not None else self.crop_size UpperCamelCase__ = get_size_dict(SCREAMING_SNAKE_CASE_ , param_name='''crop_size''' ) UpperCamelCase__ = make_list_of_images(SCREAMING_SNAKE_CASE_ ) if not valid_images(SCREAMING_SNAKE_CASE_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. UpperCamelCase__ = [to_numpy_array(SCREAMING_SNAKE_CASE_ ) for image in images] if do_resize: UpperCamelCase__ = [self.resize(image=SCREAMING_SNAKE_CASE_ , size=SCREAMING_SNAKE_CASE_ , resample=SCREAMING_SNAKE_CASE_ ) for image in images] if do_center_crop: UpperCamelCase__ = [self.center_crop(image=SCREAMING_SNAKE_CASE_ , size=SCREAMING_SNAKE_CASE_ ) for image in images] if do_rescale: UpperCamelCase__ = [self.rescale(image=SCREAMING_SNAKE_CASE_ , scale=SCREAMING_SNAKE_CASE_ ) for image in images] if do_normalize: UpperCamelCase__ = [self.normalize(image=SCREAMING_SNAKE_CASE_ , mean=SCREAMING_SNAKE_CASE_ , std=SCREAMING_SNAKE_CASE_ ) for image in images] UpperCamelCase__ = [to_channel_dimension_format(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) for image in images] UpperCamelCase__ = {'''pixel_values''': images} return BatchFeature(data=SCREAMING_SNAKE_CASE_ , tensor_type=SCREAMING_SNAKE_CASE_ )
415
1
from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class SCREAMING_SNAKE_CASE_ : """simple docstring""" def __init__( self , _lowerCAmelCase , _lowerCAmelCase=13 , _lowerCAmelCase=30 , _lowerCAmelCase=2 , _lowerCAmelCase=3 , _lowerCAmelCase=True , _lowerCAmelCase=True , _lowerCAmelCase=32 , _lowerCAmelCase=2 , _lowerCAmelCase=4 , _lowerCAmelCase=37 , _lowerCAmelCase="gelu" , _lowerCAmelCase=0.1 , _lowerCAmelCase=0.1 , _lowerCAmelCase=10 , _lowerCAmelCase=0.02 , _lowerCAmelCase=3 , _lowerCAmelCase=None , _lowerCAmelCase=2 , ): lowerCamelCase__ = parent lowerCamelCase__ = batch_size lowerCamelCase__ = image_size lowerCamelCase__ = patch_size lowerCamelCase__ = num_channels lowerCamelCase__ = is_training lowerCamelCase__ = use_labels lowerCamelCase__ = hidden_size lowerCamelCase__ = num_hidden_layers lowerCamelCase__ = num_attention_heads lowerCamelCase__ = intermediate_size lowerCamelCase__ = hidden_act lowerCamelCase__ = hidden_dropout_prob lowerCamelCase__ = attention_probs_dropout_prob lowerCamelCase__ = type_sequence_label_size lowerCamelCase__ = initializer_range lowerCamelCase__ = scope lowerCamelCase__ = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) lowerCamelCase__ = (image_size // patch_size) ** 2 lowerCamelCase__ = num_patches + 2 def __magic_name__ ( self ): lowerCamelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ = None if self.use_labels: lowerCamelCase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase__ = self.get_config() return config, pixel_values, labels def __magic_name__ ( self ): return DeiTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowerCAmelCase , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __magic_name__ ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): lowerCamelCase__ = TFDeiTModel(config=_lowerCAmelCase ) lowerCamelCase__ = model(_lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __magic_name__ ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): lowerCamelCase__ = TFDeiTForMaskedImageModeling(config=_lowerCAmelCase ) lowerCamelCase__ = model(_lowerCAmelCase ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCamelCase__ = 1 lowerCamelCase__ = TFDeiTForMaskedImageModeling(_lowerCAmelCase ) lowerCamelCase__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase__ = model(_lowerCAmelCase ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __magic_name__ ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): lowerCamelCase__ = self.type_sequence_label_size lowerCamelCase__ = TFDeiTForImageClassification(_lowerCAmelCase ) lowerCamelCase__ = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images lowerCamelCase__ = 1 lowerCamelCase__ = TFDeiTForImageClassification(_lowerCAmelCase ) lowerCamelCase__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase__ = model(_lowerCAmelCase , labels=_lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __magic_name__ ( self ): lowerCamelCase__ = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = config_and_inputs lowerCamelCase__ = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE_ ( lowercase_ , lowercase_ , unittest.TestCase ): """simple docstring""" A__ = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) A__ = ( { "feature-extraction": TFDeiTModel, "image-classification": (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) A__ = False A__ = False A__ = False A__ = False def __magic_name__ ( self ): lowerCamelCase__ = TFDeiTModelTester(self ) lowerCamelCase__ = ConfigTester(self , config_class=_lowerCAmelCase , has_text_modality=_lowerCAmelCase , hidden_size=37 ) def __magic_name__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="DeiT does not use inputs_embeds" ) def __magic_name__ ( self ): pass def __magic_name__ ( self ): lowerCamelCase__ , lowerCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ = model_class(_lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowerCAmelCase , tf.keras.layers.Dense ) ) def __magic_name__ ( self ): lowerCamelCase__ , lowerCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ = model_class(_lowerCAmelCase ) lowerCamelCase__ = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ = [*signature.parameters.keys()] lowerCamelCase__ = ["pixel_values"] self.assertListEqual(arg_names[:1] , _lowerCAmelCase ) def __magic_name__ ( self ): lowerCamelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCAmelCase ) def __magic_name__ ( self ): lowerCamelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*_lowerCAmelCase ) def __magic_name__ ( self ): lowerCamelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCAmelCase ) def __magic_name__ ( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=False ): lowerCamelCase__ = super()._prepare_for_class(_lowerCAmelCase , _lowerCAmelCase , return_labels=_lowerCAmelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def __magic_name__ ( self ): for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ = TFDeiTModel.from_pretrained(_lowerCAmelCase ) self.assertIsNotNone(_lowerCAmelCase ) def __UpperCamelCase ( ) ->int: lowerCamelCase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png") return image @require_tf @require_vision class SCREAMING_SNAKE_CASE_ ( unittest.TestCase ): """simple docstring""" @cached_property def __magic_name__ ( self ): return ( DeiTImageProcessor.from_pretrained("facebook/deit-base-distilled-patch16-224" ) if is_vision_available() else None ) @slow def __magic_name__ ( self ): lowerCamelCase__ = TFDeiTForImageClassificationWithTeacher.from_pretrained("facebook/deit-base-distilled-patch16-224" ) lowerCamelCase__ = self.default_image_processor lowerCamelCase__ = prepare_img() lowerCamelCase__ = image_processor(images=_lowerCAmelCase , return_tensors="tf" ) # forward pass lowerCamelCase__ = model(**_lowerCAmelCase ) # verify the logits lowerCamelCase__ = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCAmelCase ) lowerCamelCase__ = tf.constant([-1.0_266, 0.1_912, -1.2_861] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , _lowerCAmelCase , atol=1E-4 ) )
360
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class SCREAMING_SNAKE_CASE_ ( datasets.BeamBasedBuilder ): """simple docstring""" def __magic_name__ ( self ): return datasets.DatasetInfo( features=datasets.Features({"content": datasets.Value("string" )} ) , supervised_keys=_lowerCAmelCase , ) def __magic_name__ ( self , _lowerCAmelCase , _lowerCAmelCase ): return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_dummy_examples()} )] def __magic_name__ ( self , _lowerCAmelCase , _lowerCAmelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(_lowerCAmelCase ) class SCREAMING_SNAKE_CASE_ ( datasets.BeamBasedBuilder ): """simple docstring""" def __magic_name__ ( self ): return datasets.DatasetInfo( features=datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) , supervised_keys=_lowerCAmelCase , ) def __magic_name__ ( self , _lowerCAmelCase , _lowerCAmelCase ): return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"examples": get_test_nested_examples()} ) ] def __magic_name__ ( self , _lowerCAmelCase , _lowerCAmelCase ): import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(_lowerCAmelCase ) def __UpperCamelCase ( ) ->Optional[Any]: return [(i, {"content": content}) for i, content in enumerate(["foo", "bar", "foobar"])] def __UpperCamelCase ( ) ->Optional[Any]: return [(i, {"a": {"b": [content]}}) for i, content in enumerate(["foo", "bar", "foobar"])] class SCREAMING_SNAKE_CASE_ ( lowercase_ ): """simple docstring""" @require_beam def __magic_name__ ( self ): lowerCamelCase__ = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: lowerCamelCase__ = DummyBeamDataset(cache_dir=_lowerCAmelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(_lowerCAmelCase , builder.name , "default" , "0.0.0" , F"{builder.name}-train.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) lowerCamelCase__ = builder.as_dataset() self.assertEqual(dset["train"].num_rows , _lowerCAmelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , _lowerCAmelCase ) self.assertDictEqual(dset["train"][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(_lowerCAmelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def __magic_name__ ( self ): import apache_beam as beam lowerCamelCase__ = beam.io.parquetio.WriteToParquet lowerCamelCase__ = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: lowerCamelCase__ = DummyBeamDataset(cache_dir=_lowerCAmelCase , beam_runner="DirectRunner" ) with patch("apache_beam.io.parquetio.WriteToParquet" ) as write_parquet_mock: lowerCamelCase__ = partial(_lowerCAmelCase , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( _lowerCAmelCase , builder.name , "default" , "0.0.0" , F"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertTrue( os.path.exists( os.path.join( _lowerCAmelCase , builder.name , "default" , "0.0.0" , F"{builder.name}-train-00000-of-00002.arrow" ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({"content": datasets.Value("string" )} ) ) lowerCamelCase__ = builder.as_dataset() self.assertEqual(dset["train"].num_rows , _lowerCAmelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , _lowerCAmelCase ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset["train"]["content"] ) , sorted(["foo", "bar", "foobar"] ) ) self.assertTrue( os.path.exists(os.path.join(_lowerCAmelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset @require_beam def __magic_name__ ( self ): with tempfile.TemporaryDirectory() as tmp_cache_dir: lowerCamelCase__ = DummyBeamDataset(cache_dir=_lowerCAmelCase ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def __magic_name__ ( self ): lowerCamelCase__ = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: lowerCamelCase__ = NestedBeamDataset(cache_dir=_lowerCAmelCase , beam_runner="DirectRunner" ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(_lowerCAmelCase , builder.name , "default" , "0.0.0" , F"{builder.name}-train.arrow" ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({"a": datasets.Sequence({"b": datasets.Value("string" )} )} ) ) lowerCamelCase__ = builder.as_dataset() self.assertEqual(dset["train"].num_rows , _lowerCAmelCase ) self.assertEqual(dset["train"].info.splits["train"].num_examples , _lowerCAmelCase ) self.assertDictEqual(dset["train"][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset["train"][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(_lowerCAmelCase , builder.name , "default" , "0.0.0" , "dataset_info.json" ) ) ) del dset
360
1
import datasets __UpperCAmelCase = '''\ @InProceedings{conneau2018xnli, author = "Conneau, Alexis and Rinott, Ruty and Lample, Guillaume and Williams, Adina and Bowman, Samuel R. and Schwenk, Holger and Stoyanov, Veselin", title = "XNLI: Evaluating Cross-lingual Sentence Representations", booktitle = "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing", year = "2018", publisher = "Association for Computational Linguistics", location = "Brussels, Belgium", } ''' __UpperCAmelCase = '''\ XNLI is a subset of a few thousand examples from MNLI which has been translated into a 14 different languages (some low-ish resource). As with MNLI, the goal is to predict textual entailment (does sentence A imply/contradict/neither sentence B) and is a classification task (given two sentences, predict one of three labels). ''' __UpperCAmelCase = ''' Computes XNLI score which is just simple accuracy. Args: predictions: Predicted labels. references: Ground truth labels. Returns: \'accuracy\': accuracy Examples: >>> predictions = [0, 1] >>> references = [0, 1] >>> xnli_metric = datasets.load_metric("xnli") >>> results = xnli_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0} ''' def UpperCamelCase ( snake_case__ : str , snake_case__ : List[Any] ) -> Any: return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCAmelCase_ ( datasets.Metric ): def snake_case_ ( self ) -> Tuple: return datasets.MetricInfo( description=_DESCRIPTION, citation=_CITATION, inputs_description=_KWARGS_DESCRIPTION, features=datasets.Features( { 'predictions': datasets.Value('int64' if self.config_name != 'sts-b' else 'float32' ), 'references': datasets.Value('int64' if self.config_name != 'sts-b' else 'float32' ), } ), codebase_urls=[], reference_urls=[], format='numpy', ) def snake_case_ ( self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> List[str]: return {"accuracy": simple_accuracy(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ )}
40
"""simple docstring""" import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class lowerCamelCase__ ( lowerCamelCase_ ): @require_torch def lowerCamelCase_ ( self ): """simple docstring""" snake_case : Any = "\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n " snake_case : List[str] = "\nmname = \"hf-internal-testing/tiny-random-bert\"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task=\"fill-mask\", model=mname)\nprint(\"success\")\n " snake_case : Optional[int] = "\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError(\"Offline mode is enabled, we shouldn't access internet\")\nsocket.socket = offline_socket\n " # Force fetching the files so that we can use the cache snake_case : str = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(SCREAMING_SNAKE_CASE ) BertModel.from_pretrained(SCREAMING_SNAKE_CASE ) BertTokenizer.from_pretrained(SCREAMING_SNAKE_CASE ) pipeline(task="fill-mask" , model=SCREAMING_SNAKE_CASE ) # baseline - just load from_pretrained with normal network snake_case : str = [sys.executable, "-c", "\n".join([load, run, mock] )] # should succeed snake_case : Optional[Any] = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files snake_case : List[Any] = "1" snake_case : Any = subprocess.run(SCREAMING_SNAKE_CASE , env=SCREAMING_SNAKE_CASE , check=SCREAMING_SNAKE_CASE , capture_output=SCREAMING_SNAKE_CASE ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) @require_torch def lowerCamelCase_ ( self ): """simple docstring""" snake_case : Optional[int] = "\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n " snake_case : str = "\nmname = \"hf-internal-testing/tiny-random-bert\"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task=\"fill-mask\", model=mname)\nprint(\"success\")\n " snake_case : Optional[Any] = "\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error(\"Faking flaky internet\")\nsocket.socket = offline_socket\n " # Force fetching the files so that we can use the cache snake_case : List[str] = "hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(SCREAMING_SNAKE_CASE ) BertModel.from_pretrained(SCREAMING_SNAKE_CASE ) BertTokenizer.from_pretrained(SCREAMING_SNAKE_CASE ) pipeline(task="fill-mask" , model=SCREAMING_SNAKE_CASE ) # baseline - just load from_pretrained with normal network snake_case : Dict = [sys.executable, "-c", "\n".join([load, run, mock] )] # should succeed snake_case : Optional[Any] = self.get_env() snake_case : Tuple = subprocess.run(SCREAMING_SNAKE_CASE , env=SCREAMING_SNAKE_CASE , check=SCREAMING_SNAKE_CASE , capture_output=SCREAMING_SNAKE_CASE ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) @require_torch def lowerCamelCase_ ( self ): """simple docstring""" snake_case : Dict = "\nfrom transformers import BertConfig, BertModel, BertTokenizer\n " snake_case : Optional[int] = "\nmname = \"hf-internal-testing/tiny-random-bert-sharded\"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint(\"success\")\n " snake_case : str = "\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError(\"Offline mode is enabled\")\nsocket.socket = offline_socket\n " # baseline - just load from_pretrained with normal network snake_case : Optional[Any] = [sys.executable, "-c", "\n".join([load, run] )] # should succeed snake_case : Union[str, Any] = self.get_env() snake_case : Any = subprocess.run(SCREAMING_SNAKE_CASE , env=SCREAMING_SNAKE_CASE , check=SCREAMING_SNAKE_CASE , capture_output=SCREAMING_SNAKE_CASE ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) # next emulate no network snake_case : Optional[int] = [sys.executable, "-c", "\n".join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files snake_case : Optional[int] = "1" snake_case : List[Any] = subprocess.run(SCREAMING_SNAKE_CASE , env=SCREAMING_SNAKE_CASE , check=SCREAMING_SNAKE_CASE , capture_output=SCREAMING_SNAKE_CASE ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) @require_torch def lowerCamelCase_ ( self ): """simple docstring""" snake_case : List[Any] = "\nfrom transformers import pipeline\n " snake_case : Dict = "\nmname = \"hf-internal-testing/tiny-random-bert\"\npipe = pipeline(model=mname)\n " snake_case : int = "\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error(\"Offline mode is enabled\")\nsocket.socket = offline_socket\n " snake_case : Union[str, Any] = self.get_env() snake_case : List[Any] = "1" snake_case : Tuple = [sys.executable, "-c", "\n".join([load, mock, run] )] snake_case : Union[str, Any] = subprocess.run(SCREAMING_SNAKE_CASE , env=SCREAMING_SNAKE_CASE , check=SCREAMING_SNAKE_CASE , capture_output=SCREAMING_SNAKE_CASE ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( "You cannot infer task automatically within `pipeline` when using offline mode" , result.stderr.decode().replace("\n" , "" ) , ) @require_torch def lowerCamelCase_ ( self ): """simple docstring""" snake_case : List[Any] = "\nfrom transformers import AutoModel\n " snake_case : Union[str, Any] = "\nmname = \"hf-internal-testing/test_dynamic_model\"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint(\"success\")\n " # baseline - just load from_pretrained with normal network snake_case : Optional[Any] = [sys.executable, "-c", "\n".join([load, run] )] # should succeed snake_case : str = self.get_env() snake_case : int = subprocess.run(SCREAMING_SNAKE_CASE , env=SCREAMING_SNAKE_CASE , check=SCREAMING_SNAKE_CASE , capture_output=SCREAMING_SNAKE_CASE ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files snake_case : Any = "1" snake_case : List[str] = subprocess.run(SCREAMING_SNAKE_CASE , env=SCREAMING_SNAKE_CASE , check=SCREAMING_SNAKE_CASE , capture_output=SCREAMING_SNAKE_CASE ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() )
134
0
'''simple docstring''' from ...processing_utils import ProcessorMixin class SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCAmelCase__ = '''SpeechT5FeatureExtractor''' UpperCAmelCase__ = '''SpeechT5Tokenizer''' def __init__( self : Any , lowercase__ : List[Any] , lowercase__ : Any ) ->Optional[Any]: '''simple docstring''' super().__init__(__lowerCAmelCase , __lowerCAmelCase ) def __call__( self : Optional[Any] , *lowercase__ : int , **lowercase__ : Tuple ) ->Dict: '''simple docstring''' _UpperCamelCase : Any = kwargs.pop("audio" , __lowerCAmelCase ) _UpperCamelCase : int = kwargs.pop("text" , __lowerCAmelCase ) _UpperCamelCase : List[str] = kwargs.pop("text_target" , __lowerCAmelCase ) _UpperCamelCase : str = kwargs.pop("audio_target" , __lowerCAmelCase ) _UpperCamelCase : str = kwargs.pop("sampling_rate" , __lowerCAmelCase ) if audio is not None and text is not None: raise ValueError( "Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?" ) if audio_target is not None and text_target is not None: raise ValueError( "Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?" ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( "You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process." ) if audio is not None: _UpperCamelCase : Tuple = self.feature_extractor(__lowerCAmelCase , *__lowerCAmelCase , sampling_rate=__lowerCAmelCase , **__lowerCAmelCase ) elif text is not None: _UpperCamelCase : List[str] = self.tokenizer(__lowerCAmelCase , **__lowerCAmelCase ) else: _UpperCamelCase : Dict = None if audio_target is not None: _UpperCamelCase : Optional[Any] = self.feature_extractor(audio_target=__lowerCAmelCase , *__lowerCAmelCase , sampling_rate=__lowerCAmelCase , **__lowerCAmelCase ) _UpperCamelCase : Optional[int] = targets["input_values"] elif text_target is not None: _UpperCamelCase : str = self.tokenizer(__lowerCAmelCase , **__lowerCAmelCase ) _UpperCamelCase : str = targets["input_ids"] else: _UpperCamelCase : Tuple = None if inputs is None: return targets if targets is not None: _UpperCamelCase : Optional[int] = labels _UpperCamelCase : Dict = targets.get("attention_mask" ) if decoder_attention_mask is not None: _UpperCamelCase : List[str] = decoder_attention_mask return inputs def snake_case__ ( self : Tuple , *lowercase__ : List[Any] , **lowercase__ : Optional[Any] ) ->int: '''simple docstring''' _UpperCamelCase : List[str] = kwargs.pop("input_values" , __lowerCAmelCase ) _UpperCamelCase : Dict = kwargs.pop("input_ids" , __lowerCAmelCase ) _UpperCamelCase : Optional[int] = kwargs.pop("labels" , __lowerCAmelCase ) if input_values is not None and input_ids is not None: raise ValueError("Cannot process both `input_values` and `input_ids` inputs." ) if input_values is None and input_ids is None and labels is None: raise ValueError( "You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded." ) if input_values is not None: _UpperCamelCase : List[str] = self.feature_extractor.pad(__lowerCAmelCase , *__lowerCAmelCase , **__lowerCAmelCase ) elif input_ids is not None: _UpperCamelCase : Dict = self.tokenizer.pad(__lowerCAmelCase , **__lowerCAmelCase ) else: _UpperCamelCase : int = None if labels is not None: if "input_ids" in labels or (isinstance(__lowerCAmelCase , __lowerCAmelCase ) and "input_ids" in labels[0]): _UpperCamelCase : str = self.tokenizer.pad(__lowerCAmelCase , **__lowerCAmelCase ) _UpperCamelCase : Tuple = targets["input_ids"] else: _UpperCamelCase : Optional[Any] = self.feature_extractor.feature_size _UpperCamelCase : Tuple = self.feature_extractor.num_mel_bins _UpperCamelCase : Optional[Any] = self.feature_extractor.pad(__lowerCAmelCase , *__lowerCAmelCase , **__lowerCAmelCase ) _UpperCamelCase : Optional[int] = feature_size_hack _UpperCamelCase : Tuple = targets["input_values"] else: _UpperCamelCase : Optional[Any] = None if inputs is None: return targets if targets is not None: _UpperCamelCase : Dict = labels _UpperCamelCase : Dict = targets.get("attention_mask" ) if decoder_attention_mask is not None: _UpperCamelCase : str = decoder_attention_mask return inputs def snake_case__ ( self : int , *lowercase__ : Optional[Any] , **lowercase__ : Union[str, Any] ) ->Union[str, Any]: '''simple docstring''' return self.tokenizer.batch_decode(*__lowerCAmelCase , **__lowerCAmelCase ) def snake_case__ ( self : List[str] , *lowercase__ : List[Any] , **lowercase__ : Any ) ->str: '''simple docstring''' return self.tokenizer.decode(*__lowerCAmelCase , **__lowerCAmelCase )
707
'''simple docstring''' from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging lowerCAmelCase_ : int = logging.get_logger(__name__) lowerCAmelCase_ : Any = { """EleutherAI/gpt-j-6B""": """https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json""", # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ): '''simple docstring''' UpperCAmelCase__ = '''gptj''' UpperCAmelCase__ = { '''max_position_embeddings''': '''n_positions''', '''hidden_size''': '''n_embd''', '''num_attention_heads''': '''n_head''', '''num_hidden_layers''': '''n_layer''', } def __init__( self : Optional[Any] , lowercase__ : Union[str, Any]=50_400 , lowercase__ : Union[str, Any]=2_048 , lowercase__ : Tuple=4_096 , lowercase__ : List[str]=28 , lowercase__ : Optional[int]=16 , lowercase__ : str=64 , lowercase__ : Any=None , lowercase__ : Any="gelu_new" , lowercase__ : Union[str, Any]=0.0 , lowercase__ : Optional[Any]=0.0 , lowercase__ : Any=0.0 , lowercase__ : Tuple=1e-5 , lowercase__ : Any=0.0_2 , lowercase__ : int=True , lowercase__ : int=50_256 , lowercase__ : Any=50_256 , lowercase__ : Tuple=False , **lowercase__ : str , ) ->Optional[Any]: '''simple docstring''' _UpperCamelCase : Dict = vocab_size _UpperCamelCase : List[str] = n_positions _UpperCamelCase : Union[str, Any] = n_embd _UpperCamelCase : Union[str, Any] = n_layer _UpperCamelCase : Optional[Any] = n_head _UpperCamelCase : Dict = n_inner _UpperCamelCase : Optional[Any] = rotary_dim _UpperCamelCase : Tuple = activation_function _UpperCamelCase : List[Any] = resid_pdrop _UpperCamelCase : Any = embd_pdrop _UpperCamelCase : Optional[Any] = attn_pdrop _UpperCamelCase : Optional[Any] = layer_norm_epsilon _UpperCamelCase : Union[str, Any] = initializer_range _UpperCamelCase : Optional[int] = use_cache _UpperCamelCase : str = bos_token_id _UpperCamelCase : Any = eos_token_id super().__init__( bos_token_id=lowercase__ , eos_token_id=lowercase__ , tie_word_embeddings=lowercase__ , **lowercase__ ) class SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : Optional[int] , lowercase__ : PretrainedConfig , lowercase__ : str = "default" , lowercase__ : List[PatchingSpec] = None , lowercase__ : bool = False , ) ->Union[str, Any]: '''simple docstring''' super().__init__(lowercase__ , task=lowercase__ , patching_specs=lowercase__ , use_past=lowercase__ ) if not getattr(self._config , "pad_token_id" , lowercase__ ): # TODO: how to do that better? _UpperCamelCase : Optional[int] = 0 @property def snake_case__ ( self : List[str] ) ->Mapping[str, Mapping[int, str]]: '''simple docstring''' _UpperCamelCase : List[str] = OrderedDict({"input_ids": {0: "batch", 1: "sequence"}} ) if self.use_past: self.fill_with_past_key_values_(lowercase__ , direction="inputs" ) _UpperCamelCase : str = {0: "batch", 1: "past_sequence + sequence"} else: _UpperCamelCase : Optional[Any] = {0: "batch", 1: "sequence"} return common_inputs @property def snake_case__ ( self : int ) ->int: '''simple docstring''' return self._config.n_layer @property def snake_case__ ( self : Dict ) ->int: '''simple docstring''' return self._config.n_head def snake_case__ ( self : int , lowercase__ : PreTrainedTokenizer , lowercase__ : int = -1 , lowercase__ : int = -1 , lowercase__ : bool = False , lowercase__ : Optional[TensorType] = None , ) ->Mapping[str, Any]: '''simple docstring''' _UpperCamelCase : int = super(lowercase__ , self ).generate_dummy_inputs( lowercase__ , batch_size=lowercase__ , seq_length=lowercase__ , is_pair=lowercase__ , framework=lowercase__ ) # We need to order the input in the way they appears in the forward() _UpperCamelCase : Tuple = OrderedDict({"input_ids": common_inputs["input_ids"]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch _UpperCamelCase , _UpperCamelCase : Optional[Any] = common_inputs["input_ids"].shape # Not using the same length for past_key_values _UpperCamelCase : Optional[int] = seqlen + 2 _UpperCamelCase : str = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _UpperCamelCase : Dict = [ (torch.zeros(lowercase__ ), torch.zeros(lowercase__ )) for _ in range(self.num_layers ) ] _UpperCamelCase : str = common_inputs["attention_mask"] if self.use_past: _UpperCamelCase : int = ordered_inputs["attention_mask"].dtype _UpperCamelCase : Optional[int] = torch.cat( [ordered_inputs["attention_mask"], torch.ones(lowercase__ , lowercase__ , dtype=lowercase__ )] , dim=1 ) return ordered_inputs @property def snake_case__ ( self : Tuple ) ->int: '''simple docstring''' return 13
204
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging a : Optional[Any] = logging.get_logger(__name__) a : Tuple = { 'google/vit-base-patch16-224': 'https://huggingface.co/vit-base-patch16-224/resolve/main/config.json', # See all ViT models at https://huggingface.co/models?filter=vit } class SCREAMING_SNAKE_CASE__ ( snake_case_ ): __SCREAMING_SNAKE_CASE = 'vit' def __init__( self : Optional[int] , a_ : str=768 , a_ : Optional[int]=12 , a_ : List[Any]=12 , a_ : Optional[int]=3_072 , a_ : Any="gelu" , a_ : Tuple=0.0 , a_ : List[str]=0.0 , a_ : List[str]=0.02 , a_ : Optional[Any]=1e-12 , a_ : Tuple=224 , a_ : int=16 , a_ : Tuple=3 , a_ : Tuple=True , a_ : Optional[Any]=16 , **a_ : List[str] , ): """simple docstring""" super().__init__(**a_ ) __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = initializer_range __snake_case = layer_norm_eps __snake_case = image_size __snake_case = patch_size __snake_case = num_channels __snake_case = qkv_bias __snake_case = encoder_stride class SCREAMING_SNAKE_CASE__ ( snake_case_ ): __SCREAMING_SNAKE_CASE = version.parse("""1.11""" ) @property def A ( self : Optional[int] ): """simple docstring""" return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def A ( self : Union[str, Any] ): """simple docstring""" return 1e-4
69
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING __a :List[str] = logging.get_logger(__name__) @add_end_docstrings(snake_case_ ) class _a ( snake_case_ ): """simple docstring""" def __init__( self : Any , **UpperCAmelCase : List[str] ): super().__init__(**UpperCAmelCase ) if self.framework == "tf": raise ValueError(f'''The {self.__class__} is only available in PyTorch.''' ) requires_backends(self , "vision" ) self.check_model_type(UpperCAmelCase ) def __call__( self : Optional[int] , UpperCAmelCase : Union[str, "Image.Image", List[Dict[str, Any]]] , UpperCAmelCase : Union[str, List[str]] = None , **UpperCAmelCase : List[Any] , ): if "text_queries" in kwargs: A_ = kwargs.pop("text_queries" ) if isinstance(UpperCAmelCase , (str, Image.Image) ): A_ = {"image": image, "candidate_labels": candidate_labels} else: A_ = image A_ = super().__call__(UpperCAmelCase , **UpperCAmelCase ) return results def __A ( self : int , **UpperCAmelCase : Tuple ): A_ = {} if "threshold" in kwargs: A_ = kwargs["threshold"] if "top_k" in kwargs: A_ = kwargs["top_k"] return {}, {}, postprocess_params def __A ( self : List[str] , UpperCAmelCase : Dict ): A_ = load_image(inputs["image"] ) A_ = inputs["candidate_labels"] if isinstance(UpperCAmelCase , UpperCAmelCase ): A_ = candidate_labels.split("," ) A_ = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(UpperCAmelCase ): A_ = self.tokenizer(UpperCAmelCase , return_tensors=self.framework ) A_ = self.image_processor(UpperCAmelCase , return_tensors=self.framework ) yield { "is_last": i == len(UpperCAmelCase ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def __A ( self : str , UpperCAmelCase : int ): A_ = model_inputs.pop("target_size" ) A_ = model_inputs.pop("candidate_label" ) A_ = model_inputs.pop("is_last" ) A_ = self.model(**UpperCAmelCase ) A_ = {"target_size": target_size, "candidate_label": candidate_label, "is_last": is_last, **outputs} return model_outputs def __A ( self : Dict , UpperCAmelCase : Any , UpperCAmelCase : Optional[int]=0.1 , UpperCAmelCase : Optional[int]=None ): A_ = [] for model_output in model_outputs: A_ = model_output["candidate_label"] A_ = BaseModelOutput(UpperCAmelCase ) A_ = self.image_processor.post_process_object_detection( outputs=UpperCAmelCase , threshold=UpperCAmelCase , target_sizes=model_output["target_size"] )[0] for index in outputs["scores"].nonzero(): A_ = outputs["scores"][index].item() A_ = self._get_bounding_box(outputs["boxes"][index][0] ) A_ = {"score": score, "label": label, "box": box} results.append(UpperCAmelCase ) A_ = sorted(UpperCAmelCase , key=lambda UpperCAmelCase : x["score"] , reverse=UpperCAmelCase ) if top_k: A_ = results[:top_k] return results def __A ( self : List[str] , UpperCAmelCase : "torch.Tensor" ): if self.framework != "pt": raise ValueError("The ZeroShotObjectDetectionPipeline is only available in PyTorch." ) A_ , A_ , A_ , A_ = box.int().tolist() A_ = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
86
0
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XGLMTokenizer, XGLMTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase : Optional[int] = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): lowercase = XGLMTokenizer lowercase = XGLMTokenizerFast lowercase = True lowercase = True def UpperCAmelCase ( self ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing __UpperCamelCase = XGLMTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = '<pad>' __UpperCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCAmelCase ) , __UpperCAmelCase ) def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(len(__UpperCAmelCase ) , 1008 ) def UpperCAmelCase ( self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1008 ) def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = XGLMTokenizer(__UpperCAmelCase , keep_accents=__UpperCAmelCase ) __UpperCamelCase = tokenizer.tokenize('This is a test' ) self.assertListEqual(__UpperCAmelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __UpperCamelCase = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) __UpperCamelCase = tokenizer.convert_tokens_to_ids(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) __UpperCamelCase = tokenizer.convert_ids_to_tokens(__UpperCAmelCase ) self.assertListEqual( __UpperCAmelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) @cached_property def UpperCAmelCase ( self ): '''simple docstring''' return XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) def UpperCAmelCase ( self ): '''simple docstring''' with tempfile.NamedTemporaryFile() as f: shutil.copyfile(__UpperCAmelCase , f.name ) __UpperCamelCase = XGLMTokenizer(f.name , keep_accents=__UpperCAmelCase ) __UpperCamelCase = pickle.dumps(__UpperCAmelCase ) pickle.loads(__UpperCAmelCase ) def UpperCAmelCase ( self ): '''simple docstring''' if not self.test_rust_tokenizer: return __UpperCamelCase = self.get_tokenizer() __UpperCamelCase = self.get_rust_tokenizer() __UpperCamelCase = 'I was born in 92000, and this is falsé.' __UpperCamelCase = tokenizer.tokenize(__UpperCAmelCase ) __UpperCamelCase = rust_tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) __UpperCamelCase = tokenizer.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) __UpperCamelCase = rust_tokenizer.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) __UpperCamelCase = self.get_rust_tokenizer() __UpperCamelCase = tokenizer.encode(__UpperCAmelCase ) __UpperCamelCase = rust_tokenizer.encode(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) @slow def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = 'Hello World!' __UpperCamelCase = [2, 3_1227, 4447, 35] self.assertListEqual(__UpperCAmelCase , self.big_tokenizer.encode(__UpperCAmelCase ) ) @slow def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to unk, such as saoneuhaoesuth' ) # fmt: off __UpperCamelCase = [2, 1018, 67, 11, 1988, 2617, 5631, 278, 11, 3407, 48, 7_1630, 2_8085, 4, 3234, 157, 13, 6, 5, 6, 4, 3526, 768, 15, 659, 57, 298, 3983, 864, 129, 21, 6, 5, 1_3675, 377, 652, 7580, 1_0341, 155, 2817, 422, 1666, 7, 1674, 53, 113, 20_2277, 1_7892, 33, 60, 87, 4, 3234, 157, 61, 2667, 5_2376, 19, 88, 23, 735] # fmt: on self.assertListEqual(__UpperCAmelCase , self.big_tokenizer.encode(__UpperCAmelCase ) ) @slow def UpperCAmelCase ( self ): '''simple docstring''' __UpperCamelCase = { 'input_ids': [[2, 10_8825, 1163, 15, 8_8010, 473, 1_5898, 157, 1_3672, 1857, 312, 8, 23_8021, 1163, 53, 1_3672, 1857, 312, 8, 5_3283, 18_2396, 8, 1_8566, 16, 3_6733, 4101, 8, 230, 24_4017, 12_2553, 7, 15, 13_2597, 4, 293, 1_2511, 7610, 4, 3414, 13_2597, 9, 4, 3_2361, 362, 4, 734, 2_8512, 3_2569, 18, 4, 3_2361, 2_6096, 1_4982, 73, 1_8715, 2_1433, 23_5261, 15, 492, 1_2427, 16, 53, 1_8715, 2_1433, 6_5454, 15, 2_3659, 563, 16, 278, 597, 2843, 595, 7931, 18_2396, 6_4186, 22, 886, 595, 13_2981, 53, 2_5540, 3449, 4_3982, 3_9901, 5951, 878, 330, 4, 2_7694, 8_0269, 312, 53, 6517, 1_1780, 611, 2_0408, 5], [2, 6, 13_2597, 67, 4_2897, 33, 592, 8, 16_3729, 2_5540, 361, 13_6997, 10_9514, 17_3230, 7, 501, 60, 10_2913, 196, 5631, 235, 6_3243, 473, 6, 23_1757, 74, 5277, 7905, 53, 3095, 3_7317, 22, 454, 18_3874, 5], [2, 268, 3_1298, 4_6530, 6, 13_2935, 4_3831, 7, 597, 32, 24, 3688, 9865, 5]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] } # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__UpperCAmelCase , model_name='facebook/xglm-564M' , padding=__UpperCAmelCase , )
293
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase : List[Any] = { "configuration_xlm_roberta": [ "XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLMRobertaConfig", "XLMRobertaOnnxConfig", ], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase : Optional[Any] = ["XLMRobertaTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase : int = ["XLMRobertaTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase : Optional[int] = [ "XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST", "XLMRobertaForCausalLM", "XLMRobertaForMaskedLM", "XLMRobertaForMultipleChoice", "XLMRobertaForQuestionAnswering", "XLMRobertaForSequenceClassification", "XLMRobertaForTokenClassification", "XLMRobertaModel", "XLMRobertaPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase : List[str] = [ "TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLMRobertaForCausalLM", "TFXLMRobertaForMaskedLM", "TFXLMRobertaForMultipleChoice", "TFXLMRobertaForQuestionAnswering", "TFXLMRobertaForSequenceClassification", "TFXLMRobertaForTokenClassification", "TFXLMRobertaModel", "TFXLMRobertaPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase : List[str] = [ "FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST", "FlaxXLMRobertaForMaskedLM", "FlaxXLMRobertaForCausalLM", "FlaxXLMRobertaForMultipleChoice", "FlaxXLMRobertaForQuestionAnswering", "FlaxXLMRobertaForSequenceClassification", "FlaxXLMRobertaForTokenClassification", "FlaxXLMRobertaModel", "FlaxXLMRobertaPreTrainedModel", ] if TYPE_CHECKING: from .configuration_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaConfig, XLMRobertaOnnxConfig, ) try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta import XLMRobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta_fast import XLMRobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaForCausalLM, XLMRobertaForMaskedLM, XLMRobertaForMultipleChoice, XLMRobertaForQuestionAnswering, XLMRobertaForSequenceClassification, XLMRobertaForTokenClassification, XLMRobertaModel, XLMRobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm_roberta import ( TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMRobertaForCausalLM, TFXLMRobertaForMaskedLM, TFXLMRobertaForMultipleChoice, TFXLMRobertaForQuestionAnswering, TFXLMRobertaForSequenceClassification, TFXLMRobertaForTokenClassification, TFXLMRobertaModel, TFXLMRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xlm_roberta import ( FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxXLMRobertaForCausalLM, FlaxXLMRobertaForMaskedLM, FlaxXLMRobertaForMultipleChoice, FlaxXLMRobertaForQuestionAnswering, FlaxXLMRobertaForSequenceClassification, FlaxXLMRobertaForTokenClassification, FlaxXLMRobertaModel, FlaxXLMRobertaPreTrainedModel, ) else: import sys UpperCamelCase : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
293
1
import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def __init__( self : Optional[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : List[str]=13 , __lowerCamelCase : List[str]=7 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Dict=True , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : Optional[int]=99 , __lowerCamelCase : Tuple=32 , __lowerCamelCase : Any=5 , __lowerCamelCase : int=4 , __lowerCamelCase : Any=37 , __lowerCamelCase : Any="gelu" , __lowerCamelCase : int=0.1 , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : Optional[Any]=512 , __lowerCamelCase : Union[str, Any]=16 , __lowerCamelCase : Any=2 , __lowerCamelCase : List[Any]=0.02 , __lowerCamelCase : int=4 , ): SCREAMING_SNAKE_CASE = parent SCREAMING_SNAKE_CASE = batch_size SCREAMING_SNAKE_CASE = seq_length SCREAMING_SNAKE_CASE = is_training SCREAMING_SNAKE_CASE = use_attention_mask SCREAMING_SNAKE_CASE = use_token_type_ids SCREAMING_SNAKE_CASE = use_labels SCREAMING_SNAKE_CASE = vocab_size SCREAMING_SNAKE_CASE = hidden_size SCREAMING_SNAKE_CASE = num_hidden_layers SCREAMING_SNAKE_CASE = num_attention_heads SCREAMING_SNAKE_CASE = intermediate_size SCREAMING_SNAKE_CASE = hidden_act SCREAMING_SNAKE_CASE = hidden_dropout_prob SCREAMING_SNAKE_CASE = attention_probs_dropout_prob SCREAMING_SNAKE_CASE = max_position_embeddings SCREAMING_SNAKE_CASE = type_vocab_size SCREAMING_SNAKE_CASE = type_sequence_label_size SCREAMING_SNAKE_CASE = initializer_range SCREAMING_SNAKE_CASE = num_choices def _snake_case ( self : str ): SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE = None if self.use_attention_mask: SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) SCREAMING_SNAKE_CASE = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _snake_case ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = config_and_inputs SCREAMING_SNAKE_CASE = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class _SCREAMING_SNAKE_CASE ( __snake_case , unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def _snake_case ( self : List[str] ): SCREAMING_SNAKE_CASE = FlaxAlbertModelTester(self ) @slow def _snake_case ( self : Any ): for model_class_name in self.all_model_classes: SCREAMING_SNAKE_CASE = model_class_name.from_pretrained("albert-base-v2" ) SCREAMING_SNAKE_CASE = model(np.ones((1, 1) ) ) self.assertIsNotNone(__lowerCamelCase ) @require_flax class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @slow def _snake_case ( self : str ): SCREAMING_SNAKE_CASE = FlaxAlbertModel.from_pretrained("albert-base-v2" ) SCREAMING_SNAKE_CASE = np.array([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) SCREAMING_SNAKE_CASE = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) SCREAMING_SNAKE_CASE = model(__lowerCamelCase , attention_mask=__lowerCamelCase )[0] SCREAMING_SNAKE_CASE = (1, 11, 768) self.assertEqual(output.shape , __lowerCamelCase ) SCREAMING_SNAKE_CASE = np.array( [[[-0.6_513, 1.5_035, -0.2_766], [-0.6_515, 1.5_046, -0.2_780], [-0.6_512, 1.5_049, -0.2_784]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , __lowerCamelCase , atol=1e-4 ) )
16
import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __A : Any = logging.get_logger(__name__) __A : Any = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } __A : Optional[Any] = { 'vocab_file': {'facebook/blenderbot-3B': 'https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json'}, 'merges_file': {'facebook/blenderbot-3B': 'https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt'}, 'tokenizer_config_file': { 'facebook/blenderbot-3B': 'https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json' }, } __A : Union[str, Any] = {'facebook/blenderbot-3B': 1_2_8} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def __a ( ): SCREAMING_SNAKE_CASE = ( list(range(ord("!" ) , ord("~" ) + 1 ) ) + list(range(ord("¡" ) , ord("¬" ) + 1 ) ) + list(range(ord("®" ) , ord("ÿ" ) + 1 ) ) ) SCREAMING_SNAKE_CASE = bs[:] SCREAMING_SNAKE_CASE = 0 for b in range(2**8 ): if b not in bs: bs.append(A__ ) cs.append(2**8 + n ) n += 1 SCREAMING_SNAKE_CASE = [chr(A__ ) for n in cs] return dict(zip(A__ , A__ ) ) def __a ( A__ : List[Any] ): SCREAMING_SNAKE_CASE = set() SCREAMING_SNAKE_CASE = word[0] for char in word[1:]: pairs.add((prev_char, char) ) SCREAMING_SNAKE_CASE = char return pairs class _SCREAMING_SNAKE_CASE ( __snake_case ): '''simple docstring''' lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = ["input_ids", "attention_mask"] def __init__( self : Union[str, Any] , __lowerCamelCase : Dict , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Any="replace" , __lowerCamelCase : Optional[Any]="<s>" , __lowerCamelCase : Optional[Any]="</s>" , __lowerCamelCase : Any="</s>" , __lowerCamelCase : Union[str, Any]="<s>" , __lowerCamelCase : List[str]="<unk>" , __lowerCamelCase : Optional[Any]="<pad>" , __lowerCamelCase : Dict="<mask>" , __lowerCamelCase : Any=False , **__lowerCamelCase : Optional[Any] , ): SCREAMING_SNAKE_CASE = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else bos_token SCREAMING_SNAKE_CASE = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else eos_token SCREAMING_SNAKE_CASE = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else sep_token SCREAMING_SNAKE_CASE = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else cls_token SCREAMING_SNAKE_CASE = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else unk_token SCREAMING_SNAKE_CASE = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it SCREAMING_SNAKE_CASE = AddedToken(__lowerCamelCase , lstrip=__lowerCamelCase , rstrip=__lowerCamelCase ) if isinstance(__lowerCamelCase , __lowerCamelCase ) else mask_token super().__init__( errors=__lowerCamelCase , bos_token=__lowerCamelCase , eos_token=__lowerCamelCase , unk_token=__lowerCamelCase , sep_token=__lowerCamelCase , cls_token=__lowerCamelCase , pad_token=__lowerCamelCase , mask_token=__lowerCamelCase , add_prefix_space=__lowerCamelCase , **__lowerCamelCase , ) with open(__lowerCamelCase , encoding="utf-8" ) as vocab_handle: SCREAMING_SNAKE_CASE = json.load(__lowerCamelCase ) SCREAMING_SNAKE_CASE = {v: k for k, v in self.encoder.items()} SCREAMING_SNAKE_CASE = errors # how to handle errors in decoding SCREAMING_SNAKE_CASE = bytes_to_unicode() SCREAMING_SNAKE_CASE = {v: k for k, v in self.byte_encoder.items()} with open(__lowerCamelCase , encoding="utf-8" ) as merges_handle: SCREAMING_SNAKE_CASE = merges_handle.read().split("\n" )[1:-1] SCREAMING_SNAKE_CASE = [tuple(merge.split() ) for merge in bpe_merges] SCREAMING_SNAKE_CASE = dict(zip(__lowerCamelCase , range(len(__lowerCamelCase ) ) ) ) SCREAMING_SNAKE_CASE = {} SCREAMING_SNAKE_CASE = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions SCREAMING_SNAKE_CASE = re.compile(r"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def _snake_case ( self : str ): return len(self.encoder ) def _snake_case ( self : Union[str, Any] ): return dict(self.encoder , **self.added_tokens_encoder ) def _snake_case ( self : Dict , __lowerCamelCase : List[Any] ): if token in self.cache: return self.cache[token] SCREAMING_SNAKE_CASE = tuple(__lowerCamelCase ) SCREAMING_SNAKE_CASE = get_pairs(__lowerCamelCase ) if not pairs: return token while True: SCREAMING_SNAKE_CASE = min(__lowerCamelCase , key=lambda __lowerCamelCase : self.bpe_ranks.get(__lowerCamelCase , float("inf" ) ) ) if bigram not in self.bpe_ranks: break SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = bigram SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = 0 while i < len(__lowerCamelCase ): try: SCREAMING_SNAKE_CASE = word.index(__lowerCamelCase , __lowerCamelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) SCREAMING_SNAKE_CASE = j if word[i] == first and i < len(__lowerCamelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 SCREAMING_SNAKE_CASE = tuple(__lowerCamelCase ) SCREAMING_SNAKE_CASE = new_word if len(__lowerCamelCase ) == 1: break else: SCREAMING_SNAKE_CASE = get_pairs(__lowerCamelCase ) SCREAMING_SNAKE_CASE = " ".join(__lowerCamelCase ) SCREAMING_SNAKE_CASE = word return word def _snake_case ( self : Optional[Any] , __lowerCamelCase : List[Any] ): SCREAMING_SNAKE_CASE = [] for token in re.findall(self.pat , __lowerCamelCase ): SCREAMING_SNAKE_CASE = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(__lowerCamelCase ).split(" " ) ) return bpe_tokens def _snake_case ( self : Tuple , __lowerCamelCase : Dict ): return self.encoder.get(__lowerCamelCase , self.encoder.get(self.unk_token ) ) def _snake_case ( self : Any , __lowerCamelCase : Optional[int] ): return self.decoder.get(__lowerCamelCase ) def _snake_case ( self : Optional[int] , __lowerCamelCase : List[Any] ): SCREAMING_SNAKE_CASE = "".join(__lowerCamelCase ) SCREAMING_SNAKE_CASE = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" , errors=self.errors ) return text def _snake_case ( self : Any , __lowerCamelCase : str , __lowerCamelCase : Optional[str] = None ): if not os.path.isdir(__lowerCamelCase ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return SCREAMING_SNAKE_CASE = os.path.join( __lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) SCREAMING_SNAKE_CASE = os.path.join( __lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(__lowerCamelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=__lowerCamelCase , ensure_ascii=__lowerCamelCase ) + "\n" ) SCREAMING_SNAKE_CASE = 0 with open(__lowerCamelCase , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda __lowerCamelCase : kv[1] ): if index != token_index: logger.warning( f"Saving vocabulary to {merge_file}: BPE merge indices are not consecutive." " Please check that the tokenizer is not corrupted!" ) SCREAMING_SNAKE_CASE = token_index writer.write(" ".join(__lowerCamelCase ) + "\n" ) index += 1 return vocab_file, merge_file def _snake_case ( self : Dict , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None , __lowerCamelCase : bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowerCamelCase , token_ids_a=__lowerCamelCase , already_has_special_tokens=__lowerCamelCase ) if token_ids_a is None: return [1] + ([0] * len(__lowerCamelCase )) + [1] return [1] + ([0] * len(__lowerCamelCase )) + [1, 1] + ([0] * len(__lowerCamelCase )) + [1] def _snake_case ( self : Optional[Any] , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None ): SCREAMING_SNAKE_CASE = [self.sep_token_id] SCREAMING_SNAKE_CASE = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _snake_case ( self : Optional[int] , __lowerCamelCase : Dict , __lowerCamelCase : Any=False , **__lowerCamelCase : Union[str, Any] ): SCREAMING_SNAKE_CASE = kwargs.pop("add_prefix_space" , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(__lowerCamelCase ) > 0 and not text[0].isspace()): SCREAMING_SNAKE_CASE = " " + text return (text, kwargs) def _snake_case ( self : Any , __lowerCamelCase : List[int] , __lowerCamelCase : Optional[List[int]] = None ): return token_ids_a + [self.eos_token_id] def _snake_case ( self : int , __lowerCamelCase : "Conversation" ): SCREAMING_SNAKE_CASE = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(" " + text ) else: # Generated responses should contain them already. inputs.append(__lowerCamelCase ) SCREAMING_SNAKE_CASE = " ".join(__lowerCamelCase ) SCREAMING_SNAKE_CASE = self.encode(__lowerCamelCase ) if len(__lowerCamelCase ) > self.model_max_length: SCREAMING_SNAKE_CASE = input_ids[-self.model_max_length :] logger.warning(f"Trimmed input from conversation as it was longer than {self.model_max_length} tokens." ) return input_ids
16
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging a = logging.get_logger(__name__) a = { "naver-clova-ix/donut-base": "https://huggingface.co/naver-clova-ix/donut-base/resolve/main/config.json", # See all Donut models at https://huggingface.co/models?filter=donut-swin } class __a ( _snake_case ): __UpperCamelCase : Union[str, Any] = 'donut-swin' __UpperCamelCase : Dict = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : List[str] ,lowerCamelCase : List[Any]=224 ,lowerCamelCase : List[str]=4 ,lowerCamelCase : str=3 ,lowerCamelCase : Tuple=96 ,lowerCamelCase : Dict=[2, 2, 6, 2] ,lowerCamelCase : str=[3, 6, 12, 24] ,lowerCamelCase : List[str]=7 ,lowerCamelCase : Dict=4.0 ,lowerCamelCase : str=True ,lowerCamelCase : Union[str, Any]=0.0 ,lowerCamelCase : Union[str, Any]=0.0 ,lowerCamelCase : str=0.1 ,lowerCamelCase : Dict="gelu" ,lowerCamelCase : Union[str, Any]=False ,lowerCamelCase : Dict=0.02 ,lowerCamelCase : List[Any]=1E-5 ,**lowerCamelCase : Any ,): '''simple docstring''' super().__init__(**lowerCamelCase ) __SCREAMING_SNAKE_CASE = image_size __SCREAMING_SNAKE_CASE = patch_size __SCREAMING_SNAKE_CASE = num_channels __SCREAMING_SNAKE_CASE = embed_dim __SCREAMING_SNAKE_CASE = depths __SCREAMING_SNAKE_CASE = len(lowerCamelCase ) __SCREAMING_SNAKE_CASE = num_heads __SCREAMING_SNAKE_CASE = window_size __SCREAMING_SNAKE_CASE = mlp_ratio __SCREAMING_SNAKE_CASE = qkv_bias __SCREAMING_SNAKE_CASE = hidden_dropout_prob __SCREAMING_SNAKE_CASE = attention_probs_dropout_prob __SCREAMING_SNAKE_CASE = drop_path_rate __SCREAMING_SNAKE_CASE = hidden_act __SCREAMING_SNAKE_CASE = use_absolute_embeddings __SCREAMING_SNAKE_CASE = layer_norm_eps __SCREAMING_SNAKE_CASE = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model __SCREAMING_SNAKE_CASE = int(embed_dim * 2 ** (len(lowerCamelCase ) - 1) )
13
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a = {"configuration_sew": ["SEW_PRETRAINED_CONFIG_ARCHIVE_MAP", "SEWConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a = [ "SEW_PRETRAINED_MODEL_ARCHIVE_LIST", "SEWForCTC", "SEWForSequenceClassification", "SEWModel", "SEWPreTrainedModel", ] if TYPE_CHECKING: from .configuration_sew import SEW_PRETRAINED_CONFIG_ARCHIVE_MAP, SEWConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_sew import ( SEW_PRETRAINED_MODEL_ARCHIVE_LIST, SEWForCTC, SEWForSequenceClassification, SEWModel, SEWPreTrainedModel, ) else: import sys a = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
13
1
'''simple docstring''' import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = "https://openaipublic.azureedge.net/jukebox/models/" SCREAMING_SNAKE_CASE_ = { "jukebox-1b-lyrics": [ "5b/vqvae.pth.tar", "5b/prior_level_0.pth.tar", "5b/prior_level_1.pth.tar", "1b_lyrics/prior_level_2.pth.tar", ], "jukebox-5b-lyrics": [ "5b/vqvae.pth.tar", "5b/prior_level_0.pth.tar", "5b/prior_level_1.pth.tar", "5b_lyrics/prior_level_2.pth.tar", ], } def lowerCamelCase__ ( a__) -> Dict: """simple docstring""" if key.endswith('.model.1.bias') and len(key.split('.')) > 1_0: _snake_case : List[str] = key.replace('.model.1.bias' , '.conv1d_1.bias') elif key.endswith('.model.1.weight') and len(key.split('.')) > 1_0: _snake_case : Any = key.replace('.model.1.weight' , '.conv1d_1.weight') elif key.endswith('.model.3.bias') and len(key.split('.')) > 1_0: _snake_case : List[str] = key.replace('.model.3.bias' , '.conv1d_2.bias') elif key.endswith('.model.3.weight') and len(key.split('.')) > 1_0: _snake_case : Any = key.replace('.model.3.weight' , '.conv1d_2.weight') if "conditioner_blocks.0." in key: _snake_case : Tuple = key.replace('conditioner_blocks.0' , 'conditioner_blocks') if "prime_prior" in key: _snake_case : Any = key.replace('prime_prior' , 'encoder') if ".emb." in key and "total" not in key and "absolute" not in key and "relative" not in key: _snake_case : Optional[Any] = key.replace('.emb.' , '.') if key.endswith('k'): # replace vqvae.X.k with vqvae.X.codebook return key.replace('.k' , '.codebook') if "y_emb." in key: return key.replace('y_emb.' , 'metadata_embedding.') if "x_emb.emb." in key: _snake_case : Optional[Any] = key.replace('0.x_emb.emb' , 'embed_tokens') if "prime_state_ln" in key: return key.replace('prime_state_ln' , 'encoder.final_layer_norm') if ".ln" in key: return key.replace('.ln' , '.layer_norm') if "_ln" in key: return key.replace('_ln' , '_layer_norm') if "prime_state_proj" in key: return key.replace('prime_state_proj' , 'encoder.proj_in') if "prime_x_out" in key: return key.replace('prime_x_out' , 'encoder.lm_head') if "prior.x_out" in key: return key.replace('x_out' , 'fc_proj_out') if "x_emb" in key: return key.replace('x_emb' , 'embed_tokens') return key def lowerCamelCase__ ( a__ , a__ , a__ , a__) -> Union[str, Any]: """simple docstring""" _snake_case : List[str] = {} import re _snake_case : Tuple = re.compile(R'encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)') _snake_case : List[str] = re.compile( R'encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)') _snake_case : Union[str, Any] = re.compile(R'encoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)') _snake_case : Dict = re.compile(R'decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)') _snake_case : str = re.compile( R'decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)') _snake_case : Any = re.compile(R'decoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)') _snake_case : List[str] = re.compile(R'conditioner_blocks.(\d*).cond.model.(\d*).(\d).(bias|weight)') _snake_case : int = re.compile( R'conditioner_blocks.(\d*).cond.model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)') _snake_case : Dict = re.compile(R'conditioner_blocks.(\d*).cond.model.(\d*).(bias|weight)') for original_key, value in state_dict.items(): # rename vqvae.encoder keys if re_encoder_block_conv_in.fullmatch(a__): _snake_case : str = re_encoder_block_conv_in.match(a__) _snake_case : Union[str, Any] = regex_match.groups() _snake_case : Any = int(groups[2]) * 2 + int(groups[3]) _snake_case : Tuple = F"""encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.{groups[-1]}""" _snake_case : Tuple = re_encoder_block_conv_in.sub(a__ , a__) elif re_encoder_block_resnet.fullmatch(a__): _snake_case : Dict = re_encoder_block_resnet.match(a__) _snake_case : List[Any] = regex_match.groups() _snake_case : str = int(groups[2]) * 2 + int(groups[3]) _snake_case : Dict = {'1': 1, '3': 2}[groups[-2]] _snake_case : Any = F"""encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.""" _snake_case : Tuple = F"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" _snake_case : Any = prefix + resnet_block _snake_case : Any = re_encoder_block_resnet.sub(a__ , a__) elif re_encoder_block_proj_out.fullmatch(a__): _snake_case : Tuple = re_encoder_block_proj_out.match(a__) _snake_case : Any = regex_match.groups() _snake_case : Tuple = F"""encoders.{groups[0]}.level_blocks.{groups[1]}.proj_out.{groups[-1]}""" _snake_case : Optional[Any] = re_encoder_block_proj_out.sub(a__ , a__) # rename vqvae.decoder keys elif re_decoder_block_conv_out.fullmatch(a__): _snake_case : Tuple = re_decoder_block_conv_out.match(a__) _snake_case : Union[str, Any] = regex_match.groups() _snake_case : Tuple = int(groups[2]) * 2 + int(groups[3]) - 2 _snake_case : Optional[Any] = F"""decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.{groups[-1]}""" _snake_case : str = re_decoder_block_conv_out.sub(a__ , a__) elif re_decoder_block_resnet.fullmatch(a__): _snake_case : Tuple = re_decoder_block_resnet.match(a__) _snake_case : str = regex_match.groups() _snake_case : Dict = int(groups[2]) * 2 + int(groups[3]) - 2 _snake_case : List[str] = {'1': 1, '3': 2}[groups[-2]] _snake_case : List[str] = F"""decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.""" _snake_case : Union[str, Any] = F"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" _snake_case : Dict = prefix + resnet_block _snake_case : str = re_decoder_block_resnet.sub(a__ , a__) elif re_decoder_block_proj_in.fullmatch(a__): _snake_case : str = re_decoder_block_proj_in.match(a__) _snake_case : Optional[Any] = regex_match.groups() _snake_case : List[Any] = F"""decoders.{groups[0]}.level_blocks.{groups[1]}.proj_in.{groups[-1]}""" _snake_case : Optional[int] = re_decoder_block_proj_in.sub(a__ , a__) # rename prior cond.model to upsampler.upsample_block and resnet elif re_prior_cond_conv_out.fullmatch(a__): _snake_case : List[str] = re_prior_cond_conv_out.match(a__) _snake_case : Dict = regex_match.groups() _snake_case : Optional[int] = int(groups[1]) * 2 + int(groups[2]) - 2 _snake_case : List[str] = F"""conditioner_blocks.upsampler.upsample_block.{block_index}.{groups[-1]}""" _snake_case : Any = re_prior_cond_conv_out.sub(a__ , a__) elif re_prior_cond_resnet.fullmatch(a__): _snake_case : int = re_prior_cond_resnet.match(a__) _snake_case : Optional[Any] = regex_match.groups() _snake_case : Optional[Any] = int(groups[1]) * 2 + int(groups[2]) - 2 _snake_case : Union[str, Any] = {'1': 1, '3': 2}[groups[-2]] _snake_case : Tuple = F"""conditioner_blocks.upsampler.upsample_block.{block_index}.""" _snake_case : Any = F"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" _snake_case : Optional[Any] = prefix + resnet_block _snake_case : Dict = re_prior_cond_resnet.sub(a__ , a__) elif re_prior_cond_proj_in.fullmatch(a__): _snake_case : str = re_prior_cond_proj_in.match(a__) _snake_case : Union[str, Any] = regex_match.groups() _snake_case : List[Any] = F"""conditioner_blocks.upsampler.proj_in.{groups[-1]}""" _snake_case : int = re_prior_cond_proj_in.sub(a__ , a__) # keep original key else: _snake_case : int = original_key _snake_case : str = replace_key(a__) if F"""{key_prefix}.{key}""" not in model_state_dict or key is None: print(F"""failed converting {original_key} to {key}, does not match""") # handle missmatched shape elif value.shape != model_state_dict[F"""{key_prefix}.{key}"""].shape: _snake_case : List[str] = model_state_dict[F"""{key_prefix}.{key}"""] print(F"""{original_key}-> {key} : \nshape {val.shape} and { value.shape}, do not match""") _snake_case : Dict = original_key _snake_case : Tuple = original_key _snake_case : List[str] = value return new_dict @torch.no_grad() def lowerCamelCase__ ( a__=None , a__=None) -> Optional[int]: """simple docstring""" for file in MODEL_MAPPING[model_name]: if not os.path.isfile(F"""{pytorch_dump_folder_path}/{file.split('/')[-1]}"""): _snake_case : Tuple = requests.get(F"""{PREFIX}{file}""" , allow_redirects=a__) os.makedirs(F"""{pytorch_dump_folder_path}/""" , exist_ok=a__) open(F"""{pytorch_dump_folder_path}/{file.split('/')[-1]}""" , 'wb').write(r.content) _snake_case : Any = MODEL_MAPPING[model_name.split('/')[-1]] _snake_case : Tuple = JukeboxConfig.from_pretrained(a__) _snake_case : Dict = JukeboxModel(a__) _snake_case : Any = [] _snake_case : Optional[Any] = {} for i, dict_name in enumerate(a__): _snake_case : int = torch.load(F"""{pytorch_dump_folder_path}/{dict_name.split('/')[-1]}""")['model'] _snake_case : Optional[Any] = {} for k in old_dic.keys(): if k.endswith('.b'): _snake_case : List[str] = old_dic[k] elif k.endswith('.w'): _snake_case : Optional[Any] = old_dic[k] elif "level_2" not in dict_name and "cond.model." in k: _snake_case : Any = old_dic[k] else: _snake_case : str = old_dic[k] _snake_case : Union[str, Any] = 'vqvae' if i == 0 else F"""priors.{3 - i}""" _snake_case : List[Any] = fix_jukebox_keys(a__ , model.state_dict() , a__ , a__) weight_dict.append(a__) _snake_case : int = weight_dict.pop(0) model.vqvae.load_state_dict(a__) for i in range(len(a__)): model.priors[i].load_state_dict(weight_dict[2 - i]) Path(a__).mkdir(exist_ok=a__) with open(F"""{pytorch_dump_folder_path}/mapping.json""" , 'w') as txtfile: json.dump(a__ , a__) print(F"""Saving model {model_name} to {pytorch_dump_folder_path}""") model.save_pretrained(a__) return weight_dict if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="jukebox-5b-lyrics", type=str, help="Name of the model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default="jukebox-5b-lyrics-converted", type=str, help="Path to the output PyTorch model directory.", ) SCREAMING_SNAKE_CASE_ = parser.parse_args() convert_openai_checkpoint(args.model_name, args.pytorch_dump_folder_path)
517
'''simple docstring''' # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers SCREAMING_SNAKE_CASE_ = "3" print("Python version:", sys.version) print("transformers version:", transformers.__version__) try: import torch print("Torch version:", torch.__version__) print("Cuda available:", torch.cuda.is_available()) print("Cuda version:", torch.version.cuda) print("CuDNN version:", torch.backends.cudnn.version()) print("Number of GPUs available:", torch.cuda.device_count()) print("NCCL version:", torch.cuda.nccl.version()) except ImportError: print("Torch version:", None) try: import deepspeed print("DeepSpeed version:", deepspeed.__version__) except ImportError: print("DeepSpeed version:", None) try: import tensorflow as tf print("TensorFlow version:", tf.__version__) print("TF GPUs available:", bool(tf.config.list_physical_devices("GPU"))) print("Number of TF GPUs available:", len(tf.config.list_physical_devices("GPU"))) except ImportError: print("TensorFlow version:", None)
517
1
from __future__ import annotations import unittest from transformers import BlenderbotSmallConfig, BlenderbotSmallTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel @require_tf class UpperCAmelCase__ : snake_case_ = BlenderbotSmallConfig snake_case_ = {} snake_case_ = '''gelu''' def __init__( self , A__ , A__=13 , A__=7 , A__=True , A__=False , A__=99 , A__=32 , A__=2 , A__=4 , A__=37 , A__=0.1 , A__=0.1 , A__=20 , A__=2 , A__=1 , A__=0 , ): """simple docstring""" UpperCAmelCase_: Union[str, Any] = parent UpperCAmelCase_: Any = batch_size UpperCAmelCase_: Optional[Any] = seq_length UpperCAmelCase_: Any = is_training UpperCAmelCase_: str = use_labels UpperCAmelCase_: Dict = vocab_size UpperCAmelCase_: Any = hidden_size UpperCAmelCase_: Tuple = num_hidden_layers UpperCAmelCase_: Optional[int] = num_attention_heads UpperCAmelCase_: Optional[int] = intermediate_size UpperCAmelCase_: Dict = hidden_dropout_prob UpperCAmelCase_: List[Any] = attention_probs_dropout_prob UpperCAmelCase_: Any = max_position_embeddings UpperCAmelCase_: int = eos_token_id UpperCAmelCase_: List[Any] = pad_token_id UpperCAmelCase_: Optional[Any] = bos_token_id def snake_case_ ( self ): """simple docstring""" UpperCAmelCase_: Optional[Any] = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) UpperCAmelCase_: List[str] = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) UpperCAmelCase_: List[str] = tf.concat([input_ids, eos_tensor] , axis=1 ) UpperCAmelCase_: Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase_: Optional[int] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) UpperCAmelCase_: Optional[int] = prepare_blenderbot_small_inputs_dict(A__ , A__ , A__ ) return config, inputs_dict def snake_case_ ( self , A__ , A__ ): """simple docstring""" UpperCAmelCase_: Dict = TFBlenderbotSmallModel(config=A__ ).get_decoder() UpperCAmelCase_: List[str] = inputs_dict["input_ids"] UpperCAmelCase_: Optional[int] = input_ids[:1, :] UpperCAmelCase_: int = inputs_dict["attention_mask"][:1, :] UpperCAmelCase_: Union[str, Any] = inputs_dict["head_mask"] UpperCAmelCase_: Tuple = 1 # first forward pass UpperCAmelCase_: Union[str, Any] = model(A__ , attention_mask=A__ , head_mask=A__ , use_cache=A__ ) UpperCAmelCase_: List[Any] = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCAmelCase_: int = ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCAmelCase_: int = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and UpperCAmelCase_: Dict = tf.concat([input_ids, next_tokens] , axis=-1 ) UpperCAmelCase_: Union[str, Any] = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) UpperCAmelCase_: Union[str, Any] = model(A__ , attention_mask=A__ )[0] UpperCAmelCase_: Any = model(A__ , attention_mask=A__ , past_key_values=A__ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice UpperCAmelCase_: List[str] = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) UpperCAmelCase_: Tuple = output_from_no_past[:, -3:, random_slice_idx] UpperCAmelCase_: Optional[int] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(A__ , A__ , rtol=1E-3 ) def lowercase ( _a ,_a ,_a ,_a=None ,_a=None ,_a=None ,_a=None ,_a=None ,) -> Tuple: if attention_mask is None: UpperCAmelCase_: Union[str, Any] = tf.cast(tf.math.not_equal(_a ,config.pad_token_id ) ,tf.inta ) if decoder_attention_mask is None: UpperCAmelCase_: Optional[Any] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape ,dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] ,config.pad_token_id ) ,tf.inta ), ] ,axis=-1 ,) if head_mask is None: UpperCAmelCase_: int = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: UpperCAmelCase_: Any = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: UpperCAmelCase_: Any = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class UpperCAmelCase__ ( snake_case__ , snake_case__ , unittest.TestCase ): snake_case_ = ( (TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel) if is_tf_available() else () ) snake_case_ = (TFBlenderbotSmallForConditionalGeneration,) if is_tf_available() else () snake_case_ = ( { '''conversational''': TFBlenderbotSmallForConditionalGeneration, '''feature-extraction''': TFBlenderbotSmallModel, '''summarization''': TFBlenderbotSmallForConditionalGeneration, '''text2text-generation''': TFBlenderbotSmallForConditionalGeneration, '''translation''': TFBlenderbotSmallForConditionalGeneration, } if is_tf_available() else {} ) snake_case_ = True snake_case_ = False snake_case_ = False def snake_case_ ( self ): """simple docstring""" UpperCAmelCase_: Any = TFBlenderbotSmallModelTester(self ) UpperCAmelCase_: List[Any] = ConfigTester(self , config_class=A__ ) def snake_case_ ( self ): """simple docstring""" self.config_tester.run_common_tests() def snake_case_ ( self ): """simple docstring""" UpperCAmelCase_: Dict = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*A__ ) @require_tokenizers @require_tf class UpperCAmelCase__ ( unittest.TestCase ): snake_case_ = [ '''Social anxiety\nWow, I am never shy. Do you have anxiety?\nYes. I end up sweating and blushing and feel like ''' ''' i\'m going to throw up.\nand why is that?''' ] snake_case_ = '''facebook/blenderbot_small-90M''' @cached_property def snake_case_ ( self ): """simple docstring""" return BlenderbotSmallTokenizer.from_pretrained("facebook/blenderbot-90M" ) @cached_property def snake_case_ ( self ): """simple docstring""" UpperCAmelCase_: str = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def snake_case_ ( self ): """simple docstring""" UpperCAmelCase_: str = self.tokenizer(self.src_text , return_tensors="tf" ) UpperCAmelCase_: Tuple = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=A__ , ) UpperCAmelCase_: str = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=A__ )[0] assert generated_words in ( "i don't know. i just feel like i'm going to throw up. it's not fun.", "i'm not sure. i just feel like i've been feeling like i have to be in a certain place", "i'm not sure. i just feel like i've been in a bad situation.", )
704
from __future__ import annotations _lowerCAmelCase = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0] _lowerCAmelCase = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1] def lowercase ( _a ) -> list[float]: UpperCAmelCase_: Dict = [] UpperCAmelCase_: List[Any] = len(_a ) for i in range(_a ): UpperCAmelCase_: float = -1 for j in range(i + 1 ,_a ): if arr[i] < arr[j]: UpperCAmelCase_: List[str] = arr[j] break result.append(_a ) return result def lowercase ( _a ) -> list[float]: UpperCAmelCase_: List[Any] = [] for i, outer in enumerate(_a ): UpperCAmelCase_: float = -1 for inner in arr[i + 1 :]: if outer < inner: UpperCAmelCase_: Union[str, Any] = inner break result.append(_a ) return result def lowercase ( _a ) -> list[float]: UpperCAmelCase_: Union[str, Any] = len(_a ) UpperCAmelCase_: list[float] = [] UpperCAmelCase_: list[float] = [-1] * arr_size for index in reversed(range(_a ) ): if stack: while stack[-1] <= arr[index]: stack.pop() if not stack: break if stack: UpperCAmelCase_: Union[str, Any] = stack[-1] stack.append(arr[index] ) return result if __name__ == "__main__": from doctest import testmod from timeit import timeit testmod() print(next_greatest_element_slow(arr)) print(next_greatest_element_fast(arr)) print(next_greatest_element(arr)) _lowerCAmelCase = ( """from __main__ import arr, next_greatest_element_slow, """ """next_greatest_element_fast, next_greatest_element""" ) print( """next_greatest_element_slow():""", timeit("""next_greatest_element_slow(arr)""", setup=setup), ) print( """next_greatest_element_fast():""", timeit("""next_greatest_element_fast(arr)""", setup=setup), ) print( """ next_greatest_element():""", timeit("""next_greatest_element(arr)""", setup=setup), )
306
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A = {'''configuration_xlnet''': ['''XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLNetConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['''XLNetTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['''XLNetTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ '''XLNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLNetForMultipleChoice''', '''XLNetForQuestionAnswering''', '''XLNetForQuestionAnsweringSimple''', '''XLNetForSequenceClassification''', '''XLNetForTokenClassification''', '''XLNetLMHeadModel''', '''XLNetModel''', '''XLNetPreTrainedModel''', '''load_tf_weights_in_xlnet''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ '''TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLNetForMultipleChoice''', '''TFXLNetForQuestionAnsweringSimple''', '''TFXLNetForSequenceClassification''', '''TFXLNetForTokenClassification''', '''TFXLNetLMHeadModel''', '''TFXLNetMainLayer''', '''TFXLNetModel''', '''TFXLNetPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
52
'''simple docstring''' def lowerCamelCase__ ( a ): if number < 0: raise ValueError('number must not be negative' ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
356
0
def lowerCamelCase_(lowerCamelCase_ ) -> List[str]: UpperCAmelCase = set() # To detect a back edge, keep track of vertices currently in the recursion stack UpperCAmelCase = set() return any( node not in visited and depth_first_search(__lowercase , __lowercase , __lowercase , __lowercase ) for node in graph ) def lowerCamelCase_(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> Any: visited.add(__lowercase ) rec_stk.add(__lowercase ) for node in graph[vertex]: if node not in visited: if depth_first_search(__lowercase , __lowercase , __lowercase , __lowercase ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(__lowercase ) return False if __name__ == "__main__": from doctest import testmod testmod()
704
def lowerCamelCase_(lowerCamelCase_ = 600_851_475_143 ) -> int: try: UpperCAmelCase = int(lowerCamelCase_ ) except (TypeError, ValueError): raise TypeError("Parameter n must be int or castable to int." ) if n <= 0: raise ValueError("Parameter n must be greater than or equal to one." ) UpperCAmelCase = 1 UpperCAmelCase = 2 while i * i <= n: while n % i == 0: UpperCAmelCase = i n //= i i += 1 if n > 1: UpperCAmelCase = n return int(lowerCamelCase_ ) if __name__ == "__main__": print(F'''{solution() = }''')
457
0
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model __A : Optional[Any] = '0.12' # assumed parallelism: 8 if is_torch_available(): import torch def __a ( A__ : Any , A__ : Tuple , A__ : str=None ): if rng is None: SCREAMING_SNAKE_CASE = random.Random() SCREAMING_SNAKE_CASE = 1 for dim in shape: total_dims *= dim SCREAMING_SNAKE_CASE = [] for _ in range(A__ ): values.append(rng.randint(0 , vocab_size - 1 ) ) SCREAMING_SNAKE_CASE = np.array(A__ , dtype=jnp.intaa ).reshape(A__ ) return output def __a ( A__ : Union[str, Any] , A__ : Union[str, Any]=None ): SCREAMING_SNAKE_CASE = ids_tensor(A__ , vocab_size=2 , rng=A__ ) # make sure that at least one token is attended to for each batch SCREAMING_SNAKE_CASE = 1 return attn_mask @require_flax class _SCREAMING_SNAKE_CASE : '''simple docstring''' lowerCamelCase__ = None lowerCamelCase__ = () def _snake_case ( self : List[Any] ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 SCREAMING_SNAKE_CASE = 2 SCREAMING_SNAKE_CASE = inputs["input_ids"].shape[-1] // 2 SCREAMING_SNAKE_CASE = inputs["input_ids"][:max_batch_size, :sequence_length] SCREAMING_SNAKE_CASE = jnp.ones_like(__lowerCamelCase ) SCREAMING_SNAKE_CASE = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens SCREAMING_SNAKE_CASE = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` SCREAMING_SNAKE_CASE = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def _snake_case ( self : List[str] ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = max_length SCREAMING_SNAKE_CASE = 0 for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE = model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE = model_class.__name__[4:] # Skip the "Flax" at the beginning SCREAMING_SNAKE_CASE = getattr(__lowerCamelCase , __lowerCamelCase ) SCREAMING_SNAKE_CASE = pt_model_class(__lowerCamelCase ).eval() SCREAMING_SNAKE_CASE = load_flax_weights_in_pytorch_model(__lowerCamelCase , flax_model.params ) SCREAMING_SNAKE_CASE = flax_model.generate(__lowerCamelCase ).sequences SCREAMING_SNAKE_CASE = pt_model.generate(torch.tensor(__lowerCamelCase , dtype=torch.long ) ) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: SCREAMING_SNAKE_CASE = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() , flax_generation_outputs.tolist() ) def _snake_case ( self : List[Any] ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = max_length for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE = model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE = model.generate(__lowerCamelCase ).sequences self.assertEqual(generation_outputs.shape[-1] , __lowerCamelCase ) SCREAMING_SNAKE_CASE = jit(model.generate ) SCREAMING_SNAKE_CASE = jit_generate(__lowerCamelCase ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _snake_case ( self : Tuple ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = max_length for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE = model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE = model.generate(__lowerCamelCase ).sequences self.assertEqual(generation_outputs.shape[-1] , __lowerCamelCase ) SCREAMING_SNAKE_CASE = jit(model.generate ) SCREAMING_SNAKE_CASE = jit_generate(__lowerCamelCase ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _snake_case ( self : List[Any] ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = max_length SCREAMING_SNAKE_CASE = 2 for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE = model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE = model.generate(__lowerCamelCase ).sequences self.assertEqual(generation_outputs.shape[-1] , __lowerCamelCase ) SCREAMING_SNAKE_CASE = jit(model.generate ) SCREAMING_SNAKE_CASE = jit_generate(__lowerCamelCase ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _snake_case ( self : Any ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = max_length SCREAMING_SNAKE_CASE = 2 SCREAMING_SNAKE_CASE = 2 for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE = model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE = model.generate(__lowerCamelCase ).sequences self.assertEqual(generation_outputs.shape[0] , input_ids.shape[0] * config.num_return_sequences ) def _snake_case ( self : Dict ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = max_length SCREAMING_SNAKE_CASE = 0.8 SCREAMING_SNAKE_CASE = 10 SCREAMING_SNAKE_CASE = 0.3 SCREAMING_SNAKE_CASE = 1 SCREAMING_SNAKE_CASE = 8 SCREAMING_SNAKE_CASE = 9 for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE = model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE = model.generate(__lowerCamelCase ).sequences self.assertEqual(generation_outputs.shape[-1] , __lowerCamelCase ) SCREAMING_SNAKE_CASE = jit(model.generate ) SCREAMING_SNAKE_CASE = jit_generate(__lowerCamelCase ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _snake_case ( self : str ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE = max_length SCREAMING_SNAKE_CASE = 1 SCREAMING_SNAKE_CASE = 8 SCREAMING_SNAKE_CASE = 9 for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE = model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE = model.generate(__lowerCamelCase ).sequences self.assertEqual(generation_outputs.shape[-1] , __lowerCamelCase ) SCREAMING_SNAKE_CASE = jit(model.generate ) SCREAMING_SNAKE_CASE = jit_generate(__lowerCamelCase ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _snake_case ( self : int ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self._get_input_ids_and_config() SCREAMING_SNAKE_CASE = max_length SCREAMING_SNAKE_CASE = 2 SCREAMING_SNAKE_CASE = 1 SCREAMING_SNAKE_CASE = 8 SCREAMING_SNAKE_CASE = 9 for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE = model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE = model.generate(__lowerCamelCase ).sequences self.assertEqual(generation_outputs.shape[-1] , __lowerCamelCase ) SCREAMING_SNAKE_CASE = jit(model.generate ) SCREAMING_SNAKE_CASE = jit_generate(__lowerCamelCase ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _snake_case ( self : Any ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self._get_input_ids_and_config() # pad attention mask on the left SCREAMING_SNAKE_CASE = attention_mask.at[(0, 0)].set(0 ) SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = max_length for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE = model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE = model.generate(__lowerCamelCase , attention_mask=__lowerCamelCase ).sequences self.assertEqual(generation_outputs.shape[-1] , __lowerCamelCase ) SCREAMING_SNAKE_CASE = jit(model.generate ) SCREAMING_SNAKE_CASE = jit_generate(__lowerCamelCase , attention_mask=__lowerCamelCase ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _snake_case ( self : str ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self._get_input_ids_and_config() # pad attention mask on the left SCREAMING_SNAKE_CASE = attention_mask.at[(0, 0)].set(0 ) SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = max_length for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE = model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE = model.generate(__lowerCamelCase , attention_mask=__lowerCamelCase ).sequences self.assertEqual(generation_outputs.shape[-1] , __lowerCamelCase ) SCREAMING_SNAKE_CASE = jit(model.generate ) SCREAMING_SNAKE_CASE = jit_generate(__lowerCamelCase , attention_mask=__lowerCamelCase ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _snake_case ( self : int ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = self._get_input_ids_and_config() # pad attention mask on the left SCREAMING_SNAKE_CASE = attention_mask.at[(0, 0)].set(0 ) SCREAMING_SNAKE_CASE = 2 SCREAMING_SNAKE_CASE = max_length for model_class in self.all_generative_model_classes: SCREAMING_SNAKE_CASE = model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE = model.generate(__lowerCamelCase , attention_mask=__lowerCamelCase ).sequences self.assertEqual(generation_outputs.shape[-1] , __lowerCamelCase ) SCREAMING_SNAKE_CASE = jit(model.generate ) SCREAMING_SNAKE_CASE = jit_generate(__lowerCamelCase , attention_mask=__lowerCamelCase ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) @require_flax class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def _snake_case ( self : Dict ): SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-bert" ) SCREAMING_SNAKE_CASE = FlaxAutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-bert-flax-only" ) SCREAMING_SNAKE_CASE = "Hello world" SCREAMING_SNAKE_CASE = tokenizer(__lowerCamelCase , return_tensors="np" ).input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(__lowerCamelCase , "do_samples" ): model.generate(__lowerCamelCase , do_samples=__lowerCamelCase ) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(__lowerCamelCase , "foo" ): SCREAMING_SNAKE_CASE = {"foo": "bar"} model.generate(__lowerCamelCase , **__lowerCamelCase )
16
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL a = logging.get_logger(__name__) def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> str: '''simple docstring''' __SCREAMING_SNAKE_CASE = b.T __SCREAMING_SNAKE_CASE = np.sum(np.square(__UpperCAmelCase ) , axis=1 ) __SCREAMING_SNAKE_CASE = np.sum(np.square(__UpperCAmelCase ) , axis=0 ) __SCREAMING_SNAKE_CASE = np.matmul(__UpperCAmelCase , __UpperCAmelCase ) __SCREAMING_SNAKE_CASE = aa[:, None] - 2 * ab + ba[None, :] return d def __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase ) -> List[Any]: '''simple docstring''' __SCREAMING_SNAKE_CASE = x.reshape(-1 , 3 ) __SCREAMING_SNAKE_CASE = squared_euclidean_distance(__UpperCAmelCase , __UpperCAmelCase ) return np.argmin(__UpperCAmelCase , axis=1 ) class __a ( _snake_case ): __UpperCamelCase : Any = ['pixel_values'] def __init__( self : Any ,lowerCamelCase : Optional[Union[List[List[int]], np.ndarray]] = None ,lowerCamelCase : bool = True ,lowerCamelCase : Dict[str, int] = None ,lowerCamelCase : PILImageResampling = PILImageResampling.BILINEAR ,lowerCamelCase : bool = True ,lowerCamelCase : bool = True ,**lowerCamelCase : Optional[Any] ,): '''simple docstring''' super().__init__(**lowerCamelCase ) __SCREAMING_SNAKE_CASE = size if size is not None else {"""height""": 256, """width""": 256} __SCREAMING_SNAKE_CASE = get_size_dict(lowerCamelCase ) __SCREAMING_SNAKE_CASE = np.array(lowerCamelCase ) if clusters is not None else None __SCREAMING_SNAKE_CASE = do_resize __SCREAMING_SNAKE_CASE = size __SCREAMING_SNAKE_CASE = resample __SCREAMING_SNAKE_CASE = do_normalize __SCREAMING_SNAKE_CASE = do_color_quantize def UpperCAmelCase__ ( self : Optional[int] ,lowerCamelCase : np.ndarray ,lowerCamelCase : Dict[str, int] ,lowerCamelCase : PILImageResampling = PILImageResampling.BILINEAR ,lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ,**lowerCamelCase : int ,): '''simple docstring''' __SCREAMING_SNAKE_CASE = get_size_dict(lowerCamelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""Size dictionary must contain both height and width keys. Got {size.keys()}""" ) return resize( lowerCamelCase ,size=(size["""height"""], size["""width"""]) ,resample=lowerCamelCase ,data_format=lowerCamelCase ,**lowerCamelCase ) def UpperCAmelCase__ ( self : Optional[Any] ,lowerCamelCase : np.ndarray ,lowerCamelCase : Optional[Union[str, ChannelDimension]] = None ,): '''simple docstring''' __SCREAMING_SNAKE_CASE = rescale(image=lowerCamelCase ,scale=1 / 127.5 ,data_format=lowerCamelCase ) __SCREAMING_SNAKE_CASE = image - 1 return image def UpperCAmelCase__ ( self : Dict ,lowerCamelCase : ImageInput ,lowerCamelCase : bool = None ,lowerCamelCase : Dict[str, int] = None ,lowerCamelCase : PILImageResampling = None ,lowerCamelCase : bool = None ,lowerCamelCase : Optional[bool] = None ,lowerCamelCase : Optional[Union[List[List[int]], np.ndarray]] = None ,lowerCamelCase : Optional[Union[str, TensorType]] = None ,lowerCamelCase : Optional[Union[str, ChannelDimension]] = ChannelDimension.FIRST ,**lowerCamelCase : Any ,): '''simple docstring''' __SCREAMING_SNAKE_CASE = do_resize if do_resize is not None else self.do_resize __SCREAMING_SNAKE_CASE = size if size is not None else self.size __SCREAMING_SNAKE_CASE = get_size_dict(lowerCamelCase ) __SCREAMING_SNAKE_CASE = resample if resample is not None else self.resample __SCREAMING_SNAKE_CASE = do_normalize if do_normalize is not None else self.do_normalize __SCREAMING_SNAKE_CASE = do_color_quantize if do_color_quantize is not None else self.do_color_quantize __SCREAMING_SNAKE_CASE = clusters if clusters is not None else self.clusters __SCREAMING_SNAKE_CASE = np.array(lowerCamelCase ) __SCREAMING_SNAKE_CASE = make_list_of_images(lowerCamelCase ) if not valid_images(lowerCamelCase ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None or resample is None: raise ValueError("""Size and resample must be specified if do_resize is True.""" ) if do_color_quantize and clusters is None: raise ValueError("""Clusters must be specified if do_color_quantize is True.""" ) # All transformations expect numpy arrays. __SCREAMING_SNAKE_CASE = [to_numpy_array(lowerCamelCase ) for image in images] if do_resize: __SCREAMING_SNAKE_CASE = [self.resize(image=lowerCamelCase ,size=lowerCamelCase ,resample=lowerCamelCase ) for image in images] if do_normalize: __SCREAMING_SNAKE_CASE = [self.normalize(image=lowerCamelCase ) for image in images] if do_color_quantize: __SCREAMING_SNAKE_CASE = [to_channel_dimension_format(lowerCamelCase ,ChannelDimension.LAST ) for image in images] # color quantize from (batch_size, height, width, 3) to (batch_size, height, width) __SCREAMING_SNAKE_CASE = np.array(lowerCamelCase ) __SCREAMING_SNAKE_CASE = color_quantize(lowerCamelCase ,lowerCamelCase ).reshape(images.shape[:-1] ) # flatten to (batch_size, height*width) __SCREAMING_SNAKE_CASE = images.shape[0] __SCREAMING_SNAKE_CASE = images.reshape(lowerCamelCase ,-1 ) # We need to convert back to a list of images to keep consistent behaviour across processors. __SCREAMING_SNAKE_CASE = list(lowerCamelCase ) else: __SCREAMING_SNAKE_CASE = [to_channel_dimension_format(lowerCamelCase ,lowerCamelCase ) for image in images] __SCREAMING_SNAKE_CASE = {"""input_ids""": images} return BatchFeature(data=lowerCamelCase ,tensor_type=lowerCamelCase )
109
0
'''simple docstring''' def __UpperCamelCase ( _lowerCAmelCase = 6008_5147_5143 ) -> int: """simple docstring""" try: A : Optional[int] = int(SCREAMING_SNAKE_CASE_ ) except (TypeError, ValueError): raise TypeError("""Parameter n must be int or castable to int.""" ) if n <= 0: raise ValueError("""Parameter n must be greater than or equal to one.""" ) A : int = 1 A : List[str] = 2 while i * i <= n: while n % i == 0: A : Union[str, Any] = i n //= i i += 1 if n > 1: A : int = n return int(SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": print(F"""{solution() = }""")
703
import argparse import re from pathlib import Path import requests import torch from PIL import Image from torchvision.transforms import CenterCrop, Compose, Normalize, Resize, ToTensor from transformers import ( EfficientFormerConfig, EfficientFormerForImageClassificationWithTeacher, EfficientFormerImageProcessor, ) from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ) -> Union[str, Any]: """simple docstring""" A : Union[str, Any] = old_name if "patch_embed" in old_name: A , A , A : Optional[Any] = old_name.split(""".""" ) if layer == "0": A : Optional[int] = old_name.replace("""0""" , """convolution1""" ) elif layer == "1": A : List[str] = old_name.replace("""1""" , """batchnorm_before""" ) elif layer == "3": A : Optional[int] = old_name.replace("""3""" , """convolution2""" ) else: A : List[Any] = old_name.replace("""4""" , """batchnorm_after""" ) if "network" in old_name and re.search(R"""\d\.\d""" , _lowerCAmelCase ): A : Union[str, Any] = R"""\b\d{2}\b""" if bool(re.search(_lowerCAmelCase , _lowerCAmelCase ) ): A : int = re.search(R"""\d\.\d\d.""" , _lowerCAmelCase ).group() else: A : int = re.search(R"""\d\.\d.""" , _lowerCAmelCase ).group() if int(match[0] ) < 6: A : Any = old_name.replace(_lowerCAmelCase , """""" ) A : Dict = trimmed_name.replace("""network""" , match[0] + """.meta4D_layers.blocks.""" + match[2:-1] ) A : int = """intermediate_stages.""" + trimmed_name else: A : Tuple = old_name.replace(_lowerCAmelCase , """""" ) if int(match[2] ) < num_meta4D_last_stage: A : Optional[Any] = trimmed_name.replace("""network""" , """meta4D_layers.blocks.""" + match[2] ) else: A : int = str(int(match[2] ) - num_meta4D_last_stage ) A : Optional[Any] = trimmed_name.replace("""network""" , """meta3D_layers.blocks.""" + layer_index ) if "norm1" in old_name: A : Optional[Any] = trimmed_name.replace("""norm1""" , """layernorm1""" ) elif "norm2" in old_name: A : Optional[int] = trimmed_name.replace("""norm2""" , """layernorm2""" ) elif "fc1" in old_name: A : Dict = trimmed_name.replace("""fc1""" , """linear_in""" ) elif "fc2" in old_name: A : Optional[Any] = trimmed_name.replace("""fc2""" , """linear_out""" ) A : Dict = """last_stage.""" + trimmed_name elif "network" in old_name and re.search(R""".\d.""" , _lowerCAmelCase ): A : Union[str, Any] = old_name.replace("""network""" , """intermediate_stages""" ) if "fc" in new_name: A : Any = new_name.replace("""fc""" , """convolution""" ) elif ("norm1" in new_name) and ("layernorm1" not in new_name): A : Union[str, Any] = new_name.replace("""norm1""" , """batchnorm_before""" ) elif ("norm2" in new_name) and ("layernorm2" not in new_name): A : Tuple = new_name.replace("""norm2""" , """batchnorm_after""" ) if "proj" in new_name: A : Tuple = new_name.replace("""proj""" , """projection""" ) if "dist_head" in new_name: A : int = new_name.replace("""dist_head""" , """distillation_classifier""" ) elif "head" in new_name: A : Dict = new_name.replace("""head""" , """classifier""" ) elif "patch_embed" in new_name: A : int = """efficientformer.""" + new_name elif new_name == "norm.weight" or new_name == "norm.bias": A : Any = new_name.replace("""norm""" , """layernorm""" ) A : int = """efficientformer.""" + new_name else: A : int = """efficientformer.encoder.""" + new_name return new_name def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ) -> str: """simple docstring""" for key in checkpoint.copy().keys(): A : Union[str, Any] = checkpoint.pop(_lowerCAmelCase ) A : Optional[Any] = val return checkpoint def __UpperCamelCase ( ) -> Union[str, Any]: """simple docstring""" A : Tuple = """http://images.cocodataset.org/val2017/000000039769.jpg""" A : int = Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ) return image def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> List[Any]: """simple docstring""" A : int = torch.load(_lowerCAmelCase , map_location="""cpu""" )["""model"""] A : Tuple = EfficientFormerConfig.from_json_file(_lowerCAmelCase ) A : Optional[int] = EfficientFormerForImageClassificationWithTeacher(_lowerCAmelCase ) A : int = """_""".join(checkpoint_path.split("""/""" )[-1].split(""".""" )[0].split("""_""" )[:-1] ) A : Optional[int] = config.depths[-1] - config.num_metaad_blocks + 1 A : Dict = convert_torch_checkpoint(_lowerCAmelCase , _lowerCAmelCase ) model.load_state_dict(_lowerCAmelCase ) model.eval() A : Dict = { """bilinear""": PILImageResampling.BILINEAR, """bicubic""": PILImageResampling.BICUBIC, """nearest""": PILImageResampling.NEAREST, } # prepare image A : str = prepare_img() A : int = 256 A : Any = 224 A : List[str] = EfficientFormerImageProcessor( size={"""shortest_edge""": image_size} , crop_size={"""height""": crop_size, """width""": crop_size} , resample=pillow_resamplings["""bicubic"""] , ) A : int = processor(images=_lowerCAmelCase , return_tensors="""pt""" ).pixel_values # original processing pipeline A : Dict = Compose( [ Resize(_lowerCAmelCase , interpolation=pillow_resamplings["""bicubic"""] ), CenterCrop(_lowerCAmelCase ), ToTensor(), Normalize(_lowerCAmelCase , _lowerCAmelCase ), ] ) A : Any = image_transforms(_lowerCAmelCase ).unsqueeze(0 ) assert torch.allclose(_lowerCAmelCase , _lowerCAmelCase ) A : Any = model(_lowerCAmelCase ) A : Any = outputs.logits A : Any = (1, 1000) if "l1" in model_name: A : Tuple = torch.Tensor( [-0.1_312, 0.4_353, -1.0_499, -0.5_124, 0.4_183, -0.6_793, -1.3_777, -0.0_893, -0.7_358, -2.4_328] ) assert torch.allclose(logits[0, :10] , _lowerCAmelCase , atol=1e-3 ) assert logits.shape == expected_shape elif "l3" in model_name: A : List[Any] = torch.Tensor( [-1.3_150, -1.5_456, -1.2_556, -0.8_496, -0.7_127, -0.7_897, -0.9_728, -0.3_052, 0.3_751, -0.3_127] ) assert torch.allclose(logits[0, :10] , _lowerCAmelCase , atol=1e-3 ) assert logits.shape == expected_shape elif "l7" in model_name: A : str = torch.Tensor( [-1.0_283, -1.4_131, -0.5_644, -1.3_115, -0.5_785, -1.2_049, -0.7_528, 0.1_992, -0.3_822, -0.0_878] ) assert logits.shape == expected_shape else: raise ValueError( f'''Unknown model checkpoint: {checkpoint_path}. Supported version of efficientformer are l1, l3 and l7''' ) # Save Checkpoints Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) model.save_pretrained(_lowerCAmelCase ) print(f'''Checkpoint successfuly converted. Model saved at {pytorch_dump_path}''' ) processor.save_pretrained(_lowerCAmelCase ) print(f'''Processor successfuly saved at {pytorch_dump_path}''' ) if push_to_hub: print("""Pushing model to the hub...""" ) model.push_to_hub( repo_id=f'''Bearnardd/{pytorch_dump_path}''' , commit_message="""Add model""" , use_temp_dir=_lowerCAmelCase , ) processor.push_to_hub( repo_id=f'''Bearnardd/{pytorch_dump_path}''' , commit_message="""Add image processor""" , use_temp_dir=_lowerCAmelCase , ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_:Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--pytorch_model_path""", default=None, type=str, required=True, help="""Path to EfficientFormer pytorch checkpoint.""", ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help="""The json file for EfficientFormer model config.""", ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument("""--push_to_hub""", action="""store_true""", help="""Push model and image processor to the hub""") parser.add_argument( """--no-push_to_hub""", dest="""push_to_hub""", action="""store_false""", help="""Do not push model and image processor to the hub""", ) parser.set_defaults(push_to_hub=True) SCREAMING_SNAKE_CASE_:Union[str, Any] = parser.parse_args() convert_efficientformer_checkpoint( checkpoint_path=args.pytorch_model_path, efficientformer_config_file=args.config_file, pytorch_dump_path=args.pytorch_dump_path, push_to_hub=args.push_to_hub, )
520
0
import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor __lowercase = logging.get_logger(__name__) class _lowercase ( __lowerCamelCase ): def __init__( self : Optional[Any] , *lowerCamelCase__ : Optional[int] , **lowerCamelCase__ : List[str] ) -> None: """simple docstring""" warnings.warn( '''The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use CLIPImageProcessor instead.''' , lowerCamelCase__ , ) super().__init__(*lowerCamelCase__ , **lowerCamelCase__ )
203
import logging import os from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional from tqdm import auto as tqdm_lib __lowercase = { """debug""": logging.DEBUG, """info""": logging.INFO, """warning""": logging.WARNING, """error""": logging.ERROR, """critical""": logging.CRITICAL, } __lowercase = logging.WARNING def _lowerCamelCase ( ): '''simple docstring''' A_ = os.getenv('''DATASETS_VERBOSITY''' , SCREAMING_SNAKE_CASE ) if env_level_str: if env_level_str in log_levels: return log_levels[env_level_str] else: logging.getLogger().warning( f"Unknown option DATASETS_VERBOSITY={env_level_str}, " f"has to be one of: { ', '.join(log_levels.keys() ) }" ) return _default_log_level def _lowerCamelCase ( ): '''simple docstring''' return __name__.split('''.''' )[0] def _lowerCamelCase ( ): '''simple docstring''' return logging.getLogger(_get_library_name() ) def _lowerCamelCase ( ): '''simple docstring''' A_ = _get_library_root_logger() library_root_logger.setLevel(_get_default_logging_level() ) def _lowerCamelCase ( ): '''simple docstring''' A_ = _get_library_root_logger() library_root_logger.setLevel(logging.NOTSET ) def _lowerCamelCase ( SCREAMING_SNAKE_CASE = None ): '''simple docstring''' if name is None: A_ = _get_library_name() return logging.getLogger(SCREAMING_SNAKE_CASE ) def _lowerCamelCase ( ): '''simple docstring''' return _get_library_root_logger().getEffectiveLevel() def _lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' _get_library_root_logger().setLevel(SCREAMING_SNAKE_CASE ) def _lowerCamelCase ( ): '''simple docstring''' return set_verbosity(SCREAMING_SNAKE_CASE ) def _lowerCamelCase ( ): '''simple docstring''' return set_verbosity(SCREAMING_SNAKE_CASE ) def _lowerCamelCase ( ): '''simple docstring''' return set_verbosity(SCREAMING_SNAKE_CASE ) def _lowerCamelCase ( ): '''simple docstring''' return set_verbosity(SCREAMING_SNAKE_CASE ) def _lowerCamelCase ( ): '''simple docstring''' A_ = False def _lowerCamelCase ( ): '''simple docstring''' A_ = True # Configure the library root logger at the module level (singleton-like) _configure_library_root_logger() class _lowercase : def __init__( self : Optional[int] , *lowerCamelCase__ : int , **lowerCamelCase__ : Tuple ) -> Dict: # pylint: disable=unused-argument """simple docstring""" A_ = args[0] if args else None def __iter__( self : Optional[int] ) -> List[str]: """simple docstring""" return iter(self._iterator ) def __getattr__( self : int , lowerCamelCase__ : List[str] ) -> List[str]: """simple docstring""" def empty_fn(*lowerCamelCase__ : Optional[Any] , **lowerCamelCase__ : str ): # pylint: disable=unused-argument return return empty_fn def __enter__( self : Tuple ) -> Tuple: """simple docstring""" return self def __exit__( self : Tuple , lowerCamelCase__ : List[Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Optional[Any] ) -> Optional[int]: """simple docstring""" return __lowercase = True class _lowercase : def __call__( self : Optional[int] , *lowerCamelCase__ : Optional[int] , lowerCamelCase__ : List[str]=False , **lowerCamelCase__ : Optional[Any] ) -> Optional[Any]: """simple docstring""" if _tqdm_active and not disable: return tqdm_lib.tqdm(*lowerCamelCase__ , **lowerCamelCase__ ) else: return EmptyTqdm(*lowerCamelCase__ , **lowerCamelCase__ ) def UpperCamelCase ( self : Optional[Any] , *lowerCamelCase__ : Any , **lowerCamelCase__ : int ) -> str: """simple docstring""" A_ = None if _tqdm_active: return tqdm_lib.tqdm.set_lock(*lowerCamelCase__ , **lowerCamelCase__ ) def UpperCamelCase ( self : Optional[Any] ) -> int: """simple docstring""" if _tqdm_active: return tqdm_lib.tqdm.get_lock() __lowercase = _tqdm_cls() def _lowerCamelCase ( ): '''simple docstring''' global _tqdm_active return bool(_tqdm_active ) def _lowerCamelCase ( ): '''simple docstring''' global _tqdm_active A_ = True def _lowerCamelCase ( ): '''simple docstring''' global _tqdm_active A_ = False
203
1
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import platform import numpy as np import psutil import torch from accelerate import __version__ as version from accelerate.commands.config import default_config_file, load_config_from_file from ..utils import is_npu_available, is_xpu_available def lowercase_ ( __snake_case : Optional[Any]=None ) -> str: '''simple docstring''' if subparsers is not None: snake_case__ :Optional[Any] = subparsers.add_parser("env" ) else: snake_case__ :List[str] = argparse.ArgumentParser("Accelerate env command" ) parser.add_argument( "--config_file" , default=__snake_case , help="The config file to use for the default values in the launching script." ) if subparsers is not None: parser.set_defaults(func=__snake_case ) return parser def lowercase_ ( __snake_case : int ) -> int: '''simple docstring''' snake_case__ :Dict = torch.__version__ snake_case__ :int = torch.cuda.is_available() snake_case__ :str = is_xpu_available() snake_case__ :List[str] = is_npu_available() snake_case__ :List[str] = "Not found" # Get the default from the config file. if args.config_file is not None or os.path.isfile(__snake_case ): snake_case__ :List[Any] = load_config_from_file(args.config_file ).to_dict() snake_case__ :Union[str, Any] = { "`Accelerate` version": version, "Platform": platform.platform(), "Python version": platform.python_version(), "Numpy version": np.__version__, "PyTorch version (GPU?)": F'{pt_version} ({pt_cuda_available})', "PyTorch XPU available": str(__snake_case ), "PyTorch NPU available": str(__snake_case ), "System RAM": F'{psutil.virtual_memory().total / 10_24 ** 3:.2f} GB', } if pt_cuda_available: snake_case__ :Dict = torch.cuda.get_device_name() print("\nCopy-and-paste the text below in your GitHub issue\n" ) print("\n".join([F'- {prop}: {val}' for prop, val in info.items()] ) ) print("- `Accelerate` default config:" if args.config_file is None else "- `Accelerate` config passed:" ) snake_case__ :List[str] = ( "\n".join([F'\t- {prop}: {val}' for prop, val in accelerate_config.items()] ) if isinstance(__snake_case , __snake_case ) else F'\t{accelerate_config}' ) print(__snake_case ) snake_case__ :str = accelerate_config return info def lowercase_ ( ) -> int: '''simple docstring''' snake_case__ :Union[str, Any] = env_command_parser() snake_case__ :Union[str, Any] = parser.parse_args() env_command(__snake_case ) return 0 if __name__ == "__main__": raise SystemExit(main())
710
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available __UpperCAmelCase : Tuple = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase : List[Any] = ["BartphoTokenizer"] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys __UpperCAmelCase : Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
57
0