code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean __SCREAMING_SNAKE_CASE : Any = 0 __SCREAMING_SNAKE_CASE : Dict = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] __SCREAMING_SNAKE_CASE : Dict = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right __SCREAMING_SNAKE_CASE : Union[str, Any] = tuple[int, int] class __A : '''simple docstring''' def __init__( self : Dict , UpperCAmelCase_ : int , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Union[str, Any] , ) ->List[Any]: """simple docstring""" snake_case_ = pos_x snake_case_ = pos_y snake_case_ = (pos_y, pos_x) snake_case_ = goal_x snake_case_ = goal_y snake_case_ = g_cost snake_case_ = parent snake_case_ = self.calculate_heuristic() snake_case_ = self.g_cost + self.h_cost def lowerCAmelCase ( self : int ) ->Union[str, Any]: """simple docstring""" snake_case_ = self.pos_x - self.goal_x snake_case_ = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowerCAmelCase_ ) + abs(lowerCAmelCase_ ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self : str , UpperCAmelCase_ : List[str] ) ->Union[str, Any]: """simple docstring""" return self.f_cost < other.f_cost class __A : '''simple docstring''' def __init__( self : Tuple , UpperCAmelCase_ : int , UpperCAmelCase_ : str ) ->Dict: """simple docstring""" snake_case_ = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , lowerCAmelCase_ ) snake_case_ = Node(goal[1] , goal[0] , goal[1] , goal[0] , 99_999 , lowerCAmelCase_ ) snake_case_ = [self.start] snake_case_ = [] snake_case_ = False def lowerCAmelCase ( self : Union[str, Any] ) ->List[Any]: """simple docstring""" while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() snake_case_ = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(lowerCAmelCase_ ) self.closed_nodes.append(lowerCAmelCase_ ) snake_case_ = self.get_successors(lowerCAmelCase_ ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowerCAmelCase_ ) else: # retrieve the best current path snake_case_ = self.open_nodes.pop(self.open_nodes.index(lowerCAmelCase_ ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowerCAmelCase_ ) else: self.open_nodes.append(lowerCAmelCase_ ) return [self.start.pos] def lowerCAmelCase ( self : Dict , UpperCAmelCase_ : int ) ->List[Any]: """simple docstring""" snake_case_ = [] for action in delta: snake_case_ = parent.pos_x + action[1] snake_case_ = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowerCAmelCase_ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowerCAmelCase_ , lowerCAmelCase_ , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , lowerCAmelCase_ , ) ) return successors def lowerCAmelCase ( self : List[str] , UpperCAmelCase_ : Optional[Any] ) ->Union[str, Any]: """simple docstring""" snake_case_ = node snake_case_ = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) snake_case_ = current_node.parent path.reverse() return path class __A : '''simple docstring''' def __init__( self : List[Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Union[str, Any] ) ->Optional[Any]: """simple docstring""" snake_case_ = AStar(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ = AStar(lowerCAmelCase_ , lowerCAmelCase_ ) snake_case_ = False def lowerCAmelCase ( self : List[str] ) ->Any: """simple docstring""" while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() snake_case_ = self.fwd_astar.open_nodes.pop(0 ) snake_case_ = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowerCAmelCase_ , lowerCAmelCase_ ) self.fwd_astar.closed_nodes.append(lowerCAmelCase_ ) self.bwd_astar.closed_nodes.append(lowerCAmelCase_ ) snake_case_ = current_bwd_node snake_case_ = current_fwd_node snake_case_ = { self.fwd_astar: self.fwd_astar.get_successors(lowerCAmelCase_ ), self.bwd_astar: self.bwd_astar.get_successors(lowerCAmelCase_ ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowerCAmelCase_ ) else: # retrieve the best current path snake_case_ = astar.open_nodes.pop( astar.open_nodes.index(lowerCAmelCase_ ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowerCAmelCase_ ) else: astar.open_nodes.append(lowerCAmelCase_ ) return [self.fwd_astar.start.pos] def lowerCAmelCase ( self : Union[str, Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : int ) ->List[str]: """simple docstring""" snake_case_ = self.fwd_astar.retrace_path(lowerCAmelCase_ ) snake_case_ = self.bwd_astar.retrace_path(lowerCAmelCase_ ) bwd_path.pop() bwd_path.reverse() snake_case_ = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] __SCREAMING_SNAKE_CASE : Dict = (0, 0) __SCREAMING_SNAKE_CASE : str = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) __SCREAMING_SNAKE_CASE : Dict = time.time() __SCREAMING_SNAKE_CASE : Tuple = AStar(init, goal) __SCREAMING_SNAKE_CASE : Optional[int] = a_star.search() __SCREAMING_SNAKE_CASE : str = time.time() - start_time print(f"""AStar execution time = {end_time:f} seconds""") __SCREAMING_SNAKE_CASE : List[str] = time.time() __SCREAMING_SNAKE_CASE : int = BidirectionalAStar(init, goal) __SCREAMING_SNAKE_CASE : str = time.time() - bd_start_time print(f"""BidirectionalAStar execution time = {bd_end_time:f} seconds""")
347
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A = 1_000_000 ) -> int: _snake_case = limit + 1 _snake_case = [0] * limit for first_term in range(1 , __A ): for n in range(__A , __A , __A ): _snake_case = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a _snake_case = sum(1 for x in frequency[1:limit] if x == 10 ) return count if __name__ == "__main__": print(F'''{solution() = }''')
42
0
from __future__ import annotations from functools import lru_cache from math import ceil UpperCamelCase = 100 UpperCamelCase = set(range(3, NUM_PRIMES, 2)) primes.add(2) UpperCamelCase = 42 for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=100) def lowercase_ ( _lowerCamelCase : Any): if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} lowercase__ : List[Any] = set() lowercase__ : Any = 42 lowercase__ : Union[str, Any] = 42 for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime): ret.add(sub * prime) return ret def lowercase_ ( _lowerCamelCase : Optional[Any] = 5000): for number_to_partition in range(1 , __A): if len(partition(__A)) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(f"{solution() = }")
87
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowercase : Tuple = { "configuration_xlm": ["XLM_PRETRAINED_CONFIG_ARCHIVE_MAP", "XLMConfig", "XLMOnnxConfig"], "tokenization_xlm": ["XLMTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : int = [ "XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "XLMForMultipleChoice", "XLMForQuestionAnswering", "XLMForQuestionAnsweringSimple", "XLMForSequenceClassification", "XLMForTokenClassification", "XLMModel", "XLMPreTrainedModel", "XLMWithLMHeadModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ "TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST", "TFXLMForMultipleChoice", "TFXLMForQuestionAnsweringSimple", "TFXLMForSequenceClassification", "TFXLMForTokenClassification", "TFXLMMainLayer", "TFXLMModel", "TFXLMPreTrainedModel", "TFXLMWithLMHeadModel", ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys lowercase : Union[str, Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
from pathlib import Path import json import tempfile from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration from transformers.models.fsmt.tokenization_fsmt import VOCAB_FILES_NAMES a__ = "tiny-wmt19-en-ru" # Build # borrowed from a test a__ = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "w</w>", "r</w>", "t</w>", "lo", "low", "er</w>", "low</w>", "lowest</w>", "newer</w>", "wider</w>", "<unk>", ] a__ = dict(zip(vocab, range(len(vocab)))) a__ = ["l o 123", "lo w 1456", "e r</w> 1789", ""] with tempfile.TemporaryDirectory() as tmpdirname: a__ = Path(tmpdirname) a__ = build_dir / VOCAB_FILES_NAMES["src_vocab_file"] a__ = build_dir / VOCAB_FILES_NAMES["tgt_vocab_file"] a__ = build_dir / VOCAB_FILES_NAMES["merges_file"] with open(src_vocab_file, """w""") as fp: fp.write(json.dumps(vocab_tokens)) with open(tgt_vocab_file, """w""") as fp: fp.write(json.dumps(vocab_tokens)) with open(merges_file, """w""") as fp: fp.write("""\n""".join(merges)) a__ = FSMTTokenizer( langs=["""en""", """ru"""], src_vocab_size=len(vocab), tgt_vocab_size=len(vocab), src_vocab_file=src_vocab_file, tgt_vocab_file=tgt_vocab_file, merges_file=merges_file, ) a__ = FSMTConfig( langs=["""ru""", """en"""], src_vocab_size=10_00, tgt_vocab_size=10_00, d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) a__ = FSMTForConditionalGeneration(config) print(F'''num of params {tiny_model.num_parameters()}''') # Test a__ = tokenizer(["""Making tiny model"""], return_tensors="""pt""") a__ = tiny_model(**batch) print("""test output:""", len(outputs.logits[0])) # Save tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F'''Generated {mname_tiny}''') # Upload # transformers-cli upload tiny-wmt19-en-ru
317
'''simple docstring''' from collections import defaultdict from math import gcd def SCREAMING_SNAKE_CASE__ ( __A = 1_500_000 ) -> int: _snake_case = defaultdict(__A ) _snake_case = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1 , __A , 2 ): if gcd(__A , __A ) > 1: continue _snake_case = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(__A , limit + 1 , __A ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
from scipy.stats import spearmanr import datasets lowerCAmelCase_ = "\nThe Spearman rank-order correlation coefficient is a measure of the\nrelationship between two datasets. Like other correlation coefficients,\nthis one varies between -1 and +1 with 0 implying no correlation.\nPositive correlations imply that as data in dataset x increases, so\ndoes data in dataset y. Negative correlations imply that as x increases,\ny decreases. Correlations of -1 or +1 imply an exact monotonic relationship.\n\nUnlike the Pearson correlation, the Spearman correlation does not\nassume that both datasets are normally distributed.\n\nThe p-value roughly indicates the probability of an uncorrelated system\nproducing datasets that have a Spearman correlation at least as extreme\nas the one computed from these datasets. The p-values are not entirely\nreliable but are probably reasonable for datasets larger than 500 or so.\n" lowerCAmelCase_ = "\nArgs:\n predictions (`List[float]`): Predicted labels, as returned by a model.\n references (`List[float]`): Ground truth labels.\n return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns\n only the spearmanr score. Defaults to `False`.\nReturns:\n spearmanr (`float`): Spearman correlation coefficient.\n p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input.\nExamples:\n Example 1:\n >>> spearmanr_metric = datasets.load_metric(\"spearmanr\")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4])\n >>> print(results)\n {'spearmanr': -0.7}\n\n Example 2:\n >>> spearmanr_metric = datasets.load_metric(\"spearmanr\")\n >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5],\n ... predictions=[10, 9, 2.5, 6, 4],\n ... return_pvalue=True)\n >>> print(results['spearmanr'])\n -0.7\n >>> print(round(results['spearmanr_pvalue'], 2))\n 0.19\n" lowerCAmelCase_ = r"\\n@book{kokoska2000crc,\n title={CRC standard probability and statistics tables and formulae},\n author={Kokoska, Stephen and Zwillinger, Daniel},\n year={2000},\n publisher={Crc Press}\n}\n@article{2020SciPy-NMeth,\n author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\n title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\n journal = {Nature Methods},\n year = {2020},\n volume = {17},\n pages = {261--272},\n adsurl = {https://rdcu.be/b08Wh},\n doi = {10.1038/s41592-019-0686-2},\n}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _A ( datasets.Metric ): def __a ( self : Tuple ) -> Optional[int]: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) , reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] , ) def __a ( self : str , _A : Dict , _A : Dict , _A : Any=False ) -> Tuple: """simple docstring""" lowercase : Dict = spearmanr(lowerCAmelCase_ , lowerCAmelCase_ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
308
'''simple docstring''' import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowercase : Optional[Any] = False class __UpperCAmelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained('shi-labs/versatile-diffusion' ) # remove text_unet pipe.remove_unused_weights() pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCAmelCase_ ) _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained(lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = generator.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained( 'shi-labs/versatile-diffusion' , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' ).images _snake_case = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) _snake_case = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
42
0
"""simple docstring""" import argparse import logging import os import re import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, DataCollatorForLanguageModeling, PushToHubCallback, TFAutoModelForMaskedLM, create_optimizer, ) lowerCAmelCase : Optional[int] = logging.getLogger(__name__) lowerCAmelCase : Optional[int] = tf.data.AUTOTUNE def a__ ( ) -> int: lowerCamelCase = argparse.ArgumentParser(description="""Train a masked language model on TPU.""" ) parser.add_argument( """--pretrained_model_config""" , type=__A , default="""roberta-base""" , help="""The model config to use. Note that we don\'t copy the model\'s weights, only the config!""" , ) parser.add_argument( """--tokenizer""" , type=__A , default="""unigram-tokenizer-wikitext""" , help="""The name of the tokenizer to load. We use the pretrained tokenizer to initialize the model\'s vocab size.""" , ) parser.add_argument( """--per_replica_batch_size""" , type=__A , default=8 , help="""Batch size per TPU core.""" , ) parser.add_argument( """--no_tpu""" , action="""store_true""" , help="""If set, run on CPU and don\'t try to initialize a TPU. Useful for debugging on non-TPU instances.""" , ) parser.add_argument( """--tpu_name""" , type=__A , help="""Name of TPU resource to initialize. Should be blank on Colab, and \'local\' on TPU VMs.""" , default="""local""" , ) parser.add_argument( """--tpu_zone""" , type=__A , help="""Google cloud zone that TPU resource is located in. Only used for non-Colab TPU nodes.""" , ) parser.add_argument( """--gcp_project""" , type=__A , help="""Google cloud project name. Only used for non-Colab TPU nodes.""" ) parser.add_argument( """--bfloat16""" , action="""store_true""" , help="""Use mixed-precision bfloat16 for training. This is the recommended lower-precision format for TPU.""" , ) parser.add_argument( """--train_dataset""" , type=__A , help="""Path to training dataset to load. If the path begins with `gs://`""" """ then the dataset will be loaded from a Google Cloud Storage bucket.""" , ) parser.add_argument( """--shuffle_buffer_size""" , type=__A , default=2**18 , help="""Size of the shuffle buffer (in samples)""" , ) parser.add_argument( """--eval_dataset""" , type=__A , help="""Path to evaluation dataset to load. If the path begins with `gs://`""" """ then the dataset will be loaded from a Google Cloud Storage bucket.""" , ) parser.add_argument( """--num_epochs""" , type=__A , default=1 , help="""Number of epochs to train for.""" , ) parser.add_argument( """--learning_rate""" , type=__A , default=1E-4 , help="""Learning rate to use for training.""" , ) parser.add_argument( """--weight_decay_rate""" , type=__A , default=1E-3 , help="""Weight decay rate to use for training.""" , ) parser.add_argument( """--max_length""" , type=__A , default=5_12 , help="""Maximum length of tokenized sequences. Should match the setting used in prepare_tfrecord_shards.py""" , ) parser.add_argument( """--mlm_probability""" , type=__A , default=0.15 , help="""Fraction of tokens to mask during training.""" , ) parser.add_argument("""--output_dir""" , type=__A , required=__A , help="""Path to save model checkpoints to.""" ) parser.add_argument("""--hub_model_id""" , type=__A , help="""Model ID to upload to on the Hugging Face Hub.""" ) lowerCamelCase = parser.parse_args() return args def a__ ( snake_case__ ) -> Dict: try: if args.tpu_name: lowerCamelCase = tf.distribute.cluster_resolver.TPUClusterResolver( args.tpu_name , zone=args.tpu_zone , project=args.gcp_project ) else: lowerCamelCase = tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: raise RuntimeError( """Couldn\'t connect to TPU! Most likely you need to specify --tpu_name, --tpu_zone, or """ """--gcp_project. When running on a TPU VM, use --tpu_name local.""" ) tf.config.experimental_connect_to_cluster(__A ) tf.tpu.experimental.initialize_tpu_system(__A ) return tpu def a__ ( snake_case__ ) -> int: lowerCamelCase = 0 for file in file_list: lowerCamelCase = file.split("""/""" )[-1] lowerCamelCase = re.search(R"""-\d+-(\d+)\.tfrecord""" , __A ).group(1 ) lowerCamelCase = int(__A ) num_samples += sample_count return num_samples def a__ ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__=None ) -> int: lowerCamelCase = count_samples(__A ) lowerCamelCase = tf.data.Dataset.from_tensor_slices(__A ) if shuffle: lowerCamelCase = dataset.shuffle(len(__A ) ) lowerCamelCase = tf.data.TFRecordDataset(__A , num_parallel_reads=__A ) # TF can't infer the total sample count because it doesn't read all the records yet, so we assert it here lowerCamelCase = dataset.apply(tf.data.experimental.assert_cardinality(__A ) ) lowerCamelCase = dataset.map(__A , num_parallel_calls=__A ) if shuffle: assert shuffle_buffer_size is not None lowerCamelCase = dataset.shuffle(args.shuffle_buffer_size ) lowerCamelCase = dataset.batch(__A , drop_remainder=__A ) lowerCamelCase = dataset.map(__A , num_parallel_calls=__A ) lowerCamelCase = dataset.prefetch(__A ) return dataset def a__ ( snake_case__ ) -> str: if not args.no_tpu: lowerCamelCase = initialize_tpu(__A ) lowerCamelCase = tf.distribute.TPUStrategy(__A ) else: lowerCamelCase = tf.distribute.OneDeviceStrategy(device="""/gpu:0""" ) if args.bfloataa: tf.keras.mixed_precision.set_global_policy("""mixed_bfloat16""" ) lowerCamelCase = AutoTokenizer.from_pretrained(args.tokenizer ) lowerCamelCase = AutoConfig.from_pretrained(args.pretrained_model_config ) lowerCamelCase = tokenizer.vocab_size lowerCamelCase = tf.io.gfile.glob(os.path.join(args.train_dataset , """*.tfrecord""" ) ) if not training_records: raise ValueError(F'No .tfrecord files found in {args.train_dataset}.' ) lowerCamelCase = tf.io.gfile.glob(os.path.join(args.eval_dataset , """*.tfrecord""" ) ) if not eval_records: raise ValueError(F'No .tfrecord files found in {args.eval_dataset}.' ) lowerCamelCase = count_samples(__A ) lowerCamelCase = num_train_samples // (args.per_replica_batch_size * strategy.num_replicas_in_sync) lowerCamelCase = steps_per_epoch * args.num_epochs with strategy.scope(): lowerCamelCase = TFAutoModelForMaskedLM.from_config(__A ) model(model.dummy_inputs ) # Pass some dummy inputs through the model to ensure all the weights are built lowerCamelCase , lowerCamelCase = create_optimizer( num_train_steps=__A , num_warmup_steps=total_train_steps // 20 , init_lr=args.learning_rate , weight_decay_rate=args.weight_decay_rate , ) # Transformers models compute the right loss for their task by default when labels are passed, and will # use this for training unless you specify your own loss function in compile(). model.compile(optimizer=__A , metrics=["""accuracy"""] ) def decode_fn(snake_case__ ): lowerCamelCase = { """input_ids""": tf.io.FixedLenFeature(dtype=tf.intaa , shape=(args.max_length,) ), """attention_mask""": tf.io.FixedLenFeature(dtype=tf.intaa , shape=(args.max_length,) ), } return tf.io.parse_single_example(__A , __A ) # Many of the data collators in Transformers are TF-compilable when return_tensors == "tf", so we can # use their methods in our data pipeline. lowerCamelCase = DataCollatorForLanguageModeling( tokenizer=__A , mlm_probability=args.mlm_probability , mlm=__A , return_tensors="""tf""" ) def mask_with_collator(snake_case__ ): # TF really needs an isin() function lowerCamelCase = ( ~tf.cast(batch["""attention_mask"""] , tf.bool ) | (batch["""input_ids"""] == tokenizer.cls_token_id) | (batch["""input_ids"""] == tokenizer.sep_token_id) ) lowerCamelCase , lowerCamelCase = data_collator.tf_mask_tokens( batch["""input_ids"""] , vocab_size=len(__A ) , mask_token_id=tokenizer.mask_token_id , special_tokens_mask=__A , ) return batch lowerCamelCase = args.per_replica_batch_size * strategy.num_replicas_in_sync lowerCamelCase = prepare_dataset( __A , decode_fn=__A , mask_fn=__A , batch_size=__A , shuffle=__A , shuffle_buffer_size=args.shuffle_buffer_size , ) lowerCamelCase = prepare_dataset( __A , decode_fn=__A , mask_fn=__A , batch_size=__A , shuffle=__A , ) lowerCamelCase = [] if args.hub_model_id: callbacks.append( PushToHubCallback(output_dir=args.output_dir , hub_model_id=args.hub_model_id , tokenizer=__A ) ) model.fit( __A , validation_data=__A , epochs=args.num_epochs , callbacks=__A , ) model.save_pretrained(args.output_dir ) if __name__ == "__main__": lowerCAmelCase : Any = parse_args() main(args)
291
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A = 100 ) -> int: _snake_case = n * (n + 1) * (2 * n + 1) / 6 _snake_case = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(F'''{solution() = }''')
42
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) a : Any = { "configuration_convnext": ["CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ConvNextConfig", "ConvNextOnnxConfig"] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[int] = ["ConvNextFeatureExtractor"] a : Dict = ["ConvNextImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Any = [ "CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST", "ConvNextForImageClassification", "ConvNextModel", "ConvNextPreTrainedModel", "ConvNextBackbone", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Tuple = [ "TFConvNextForImageClassification", "TFConvNextModel", "TFConvNextPreTrainedModel", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys a : Dict = _LazyModule(__name__, globals()["__file__"], _import_structure)
311
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary # Register SEW's fairseq modules from sew_asapp import tasks # noqa: F401 from transformers import ( SEWConfig, SEWForCTC, SEWModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() lowercase : str = logging.get_logger(__name__) lowercase : Union[str, Any] = { "post_extract_proj": "feature_projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.upsample.0": "encoder.upsample.projection", "encoder.layer_norm": "encoder.layer_norm", "w2v_model.layer_norm": "layer_norm", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", } def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A , __A ) -> Dict: for attribute in key.split('.' ): _snake_case = getattr(__A , __A ) if weight_type is not None: _snake_case = getattr(__A , __A ).shape else: _snake_case = hf_pointer.shape assert hf_shape == value.shape, ( F'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' F' {value.shape} for {full_name}' ) if weight_type == "weight": _snake_case = value elif weight_type == "weight_g": _snake_case = value elif weight_type == "weight_v": _snake_case = value elif weight_type == "bias": _snake_case = value else: _snake_case = value logger.info(F'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Any: _snake_case = [] _snake_case = fairseq_model.state_dict() _snake_case = hf_model.sew.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): _snake_case = False if "conv_layers" in name: load_conv_layer( __A , __A , __A , __A , hf_model.config.feat_extract_norm == 'group' , ) _snake_case = True else: for key, mapped_key in MAPPING.items(): _snake_case = 'sew.' + mapped_key if (is_finetuned and mapped_key != 'lm_head') else mapped_key if key in name or key.split('w2v_model.' )[-1] == name.split('.' )[0]: _snake_case = True if "*" in mapped_key: _snake_case = name.split(__A )[0].split('.' )[-2] _snake_case = mapped_key.replace('*' , __A ) if "weight_g" in name: _snake_case = 'weight_g' elif "weight_v" in name: _snake_case = 'weight_v' elif "weight" in name: _snake_case = 'weight' elif "bias" in name: _snake_case = 'bias' else: _snake_case = None set_recursively(__A , __A , __A , __A , __A ) continue if not is_used: unused_weights.append(__A ) logger.warning(F'Unused weights: {unused_weights}' ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A , __A ) -> int: _snake_case = full_name.split('conv_layers.' )[-1] _snake_case = name.split('.' ) _snake_case = int(items[0] ) _snake_case = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F'{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was' " found." ) _snake_case = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F'{full_name} has size {value.shape}, but' F' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) _snake_case = value logger.info(F'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(__A ) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str: _snake_case = SEWConfig() if is_finetuned: _snake_case = model.wav_encoder.wav_model.cfg else: _snake_case = model.cfg _snake_case = fs_config.conv_bias _snake_case = eval(fs_config.conv_feature_layers ) _snake_case = [x[0] for x in conv_layers] _snake_case = [x[1] for x in conv_layers] _snake_case = [x[2] for x in conv_layers] _snake_case = 'gelu' _snake_case = 'layer' if fs_config.extractor_mode == 'layer_norm' else 'group' _snake_case = 0.0 _snake_case = fs_config.activation_fn.name _snake_case = fs_config.encoder_embed_dim _snake_case = 0.0_2 _snake_case = fs_config.encoder_ffn_embed_dim _snake_case = 1e-5 _snake_case = fs_config.encoder_layerdrop _snake_case = fs_config.encoder_attention_heads _snake_case = fs_config.conv_pos_groups _snake_case = fs_config.conv_pos _snake_case = len(__A ) _snake_case = fs_config.encoder_layers _snake_case = fs_config.squeeze_factor # take care of any params that are overridden by the Wav2VecCtc model if is_finetuned: _snake_case = model.cfg _snake_case = fs_config.final_dropout _snake_case = fs_config.layerdrop _snake_case = fs_config.activation_dropout _snake_case = fs_config.mask_prob > 0 or fs_config.mask_channel_prob > 0 _snake_case = fs_config.attention_dropout _snake_case = fs_config.dropout_input _snake_case = fs_config.dropout _snake_case = fs_config.mask_channel_length _snake_case = fs_config.mask_channel_prob _snake_case = fs_config.mask_length _snake_case = fs_config.mask_prob _snake_case = 'Wav2Vec2FeatureExtractor' _snake_case = 'Wav2Vec2CTCTokenizer' return config @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A=None , __A=None , __A=True ) -> List[str]: if is_finetuned: _snake_case , _snake_case , _snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'data': '/'.join(dict_path.split('/' )[:-1] )} ) else: _snake_case , _snake_case , _snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) if config_path is not None: _snake_case = SEWConfig.from_pretrained(__A ) else: _snake_case = convert_config(model[0] , __A ) _snake_case = model[0].eval() _snake_case = True if config.feat_extract_norm == 'layer' else False _snake_case = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0 , do_normalize=__A , return_attention_mask=__A , ) if is_finetuned: if dict_path: _snake_case = Dictionary.load(__A ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _snake_case = target_dict.pad_index _snake_case = target_dict.bos_index _snake_case = target_dict.pad_index _snake_case = target_dict.bos_index _snake_case = target_dict.eos_index _snake_case = len(target_dict.symbols ) _snake_case = os.path.join(__A , 'vocab.json' ) if not os.path.isdir(__A ): logger.error('--pytorch_dump_folder_path ({}) should be a directory'.format(__A ) ) return os.makedirs(__A , exist_ok=__A ) with open(__A , 'w' , encoding='utf-8' ) as vocab_handle: json.dump(target_dict.indices , __A ) _snake_case = WavaVecaCTCTokenizer( __A , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='|' , do_lower_case=__A , ) _snake_case = WavaVecaProcessor(feature_extractor=__A , tokenizer=__A ) processor.save_pretrained(__A ) _snake_case = SEWForCTC(__A ) else: _snake_case = SEWModel(__A ) feature_extractor.save_pretrained(__A ) recursively_load_weights(__A , __A , __A ) hf_model.save_pretrained(__A ) if __name__ == "__main__": lowercase : int = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--is_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) lowercase : Union[str, Any] = parser.parse_args() convert_sew_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, args.is_finetuned )
42
0
"""simple docstring""" import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() lowerCAmelCase__ = logging.get_logger('''transformers.models.speecht5''') def a__ ( SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : List[str] , SCREAMING_SNAKE_CASE : Optional[int] ): '''simple docstring''' hf_model.apply_weight_norm() lowerCAmelCase : List[Any] = checkpoint["input_conv.weight_g"] lowerCAmelCase : str = checkpoint["input_conv.weight_v"] lowerCAmelCase : List[str] = checkpoint["input_conv.bias"] for i in range(len(config.upsample_rates ) ): lowerCAmelCase : List[Any] = checkpoint[f"""upsamples.{i}.1.weight_g"""] lowerCAmelCase : List[str] = checkpoint[f"""upsamples.{i}.1.weight_v"""] lowerCAmelCase : str = checkpoint[f"""upsamples.{i}.1.bias"""] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): lowerCAmelCase : str = checkpoint[f"""blocks.{i}.convs1.{j}.1.weight_g"""] lowerCAmelCase : Tuple = checkpoint[f"""blocks.{i}.convs1.{j}.1.weight_v"""] lowerCAmelCase : Any = checkpoint[f"""blocks.{i}.convs1.{j}.1.bias"""] lowerCAmelCase : Union[str, Any] = checkpoint[f"""blocks.{i}.convs2.{j}.1.weight_g"""] lowerCAmelCase : List[str] = checkpoint[f"""blocks.{i}.convs2.{j}.1.weight_v"""] lowerCAmelCase : int = checkpoint[f"""blocks.{i}.convs2.{j}.1.bias"""] lowerCAmelCase : Optional[Any] = checkpoint["output_conv.1.weight_g"] lowerCAmelCase : int = checkpoint["output_conv.1.weight_v"] lowerCAmelCase : int = checkpoint["output_conv.1.bias"] hf_model.remove_weight_norm() @torch.no_grad() def a__ ( SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : Any , SCREAMING_SNAKE_CASE : List[str]=None , SCREAMING_SNAKE_CASE : Tuple=None , ): '''simple docstring''' if config_path is not None: lowerCAmelCase : Dict = SpeechTaHifiGanConfig.from_pretrained(__A ) else: lowerCAmelCase : List[str] = SpeechTaHifiGanConfig() lowerCAmelCase : Union[str, Any] = SpeechTaHifiGan(__A ) lowerCAmelCase : Tuple = torch.load(__A ) load_weights(orig_checkpoint["model"]["generator"] , __A , __A ) lowerCAmelCase : str = np.load(__A ) lowerCAmelCase : str = stats[0].reshape(-1 ) lowerCAmelCase : int = stats[1].reshape(-1 ) lowerCAmelCase : Tuple = torch.from_numpy(__A ).float() lowerCAmelCase : Dict = torch.from_numpy(__A ).float() model.save_pretrained(__A ) if repo_id: print("Pushing to the hub..." ) model.push_to_hub(__A ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to original checkpoint''') parser.add_argument('''--stats_path''', required=True, default=None, type=str, help='''Path to stats.npy file''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) lowerCAmelCase__ = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
108
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : int = logging.get_logger(__name__) lowercase : Union[str, Any] = { "xlnet-base-cased": "https://huggingface.co/xlnet-base-cased/resolve/main/config.json", "xlnet-large-cased": "https://huggingface.co/xlnet-large-cased/resolve/main/config.json", } class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = """xlnet""" __lowercase = ["""mems"""] __lowercase = { """n_token""": """vocab_size""", # Backward compatibility """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowerCAmelCase_=3_20_00 , lowerCAmelCase_=10_24 , lowerCAmelCase_=24 , lowerCAmelCase_=16 , lowerCAmelCase_=40_96 , lowerCAmelCase_="gelu" , lowerCAmelCase_=True , lowerCAmelCase_="bi" , lowerCAmelCase_=0.02 , lowerCAmelCase_=1E-12 , lowerCAmelCase_=0.1 , lowerCAmelCase_=5_12 , lowerCAmelCase_=None , lowerCAmelCase_=True , lowerCAmelCase_=False , lowerCAmelCase_=False , lowerCAmelCase_=-1 , lowerCAmelCase_=False , lowerCAmelCase_="last" , lowerCAmelCase_=True , lowerCAmelCase_="tanh" , lowerCAmelCase_=0.1 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=5 , lowerCAmelCase_=1 , lowerCAmelCase_=2 , **lowerCAmelCase_ , ): """simple docstring""" _snake_case = vocab_size _snake_case = d_model _snake_case = n_layer _snake_case = n_head if d_model % n_head != 0: raise ValueError(F'\'d_model % n_head\' ({d_model % n_head}) should be equal to 0' ) if "d_head" in kwargs: if kwargs["d_head"] != d_model // n_head: raise ValueError( F'`d_head` ({kwargs["d_head"]}) should be equal to `d_model // n_head` ({d_model // n_head})' ) _snake_case = d_model // n_head _snake_case = ff_activation _snake_case = d_inner _snake_case = untie_r _snake_case = attn_type _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = dropout _snake_case = mem_len _snake_case = reuse_len _snake_case = bi_data _snake_case = clamp_len _snake_case = same_length _snake_case = summary_type _snake_case = summary_use_proj _snake_case = summary_activation _snake_case = summary_last_dropout _snake_case = start_n_top _snake_case = end_n_top _snake_case = bos_token_id _snake_case = pad_token_id _snake_case = eos_token_id if "use_cache" in kwargs: warnings.warn( 'The `use_cache` argument is deprecated and will be removed in a future version, use `use_mems_eval`' ' instead.' , lowerCAmelCase_ , ) _snake_case = kwargs['use_cache'] _snake_case = use_mems_eval _snake_case = use_mems_train super().__init__(pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ ) @property def lowerCamelCase ( self ): """simple docstring""" logger.info(F'The model {self.model_type} is one of the few models that has no sequence length limit.' ) return -1 @max_position_embeddings.setter def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" raise NotImplementedError( F'The model {self.model_type} is one of the few models that has no sequence length limit.' )
42
0
"""simple docstring""" import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging a = logging.get_logger(__name__) a = "▁" a = { "vocab_file": "vocab.json", "spm_file": "sentencepiece.bpe.model", "tokenizer_config_file": "tokenizer_config.json", } a = { "vocab_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json", }, "spm_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model", }, "tokenizer_config_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json", }, } a = { "facebook/m2m100_418M": 1_024, } # fmt: off a = { "m2m100": ["af", "am", "ar", "ast", "az", "ba", "be", "bg", "bn", "br", "bs", "ca", "ceb", "cs", "cy", "da", "de", "el", "en", "es", "et", "fa", "ff", "fi", "fr", "fy", "ga", "gd", "gl", "gu", "ha", "he", "hi", "hr", "ht", "hu", "hy", "id", "ig", "ilo", "is", "it", "ja", "jv", "ka", "kk", "km", "kn", "ko", "lb", "lg", "ln", "lo", "lt", "lv", "mg", "mk", "ml", "mn", "mr", "ms", "my", "ne", "nl", "no", "ns", "oc", "or", "pa", "pl", "ps", "pt", "ro", "ru", "sd", "si", "sk", "sl", "so", "sq", "sr", "ss", "su", "sv", "sw", "ta", "th", "tl", "tn", "tr", "uk", "ur", "uz", "vi", "wo", "xh", "yi", "yo", "zh", "zu"], "wmt21": ["en", "ha", "is", "ja", "cs", "ru", "zh", "de"] } class lowercase_ ( _lowerCamelCase ): '''simple docstring''' UpperCAmelCase : str = VOCAB_FILES_NAMES UpperCAmelCase : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase : Dict = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase : List[str] = ['''input_ids''', '''attention_mask'''] UpperCAmelCase : Optional[Any] = [] UpperCAmelCase : Optional[Any] = [] def __init__( self : Tuple , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Any=None , _UpperCAmelCase : str=None , _UpperCAmelCase : Tuple="<s>" , _UpperCAmelCase : Union[str, Any]="</s>" , _UpperCAmelCase : Union[str, Any]="</s>" , _UpperCAmelCase : Optional[int]="<pad>" , _UpperCAmelCase : List[Any]="<unk>" , _UpperCAmelCase : List[str]="m2m100" , _UpperCAmelCase : List[Any] = None , _UpperCAmelCase : List[Any]=8 , **_UpperCAmelCase : str , ): _A = {} if sp_model_kwargs is None else sp_model_kwargs _A = language_codes _A = FAIRSEQ_LANGUAGE_CODES[language_codes] _A = {lang_code: F'''__{lang_code}__''' for lang_code in fairseq_language_code} _A = kwargs.get('additional_special_tokens' , [] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(lowerCAmelCase_ ) for lang_code in fairseq_language_code if self.get_lang_token(lowerCAmelCase_ ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=lowerCAmelCase_ , tgt_lang=lowerCAmelCase_ , bos_token=lowerCAmelCase_ , eos_token=lowerCAmelCase_ , sep_token=lowerCAmelCase_ , unk_token=lowerCAmelCase_ , pad_token=lowerCAmelCase_ , language_codes=lowerCAmelCase_ , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=lowerCAmelCase_ , **lowerCAmelCase_ , ) _A = vocab_file _A = load_json(lowerCAmelCase_ ) _A = {v: k for k, v in self.encoder.items()} _A = spm_file _A = load_spm(lowerCAmelCase_ , self.sp_model_kwargs ) _A = len(self.encoder ) _A = { self.get_lang_token(lowerCAmelCase_ ): self.encoder_size + i for i, lang_code in enumerate(lowerCAmelCase_ ) } _A = {lang_code: self.encoder_size + i for i, lang_code in enumerate(lowerCAmelCase_ )} _A = {v: k for k, v in self.lang_token_to_id.items()} _A = src_lang if src_lang is not None else 'en' _A = tgt_lang _A = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) _A = num_madeup_words @property def lowerCAmelCase_ ( self : int ): return len(self.encoder ) + len(self.lang_token_to_id ) @property def lowerCAmelCase_ ( self : Any ): return self._src_lang @src_lang.setter def lowerCAmelCase_ ( self : List[Any] , _UpperCAmelCase : Dict ): _A = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowerCAmelCase_ ( self : Dict , _UpperCAmelCase : Dict ): return self.sp_model.encode(lowerCAmelCase_ , out_type=lowerCAmelCase_ ) def lowerCAmelCase_ ( self : int , _UpperCAmelCase : Optional[Any] ): if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(lowerCAmelCase_ , self.encoder[self.unk_token] ) def lowerCAmelCase_ ( self : Optional[Any] , _UpperCAmelCase : str ): if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(lowerCAmelCase_ , self.unk_token ) def lowerCAmelCase_ ( self : Union[str, Any] , _UpperCAmelCase : List[str] ): _A = [] _A = '' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(lowerCAmelCase_ ) + token _A = [] else: current_sub_tokens.append(lowerCAmelCase_ ) out_string += self.sp_model.decode(lowerCAmelCase_ ) return out_string.strip() def lowerCAmelCase_ ( self : Dict , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Dict = None , _UpperCAmelCase : Union[str, Any] = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase_ , token_ids_a=lowerCAmelCase_ , already_has_special_tokens=lowerCAmelCase_ ) _A = [1] * len(self.prefix_tokens ) _A = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(lowerCAmelCase_ )) + suffix_ones return prefix_ones + ([0] * len(lowerCAmelCase_ )) + ([0] * len(lowerCAmelCase_ )) + suffix_ones def lowerCAmelCase_ ( self : List[Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Tuple = None ): if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowerCAmelCase_ ( self : Union[str, Any] ): _A = {self.convert_ids_to_tokens(lowerCAmelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Union[str, Any] ): _A = self.__dict__.copy() _A = None return state def __setstate__( self : Any , _UpperCAmelCase : Optional[int] ): _A = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): _A = {} _A = load_spm(self.spm_file , self.sp_model_kwargs ) def lowerCAmelCase_ ( self : Tuple , _UpperCAmelCase : Any , _UpperCAmelCase : List[str] = None ): _A = Path(lowerCAmelCase_ ) if not save_dir.is_dir(): raise OSError(F'''{save_directory} should be a directory''' ) _A = save_dir / ( (filename_prefix + '-' if filename_prefix else '') + self.vocab_files_names['vocab_file'] ) _A = save_dir / ( (filename_prefix + '-' if filename_prefix else '') + self.vocab_files_names['spm_file'] ) save_json(self.encoder , lowerCAmelCase_ ) if os.path.abspath(self.spm_file ) != os.path.abspath(lowerCAmelCase_ ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , lowerCAmelCase_ ) elif not os.path.isfile(self.spm_file ): with open(lowerCAmelCase_ , 'wb' ) as fi: _A = self.sp_model.serialized_model_proto() fi.write(lowerCAmelCase_ ) return (str(lowerCAmelCase_ ), str(lowerCAmelCase_ )) def lowerCAmelCase_ ( self : Union[str, Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : str = "en" , _UpperCAmelCase : Any = None , _UpperCAmelCase : List[str] = "ro" , **_UpperCAmelCase : Tuple , ): _A = src_lang _A = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(lowerCAmelCase_ , lowerCAmelCase_ , **lowerCAmelCase_ ) def lowerCAmelCase_ ( self : Union[str, Any] , _UpperCAmelCase : List[str] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Any , **_UpperCAmelCase : int ): if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) _A = src_lang _A = self(lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ , **lowerCAmelCase_ ) _A = self.get_lang_id(lowerCAmelCase_ ) _A = tgt_lang_id return inputs def lowerCAmelCase_ ( self : Optional[Any] ): self.set_src_lang_special_tokens(self.src_lang ) def lowerCAmelCase_ ( self : str ): self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowerCAmelCase_ ( self : Optional[Any] , _UpperCAmelCase : Any ): _A = self.get_lang_token(lowerCAmelCase_ ) _A = self.lang_token_to_id[lang_token] _A = [self.cur_lang_id] _A = [self.eos_token_id] def lowerCAmelCase_ ( self : str , _UpperCAmelCase : List[str] ): _A = self.get_lang_token(lowerCAmelCase_ ) _A = self.lang_token_to_id[lang_token] _A = [self.cur_lang_id] _A = [self.eos_token_id] def lowerCAmelCase_ ( self : Dict , _UpperCAmelCase : int ): return self.lang_code_to_token[lang] def lowerCAmelCase_ ( self : List[Any] , _UpperCAmelCase : Optional[Any] ): _A = self.get_lang_token(lowerCAmelCase_ ) return self.lang_token_to_id[lang_token] def _snake_case ( _snake_case : Any , _snake_case : Optional[int] ) -> sentencepiece.SentencePieceProcessor: '''simple docstring''' _A = sentencepiece.SentencePieceProcessor(**__A ) spm.Load(str(__A ) ) return spm def _snake_case ( _snake_case : Tuple ) -> Union[Dict, List]: '''simple docstring''' with open(__A , 'r' ) as f: return json.load(__A ) def _snake_case ( _snake_case : Tuple , _snake_case : Any ) -> None: '''simple docstring''' with open(__A , 'w' ) as f: json.dump(__A , __A , indent=2 )
315
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCamembertModel @require_tf @require_sentencepiece @require_tokenizers class __UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = TFCamembertModel.from_pretrained('jplu/tf-camembert-base' ) _snake_case = tf.convert_to_tensor( [[5, 1_21, 11, 6_60, 16, 7_30, 2_55_43, 1_10, 83, 6]] , dtype=tf.intaa , ) # J'aime le camembert !" _snake_case = model(lowerCAmelCase_ )['last_hidden_state'] _snake_case = tf.TensorShape((1, 10, 7_68) ) self.assertEqual(output.shape , lowerCAmelCase_ ) # compare the actual values for a slice. _snake_case = tf.convert_to_tensor( [[[-0.0254, 0.0235, 0.1027], [0.0606, -0.1811, -0.0418], [-0.1561, -0.1127, 0.2687]]] , dtype=tf.floataa , ) # camembert = torch.hub.load('pytorch/fairseq', 'camembert.v0') # camembert.eval() # expected_slice = roberta.model.forward(input_ids)[0][:, :3, :3].detach() self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
42
0
def _A ( _lowercase = 3 , _lowercase = 7 , _lowercase = 1_00_00_00 ) -> int: """simple docstring""" __UpperCamelCase = 0 __UpperCamelCase = 1 for current_denominator in range(1 , limit + 1 ): __UpperCamelCase = current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: __UpperCamelCase = current_numerator __UpperCamelCase = current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=1_0_0_0_0_0_0))
310
'''simple docstring''' from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
42
0
from dataclasses import dataclass from typing import Optional, Tuple import torch from torch import nn from transformers import RobertaPreTrainedModel, XLMRobertaConfig, XLMRobertaModel from transformers.utils import ModelOutput @dataclass class lowercase__ ( _lowerCamelCase): UpperCamelCase_ = None UpperCamelCase_ = None UpperCamelCase_ = None UpperCamelCase_ = None class lowercase__ ( _lowerCamelCase): def __init__( self : Optional[Any] , UpperCamelCase__ : Any=1 , UpperCamelCase__ : Union[str, Any]=0 , UpperCamelCase__ : Optional[int]=2 , UpperCamelCase__ : Dict=512 , UpperCamelCase__ : List[str]="cls" , UpperCamelCase__ : Tuple=False , UpperCamelCase__ : Optional[int]=True , **UpperCamelCase__ : int , ): '''simple docstring''' super().__init__(pad_token_id=lowerCAmelCase_ , bos_token_id=lowerCAmelCase_ , eos_token_id=lowerCAmelCase_ , **lowerCAmelCase_ ) SCREAMING_SNAKE_CASE : Tuple = project_dim SCREAMING_SNAKE_CASE : Dict = pooler_fn SCREAMING_SNAKE_CASE : Union[str, Any] = learn_encoder SCREAMING_SNAKE_CASE : int = use_attention_mask class lowercase__ ( _lowerCamelCase): UpperCamelCase_ = [r"""pooler""", r"""logit_scale"""] UpperCamelCase_ = [r"""position_ids""", r"""predictions.decoder.bias"""] UpperCamelCase_ = """roberta""" UpperCamelCase_ = RobertaSeriesConfig def __init__( self : Union[str, Any] , UpperCamelCase__ : Union[str, Any] ): '''simple docstring''' super().__init__(lowerCAmelCase_ ) SCREAMING_SNAKE_CASE : int = XLMRobertaModel(lowerCAmelCase_ ) SCREAMING_SNAKE_CASE : int = nn.Linear(config.hidden_size , config.project_dim ) SCREAMING_SNAKE_CASE : int = getattr(lowerCAmelCase_ , '''has_pre_transformation''' , lowerCAmelCase_ ) if self.has_pre_transformation: SCREAMING_SNAKE_CASE : Union[str, Any] = nn.Linear(config.hidden_size , config.project_dim ) SCREAMING_SNAKE_CASE : Optional[int] = nn.LayerNorm(config.hidden_size , eps=config.layer_norm_eps ) self.post_init() def __A ( self : Tuple , UpperCamelCase__ : int = None , UpperCamelCase__ : int = None , UpperCamelCase__ : Optional[int] = None , UpperCamelCase__ : int = None , UpperCamelCase__ : str = None , UpperCamelCase__ : str = None , UpperCamelCase__ : int = None , UpperCamelCase__ : Tuple = None , UpperCamelCase__ : str = None , UpperCamelCase__ : int = None , UpperCamelCase__ : Dict = None , ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = return_dict if return_dict is not None else self.config.use_return_dict SCREAMING_SNAKE_CASE : Tuple = self.base_model( input_ids=lowerCAmelCase_ , attention_mask=lowerCAmelCase_ , token_type_ids=lowerCAmelCase_ , position_ids=lowerCAmelCase_ , head_mask=lowerCAmelCase_ , inputs_embeds=lowerCAmelCase_ , encoder_hidden_states=lowerCAmelCase_ , encoder_attention_mask=lowerCAmelCase_ , output_attentions=lowerCAmelCase_ , output_hidden_states=True if self.has_pre_transformation else output_hidden_states , return_dict=lowerCAmelCase_ , ) if self.has_pre_transformation: SCREAMING_SNAKE_CASE : List[str] = outputs['''hidden_states'''][-2] SCREAMING_SNAKE_CASE : List[str] = self.pre_LN(lowerCAmelCase_ ) SCREAMING_SNAKE_CASE : str = self.transformation_pre(lowerCAmelCase_ ) return TransformationModelOutput( projection_state=lowerCAmelCase_ , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , ) else: SCREAMING_SNAKE_CASE : int = self.transformation(outputs.last_hidden_state ) return TransformationModelOutput( projection_state=lowerCAmelCase_ , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
182
'''simple docstring''' import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() lowercase : List[str] = logging.get_logger("transformers.models.speecht5") def SCREAMING_SNAKE_CASE__ ( __A , __A , __A ) -> Dict: hf_model.apply_weight_norm() _snake_case = checkpoint['input_conv.weight_g'] _snake_case = checkpoint['input_conv.weight_v'] _snake_case = checkpoint['input_conv.bias'] for i in range(len(config.upsample_rates ) ): _snake_case = checkpoint[F'upsamples.{i}.1.weight_g'] _snake_case = checkpoint[F'upsamples.{i}.1.weight_v'] _snake_case = checkpoint[F'upsamples.{i}.1.bias'] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.weight_g'] _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.weight_v'] _snake_case = checkpoint[F'blocks.{i}.convs1.{j}.1.bias'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.weight_g'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.weight_v'] _snake_case = checkpoint[F'blocks.{i}.convs2.{j}.1.bias'] _snake_case = checkpoint['output_conv.1.weight_g'] _snake_case = checkpoint['output_conv.1.weight_v'] _snake_case = checkpoint['output_conv.1.bias'] hf_model.remove_weight_norm() @torch.no_grad() def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , __A=None , __A=None , ) -> List[Any]: if config_path is not None: _snake_case = SpeechTaHifiGanConfig.from_pretrained(__A ) else: _snake_case = SpeechTaHifiGanConfig() _snake_case = SpeechTaHifiGan(__A ) _snake_case = torch.load(__A ) load_weights(orig_checkpoint['model']['generator'] , __A , __A ) _snake_case = np.load(__A ) _snake_case = stats[0].reshape(-1 ) _snake_case = stats[1].reshape(-1 ) _snake_case = torch.from_numpy(__A ).float() _snake_case = torch.from_numpy(__A ).float() model.save_pretrained(__A ) if repo_id: print('Pushing to the hub...' ) model.push_to_hub(__A ) if __name__ == "__main__": lowercase : Dict = argparse.ArgumentParser() parser.add_argument("--checkpoint_path", required=True, default=None, type=str, help="Path to original checkpoint") parser.add_argument("--stats_path", required=True, default=None, type=str, help="Path to stats.npy file") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--pytorch_dump_folder_path", required=True, default=None, type=str, help="Path to the output PyTorch model." ) parser.add_argument( "--push_to_hub", default=None, type=str, help="Where to upload the converted model on the 🤗 hub." ) lowercase : List[Any] = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
42
0
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING snake_case_ : List[str] = logging.get_logger(__name__) snake_case_ : str = { "SenseTime/deformable-detr": "https://huggingface.co/sensetime/deformable-detr/resolve/main/config.json", # See all Deformable DETR models at https://huggingface.co/models?filter=deformable-detr } class lowercase__ ( _lowerCamelCase ): lowercase__ = """deformable_detr""" lowercase__ = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self : Optional[Any] ,lowerCamelCase__ : List[str]=True ,lowerCamelCase__ : Optional[int]=None ,lowerCamelCase__ : Optional[int]=3 ,lowerCamelCase__ : List[str]=300 ,lowerCamelCase__ : Tuple=1024 ,lowerCamelCase__ : Optional[int]=6 ,lowerCamelCase__ : Dict=1024 ,lowerCamelCase__ : int=8 ,lowerCamelCase__ : str=6 ,lowerCamelCase__ : Optional[int]=1024 ,lowerCamelCase__ : Dict=8 ,lowerCamelCase__ : Tuple=0.0 ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : List[Any]="relu" ,lowerCamelCase__ : Any=256 ,lowerCamelCase__ : List[Any]=0.1 ,lowerCamelCase__ : Optional[Any]=0.0 ,lowerCamelCase__ : List[Any]=0.0 ,lowerCamelCase__ : Tuple=0.0_2 ,lowerCamelCase__ : Dict=1.0 ,lowerCamelCase__ : Optional[Any]=True ,lowerCamelCase__ : List[Any]=False ,lowerCamelCase__ : List[str]="sine" ,lowerCamelCase__ : Any="resnet50" ,lowerCamelCase__ : int=True ,lowerCamelCase__ : Optional[Any]=False ,lowerCamelCase__ : Union[str, Any]=4 ,lowerCamelCase__ : int=4 ,lowerCamelCase__ : Dict=4 ,lowerCamelCase__ : List[str]=False ,lowerCamelCase__ : List[str]=300 ,lowerCamelCase__ : Optional[Any]=False ,lowerCamelCase__ : int=1 ,lowerCamelCase__ : Any=5 ,lowerCamelCase__ : List[str]=2 ,lowerCamelCase__ : Dict=1 ,lowerCamelCase__ : Dict=1 ,lowerCamelCase__ : str=5 ,lowerCamelCase__ : Any=2 ,lowerCamelCase__ : Any=0.1 ,lowerCamelCase__ : Optional[Any]=0.2_5 ,lowerCamelCase__ : Optional[Any]=False ,**lowerCamelCase__ : int ,): '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError('You can\'t specify both `backbone_config` and `use_timm_backbone`.' ) if not use_timm_backbone: if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) _UpperCamelCase : str = CONFIG_MAPPING['resnet'](out_features=['stage4'] ) elif isinstance(lowerCAmelCase_ ,lowerCAmelCase_ ): _UpperCamelCase : Any = backbone_config.get('model_type' ) _UpperCamelCase : int = CONFIG_MAPPING[backbone_model_type] _UpperCamelCase : Optional[int] = config_class.from_dict(lowerCAmelCase_ ) _UpperCamelCase : Optional[int] = use_timm_backbone _UpperCamelCase : str = backbone_config _UpperCamelCase : Optional[int] = num_channels _UpperCamelCase : Union[str, Any] = num_queries _UpperCamelCase : str = max_position_embeddings _UpperCamelCase : int = d_model _UpperCamelCase : Optional[int] = encoder_ffn_dim _UpperCamelCase : Any = encoder_layers _UpperCamelCase : Union[str, Any] = encoder_attention_heads _UpperCamelCase : Dict = decoder_ffn_dim _UpperCamelCase : Any = decoder_layers _UpperCamelCase : str = decoder_attention_heads _UpperCamelCase : Any = dropout _UpperCamelCase : Optional[Any] = attention_dropout _UpperCamelCase : Optional[Any] = activation_dropout _UpperCamelCase : List[Any] = activation_function _UpperCamelCase : Any = init_std _UpperCamelCase : Dict = init_xavier_std _UpperCamelCase : Optional[int] = encoder_layerdrop _UpperCamelCase : Union[str, Any] = auxiliary_loss _UpperCamelCase : Dict = position_embedding_type _UpperCamelCase : List[Any] = backbone _UpperCamelCase : Tuple = use_pretrained_backbone _UpperCamelCase : List[Any] = dilation # deformable attributes _UpperCamelCase : Tuple = num_feature_levels _UpperCamelCase : List[Any] = encoder_n_points _UpperCamelCase : Tuple = decoder_n_points _UpperCamelCase : Dict = two_stage _UpperCamelCase : Dict = two_stage_num_proposals _UpperCamelCase : int = with_box_refine if two_stage is True and with_box_refine is False: raise ValueError('If two_stage is True, with_box_refine must be True.' ) # Hungarian matcher _UpperCamelCase : Dict = class_cost _UpperCamelCase : Optional[int] = bbox_cost _UpperCamelCase : Optional[int] = giou_cost # Loss coefficients _UpperCamelCase : List[str] = mask_loss_coefficient _UpperCamelCase : List[Any] = dice_loss_coefficient _UpperCamelCase : Optional[int] = bbox_loss_coefficient _UpperCamelCase : List[Any] = giou_loss_coefficient _UpperCamelCase : List[str] = eos_coefficient _UpperCamelCase : Tuple = focal_alpha _UpperCamelCase : Dict = disable_custom_kernels super().__init__(is_encoder_decoder=lowerCAmelCase_ ,**lowerCAmelCase_ ) @property def UpperCamelCase_ ( self : Any ): '''simple docstring''' return self.encoder_attention_heads @property def UpperCamelCase_ ( self : Any ): '''simple docstring''' return self.d_model def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : Optional[Any] = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: _UpperCamelCase : Optional[int] = self.backbone_config.to_dict() _UpperCamelCase : Optional[Any] = self.__class__.model_type return output
83
'''simple docstring''' from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = 42 class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_=3 , lowerCAmelCase_=3 , lowerCAmelCase_=("DownEncoderBlock2D",) , lowerCAmelCase_=(64,) , lowerCAmelCase_=2 , lowerCAmelCase_=32 , lowerCAmelCase_="silu" , lowerCAmelCase_=True , ): """simple docstring""" super().__init__() _snake_case = layers_per_block _snake_case = torch.nn.Convad( lowerCAmelCase_ , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) _snake_case = None _snake_case = nn.ModuleList([] ) # down _snake_case = block_out_channels[0] for i, down_block_type in enumerate(lowerCAmelCase_ ): _snake_case = output_channel _snake_case = block_out_channels[i] _snake_case = i == len(lowerCAmelCase_ ) - 1 _snake_case = get_down_block( lowerCAmelCase_ , num_layers=self.layers_per_block , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) self.down_blocks.append(lowerCAmelCase_ ) # mid _snake_case = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift='default' , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # out _snake_case = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=lowerCAmelCase_ , eps=1E-6 ) _snake_case = nn.SiLU() _snake_case = 2 * out_channels if double_z else out_channels _snake_case = nn.Convad(block_out_channels[-1] , lowerCAmelCase_ , 3 , padding=1 ) _snake_case = False def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = x _snake_case = self.conv_in(lowerCAmelCase_ ) if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase_ ): def custom_forward(*lowerCAmelCase_ ): return module(*lowerCAmelCase_ ) return custom_forward # down if is_torch_version('>=' , '1.11.0' ): for down_block in self.down_blocks: _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: for down_block in self.down_blocks: _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ ) # middle _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , lowerCAmelCase_ ) else: # down for down_block in self.down_blocks: _snake_case = down_block(lowerCAmelCase_ ) # middle _snake_case = self.mid_block(lowerCAmelCase_ ) # post-process _snake_case = self.conv_norm_out(lowerCAmelCase_ ) _snake_case = self.conv_act(lowerCAmelCase_ ) _snake_case = self.conv_out(lowerCAmelCase_ ) return sample class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_=3 , lowerCAmelCase_=3 , lowerCAmelCase_=("UpDecoderBlock2D",) , lowerCAmelCase_=(64,) , lowerCAmelCase_=2 , lowerCAmelCase_=32 , lowerCAmelCase_="silu" , lowerCAmelCase_="group" , ): """simple docstring""" super().__init__() _snake_case = layers_per_block _snake_case = nn.Convad( lowerCAmelCase_ , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) _snake_case = None _snake_case = nn.ModuleList([] ) _snake_case = in_channels if norm_type == 'spatial' else None # mid _snake_case = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift='default' if norm_type == 'group' else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # up _snake_case = list(reversed(lowerCAmelCase_ ) ) _snake_case = reversed_block_out_channels[0] for i, up_block_type in enumerate(lowerCAmelCase_ ): _snake_case = output_channel _snake_case = reversed_block_out_channels[i] _snake_case = i == len(lowerCAmelCase_ ) - 1 _snake_case = get_up_block( lowerCAmelCase_ , num_layers=self.layers_per_block + 1 , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , prev_output_channel=lowerCAmelCase_ , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , resnet_time_scale_shift=lowerCAmelCase_ , ) self.up_blocks.append(lowerCAmelCase_ ) _snake_case = output_channel # out if norm_type == "spatial": _snake_case = SpatialNorm(block_out_channels[0] , lowerCAmelCase_ ) else: _snake_case = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=lowerCAmelCase_ , eps=1E-6 ) _snake_case = nn.SiLU() _snake_case = nn.Convad(block_out_channels[0] , lowerCAmelCase_ , 3 , padding=1 ) _snake_case = False def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=None ): """simple docstring""" _snake_case = z _snake_case = self.conv_in(lowerCAmelCase_ ) _snake_case = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(lowerCAmelCase_ ): def custom_forward(*lowerCAmelCase_ ): return module(*lowerCAmelCase_ ) return custom_forward if is_torch_version('>=' , '1.11.0' ): # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: # middle _snake_case = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ ) else: # middle _snake_case = self.mid_block(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: _snake_case = up_block(lowerCAmelCase_ , lowerCAmelCase_ ) # post-process if latent_embeds is None: _snake_case = self.conv_norm_out(lowerCAmelCase_ ) else: _snake_case = self.conv_norm_out(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self.conv_act(lowerCAmelCase_ ) _snake_case = self.conv_out(lowerCAmelCase_ ) return sample class __UpperCAmelCase ( nn.Module ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None , lowerCAmelCase_="random" , lowerCAmelCase_=False , lowerCAmelCase_=True ): """simple docstring""" super().__init__() _snake_case = n_e _snake_case = vq_embed_dim _snake_case = beta _snake_case = legacy _snake_case = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) _snake_case = remap if self.remap is not None: self.register_buffer('used' , torch.tensor(np.load(self.remap ) ) ) _snake_case = self.used.shape[0] _snake_case = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": _snake_case = self.re_embed _snake_case = self.re_embed + 1 print( F'Remapping {self.n_e} indices to {self.re_embed} indices. ' F'Using {self.unknown_index} for unknown indices.' ) else: _snake_case = n_e _snake_case = sane_index_shape def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = inds.shape assert len(lowerCAmelCase_ ) > 1 _snake_case = inds.reshape(ishape[0] , -1 ) _snake_case = self.used.to(lowerCAmelCase_ ) _snake_case = (inds[:, :, None] == used[None, None, ...]).long() _snake_case = match.argmax(-1 ) _snake_case = match.sum(2 ) < 1 if self.unknown_index == "random": _snake_case = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: _snake_case = self.unknown_index return new.reshape(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = inds.shape assert len(lowerCAmelCase_ ) > 1 _snake_case = inds.reshape(ishape[0] , -1 ) _snake_case = self.used.to(lowerCAmelCase_ ) if self.re_embed > self.used.shape[0]: # extra token _snake_case = 0 # simply set to zero _snake_case = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , lowerCAmelCase_ ) return back.reshape(lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" _snake_case = z.permute(0 , 2 , 3 , 1 ).contiguous() _snake_case = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z _snake_case = torch.argmin(torch.cdist(lowerCAmelCase_ , self.embedding.weight ) , dim=1 ) _snake_case = self.embedding(lowerCAmelCase_ ).view(z.shape ) _snake_case = None _snake_case = None # compute loss for embedding if not self.legacy: _snake_case = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: _snake_case = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients _snake_case = z + (z_q - z).detach() # reshape back to match original input shape _snake_case = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: _snake_case = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis _snake_case = self.remap_to_used(lowerCAmelCase_ ) _snake_case = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: _snake_case = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ ): """simple docstring""" if self.remap is not None: _snake_case = indices.reshape(shape[0] , -1 ) # add batch axis _snake_case = self.unmap_to_all(lowerCAmelCase_ ) _snake_case = indices.reshape(-1 ) # flatten again # get quantized latent vectors _snake_case = self.embedding(lowerCAmelCase_ ) if shape is not None: _snake_case = z_q.view(lowerCAmelCase_ ) # reshape back to match original input shape _snake_case = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class __UpperCAmelCase ( _lowerCamelCase ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_=False ): """simple docstring""" _snake_case = parameters _snake_case , _snake_case = torch.chunk(lowerCAmelCase_ , 2 , dim=1 ) _snake_case = torch.clamp(self.logvar , -30.0 , 20.0 ) _snake_case = deterministic _snake_case = torch.exp(0.5 * self.logvar ) _snake_case = torch.exp(self.logvar ) if self.deterministic: _snake_case = _snake_case = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def lowerCamelCase ( self , lowerCAmelCase_ = None ): """simple docstring""" _snake_case = randn_tensor( self.mean.shape , generator=lowerCAmelCase_ , device=self.parameters.device , dtype=self.parameters.dtype ) _snake_case = self.mean + self.std * sample return x def lowerCamelCase ( self , lowerCAmelCase_=None ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_=[1, 2, 3] ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) _snake_case = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=lowerCAmelCase_ ) def lowerCamelCase ( self ): """simple docstring""" return self.mean
42
0
"""simple docstring""" def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> float: if density <= 0: raise ValueError("""Impossible fluid density""" ) if bulk_modulus <= 0: raise ValueError("""Impossible bulk modulus""" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
347
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
42
0
from math import factorial def lowercase_ ( _lowerCamelCase : Union[str, Any] , _lowerCamelCase : Dict , _lowerCamelCase : int): if successes > trials: raise ValueError("successes must be lower or equal to trials") if trials < 0 or successes < 0: raise ValueError("the function is defined for non-negative integers") if not isinstance(__A , __A) or not isinstance(__A , __A): raise ValueError("the function is defined for non-negative integers") if not 0 < prob < 1: raise ValueError("prob has to be in range of 1 - 0") lowercase__ : List[str] = (prob**successes) * ((1 - prob) ** (trials - successes)) # Calculate the binomial coefficient: n! / k!(n-k)! lowercase__ : str = float(factorial(__A)) coefficient /= factorial(__A) * factorial(trials - successes) return probability * coefficient if __name__ == "__main__": from doctest import testmod testmod() print('''Probability of 2 successes out of 4 trails''') print('''with probability of 0.75 is:''', end=''' ''') print(binomial_distribution(2, 4, 0.75))
87
'''simple docstring''' import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline lowercase : Optional[Any] = argparse.ArgumentParser("Stable Diffusion script with intel optimization", add_help=False) parser.add_argument("--dpm", action="store_true", help="Enable DPMSolver or not") parser.add_argument("--steps", default=None, type=int, help="Num inference steps") lowercase : Tuple = parser.parse_args() lowercase : Optional[int] = "cpu" lowercase : Optional[Any] = "a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings" lowercase : Optional[int] = "path-to-your-trained-model" lowercase : List[str] = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: lowercase : str = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) lowercase : Dict = pipe.to(device) # to channels last lowercase : Optional[Any] = pipe.unet.to(memory_format=torch.channels_last) lowercase : int = pipe.vae.to(memory_format=torch.channels_last) lowercase : Optional[Any] = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: lowercase : Optional[int] = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex lowercase : Any = torch.randn(2, 4, 64, 64) lowercase : Optional[int] = torch.rand(1) * 999 lowercase : Optional[Any] = torch.randn(2, 77, 768) lowercase : Optional[Any] = (sample, timestep, encoder_hidden_status) try: lowercase : List[Any] = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: lowercase : List[str] = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) lowercase : Tuple = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) lowercase : Optional[Any] = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: lowercase : Tuple = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute lowercase : List[str] = 666 lowercase : Tuple = torch.Generator(device).manual_seed(seed) lowercase : Union[str, Any] = {"generator": generator} if args.steps is not None: lowercase : Dict = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): lowercase : List[str] = pipe(prompt, **generate_kwargs).images[0] # save image image.save("generated.png")
42
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a__ = { "configuration_rembert": ["REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "RemBertConfig", "RemBertOnnxConfig"] } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ = ["RemBertTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ = ["RemBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ = [ "REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "RemBertForCausalLM", "RemBertForMaskedLM", "RemBertForMultipleChoice", "RemBertForQuestionAnswering", "RemBertForSequenceClassification", "RemBertForTokenClassification", "RemBertLayer", "RemBertModel", "RemBertPreTrainedModel", "load_tf_weights_in_rembert", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ = [ "TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFRemBertForCausalLM", "TFRemBertForMaskedLM", "TFRemBertForMultipleChoice", "TFRemBertForQuestionAnswering", "TFRemBertForSequenceClassification", "TFRemBertForTokenClassification", "TFRemBertLayer", "TFRemBertModel", "TFRemBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_rembert import REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RemBertConfig, RemBertOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert import RemBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert_fast import RemBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rembert import ( REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, RemBertForCausalLM, RemBertForMaskedLM, RemBertForMultipleChoice, RemBertForQuestionAnswering, RemBertForSequenceClassification, RemBertForTokenClassification, RemBertLayer, RemBertModel, RemBertPreTrainedModel, load_tf_weights_in_rembert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rembert import ( TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFRemBertForCausalLM, TFRemBertForMaskedLM, TFRemBertForMultipleChoice, TFRemBertForQuestionAnswering, TFRemBertForSequenceClassification, TFRemBertForTokenClassification, TFRemBertLayer, TFRemBertModel, TFRemBertPreTrainedModel, ) else: import sys a__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
317
'''simple docstring''' from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class __UpperCAmelCase ( _lowerCamelCase ): def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" return 0.0 def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[int | float, int | float]: _snake_case = min([-20, np.min(fft_results[1 : samplerate // 2 - 1] )] ) _snake_case = max([20, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> None: _snake_case = 512 _snake_case = [1] + [0] * (size - 1) _snake_case = [filter_type.process(__A ) for item in inputs] _snake_case = [0] * (samplerate - size) # zero-padding outputs += filler _snake_case = np.abs(np.fft.fft(__A ) ) _snake_case = 20 * np.logaa(__A ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) # Display within reasonable bounds _snake_case = get_bounds(__A , __A ) plt.ylim(max([-80, bounds[0]] ) , min([80, bounds[1]] ) ) plt.ylabel('Gain (dB)' ) plt.plot(__A ) plt.show() def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> None: _snake_case = 512 _snake_case = [1] + [0] * (size - 1) _snake_case = [filter_type.process(__A ) for item in inputs] _snake_case = [0] * (samplerate - size) # zero-padding outputs += filler _snake_case = np.angle(np.fft.fft(__A ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel('Frequency (Hz)' ) plt.xscale('log' ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel('Phase shift (Radians)' ) plt.plot(np.unwrap(__A , -2 * pi ) ) plt.show()
42
0
from __future__ import annotations lowerCAmelCase_ = 8.988E9 # units = N * m^s * C^-2 def snake_case( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ) -> dict[str, float]: '''simple docstring''' lowercase : List[Any] = abs(chargea * chargea ) if (force, chargea, chargea, distance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if distance < 0: raise ValueError('''Distance cannot be negative''' ) if force == 0: lowercase : Any = COULOMBS_CONSTANT * charge_product / (distance**2) return {"force": force} elif chargea == 0: lowercase : Tuple = abs(__A ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge1": chargea} elif chargea == 0: lowercase : str = abs(__A ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge2": chargea} elif distance == 0: lowercase : Optional[Any] = (COULOMBS_CONSTANT * charge_product / abs(__A )) ** 0.5 return {"distance": distance} raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
308
'''simple docstring''' import tensorflow as tf from ...tf_utils import shape_list class __UpperCAmelCase ( tf.keras.layers.Layer ): def __init__( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=1 , lowerCAmelCase_=False , **lowerCAmelCase_ ): """simple docstring""" super().__init__(**lowerCAmelCase_ ) _snake_case = vocab_size _snake_case = d_embed _snake_case = d_proj _snake_case = cutoffs + [vocab_size] _snake_case = [0] + self.cutoffs _snake_case = div_val _snake_case = self.cutoffs[0] _snake_case = len(self.cutoffs ) - 1 _snake_case = self.shortlist_size + self.n_clusters _snake_case = keep_order _snake_case = [] _snake_case = [] def lowerCamelCase ( self , lowerCAmelCase_ ): """simple docstring""" if self.n_clusters > 0: _snake_case = self.add_weight( shape=(self.n_clusters, self.d_embed) , initializer='zeros' , trainable=lowerCAmelCase_ , name='cluster_weight' ) _snake_case = self.add_weight( shape=(self.n_clusters,) , initializer='zeros' , trainable=lowerCAmelCase_ , name='cluster_bias' ) if self.div_val == 1: for i in range(len(self.cutoffs ) ): if self.d_proj != self.d_embed: _snake_case = self.add_weight( shape=(self.d_embed, self.d_proj) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_projs_._{i}' , ) self.out_projs.append(lowerCAmelCase_ ) else: self.out_projs.append(lowerCAmelCase_ ) _snake_case = self.add_weight( shape=(self.vocab_size, self.d_embed) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._weight' , ) _snake_case = self.add_weight( shape=(self.vocab_size,) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._bias' , ) self.out_layers.append((weight, bias) ) else: for i in range(len(self.cutoffs ) ): _snake_case , _snake_case = self.cutoff_ends[i], self.cutoff_ends[i + 1] _snake_case = self.d_embed // (self.div_val**i) _snake_case = self.add_weight( shape=(d_emb_i, self.d_proj) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_projs_._{i}' ) self.out_projs.append(lowerCAmelCase_ ) _snake_case = self.add_weight( shape=(r_idx - l_idx, d_emb_i) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._weight' , ) _snake_case = self.add_weight( shape=(r_idx - l_idx,) , initializer='zeros' , trainable=lowerCAmelCase_ , name=F'out_layers_._{i}_._bias' , ) self.out_layers.append((weight, bias) ) super().build(lowerCAmelCase_ ) @staticmethod def lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=None ): """simple docstring""" _snake_case = x if proj is not None: _snake_case = tf.einsum('ibd,ed->ibe' , lowerCAmelCase_ , lowerCAmelCase_ ) return tf.einsum('ibd,nd->ibn' , lowerCAmelCase_ , lowerCAmelCase_ ) + b @staticmethod def lowerCamelCase ( lowerCAmelCase_ , lowerCAmelCase_ ): """simple docstring""" _snake_case = shape_list(lowerCAmelCase_ ) _snake_case = tf.range(lp_size[0] , dtype=target.dtype ) _snake_case = tf.stack([r, target] , 1 ) return tf.gather_nd(lowerCAmelCase_ , lowerCAmelCase_ ) def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_=True , lowerCAmelCase_=False ): """simple docstring""" _snake_case = 0 if self.n_clusters == 0: _snake_case = self._logit(lowerCAmelCase_ , self.out_layers[0][0] , self.out_layers[0][1] , self.out_projs[0] ) if target is not None: _snake_case = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=lowerCAmelCase_ , logits=lowerCAmelCase_ ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ , axis=-1 ) else: _snake_case = shape_list(lowerCAmelCase_ ) _snake_case = [] _snake_case = tf.zeros(hidden_sizes[:2] ) for i in range(len(self.cutoffs ) ): _snake_case , _snake_case = self.cutoff_ends[i], self.cutoff_ends[i + 1] if target is not None: _snake_case = (target >= l_idx) & (target < r_idx) _snake_case = tf.where(lowerCAmelCase_ ) _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) - l_idx if self.div_val == 1: _snake_case = self.out_layers[0][0][l_idx:r_idx] _snake_case = self.out_layers[0][1][l_idx:r_idx] else: _snake_case = self.out_layers[i][0] _snake_case = self.out_layers[i][1] if i == 0: _snake_case = tf.concat([cur_W, self.cluster_weight] , 0 ) _snake_case = tf.concat([cur_b, self.cluster_bias] , 0 ) _snake_case = self._logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , self.out_projs[0] ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ ) out.append(head_logprob[..., : self.cutoffs[0]] ) if target is not None: _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self._gather_logprob(lowerCAmelCase_ , lowerCAmelCase_ ) else: _snake_case = self._logit(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , self.out_projs[i] ) _snake_case = tf.nn.log_softmax(lowerCAmelCase_ ) _snake_case = self.cutoffs[0] + i - 1 # No probability for the head cluster _snake_case = head_logprob[..., cluster_prob_idx, None] + tail_logprob out.append(lowerCAmelCase_ ) if target is not None: _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = tf.boolean_mask(lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = self._gather_logprob(lowerCAmelCase_ , lowerCAmelCase_ ) cur_logprob += cur_head_logprob[:, self.cutoff_ends[1] + i - 1] if target is not None: loss += tf.scatter_nd(lowerCAmelCase_ , -cur_logprob , shape_list(lowerCAmelCase_ ) ) _snake_case = tf.concat(lowerCAmelCase_ , axis=-1 ) if target is not None: if return_mean: _snake_case = tf.reduce_mean(lowerCAmelCase_ ) # Add the training-time loss value to the layer using `self.add_loss()`. self.add_loss(lowerCAmelCase_ ) # Log the loss as a metric (we could log arbitrary metrics, # including different metrics for training and inference. self.add_metric(lowerCAmelCase_ , name=self.name , aggregation='mean' if return_mean else '' ) return out
42
0
"""simple docstring""" from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class __magic_name__ ( _lowerCamelCase ): '''simple docstring''' __UpperCamelCase = 42 class __magic_name__ ( nn.Module ): '''simple docstring''' def __init__( self , _a=3 , _a=3 , _a=("DownEncoderBlock2D",) , _a=(64,) , _a=2 , _a=32 , _a="silu" , _a=True , ): """simple docstring""" super().__init__() lowerCamelCase = layers_per_block lowerCamelCase = torch.nn.Convad( lowerCAmelCase_ , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) lowerCamelCase = None lowerCamelCase = nn.ModuleList([] ) # down lowerCamelCase = block_out_channels[0] for i, down_block_type in enumerate(lowerCAmelCase_ ): lowerCamelCase = output_channel lowerCamelCase = block_out_channels[i] lowerCamelCase = i == len(lowerCAmelCase_ ) - 1 lowerCamelCase = get_down_block( lowerCAmelCase_ , num_layers=self.layers_per_block , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , add_downsample=not is_final_block , resnet_eps=1e-6 , downsample_padding=0 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) self.down_blocks.append(lowerCAmelCase_ ) # mid lowerCamelCase = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1e-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift="""default""" , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # out lowerCamelCase = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=lowerCAmelCase_ , eps=1e-6 ) lowerCamelCase = nn.SiLU() lowerCamelCase = 2 * out_channels if double_z else out_channels lowerCamelCase = nn.Convad(block_out_channels[-1] , lowerCAmelCase_ , 3 , padding=1 ) lowerCamelCase = False def _lowerCAmelCase ( self , _a ): """simple docstring""" lowerCamelCase = x lowerCamelCase = self.conv_in(lowerCAmelCase_ ) if self.training and self.gradient_checkpointing: def create_custom_forward(_a ): def custom_forward(*_a ): return module(*lowerCAmelCase_ ) return custom_forward # down if is_torch_version(""">=""" , """1.11.0""" ): for down_block in self.down_blocks: lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) # middle lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: for down_block in self.down_blocks: lowerCamelCase = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ ) # middle lowerCamelCase = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , lowerCAmelCase_ ) else: # down for down_block in self.down_blocks: lowerCamelCase = down_block(lowerCAmelCase_ ) # middle lowerCamelCase = self.mid_block(lowerCAmelCase_ ) # post-process lowerCamelCase = self.conv_norm_out(lowerCAmelCase_ ) lowerCamelCase = self.conv_act(lowerCAmelCase_ ) lowerCamelCase = self.conv_out(lowerCAmelCase_ ) return sample class __magic_name__ ( nn.Module ): '''simple docstring''' def __init__( self , _a=3 , _a=3 , _a=("UpDecoderBlock2D",) , _a=(64,) , _a=2 , _a=32 , _a="silu" , _a="group" , ): """simple docstring""" super().__init__() lowerCamelCase = layers_per_block lowerCamelCase = nn.Convad( lowerCAmelCase_ , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) lowerCamelCase = None lowerCamelCase = nn.ModuleList([] ) lowerCamelCase = in_channels if norm_type == """spatial""" else None # mid lowerCamelCase = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1e-6 , resnet_act_fn=lowerCAmelCase_ , output_scale_factor=1 , resnet_time_scale_shift="""default""" if norm_type == """group""" else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , ) # up lowerCamelCase = list(reversed(lowerCAmelCase_ ) ) lowerCamelCase = reversed_block_out_channels[0] for i, up_block_type in enumerate(lowerCAmelCase_ ): lowerCamelCase = output_channel lowerCamelCase = reversed_block_out_channels[i] lowerCamelCase = i == len(lowerCAmelCase_ ) - 1 lowerCamelCase = get_up_block( lowerCAmelCase_ , num_layers=self.layers_per_block + 1 , in_channels=lowerCAmelCase_ , out_channels=lowerCAmelCase_ , prev_output_channel=lowerCAmelCase_ , add_upsample=not is_final_block , resnet_eps=1e-6 , resnet_act_fn=lowerCAmelCase_ , resnet_groups=lowerCAmelCase_ , attention_head_dim=lowerCAmelCase_ , temb_channels=lowerCAmelCase_ , resnet_time_scale_shift=lowerCAmelCase_ , ) self.up_blocks.append(lowerCAmelCase_ ) lowerCamelCase = output_channel # out if norm_type == "spatial": lowerCamelCase = SpatialNorm(block_out_channels[0] , lowerCAmelCase_ ) else: lowerCamelCase = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=lowerCAmelCase_ , eps=1e-6 ) lowerCamelCase = nn.SiLU() lowerCamelCase = nn.Convad(block_out_channels[0] , lowerCAmelCase_ , 3 , padding=1 ) lowerCamelCase = False def _lowerCAmelCase ( self , _a , _a=None ): """simple docstring""" lowerCamelCase = z lowerCamelCase = self.conv_in(lowerCAmelCase_ ) lowerCamelCase = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(_a ): def custom_forward(*_a ): return module(*lowerCAmelCase_ ) return custom_forward if is_torch_version(""">=""" , """1.11.0""" ): # middle lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) lowerCamelCase = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ , use_reentrant=lowerCAmelCase_ ) else: # middle lowerCamelCase = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , lowerCAmelCase_ , lowerCAmelCase_ ) lowerCamelCase = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: lowerCamelCase = torch.utils.checkpoint.checkpoint(create_custom_forward(lowerCAmelCase_ ) , lowerCAmelCase_ , lowerCAmelCase_ ) else: # middle lowerCamelCase = self.mid_block(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCamelCase = sample.to(lowerCAmelCase_ ) # up for up_block in self.up_blocks: lowerCamelCase = up_block(lowerCAmelCase_ , lowerCAmelCase_ ) # post-process if latent_embeds is None: lowerCamelCase = self.conv_norm_out(lowerCAmelCase_ ) else: lowerCamelCase = self.conv_norm_out(lowerCAmelCase_ , lowerCAmelCase_ ) lowerCamelCase = self.conv_act(lowerCAmelCase_ ) lowerCamelCase = self.conv_out(lowerCAmelCase_ ) return sample class __magic_name__ ( nn.Module ): '''simple docstring''' def __init__( self , _a , _a , _a , _a=None , _a="random" , _a=False , _a=True ): """simple docstring""" super().__init__() lowerCamelCase = n_e lowerCamelCase = vq_embed_dim lowerCamelCase = beta lowerCamelCase = legacy lowerCamelCase = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) lowerCamelCase = remap if self.remap is not None: self.register_buffer("""used""" , torch.tensor(np.load(self.remap ) ) ) lowerCamelCase = self.used.shape[0] lowerCamelCase = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": lowerCamelCase = self.re_embed lowerCamelCase = self.re_embed + 1 print( f'Remapping {self.n_e} indices to {self.re_embed} indices. ' f'Using {self.unknown_index} for unknown indices.' ) else: lowerCamelCase = n_e lowerCamelCase = sane_index_shape def _lowerCAmelCase ( self , _a ): """simple docstring""" lowerCamelCase = inds.shape assert len(lowerCAmelCase_ ) > 1 lowerCamelCase = inds.reshape(ishape[0] , -1 ) lowerCamelCase = self.used.to(lowerCAmelCase_ ) lowerCamelCase = (inds[:, :, None] == used[None, None, ...]).long() lowerCamelCase = match.argmax(-1 ) lowerCamelCase = match.sum(2 ) < 1 if self.unknown_index == "random": lowerCamelCase = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: lowerCamelCase = self.unknown_index return new.reshape(lowerCAmelCase_ ) def _lowerCAmelCase ( self , _a ): """simple docstring""" lowerCamelCase = inds.shape assert len(lowerCAmelCase_ ) > 1 lowerCamelCase = inds.reshape(ishape[0] , -1 ) lowerCamelCase = self.used.to(lowerCAmelCase_ ) if self.re_embed > self.used.shape[0]: # extra token lowerCamelCase = 0 # simply set to zero lowerCamelCase = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , lowerCAmelCase_ ) return back.reshape(lowerCAmelCase_ ) def _lowerCAmelCase ( self , _a ): """simple docstring""" lowerCamelCase = z.permute(0 , 2 , 3 , 1 ).contiguous() lowerCamelCase = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z lowerCamelCase = torch.argmin(torch.cdist(lowerCAmelCase_ , self.embedding.weight ) , dim=1 ) lowerCamelCase = self.embedding(lowerCAmelCase_ ).view(z.shape ) lowerCamelCase = None lowerCamelCase = None # compute loss for embedding if not self.legacy: lowerCamelCase = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: lowerCamelCase = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients lowerCamelCase = z + (z_q - z).detach() # reshape back to match original input shape lowerCamelCase = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: lowerCamelCase = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis lowerCamelCase = self.remap_to_used(lowerCAmelCase_ ) lowerCamelCase = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: lowerCamelCase = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def _lowerCAmelCase ( self , _a , _a ): """simple docstring""" if self.remap is not None: lowerCamelCase = indices.reshape(shape[0] , -1 ) # add batch axis lowerCamelCase = self.unmap_to_all(lowerCAmelCase_ ) lowerCamelCase = indices.reshape(-1 ) # flatten again # get quantized latent vectors lowerCamelCase = self.embedding(lowerCAmelCase_ ) if shape is not None: lowerCamelCase = z_q.view(lowerCAmelCase_ ) # reshape back to match original input shape lowerCamelCase = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class __magic_name__ ( _lowerCamelCase ): '''simple docstring''' def __init__( self , _a , _a=False ): """simple docstring""" lowerCamelCase = parameters lowerCamelCase , lowerCamelCase = torch.chunk(lowerCAmelCase_ , 2 , dim=1 ) lowerCamelCase = torch.clamp(self.logvar , -30.0 , 20.0 ) lowerCamelCase = deterministic lowerCamelCase = torch.exp(0.5 * self.logvar ) lowerCamelCase = torch.exp(self.logvar ) if self.deterministic: lowerCamelCase = lowerCamelCase = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def _lowerCAmelCase ( self , _a = None ): """simple docstring""" lowerCamelCase = randn_tensor( self.mean.shape , generator=lowerCAmelCase_ , device=self.parameters.device , dtype=self.parameters.dtype ) lowerCamelCase = self.mean + self.std * sample return x def _lowerCAmelCase ( self , _a=None ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def _lowerCAmelCase ( self , _a , _a=[1, 2, 3] ): """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) lowerCamelCase = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=lowerCAmelCase_ ) def _lowerCAmelCase ( self ): """simple docstring""" return self.mean
291
'''simple docstring''' from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. lowercase : Dict = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. lowercase : Optional[int] = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. lowercase : Optional[Any] = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[str, float]: _snake_case = len([g for position, g in enumerate(__A ) if g == main_target[position]] ) return (item, float(__A )) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> tuple[str, str]: _snake_case = random.randint(0 , len(__A ) - 1 ) _snake_case = parent_a[:random_slice] + parent_a[random_slice:] _snake_case = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def SCREAMING_SNAKE_CASE__ ( __A , __A ) -> str: _snake_case = list(__A ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: _snake_case = random.choice(__A ) return "".join(__A ) def SCREAMING_SNAKE_CASE__ ( __A , __A , __A , ) -> list[str]: _snake_case = [] # Generate more children proportionally to the fitness score. _snake_case = int(parent_a[1] * 100 ) + 1 _snake_case = 10 if child_n >= 10 else child_n for _ in range(__A ): _snake_case = population_score[random.randint(0 , __A )][0] _snake_case , _snake_case = crossover(parent_a[0] , __A ) # Append new string to the population list. pop.append(mutate(__A , __A ) ) pop.append(mutate(__A , __A ) ) return pop def SCREAMING_SNAKE_CASE__ ( __A , __A , __A = True ) -> tuple[int, int, str]: # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: _snake_case = F'{N_POPULATION} must be bigger than {N_SELECTED}' raise ValueError(__A ) # Verify that the target contains no genes besides the ones inside genes variable. _snake_case = sorted({c for c in target if c not in genes} ) if not_in_genes_list: _snake_case = F'{not_in_genes_list} is not in genes list, evolution cannot converge' raise ValueError(__A ) # Generate random starting population. _snake_case = [] for _ in range(__A ): population.append(''.join([random.choice(__A ) for i in range(len(__A ) )] ) ) # Just some logs to know what the algorithms is doing. _snake_case , _snake_case = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(__A ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. _snake_case = [evaluate(__A , __A ) for item in population] # Check if there is a matching evolution. _snake_case = sorted(__A , key=lambda __A : x[1] , reverse=__A ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( F'\nGeneration: {generation}' F'\nTotal Population:{total_population}' F'\nBest score: {population_score[0][1]}' F'\nBest string: {population_score[0][0]}' ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. _snake_case = population[: int(N_POPULATION / 3 )] population.clear() population.extend(__A ) # Normalize population score to be between 0 and 1. _snake_case = [ (item, score / len(__A )) for item, score in population_score ] # This is selection for i in range(__A ): population.extend(select(population_score[int(__A )] , __A , __A ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(__A ) > N_POPULATION: break if __name__ == "__main__": lowercase : str = ( "This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!" ) lowercase : str = list( " ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm" "nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\" ) lowercase , lowercase , lowercase : Tuple = basic(target_str, genes_list) print( F'''\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}''' )
42
0
'''simple docstring''' from __future__ import annotations def lowercase ( __magic_name__ , __magic_name__ , __magic_name__ , __magic_name__ ): # noqa: E741 '''simple docstring''' while r - l > 1: UpperCAmelCase : List[str] = (l + r) // 2 if v[m] >= key: UpperCAmelCase : int = m else: UpperCAmelCase : int = m # noqa: E741 return r def lowercase ( __magic_name__ ): '''simple docstring''' if len(__A ) == 0: return 0 UpperCAmelCase : Optional[Any] = [0] * len(__A ) UpperCAmelCase : Optional[Any] = 1 UpperCAmelCase : Tuple = v[0] for i in range(1 , len(__A ) ): if v[i] < tail[0]: UpperCAmelCase : List[Any] = v[i] elif v[i] > tail[length - 1]: UpperCAmelCase : str = v[i] length += 1 else: UpperCAmelCase : str = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
311
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase : Any = { "configuration_chinese_clip": [ "CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "ChineseCLIPConfig", "ChineseCLIPOnnxConfig", "ChineseCLIPTextConfig", "ChineseCLIPVisionConfig", ], "processing_chinese_clip": ["ChineseCLIPProcessor"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[Any] = ["ChineseCLIPFeatureExtractor"] lowercase : List[Any] = ["ChineseCLIPImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Any = [ "CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST", "ChineseCLIPModel", "ChineseCLIPPreTrainedModel", "ChineseCLIPTextModel", "ChineseCLIPVisionModel", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowercase : List[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
42
0
"""simple docstring""" import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated lowerCAmelCase__ = collections.namedtuple('''_Datasets''', ['''train''', '''validation''', '''test''']) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ lowerCAmelCase__ = "https://storage.googleapis.com/cvdf-datasets/mnist/" def a__ ( SCREAMING_SNAKE_CASE : Optional[Any] ): '''simple docstring''' lowerCAmelCase : str = numpy.dtype(numpy.uintaa ).newbyteorder(">" ) return numpy.frombuffer(bytestream.read(4 ) , dtype=__A )[0] @deprecated(__A , "Please use tf.data to implement this functionality." ) def a__ ( SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' print("Extracting" , f.name ) with gzip.GzipFile(fileobj=__A ) as bytestream: lowerCAmelCase : Dict = _readaa(__A ) if magic != 2_0_5_1: raise ValueError( "Invalid magic number %d in MNIST image file: %s" % (magic, f.name) ) lowerCAmelCase : Optional[Any] = _readaa(__A ) lowerCAmelCase : Optional[int] = _readaa(__A ) lowerCAmelCase : Tuple = _readaa(__A ) lowerCAmelCase : int = bytestream.read(rows * cols * num_images ) lowerCAmelCase : Union[str, Any] = numpy.frombuffer(__A , dtype=numpy.uinta ) lowerCAmelCase : Any = data.reshape(__A , __A , __A , 1 ) return data @deprecated(__A , "Please use tf.one_hot on tensors." ) def a__ ( SCREAMING_SNAKE_CASE : List[Any] , SCREAMING_SNAKE_CASE : Any ): '''simple docstring''' lowerCAmelCase : Optional[int] = labels_dense.shape[0] lowerCAmelCase : Tuple = numpy.arange(__A ) * num_classes lowerCAmelCase : int = numpy.zeros((num_labels, num_classes) ) lowerCAmelCase : Optional[Any] = 1 return labels_one_hot @deprecated(__A , "Please use tf.data to implement this functionality." ) def a__ ( SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : Any=False , SCREAMING_SNAKE_CASE : Union[str, Any]=1_0 ): '''simple docstring''' print("Extracting" , f.name ) with gzip.GzipFile(fileobj=__A ) as bytestream: lowerCAmelCase : List[str] = _readaa(__A ) if magic != 2_0_4_9: raise ValueError( "Invalid magic number %d in MNIST label file: %s" % (magic, f.name) ) lowerCAmelCase : Optional[Any] = _readaa(__A ) lowerCAmelCase : Any = bytestream.read(__A ) lowerCAmelCase : Dict = numpy.frombuffer(__A , dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(__A , __A ) return labels class SCREAMING_SNAKE_CASE__ : """simple docstring""" @deprecated( lowerCAmelCase_ , "Please use alternatives such as official/mnist/_DataSet.py" " from tensorflow/models." , ) def __init__( self , snake_case__ , snake_case__ , snake_case__=False , snake_case__=False , snake_case__=dtypes.floataa , snake_case__=True , snake_case__=None , ): """simple docstring""" lowerCAmelCase , lowerCAmelCase : Any = random_seed.get_seed(lowerCAmelCase_ ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) lowerCAmelCase : Dict = dtypes.as_dtype(lowerCAmelCase_ ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError("Invalid image dtype %r, expected uint8 or float32" % dtype ) if fake_data: lowerCAmelCase : List[Any] = 10_000 lowerCAmelCase : List[Any] = one_hot else: assert ( images.shape[0] == labels.shape[0] ), f"""images.shape: {images.shape} labels.shape: {labels.shape}""" lowerCAmelCase : Dict = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 lowerCAmelCase : Optional[int] = images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. lowerCAmelCase : Any = images.astype(numpy.floataa ) lowerCAmelCase : str = numpy.multiply(lowerCAmelCase_ , 1.0 / 255.0 ) lowerCAmelCase : Tuple = images lowerCAmelCase : List[str] = labels lowerCAmelCase : Tuple = 0 lowerCAmelCase : Dict = 0 @property def lowercase__ ( self ): """simple docstring""" return self._images @property def lowercase__ ( self ): """simple docstring""" return self._labels @property def lowercase__ ( self ): """simple docstring""" return self._num_examples @property def lowercase__ ( self ): """simple docstring""" return self._epochs_completed def lowercase__ ( self , snake_case__ , snake_case__=False , snake_case__=True ): """simple docstring""" if fake_data: lowerCAmelCase : Any = [1] * 784 lowerCAmelCase : int = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(lowerCAmelCase_ )], [fake_label for _ in range(lowerCAmelCase_ )], ) lowerCAmelCase : Tuple = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: lowerCAmelCase : Optional[Any] = numpy.arange(self._num_examples ) numpy.random.shuffle(lowerCAmelCase_ ) lowerCAmelCase : str = self.images[perma] lowerCAmelCase : List[str] = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch lowerCAmelCase : Union[str, Any] = self._num_examples - start lowerCAmelCase : Union[str, Any] = self._images[start : self._num_examples] lowerCAmelCase : Dict = self._labels[start : self._num_examples] # Shuffle the data if shuffle: lowerCAmelCase : Union[str, Any] = numpy.arange(self._num_examples ) numpy.random.shuffle(lowerCAmelCase_ ) lowerCAmelCase : int = self.images[perm] lowerCAmelCase : List[str] = self.labels[perm] # Start next epoch lowerCAmelCase : Optional[int] = 0 lowerCAmelCase : Optional[Any] = batch_size - rest_num_examples lowerCAmelCase : Tuple = self._index_in_epoch lowerCAmelCase : Dict = self._images[start:end] lowerCAmelCase : List[str] = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size lowerCAmelCase : Optional[int] = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(__A , "Please write your own downloading logic." ) def a__ ( SCREAMING_SNAKE_CASE : Tuple , SCREAMING_SNAKE_CASE : Dict , SCREAMING_SNAKE_CASE : int ): '''simple docstring''' if not gfile.Exists(__A ): gfile.MakeDirs(__A ) lowerCAmelCase : Union[str, Any] = os.path.join(__A , __A ) if not gfile.Exists(__A ): urllib.request.urlretrieve(__A , __A ) # noqa: S310 with gfile.GFile(__A ) as f: lowerCAmelCase : Tuple = f.size() print("Successfully downloaded" , __A , __A , "bytes." ) return filepath @deprecated( __A , "Please use alternatives such as:" " tensorflow_datasets.load(\'mnist\')" ) def a__ ( SCREAMING_SNAKE_CASE : List[str] , SCREAMING_SNAKE_CASE : Dict=False , SCREAMING_SNAKE_CASE : Tuple=False , SCREAMING_SNAKE_CASE : Any=dtypes.floataa , SCREAMING_SNAKE_CASE : Any=True , SCREAMING_SNAKE_CASE : str=5_0_0_0 , SCREAMING_SNAKE_CASE : Union[str, Any]=None , SCREAMING_SNAKE_CASE : List[str]=DEFAULT_SOURCE_URL , ): '''simple docstring''' if fake_data: def fake(): return _DataSet( [] , [] , fake_data=__A , one_hot=__A , dtype=__A , seed=__A ) lowerCAmelCase : Tuple = fake() lowerCAmelCase : Optional[Any] = fake() lowerCAmelCase : Any = fake() return _Datasets(train=__A , validation=__A , test=__A ) if not source_url: # empty string check lowerCAmelCase : List[Any] = DEFAULT_SOURCE_URL lowerCAmelCase : Optional[Any] = "train-images-idx3-ubyte.gz" lowerCAmelCase : Union[str, Any] = "train-labels-idx1-ubyte.gz" lowerCAmelCase : List[str] = "t10k-images-idx3-ubyte.gz" lowerCAmelCase : Optional[int] = "t10k-labels-idx1-ubyte.gz" lowerCAmelCase : str = _maybe_download( __A , __A , source_url + train_images_file ) with gfile.Open(__A , "rb" ) as f: lowerCAmelCase : Tuple = _extract_images(__A ) lowerCAmelCase : Optional[Any] = _maybe_download( __A , __A , source_url + train_labels_file ) with gfile.Open(__A , "rb" ) as f: lowerCAmelCase : Union[str, Any] = _extract_labels(__A , one_hot=__A ) lowerCAmelCase : Optional[int] = _maybe_download( __A , __A , source_url + test_images_file ) with gfile.Open(__A , "rb" ) as f: lowerCAmelCase : Dict = _extract_images(__A ) lowerCAmelCase : Any = _maybe_download( __A , __A , source_url + test_labels_file ) with gfile.Open(__A , "rb" ) as f: lowerCAmelCase : int = _extract_labels(__A , one_hot=__A ) if not 0 <= validation_size <= len(__A ): lowerCAmelCase : List[str] = ( "Validation size should be between 0 and " f"""{len(__A )}. Received: {validation_size}.""" ) raise ValueError(__A ) lowerCAmelCase : str = train_images[:validation_size] lowerCAmelCase : Union[str, Any] = train_labels[:validation_size] lowerCAmelCase : Any = train_images[validation_size:] lowerCAmelCase : Optional[int] = train_labels[validation_size:] lowerCAmelCase : str = {"dtype": dtype, "reshape": reshape, "seed": seed} lowerCAmelCase : str = _DataSet(__A , __A , **__A ) lowerCAmelCase : int = _DataSet(__A , __A , **__A ) lowerCAmelCase : int = _DataSet(__A , __A , **__A ) return _Datasets(train=__A , validation=__A , test=__A )
108
'''simple docstring''' def SCREAMING_SNAKE_CASE__ ( __A ) -> str: _snake_case = 1 _snake_case = 2 while i * i <= n: _snake_case = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def SCREAMING_SNAKE_CASE__ ( ) -> List[str]: _snake_case = 1 _snake_case = 1 while True: i += 1 t_num += i if count_divisors(__A ) > 500: break return t_num if __name__ == "__main__": print(solution())
42
0
"""simple docstring""" import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class lowercase_ ( _lowerCamelCase ): '''simple docstring''' def __init__( self : Optional[Any] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Any ): _A = dataset _A = process _A = params def __len__( self : Optional[Any] ): return len(self.dataset ) def __getitem__( self : List[Any] , _UpperCAmelCase : Optional[int] ): _A = self.dataset[i] _A = self.process(lowerCAmelCase_ , **self.params ) return processed class lowercase_ ( _lowerCamelCase ): '''simple docstring''' def __init__( self : Any , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : List[Any]=None ): _A = loader _A = infer _A = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether _A = None _A = loader_batch_size # Internal bookkeeping _A = None _A = None def __len__( self : List[Any] ): return len(self.loader ) def __iter__( self : List[str] ): _A = iter(self.loader ) return self def lowerCAmelCase_ ( self : Dict ): if isinstance(self._loader_batch_data , torch.Tensor ): # Batch data is simple tensor, just fetch the slice _A = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) _A = {} for k, element in self._loader_batch_data.items(): if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): # Convert ModelOutput to tuple first _A = element.to_tuple() if isinstance(element[0] , torch.Tensor ): _A = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): _A = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] , torch.Tensor ): _A = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): _A = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around _A = None elif isinstance(element[self._loader_batch_index] , torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _A = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] , np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _A = np.expand_dims(element[self._loader_batch_index] , 0 ) else: # This is typically a list, so no need to `unsqueeze`. _A = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 _A = self._loader_batch_data.__class__(lowerCAmelCase_ ) self._loader_batch_index += 1 return result def lowerCAmelCase_ ( self : List[Any] ): if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch _A = next(self.iterator ) _A = self.infer(lowerCAmelCase_ , **self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(lowerCAmelCase_ , torch.Tensor ): _A = processed else: _A = list(processed.keys() )[0] _A = processed[key] if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _A = len(lowerCAmelCase_ ) else: _A = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _A = observed_batch_size # Setting internal index to unwrap the batch _A = processed _A = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class lowercase_ ( _lowerCamelCase ): '''simple docstring''' def __init__( self : List[Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : int , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any]=None ): super().__init__(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) def __iter__( self : List[str] ): _A = iter(self.loader ) _A = None return self def lowerCAmelCase_ ( self : List[str] ): if self.subiterator is None: _A = self.infer(next(self.iterator ) , **self.params ) try: # Try to return next item _A = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators _A = self.infer(next(self.iterator ) , **self.params ) _A = next(self.subiterator ) return processed class lowercase_ ( _lowerCamelCase ): '''simple docstring''' def __iter__( self : Union[str, Any] ): _A = iter(self.loader ) return self def lowerCAmelCase_ ( self : str ): _A = False _A = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: _A = self.loader_batch_item() _A = item.pop('is_last' ) accumulator.append(lowerCAmelCase_ ) if is_last: return accumulator while not is_last: _A = self.infer(next(self.iterator ) , **self.params ) if self.loader_batch_size is not None: if isinstance(lowerCAmelCase_ , torch.Tensor ): _A = processed else: _A = list(processed.keys() )[0] _A = processed[key] if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): _A = len(lowerCAmelCase_ ) else: _A = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _A = observed_batch_size _A = processed _A = 0 while self._loader_batch_index < self.loader_batch_size: _A = self.loader_batch_item() _A = item.pop('is_last' ) accumulator.append(lowerCAmelCase_ ) if is_last: return accumulator else: _A = processed _A = item.pop('is_last' ) accumulator.append(lowerCAmelCase_ ) return accumulator class lowercase_ ( _lowerCamelCase ): '''simple docstring''' def __init__( self : Tuple , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : List[str] ): _A = dataset _A = key def __len__( self : Union[str, Any] ): return len(self.dataset ) def __getitem__( self : Any , _UpperCAmelCase : List[str] ): return self.dataset[i][self.key] class lowercase_ ( _lowerCamelCase ): '''simple docstring''' def __init__( self : Dict , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : Union[str, Any] ): _A = dataset _A = keya _A = keya def __len__( self : Union[str, Any] ): return len(self.dataset ) def __getitem__( self : Dict , _UpperCAmelCase : str ): return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
315
'''simple docstring''' import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class __UpperCAmelCase ( _lowerCamelCase ): @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError("Offline mode is enabled, we shouldn\'t access internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _snake_case = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCAmelCase_ ) BertModel.from_pretrained(lowerCAmelCase_ ) BertTokenizer.from_pretrained(lowerCAmelCase_ ) pipeline(task='fill-mask' , model=lowerCAmelCase_ ) # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _snake_case = self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task="fill-mask", model=mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Faking flaky internet")\nsocket.socket = offline_socket\n ' # Force fetching the files so that we can use the cache _snake_case = 'hf-internal-testing/tiny-random-bert' BertConfig.from_pretrained(lowerCAmelCase_ ) BertModel.from_pretrained(lowerCAmelCase_ ) BertTokenizer.from_pretrained(lowerCAmelCase_ ) pipeline(task='fill-mask' , model=lowerCAmelCase_ ) # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run, mock] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import BertConfig, BertModel, BertTokenizer\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert-sharded"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint("success")\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError("Offline mode is enabled")\nsocket.socket = offline_socket\n ' # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # next emulate no network _snake_case = [sys.executable, '-c', '\n'.join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import pipeline\n ' _snake_case = '\nmname = "hf-internal-testing/tiny-random-bert"\npipe = pipeline(model=mname)\n ' _snake_case = '\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error("Offline mode is enabled")\nsocket.socket = offline_socket\n ' _snake_case = self.get_env() _snake_case = '1' _snake_case = [sys.executable, '-c', '\n'.join([load, mock, run] )] _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( 'You cannot infer task automatically within `pipeline` when using offline mode' , result.stderr.decode().replace('\n' , '' ) , ) @require_torch def lowerCamelCase ( self ): """simple docstring""" _snake_case = '\nfrom transformers import AutoModel\n ' _snake_case = '\nmname = "hf-internal-testing/test_dynamic_model"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint("success")\n ' # baseline - just load from_pretrained with normal network _snake_case = [sys.executable, '-c', '\n'.join([load, run] )] # should succeed _snake_case = self.get_env() _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files _snake_case = '1' _snake_case = subprocess.run(lowerCAmelCase_ , env=lowerCAmelCase_ , check=lowerCAmelCase_ , capture_output=lowerCAmelCase_ ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn('success' , result.stdout.decode() )
42
0
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class __lowerCamelCase (_lowerCamelCase ): _lowercase = ["""image_processor""", """tokenizer"""] _lowercase = """ChineseCLIPImageProcessor""" _lowercase = ("""BertTokenizer""", """BertTokenizerFast""") def __init__( self: Tuple,A_: Optional[int]=None,A_: Dict=None,**A_: Union[str, Any] ): '''simple docstring''' __UpperCamelCase = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.',lowerCAmelCase_,) __UpperCamelCase = kwargs.pop('feature_extractor' ) __UpperCamelCase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(lowerCAmelCase_,lowerCAmelCase_ ) __UpperCamelCase = self.image_processor def __call__( self: str,A_: int=None,A_: Tuple=None,A_: Union[str, Any]=None,**A_: Dict ): '''simple docstring''' if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: __UpperCamelCase = self.tokenizer(lowerCAmelCase_,return_tensors=lowerCAmelCase_,**lowerCAmelCase_ ) if images is not None: __UpperCamelCase = self.image_processor(lowerCAmelCase_,return_tensors=lowerCAmelCase_,**lowerCAmelCase_ ) if text is not None and images is not None: __UpperCamelCase = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**lowerCAmelCase_ ),tensor_type=lowerCAmelCase_ ) def snake_case_ ( self: Optional[int],*A_: Union[str, Any],**A_: Optional[int] ): '''simple docstring''' return self.tokenizer.batch_decode(*lowerCAmelCase_,**lowerCAmelCase_ ) def snake_case_ ( self: List[Any],*A_: Any,**A_: str ): '''simple docstring''' return self.tokenizer.decode(*lowerCAmelCase_,**lowerCAmelCase_ ) @property def snake_case_ ( self: Optional[int] ): '''simple docstring''' __UpperCamelCase = self.tokenizer.model_input_names __UpperCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def snake_case_ ( self: List[Any] ): '''simple docstring''' warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.',lowerCAmelCase_,) return self.image_processor_class
310
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device if is_torch_available(): from transformers import AutoModelForSeqaSeqLM, AutoTokenizer @require_torch @require_sentencepiece @require_tokenizers class __UpperCAmelCase ( unittest.TestCase ): @slow def lowerCamelCase ( self ): """simple docstring""" _snake_case = AutoModelForSeqaSeqLM.from_pretrained('google/mt5-small' , return_dict=lowerCAmelCase_ ).to(lowerCAmelCase_ ) _snake_case = AutoTokenizer.from_pretrained('google/mt5-small' ) _snake_case = tokenizer('Hello there' , return_tensors='pt' ).input_ids _snake_case = tokenizer('Hi I am' , return_tensors='pt' ).input_ids _snake_case = model(input_ids.to(lowerCAmelCase_ ) , labels=labels.to(lowerCAmelCase_ ) ).loss _snake_case = -(labels.shape[-1] * loss.item()) _snake_case = -84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
42
0
import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class lowerCamelCase_ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : Any = CTRLTokenizer a__ : Any = False a__ : Tuple = False def UpperCamelCase__ ( self) -> Tuple: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __UpperCamelCase :int = ['''adapt''', '''re@@''', '''a@@''', '''apt''', '''c@@''', '''t''', '''<unk>'''] __UpperCamelCase :Union[str, Any] = dict(zip(__lowercase , range(len(__lowercase)))) __UpperCamelCase :Dict = ['''#version: 0.2''', '''a p''', '''ap t</w>''', '''r e''', '''a d''', '''ad apt</w>''', ''''''] __UpperCamelCase :List[str] = {'''unk_token''': '''<unk>'''} __UpperCamelCase :Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file''']) __UpperCamelCase :Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file''']) with open(self.vocab_file , '''w''' , encoding='''utf-8''') as fp: fp.write(json.dumps(__lowercase) + '''\n''') with open(self.merges_file , '''w''' , encoding='''utf-8''') as fp: fp.write('''\n'''.join(__lowercase)) def UpperCamelCase__ ( self , **__lowercase) -> Dict: kwargs.update(self.special_tokens_map) return CTRLTokenizer.from_pretrained(self.tmpdirname , **__lowercase) def UpperCamelCase__ ( self , __lowercase) -> Optional[int]: __UpperCamelCase :str = '''adapt react readapt apt''' __UpperCamelCase :Optional[Any] = '''adapt react readapt apt''' return input_text, output_text def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :Any = CTRLTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map) __UpperCamelCase :Dict = '''adapt react readapt apt''' __UpperCamelCase :Optional[Any] = '''adapt re@@ a@@ c@@ t re@@ adapt apt'''.split() __UpperCamelCase :Any = tokenizer.tokenize(__lowercase) self.assertListEqual(__lowercase , __lowercase) __UpperCamelCase :Optional[Any] = tokens + [tokenizer.unk_token] __UpperCamelCase :List[str] = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowercase) , __lowercase)
43
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class lowerCamelCase_ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : str = TextToVideoSDPipeline a__ : Union[str, Any] = TEXT_TO_IMAGE_PARAMS a__ : Tuple = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. a__ : int = frozenset( [ """num_inference_steps""", """generator""", """latents""", """return_dict""", """callback""", """callback_steps""", ] ) def UpperCamelCase__ ( self) -> Optional[Any]: torch.manual_seed(0) __UpperCamelCase :str = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''DownBlock3D''') , up_block_types=('''UpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''') , cross_attention_dim=32 , attention_head_dim=4 , ) __UpperCamelCase :Optional[int] = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=__lowercase , set_alpha_to_one=__lowercase , ) torch.manual_seed(0) __UpperCamelCase :Optional[int] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0) __UpperCamelCase :Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act='''gelu''' , projection_dim=512 , ) __UpperCamelCase :Optional[Any] = CLIPTextModel(__lowercase) __UpperCamelCase :Optional[int] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') __UpperCamelCase :Union[str, Any] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0) -> Optional[int]: if str(__lowercase).startswith('''mps'''): __UpperCamelCase :List[Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :Tuple = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :int = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Optional[int] = self.get_dummy_components() __UpperCamelCase :Dict = TextToVideoSDPipeline(**__lowercase) __UpperCamelCase :Any = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :Optional[Any] = self.get_dummy_inputs(__lowercase) __UpperCamelCase :int = '''np''' __UpperCamelCase :List[str] = sd_pipe(**__lowercase).frames __UpperCamelCase :Optional[Any] = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) __UpperCamelCase :str = np.array([1_58.0, 1_60.0, 1_53.0, 1_25.0, 1_00.0, 1_21.0, 1_11.0, 93.0, 1_13.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase__ ( self) -> Tuple: self._test_attention_slicing_forward_pass(test_mean_pixel_difference=__lowercase , expected_max_diff=3E-3) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase__ ( self) -> Optional[int]: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__lowercase , expected_max_diff=1E-2) @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCamelCase__ ( self) -> Union[str, Any]: pass @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCamelCase__ ( self) -> Dict: pass @unittest.skip(reason='''`num_images_per_prompt` argument is not supported for this pipeline.''') def UpperCamelCase__ ( self) -> str: pass def UpperCamelCase__ ( self) -> List[str]: return super().test_progress_bar() @slow @skip_mps class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Union[str, Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy''') __UpperCamelCase :List[str] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''') __UpperCamelCase :Union[str, Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) __UpperCamelCase :str = pipe.to('''cuda''') __UpperCamelCase :Optional[Any] = '''Spiderman is surfing''' __UpperCamelCase :Union[str, Any] = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :List[Any] = pipe(__lowercase , generator=__lowercase , num_inference_steps=25 , output_type='''pt''').frames __UpperCamelCase :Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video).mean() < 5E-2 def UpperCamelCase__ ( self) -> int: __UpperCamelCase :str = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy''') __UpperCamelCase :Union[str, Any] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''') __UpperCamelCase :str = pipe.to('''cuda''') __UpperCamelCase :Union[str, Any] = '''Spiderman is surfing''' __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :List[Any] = pipe(__lowercase , generator=__lowercase , num_inference_steps=2 , output_type='''pt''').frames __UpperCamelCase :Optional[Any] = video_frames.cpu().numpy() assert np.abs(expected_video - video).mean() < 5E-2
43
1
import datasets __lowercase = '''\ @InProceedings{conneau2018xnli, author = "Conneau, Alexis and Rinott, Ruty and Lample, Guillaume and Williams, Adina and Bowman, Samuel R. and Schwenk, Holger and Stoyanov, Veselin", title = "XNLI: Evaluating Cross-lingual Sentence Representations", booktitle = "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing", year = "2018", publisher = "Association for Computational Linguistics", location = "Brussels, Belgium", } ''' __lowercase = '''\ XNLI is a subset of a few thousand examples from MNLI which has been translated into a 14 different languages (some low-ish resource). As with MNLI, the goal is to predict textual entailment (does sentence A imply/contradict/neither sentence B) and is a classification task (given two sentences, predict one of three labels). ''' __lowercase = ''' Computes XNLI score which is just simple accuracy. Args: predictions: Predicted labels. references: Ground truth labels. Returns: \'accuracy\': accuracy Examples: >>> predictions = [0, 1] >>> references = [0, 1] >>> xnli_metric = datasets.load_metric("xnli") >>> results = xnli_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0} ''' def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCamelCase_ ( datasets.Metric ): '''simple docstring''' def UpperCamelCase__ ( self) -> Optional[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''int64''' if self.config_name != '''sts-b''' else '''float32'''), '''references''': datasets.Value('''int64''' if self.config_name != '''sts-b''' else '''float32'''), }) , codebase_urls=[] , reference_urls=[] , format='''numpy''' , ) def UpperCamelCase__ ( self , __lowercase , __lowercase) -> Optional[Any]: return {"accuracy": simple_accuracy(__lowercase , __lowercase)}
43
def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = [0 for i in range(len(SCREAMING_SNAKE_CASE ) )] # initialize interval's left pointer and right pointer __UpperCamelCase , __UpperCamelCase :str = 0, 0 for i in range(1 , len(SCREAMING_SNAKE_CASE ) ): # case when current index is inside the interval if i <= right_pointer: __UpperCamelCase :Union[str, Any] = min(right_pointer - i + 1 , z_result[i - left_pointer] ) __UpperCamelCase :Tuple = min_edge while go_next(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: __UpperCamelCase , __UpperCamelCase :Union[str, Any] = i, i + z_result[i] - 1 return z_result def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' return i + z_result[i] < len(SCREAMING_SNAKE_CASE ) and s[z_result[i]] == s[i + z_result[i]] def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string __UpperCamelCase :Tuple = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(SCREAMING_SNAKE_CASE ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
43
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowercase = { '''configuration_biogpt''': ['''BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BioGptConfig'''], '''tokenization_biogpt''': ['''BioGptTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BioGptForCausalLM''', '''BioGptForTokenClassification''', '''BioGptForSequenceClassification''', '''BioGptModel''', '''BioGptPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
43
import math from typing import Any, Callable, List, Optional, Tuple, Union import numpy as np import torch from ...models import TaFilmDecoder from ...schedulers import DDPMScheduler from ...utils import is_onnx_available, logging, randn_tensor if is_onnx_available(): from ..onnx_utils import OnnxRuntimeModel from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline from .continous_encoder import SpectrogramContEncoder from .notes_encoder import SpectrogramNotesEncoder __lowercase = logging.get_logger(__name__) # pylint: disable=invalid-name __lowercase = 256 class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Tuple = ["""melgan"""] def __init__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> None: super().__init__() # From MELGAN __UpperCamelCase :int = math.log(1E-5) # Matches MelGAN training. __UpperCamelCase :int = 4.0 # Largest value for most examples __UpperCamelCase :str = 128 self.register_modules( notes_encoder=__lowercase , continuous_encoder=__lowercase , decoder=__lowercase , scheduler=__lowercase , melgan=__lowercase , ) def UpperCamelCase__ ( self , __lowercase , __lowercase=(-1.0, 1.0) , __lowercase=False) -> Dict: __UpperCamelCase , __UpperCamelCase :str = output_range if clip: __UpperCamelCase :Union[str, Any] = torch.clip(__lowercase , self.min_value , self.max_value) # Scale to [0, 1]. __UpperCamelCase :Union[str, Any] = (features - self.min_value) / (self.max_value - self.min_value) # Scale to [min_out, max_out]. return zero_one * (max_out - min_out) + min_out def UpperCamelCase__ ( self , __lowercase , __lowercase=(-1.0, 1.0) , __lowercase=False) -> Optional[int]: __UpperCamelCase , __UpperCamelCase :int = input_range __UpperCamelCase :Optional[int] = torch.clip(__lowercase , __lowercase , __lowercase) if clip else outputs # Scale to [0, 1]. __UpperCamelCase :List[str] = (outputs - min_out) / (max_out - min_out) # Scale to [self.min_value, self.max_value]. return zero_one * (self.max_value - self.min_value) + self.min_value def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase) -> List[Any]: __UpperCamelCase :List[str] = input_tokens > 0 __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self.notes_encoder( encoder_input_tokens=__lowercase , encoder_inputs_mask=__lowercase) __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self.continuous_encoder( encoder_inputs=__lowercase , encoder_inputs_mask=__lowercase) return [(tokens_encoded, tokens_mask), (continuous_encoded, continuous_mask)] def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase) -> str: __UpperCamelCase :Optional[int] = noise_time if not torch.is_tensor(__lowercase): __UpperCamelCase :str = torch.tensor([timesteps] , dtype=torch.long , device=input_tokens.device) elif torch.is_tensor(__lowercase) and len(timesteps.shape) == 0: __UpperCamelCase :Dict = timesteps[None].to(input_tokens.device) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML __UpperCamelCase :List[str] = timesteps * torch.ones(input_tokens.shape[0] , dtype=timesteps.dtype , device=timesteps.device) __UpperCamelCase :Tuple = self.decoder( encodings_and_masks=__lowercase , decoder_input_tokens=__lowercase , decoder_noise_time=__lowercase) return logits @torch.no_grad() def __call__( self , __lowercase , __lowercase = None , __lowercase = 100 , __lowercase = True , __lowercase = "numpy" , __lowercase = None , __lowercase = 1 , ) -> Union[AudioPipelineOutput, Tuple]: if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__lowercase , __lowercase) or callback_steps <= 0) ): raise ValueError( f"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" f""" {type(__lowercase)}.""") __UpperCamelCase :Union[str, Any] = np.zeros([1, TARGET_FEATURE_LENGTH, self.n_dims] , dtype=np.floataa) __UpperCamelCase :Union[str, Any] = np.zeros([1, 0, self.n_dims] , np.floataa) __UpperCamelCase :Union[str, Any] = torch.ones((1, TARGET_FEATURE_LENGTH) , dtype=__lowercase , device=self.device) for i, encoder_input_tokens in enumerate(__lowercase): if i == 0: __UpperCamelCase :int = torch.from_numpy(pred_mel[:1].copy()).to( device=self.device , dtype=self.decoder.dtype) # The first chunk has no previous context. __UpperCamelCase :int = torch.zeros((1, TARGET_FEATURE_LENGTH) , dtype=__lowercase , device=self.device) else: # The full song pipeline does not feed in a context feature, so the mask # will be all 0s after the feature converter. Because we know we're # feeding in a full context chunk from the previous prediction, set it # to all 1s. __UpperCamelCase :Tuple = ones __UpperCamelCase :Optional[Any] = self.scale_features( __lowercase , output_range=[-1.0, 1.0] , clip=__lowercase) __UpperCamelCase :int = self.encode( input_tokens=torch.IntTensor([encoder_input_tokens]).to(device=self.device) , continuous_inputs=__lowercase , continuous_mask=__lowercase , ) # Sample encoder_continuous_inputs shaped gaussian noise to begin loop __UpperCamelCase :int = randn_tensor( shape=encoder_continuous_inputs.shape , generator=__lowercase , device=self.device , dtype=self.decoder.dtype , ) # set step values self.scheduler.set_timesteps(__lowercase) # Denoising diffusion loop for j, t in enumerate(self.progress_bar(self.scheduler.timesteps)): __UpperCamelCase :Optional[int] = self.decode( encodings_and_masks=__lowercase , input_tokens=__lowercase , noise_time=t / self.scheduler.config.num_train_timesteps , ) # Compute previous output: x_t -> x_t-1 __UpperCamelCase :int = self.scheduler.step(__lowercase , __lowercase , __lowercase , generator=__lowercase).prev_sample __UpperCamelCase :Tuple = self.scale_to_features(__lowercase , input_range=[-1.0, 1.0]) __UpperCamelCase :List[Any] = mel[:1] __UpperCamelCase :Optional[Any] = mel.cpu().float().numpy() __UpperCamelCase :Any = np.concatenate([full_pred_mel, pred_mel[:1]] , axis=1) # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__lowercase , __lowercase) logger.info('''Generated segment''' , __lowercase) if output_type == "numpy" and not is_onnx_available(): raise ValueError( '''Cannot return output in \'np\' format if ONNX is not available. Make sure to have ONNX installed or set \'output_type\' to \'mel\'.''') elif output_type == "numpy" and self.melgan is None: raise ValueError( '''Cannot return output in \'np\' format if melgan component is not defined. Make sure to define `self.melgan` or set \'output_type\' to \'mel\'.''') if output_type == "numpy": __UpperCamelCase :Optional[Any] = self.melgan(input_features=full_pred_mel.astype(np.floataa)) else: __UpperCamelCase :List[str] = full_pred_mel if not return_dict: return (output,) return AudioPipelineOutput(audios=__lowercase)
43
1
def lowerCamelCase ( SCREAMING_SNAKE_CASE = 100 ): '''simple docstring''' __UpperCamelCase :Any = 0 __UpperCamelCase :int = 0 for i in range(1 , n + 1 ): sum_of_squares += i**2 sum_of_ints += i return sum_of_ints**2 - sum_of_squares if __name__ == "__main__": print(F'{solution() = }')
43
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( HubertConfig, HubertForCTC, HubertModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __lowercase = logging.get_logger(__name__) __lowercase = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' for attribute in key.split('''.''' ): __UpperCamelCase :str = getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if weight_type is not None: __UpperCamelCase :Any = getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).shape else: __UpperCamelCase :Union[str, Any] = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __UpperCamelCase :str = value elif weight_type == "weight_g": __UpperCamelCase :List[str] = value elif weight_type == "weight_v": __UpperCamelCase :str = value elif weight_type == "bias": __UpperCamelCase :Union[str, Any] = value else: __UpperCamelCase :str = value logger.info(f"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = [] __UpperCamelCase :int = fairseq_model.state_dict() __UpperCamelCase :List[Any] = hf_model.hubert.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): __UpperCamelCase :List[Any] = False if "conv_layers" in name: load_conv_layer( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , hf_model.config.feat_extract_norm == '''group''' , ) __UpperCamelCase :List[str] = True else: for key, mapped_key in MAPPING.items(): __UpperCamelCase :Dict = '''hubert.''' + mapped_key if (is_finetuned and mapped_key != '''lm_head''') else mapped_key if key in name or (key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0] and not is_finetuned): __UpperCamelCase :Optional[Any] = True if "*" in mapped_key: __UpperCamelCase :List[str] = name.split(SCREAMING_SNAKE_CASE )[0].split('''.''' )[-2] __UpperCamelCase :Optional[int] = mapped_key.replace('''*''' , SCREAMING_SNAKE_CASE ) if "weight_g" in name: __UpperCamelCase :int = '''weight_g''' elif "weight_v" in name: __UpperCamelCase :List[Any] = '''weight_v''' elif "weight" in name: __UpperCamelCase :Dict = '''weight''' elif "bias" in name: __UpperCamelCase :Dict = '''bias''' else: __UpperCamelCase :Dict = None set_recursively(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(SCREAMING_SNAKE_CASE ) logger.warning(f"""Unused weights: {unused_weights}""" ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = full_name.split('''conv_layers.''' )[-1] __UpperCamelCase :Optional[int] = name.split('''.''' ) __UpperCamelCase :str = int(items[0] ) __UpperCamelCase :List[Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __UpperCamelCase :Dict = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __UpperCamelCase :Any = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) __UpperCamelCase :int = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) __UpperCamelCase :Union[str, Any] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(SCREAMING_SNAKE_CASE ) @torch.no_grad() def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=True ): '''simple docstring''' if config_path is not None: __UpperCamelCase :Tuple = HubertConfig.from_pretrained(SCREAMING_SNAKE_CASE ) else: __UpperCamelCase :Optional[int] = HubertConfig() if is_finetuned: if dict_path: __UpperCamelCase :Optional[int] = Dictionary.load(SCREAMING_SNAKE_CASE ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __UpperCamelCase :Optional[int] = target_dict.pad_index __UpperCamelCase :Dict = target_dict.bos_index __UpperCamelCase :str = target_dict.eos_index __UpperCamelCase :Dict = len(target_dict.symbols ) __UpperCamelCase :List[Any] = os.path.join(SCREAMING_SNAKE_CASE , '''vocab.json''' ) if not os.path.isdir(SCREAMING_SNAKE_CASE ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(SCREAMING_SNAKE_CASE ) ) return os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) with open(SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(target_dict.indices , SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = WavaVecaCTCTokenizer( SCREAMING_SNAKE_CASE , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :Union[str, Any] = True if config.feat_extract_norm == '''layer''' else False __UpperCamelCase :Any = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0 , do_normalize=SCREAMING_SNAKE_CASE , return_attention_mask=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :Any = WavaVecaProcessor(feature_extractor=SCREAMING_SNAKE_CASE , tokenizer=SCREAMING_SNAKE_CASE ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[str] = HubertForCTC(SCREAMING_SNAKE_CASE ) else: __UpperCamelCase :str = HubertModel(SCREAMING_SNAKE_CASE ) if is_finetuned: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :int = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Any = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) __UpperCamelCase :Dict = model[0].eval() recursively_load_weights(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) hf_wavavec.save_pretrained(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) __lowercase = parser.parse_args() convert_hubert_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
43
1
import logging from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import arg_to_scheduler from transformers import TrainingArguments __lowercase = logging.getLogger(__name__) @dataclass class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[float] = field( default=0.0 , metadata={"""help""": """The label smoothing epsilon to apply (if not zero)."""} ) a__ : bool = field(default=UpperCAmelCase_ , metadata={"""help""": """Whether to SortishSamler or not."""} ) a__ : bool = field( default=UpperCAmelCase_ , metadata={"""help""": """Whether to use generate to calculate generative metrics (ROUGE, BLEU)."""} ) a__ : bool = field(default=UpperCAmelCase_ , metadata={"""help""": """whether to use adafactor"""} ) a__ : Optional[float] = field( default=UpperCAmelCase_ , metadata={"""help""": """Encoder layer dropout probability. Goes into model.config."""} ) a__ : Optional[float] = field( default=UpperCAmelCase_ , metadata={"""help""": """Decoder layer dropout probability. Goes into model.config."""} ) a__ : Optional[float] = field(default=UpperCAmelCase_ , metadata={"""help""": """Dropout probability. Goes into model.config."""} ) a__ : Optional[float] = field( default=UpperCAmelCase_ , metadata={"""help""": """Attention dropout probability. Goes into model.config."""} ) a__ : Optional[str] = field( default="""linear""" , metadata={"""help""": F'''Which lr scheduler to use. Selected in {sorted(arg_to_scheduler.keys() )}'''} , )
43
import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters __lowercase = (720, 1280) # Height, Width __lowercase = (0.4, 0.6) # if height or width lower than this scale, drop it. __lowercase = 1 / 100 __lowercase = '''''' __lowercase = '''''' __lowercase = '''''' __lowercase = 250 def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase , __UpperCamelCase :List[Any] = get_dataset(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for index in range(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = random.sample(range(len(SCREAMING_SNAKE_CASE ) ) , 4 ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :str = update_image_and_anno( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , filter_scale=SCREAMING_SNAKE_CASE , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __UpperCamelCase :List[Any] = random_chars(32 ) __UpperCamelCase :List[str] = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __UpperCamelCase :Tuple = f"""{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}""" cva.imwrite(f"""{file_root}.jpg""" , SCREAMING_SNAKE_CASE , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(f"""Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}""" ) __UpperCamelCase :Optional[Any] = [] for anno in new_annos: __UpperCamelCase :int = anno[3] - anno[1] __UpperCamelCase :Optional[int] = anno[4] - anno[2] __UpperCamelCase :int = anno[1] + width / 2 __UpperCamelCase :List[str] = anno[2] + height / 2 __UpperCamelCase :str = f"""{anno[0]} {x_center} {y_center} {width} {height}""" annos_list.append(SCREAMING_SNAKE_CASE ) with open(f"""{file_root}.txt""" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :str = [] __UpperCamelCase :str = [] for label_file in glob.glob(os.path.join(SCREAMING_SNAKE_CASE , '''*.txt''' ) ): __UpperCamelCase :Any = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(SCREAMING_SNAKE_CASE ) as in_file: __UpperCamelCase :str = in_file.readlines() __UpperCamelCase :Optional[int] = os.path.join(SCREAMING_SNAKE_CASE , f"""{label_name}.jpg""" ) __UpperCamelCase :int = [] for obj_list in obj_lists: __UpperCamelCase :Optional[int] = obj_list.rstrip('''\n''' ).split(''' ''' ) __UpperCamelCase :Any = float(obj[1] ) - float(obj[3] ) / 2 __UpperCamelCase :List[str] = float(obj[2] ) - float(obj[4] ) / 2 __UpperCamelCase :Dict = float(obj[1] ) + float(obj[3] ) / 2 __UpperCamelCase :List[str] = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(SCREAMING_SNAKE_CASE ) labels.append(SCREAMING_SNAKE_CASE ) return img_paths, labels def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 0.0 , ): '''simple docstring''' __UpperCamelCase :List[str] = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __UpperCamelCase :List[Any] = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __UpperCamelCase :int = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __UpperCamelCase :Optional[int] = int(scale_x * output_size[1] ) __UpperCamelCase :Any = int(scale_y * output_size[0] ) __UpperCamelCase :List[str] = [] __UpperCamelCase :Dict = [] for i, index in enumerate(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Any = all_img_list[index] path_list.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = all_annos[index] __UpperCamelCase :Union[str, Any] = cva.imread(SCREAMING_SNAKE_CASE ) if i == 0: # top-left __UpperCamelCase :str = cva.resize(SCREAMING_SNAKE_CASE , (divid_point_x, divid_point_y) ) __UpperCamelCase :Union[str, Any] = img for bbox in img_annos: __UpperCamelCase :Union[str, Any] = bbox[1] * scale_x __UpperCamelCase :Optional[Any] = bbox[2] * scale_y __UpperCamelCase :int = bbox[3] * scale_x __UpperCamelCase :Union[str, Any] = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __UpperCamelCase :str = cva.resize(SCREAMING_SNAKE_CASE , (output_size[1] - divid_point_x, divid_point_y) ) __UpperCamelCase :List[str] = img for bbox in img_annos: __UpperCamelCase :str = scale_x + bbox[1] * (1 - scale_x) __UpperCamelCase :Dict = bbox[2] * scale_y __UpperCamelCase :Optional[Any] = scale_x + bbox[3] * (1 - scale_x) __UpperCamelCase :List[Any] = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __UpperCamelCase :str = cva.resize(SCREAMING_SNAKE_CASE , (divid_point_x, output_size[0] - divid_point_y) ) __UpperCamelCase :Optional[int] = img for bbox in img_annos: __UpperCamelCase :Tuple = bbox[1] * scale_x __UpperCamelCase :Optional[Any] = scale_y + bbox[2] * (1 - scale_y) __UpperCamelCase :Tuple = bbox[3] * scale_x __UpperCamelCase :Dict = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __UpperCamelCase :Optional[int] = cva.resize( SCREAMING_SNAKE_CASE , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __UpperCamelCase :Optional[int] = img for bbox in img_annos: __UpperCamelCase :Optional[Any] = scale_x + bbox[1] * (1 - scale_x) __UpperCamelCase :Optional[int] = scale_y + bbox[2] * (1 - scale_y) __UpperCamelCase :Optional[Any] = scale_x + bbox[3] * (1 - scale_x) __UpperCamelCase :int = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __UpperCamelCase :List[Any] = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' assert number_char > 1, "The number of character should greater than 1" __UpperCamelCase :Optional[Any] = ascii_lowercase + digits return "".join(random.choice(SCREAMING_SNAKE_CASE ) for _ in range(SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": main() print('''DONE ✅''')
43
1
import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCamelCase_ : '''simple docstring''' def __init__( self , __lowercase , __lowercase=13 , __lowercase=7 , __lowercase=True , __lowercase=True , __lowercase=False , __lowercase=True , __lowercase=99 , __lowercase=32 , __lowercase=5 , __lowercase=4 , __lowercase=37 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=512 , __lowercase=16 , __lowercase=2 , __lowercase=0.02 , __lowercase=3 , __lowercase=4 , __lowercase=None , ) -> Optional[Any]: __UpperCamelCase :Tuple = parent __UpperCamelCase :Union[str, Any] = batch_size __UpperCamelCase :int = seq_length __UpperCamelCase :Optional[Any] = is_training __UpperCamelCase :Any = use_input_mask __UpperCamelCase :Any = use_token_type_ids __UpperCamelCase :str = use_labels __UpperCamelCase :int = vocab_size __UpperCamelCase :Optional[int] = hidden_size __UpperCamelCase :str = num_hidden_layers __UpperCamelCase :Union[str, Any] = num_attention_heads __UpperCamelCase :Dict = intermediate_size __UpperCamelCase :Tuple = hidden_act __UpperCamelCase :Dict = hidden_dropout_prob __UpperCamelCase :str = attention_probs_dropout_prob __UpperCamelCase :Union[str, Any] = max_position_embeddings __UpperCamelCase :Dict = type_vocab_size __UpperCamelCase :Optional[int] = type_sequence_label_size __UpperCamelCase :Tuple = initializer_range __UpperCamelCase :Optional[Any] = num_labels __UpperCamelCase :Any = num_choices __UpperCamelCase :Dict = scope def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __UpperCamelCase :Union[str, Any] = None if self.use_input_mask: __UpperCamelCase :Dict = random_attention_mask([self.batch_size, self.seq_length]) __UpperCamelCase :str = None if self.use_token_type_ids: __UpperCamelCase :Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) __UpperCamelCase :List[str] = None __UpperCamelCase :Any = None __UpperCamelCase :str = None if self.use_labels: __UpperCamelCase :int = ids_tensor([self.batch_size] , self.type_sequence_label_size) __UpperCamelCase :Any = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __UpperCamelCase :str = ids_tensor([self.batch_size] , self.num_choices) __UpperCamelCase :str = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase__ ( self) -> Union[str, Any]: return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowercase , initializer_range=self.initializer_range , ) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase) -> List[Any]: __UpperCamelCase :Any = BioGptModel(config=__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :int = model(__lowercase , attention_mask=__lowercase) __UpperCamelCase :int = model(__lowercase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> str: __UpperCamelCase :Dict = BioGptForCausalLM(config=__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :Any = model(__lowercase , attention_mask=__lowercase , token_type_ids=__lowercase , labels=__lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , *__lowercase) -> str: __UpperCamelCase :Optional[Any] = BioGptModel(config=__lowercase) model.to(__lowercase) model.eval() # create attention mask __UpperCamelCase :Tuple = torch.ones(input_ids.shape , dtype=torch.long , device=__lowercase) __UpperCamelCase :Optional[Any] = self.seq_length // 2 __UpperCamelCase :Tuple = 0 # first forward pass __UpperCamelCase , __UpperCamelCase :Optional[int] = model(__lowercase , attention_mask=__lowercase).to_tuple() # create hypothetical next token and extent to next_input_ids __UpperCamelCase :List[Any] = ids_tensor((self.batch_size, 1) , config.vocab_size) # change a random masked slice from input_ids __UpperCamelCase :Union[str, Any] = ids_tensor((1,) , __lowercase).item() + 1 __UpperCamelCase :str = ids_tensor((self.batch_size, 1) , config.vocab_size).squeeze(-1) __UpperCamelCase :int = random_other_next_tokens # append to next input_ids and attn_mask __UpperCamelCase :Tuple = torch.cat([input_ids, next_tokens] , dim=-1) __UpperCamelCase :List[str] = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=__lowercase)] , dim=1 , ) # get two different outputs __UpperCamelCase :List[str] = model(__lowercase , attention_mask=__lowercase)['''last_hidden_state'''] __UpperCamelCase :Optional[int] = model(__lowercase , past_key_values=__lowercase , attention_mask=__lowercase)['''last_hidden_state'''] # select random slice __UpperCamelCase :str = ids_tensor((1,) , output_from_past.shape[-1]).item() __UpperCamelCase :str = output_from_no_past[:, -1, random_slice_idx].detach() __UpperCamelCase :Tuple = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__lowercase , __lowercase , atol=1E-3)) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , *__lowercase) -> List[str]: __UpperCamelCase :str = BioGptModel(config=__lowercase).to(__lowercase).eval() __UpperCamelCase :Optional[Any] = torch.ones(input_ids.shape , dtype=torch.long , device=__lowercase) # first forward pass __UpperCamelCase :Dict = model(__lowercase , attention_mask=__lowercase , use_cache=__lowercase) __UpperCamelCase , __UpperCamelCase :Optional[int] = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids __UpperCamelCase :Optional[Any] = ids_tensor((self.batch_size, 3) , config.vocab_size) __UpperCamelCase :List[str] = ids_tensor((self.batch_size, 3) , 2) # append to next input_ids and __UpperCamelCase :Optional[Any] = torch.cat([input_ids, next_tokens] , dim=-1) __UpperCamelCase :Union[str, Any] = torch.cat([attention_mask, next_attn_mask] , dim=-1) __UpperCamelCase :Tuple = model(__lowercase , attention_mask=__lowercase)['''last_hidden_state'''] __UpperCamelCase :Optional[Any] = model(__lowercase , attention_mask=__lowercase , past_key_values=__lowercase)[ '''last_hidden_state''' ] # select random slice __UpperCamelCase :Tuple = ids_tensor((1,) , output_from_past.shape[-1]).item() __UpperCamelCase :int = output_from_no_past[:, -3:, random_slice_idx].detach() __UpperCamelCase :Optional[int] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1]) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__lowercase , __lowercase , atol=1E-3)) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , *__lowercase , __lowercase=False) -> List[str]: __UpperCamelCase :List[Any] = BioGptForCausalLM(__lowercase) model.to(__lowercase) if gradient_checkpointing: model.gradient_checkpointing_enable() __UpperCamelCase :str = model(__lowercase , labels=__lowercase) self.parent.assertEqual(result.loss.shape , ()) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) result.loss.backward() def UpperCamelCase__ ( self , __lowercase , *__lowercase) -> int: __UpperCamelCase :List[str] = BioGptModel(__lowercase) __UpperCamelCase :int = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key]) - model_std) , 0.0_01) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key]) - 0.0) , 0.01) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , *__lowercase) -> Dict: __UpperCamelCase :List[str] = self.num_labels __UpperCamelCase :List[Any] = BioGptForTokenClassification(__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :List[str] = model(__lowercase , attention_mask=__lowercase , token_type_ids=__lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels)) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :List[Any] = self.prepare_config_and_inputs() ( ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ) :Union[str, Any] = config_and_inputs __UpperCamelCase :Any = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : Optional[int] = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) a__ : str = (BioGptForCausalLM,) if is_torch_available() else () a__ : str = ( { """feature-extraction""": BioGptModel, """text-classification""": BioGptForSequenceClassification, """text-generation""": BioGptForCausalLM, """token-classification""": BioGptForTokenClassification, """zero-shot""": BioGptForSequenceClassification, } if is_torch_available() else {} ) a__ : Any = False def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Dict = BioGptModelTester(self) __UpperCamelCase :List[Any] = ConfigTester(self , config_class=__lowercase , hidden_size=37) def UpperCamelCase__ ( self) -> List[Any]: self.config_tester.run_common_tests() def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :List[Any] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __UpperCamelCase :Optional[Any] = type self.model_tester.create_and_check_model(*__lowercase) def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*__lowercase) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*__lowercase , gradient_checkpointing=__lowercase) def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*__lowercase) def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*__lowercase) def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*__lowercase) @slow def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Union[str, Any] = BioGptForCausalLM.from_pretrained('''microsoft/biogpt''') model.to(__lowercase) __UpperCamelCase :Dict = BioGptTokenizer.from_pretrained('''microsoft/biogpt''') __UpperCamelCase :Optional[int] = '''left''' # Define PAD Token = EOS Token = 50256 __UpperCamelCase :List[str] = tokenizer.eos_token __UpperCamelCase :Dict = model.config.eos_token_id # use different length sentences to test batching __UpperCamelCase :str = [ '''Hello, my dog is a little''', '''Today, I''', ] __UpperCamelCase :Optional[Any] = tokenizer(__lowercase , return_tensors='''pt''' , padding=__lowercase) __UpperCamelCase :str = inputs['''input_ids'''].to(__lowercase) __UpperCamelCase :Any = model.generate( input_ids=__lowercase , attention_mask=inputs['''attention_mask'''].to(__lowercase) , ) __UpperCamelCase :str = tokenizer(sentences[0] , return_tensors='''pt''').input_ids.to(__lowercase) __UpperCamelCase :Union[str, Any] = model.generate(input_ids=__lowercase) __UpperCamelCase :Dict = inputs_non_padded.shape[-1] - inputs['''attention_mask'''][-1].long().sum().cpu().item() __UpperCamelCase :List[Any] = tokenizer(sentences[1] , return_tensors='''pt''').input_ids.to(__lowercase) __UpperCamelCase :Union[str, Any] = model.generate(input_ids=__lowercase , max_length=model.config.max_length - num_paddings) __UpperCamelCase :str = tokenizer.batch_decode(__lowercase , skip_special_tokens=__lowercase) __UpperCamelCase :str = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__lowercase) __UpperCamelCase :int = tokenizer.decode(output_padded[0] , skip_special_tokens=__lowercase) __UpperCamelCase :Optional[Any] = [ '''Hello, my dog is a little bit bigger than a little bit.''', '''Today, I have a good idea of how to use the information''', ] self.assertListEqual(__lowercase , __lowercase) self.assertListEqual(__lowercase , [non_padded_sentence, padded_sentence]) @slow def UpperCamelCase__ ( self) -> int: for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase :Optional[int] = BioGptModel.from_pretrained(__lowercase) self.assertIsNotNone(__lowercase) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase , __UpperCamelCase :str = self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase :List[str] = 3 __UpperCamelCase :Optional[int] = input_dict['''input_ids'''] __UpperCamelCase :Dict = input_ids.ne(1).to(__lowercase) __UpperCamelCase :Dict = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size) __UpperCamelCase :Optional[Any] = BioGptForSequenceClassification(__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :Dict = model(__lowercase , attention_mask=__lowercase , labels=__lowercase) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase , __UpperCamelCase :Any = self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase :Tuple = 3 __UpperCamelCase :Optional[Any] = '''multi_label_classification''' __UpperCamelCase :Union[str, Any] = input_dict['''input_ids'''] __UpperCamelCase :Optional[int] = input_ids.ne(1).to(__lowercase) __UpperCamelCase :Optional[Any] = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size).to(torch.float) __UpperCamelCase :Tuple = BioGptForSequenceClassification(__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :Union[str, Any] = model(__lowercase , attention_mask=__lowercase , labels=__lowercase) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) @require_torch class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Dict = BioGptForCausalLM.from_pretrained('''microsoft/biogpt''') __UpperCamelCase :str = torch.tensor([[2, 4_805, 9, 656, 21]]) __UpperCamelCase :int = model(__lowercase)[0] __UpperCamelCase :List[str] = 42_384 __UpperCamelCase :str = torch.Size((1, 5, vocab_size)) self.assertEqual(output.shape , __lowercase) __UpperCamelCase :List[Any] = torch.tensor( [[[-9.52_36, -9.89_18, 10.45_57], [-11.04_69, -9.64_23, 8.10_22], [-8.86_64, -7.88_26, 5.53_25]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowercase , atol=1E-4)) @slow def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Tuple = BioGptTokenizer.from_pretrained('''microsoft/biogpt''') __UpperCamelCase :Optional[Any] = BioGptForCausalLM.from_pretrained('''microsoft/biogpt''') model.to(__lowercase) torch.manual_seed(0) __UpperCamelCase :List[str] = tokenizer('''COVID-19 is''' , return_tensors='''pt''').to(__lowercase) __UpperCamelCase :Optional[int] = model.generate( **__lowercase , min_length=100 , max_length=1_024 , num_beams=5 , early_stopping=__lowercase , ) __UpperCamelCase :List[str] = tokenizer.decode(output_ids[0] , skip_special_tokens=__lowercase) __UpperCamelCase :str = ( '''COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the''' ''' causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and''' ''' territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),''' ''' and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and''' ''' more than 800,000 deaths.''' ) self.assertEqual(__lowercase , __lowercase)
43
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = { '''facebook/wav2vec2-base-960h''': '''https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/config.json''', # See all Wav2Vec2 models at https://huggingface.co/models?filter=wav2vec2 } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Union[str, Any] = """wav2vec2""" def __init__( self , __lowercase=32 , __lowercase=768 , __lowercase=12 , __lowercase=12 , __lowercase=3_072 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.02 , __lowercase=1E-5 , __lowercase="group" , __lowercase="gelu" , __lowercase=(512, 512, 512, 512, 512, 512, 512) , __lowercase=(5, 2, 2, 2, 2, 2, 2) , __lowercase=(10, 3, 3, 3, 3, 2, 2) , __lowercase=False , __lowercase=128 , __lowercase=16 , __lowercase=False , __lowercase=True , __lowercase=0.05 , __lowercase=10 , __lowercase=2 , __lowercase=0.0 , __lowercase=10 , __lowercase=0 , __lowercase=320 , __lowercase=2 , __lowercase=0.1 , __lowercase=100 , __lowercase=256 , __lowercase=256 , __lowercase=0.1 , __lowercase="sum" , __lowercase=False , __lowercase=False , __lowercase=256 , __lowercase=(512, 512, 512, 512, 1_500) , __lowercase=(5, 3, 3, 1, 1) , __lowercase=(1, 2, 3, 1, 1) , __lowercase=512 , __lowercase=0 , __lowercase=1 , __lowercase=2 , __lowercase=False , __lowercase=3 , __lowercase=2 , __lowercase=3 , __lowercase=None , __lowercase=None , **__lowercase , ) -> int: super().__init__(**__lowercase , pad_token_id=__lowercase , bos_token_id=__lowercase , eos_token_id=__lowercase) __UpperCamelCase :Any = hidden_size __UpperCamelCase :int = feat_extract_norm __UpperCamelCase :Tuple = feat_extract_activation __UpperCamelCase :Union[str, Any] = list(__lowercase) __UpperCamelCase :List[Any] = list(__lowercase) __UpperCamelCase :int = list(__lowercase) __UpperCamelCase :List[Any] = conv_bias __UpperCamelCase :Optional[int] = num_conv_pos_embeddings __UpperCamelCase :Dict = num_conv_pos_embedding_groups __UpperCamelCase :Any = len(self.conv_dim) __UpperCamelCase :List[str] = num_hidden_layers __UpperCamelCase :int = intermediate_size __UpperCamelCase :str = hidden_act __UpperCamelCase :Any = num_attention_heads __UpperCamelCase :int = hidden_dropout __UpperCamelCase :Tuple = attention_dropout __UpperCamelCase :List[str] = activation_dropout __UpperCamelCase :Optional[Any] = feat_proj_dropout __UpperCamelCase :Any = final_dropout __UpperCamelCase :Any = layerdrop __UpperCamelCase :str = layer_norm_eps __UpperCamelCase :Optional[Any] = initializer_range __UpperCamelCase :List[str] = vocab_size __UpperCamelCase :str = do_stable_layer_norm __UpperCamelCase :Union[str, Any] = use_weighted_layer_sum if ( (len(self.conv_stride) != self.num_feat_extract_layers) or (len(self.conv_kernel) != self.num_feat_extract_layers) or (len(self.conv_dim) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' f""" {len(self.conv_dim)}`, `len(config.conv_stride) = {len(self.conv_stride)}`,""" f""" `len(config.conv_kernel) = {len(self.conv_kernel)}`.""") # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __UpperCamelCase :List[Any] = apply_spec_augment __UpperCamelCase :Tuple = mask_time_prob __UpperCamelCase :int = mask_time_length __UpperCamelCase :Dict = mask_time_min_masks __UpperCamelCase :str = mask_feature_prob __UpperCamelCase :List[str] = mask_feature_length __UpperCamelCase :Union[str, Any] = mask_feature_min_masks # parameters for pretraining with codevector quantized representations __UpperCamelCase :Optional[Any] = num_codevectors_per_group __UpperCamelCase :List[Any] = num_codevector_groups __UpperCamelCase :Tuple = contrastive_logits_temperature __UpperCamelCase :Optional[int] = feat_quantizer_dropout __UpperCamelCase :Optional[int] = num_negatives __UpperCamelCase :List[Any] = codevector_dim __UpperCamelCase :str = proj_codevector_dim __UpperCamelCase :List[str] = diversity_loss_weight # ctc loss __UpperCamelCase :Tuple = ctc_loss_reduction __UpperCamelCase :Tuple = ctc_zero_infinity # adapter __UpperCamelCase :List[str] = add_adapter __UpperCamelCase :Tuple = adapter_kernel_size __UpperCamelCase :str = adapter_stride __UpperCamelCase :Tuple = num_adapter_layers __UpperCamelCase :Tuple = output_hidden_size or hidden_size __UpperCamelCase :Optional[Any] = adapter_attn_dim # SequenceClassification-specific parameter. Feel free to ignore for other classes. __UpperCamelCase :Optional[Any] = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. __UpperCamelCase :Optional[int] = list(__lowercase) __UpperCamelCase :List[Any] = list(__lowercase) __UpperCamelCase :List[Any] = list(__lowercase) __UpperCamelCase :str = xvector_output_dim @property def UpperCamelCase__ ( self) -> List[str]: return functools.reduce(operator.mul , self.conv_stride , 1)
43
1
import sys import turtle def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' return (pa[0] + pa[0]) / 2, (pa[1] + pa[1]) / 2 def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ): '''simple docstring''' my_pen.up() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.down() my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) my_pen.goto(vertexa[0] , vertexa[1] ) if depth == 0: return triangle(SCREAMING_SNAKE_CASE , get_mid(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , get_mid(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , depth - 1 ) triangle(SCREAMING_SNAKE_CASE , get_mid(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , get_mid(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , depth - 1 ) triangle(SCREAMING_SNAKE_CASE , get_mid(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , get_mid(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , depth - 1 ) if __name__ == "__main__": if len(sys.argv) != 2: raise ValueError( '''Correct format for using this script: ''' '''python fractals.py <int:depth_for_fractal>''' ) __lowercase = turtle.Turtle() my_pen.ht() my_pen.speed(5) my_pen.pencolor('''red''') __lowercase = [(-175, -125), (0, 175), (175, -125)] # vertices of triangle triangle(vertices[0], vertices[1], vertices[2], int(sys.argv[1]))
43
from typing import List, Optional, Union import numpy as np import PIL.Image from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, PILImageResampling, get_image_size, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __lowercase = logging.get_logger(__name__) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[Any] = ["""pixel_values"""] def __init__( self , __lowercase = True , __lowercase = 32 , __lowercase=PILImageResampling.BILINEAR , __lowercase = True , **__lowercase , ) -> None: __UpperCamelCase :Optional[int] = do_resize __UpperCamelCase :Any = do_rescale __UpperCamelCase :str = size_divisor __UpperCamelCase :Dict = resample super().__init__(**__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase = None , **__lowercase) -> np.ndarray: __UpperCamelCase , __UpperCamelCase :int = get_image_size(__lowercase) # Rounds the height and width down to the closest multiple of size_divisor __UpperCamelCase :List[Any] = height // size_divisor * size_divisor __UpperCamelCase :List[str] = width // size_divisor * size_divisor __UpperCamelCase :str = resize(__lowercase , (new_h, new_w) , resample=__lowercase , data_format=__lowercase , **__lowercase) return image def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase = None , **__lowercase) -> np.ndarray: return rescale(image=__lowercase , scale=__lowercase , data_format=__lowercase , **__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase = None , __lowercase = None , __lowercase=None , __lowercase = None , __lowercase = None , __lowercase = ChannelDimension.FIRST , **__lowercase , ) -> BatchFeature: __UpperCamelCase :Union[str, Any] = do_resize if do_resize is not None else self.do_resize __UpperCamelCase :Tuple = do_rescale if do_rescale is not None else self.do_rescale __UpperCamelCase :List[str] = size_divisor if size_divisor is not None else self.size_divisor __UpperCamelCase :List[Any] = resample if resample is not None else self.resample if do_resize and size_divisor is None: raise ValueError('''size_divisor is required for resizing''') __UpperCamelCase :List[Any] = make_list_of_images(__lowercase) if not valid_images(__lowercase): raise ValueError('''Invalid image(s)''') # All transformations expect numpy arrays. __UpperCamelCase :Optional[Any] = [to_numpy_array(__lowercase) for img in images] if do_resize: __UpperCamelCase :List[str] = [self.resize(__lowercase , size_divisor=__lowercase , resample=__lowercase) for image in images] if do_rescale: __UpperCamelCase :Dict = [self.rescale(__lowercase , scale=1 / 255) for image in images] __UpperCamelCase :str = [to_channel_dimension_format(__lowercase , __lowercase) for image in images] __UpperCamelCase :int = {'''pixel_values''': images} return BatchFeature(data=__lowercase , tensor_type=__lowercase)
43
1
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :List[str] = ArgumentParser( description=( '''PyTorch TPU distributed training launch helper utility that will spawn up multiple distributed processes''' ) ) # Optional arguments for the launch helper parser.add_argument('''--num_cores''' , type=SCREAMING_SNAKE_CASE , default=1 , help='''Number of TPU cores to use (1 or 8).''' ) # positional parser.add_argument( '''training_script''' , type=SCREAMING_SNAKE_CASE , help=( '''The full path to the single TPU training ''' '''program/script to be launched in parallel, ''' '''followed by all the arguments for the ''' '''training script''' ) , ) # rest from the training program parser.add_argument('''training_script_args''' , nargs=SCREAMING_SNAKE_CASE ) return parser.parse_args() def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :str = parse_args() # Import training_script as a module. __UpperCamelCase :Tuple = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) __UpperCamelCase :Dict = script_fpath.stem __UpperCamelCase :Optional[Any] = importlib.import_module(SCREAMING_SNAKE_CASE ) # Patch sys.argv __UpperCamelCase :Optional[int] = [args.training_script] + args.training_script_args + ['''--tpu_num_cores''', str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
43
from __future__ import annotations from PIL import Image # Define glider example __lowercase = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example __lowercase = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :int = [] for i in range(len(SCREAMING_SNAKE_CASE ) ): __UpperCamelCase :Dict = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours __UpperCamelCase :List[str] = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(SCREAMING_SNAKE_CASE ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(SCREAMING_SNAKE_CASE ) - 1: neighbour_count += cells[i + 1][j] if i < len(SCREAMING_SNAKE_CASE ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. __UpperCamelCase :List[str] = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(SCREAMING_SNAKE_CASE ) return next_generation def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = [] for _ in range(SCREAMING_SNAKE_CASE ): # Create output image __UpperCamelCase :Dict = Image.new('''RGB''' , (len(cells[0] ), len(SCREAMING_SNAKE_CASE )) ) __UpperCamelCase :Any = img.load() # Save cells to image for x in range(len(SCREAMING_SNAKE_CASE ) ): for y in range(len(cells[0] ) ): __UpperCamelCase :Optional[Any] = 255 - cells[y][x] * 255 __UpperCamelCase :int = (colour, colour, colour) # Save image images.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = new_generation(SCREAMING_SNAKE_CASE ) return images if __name__ == "__main__": __lowercase = generate_images(GLIDER, 16) images[0].save('''out.gif''', save_all=True, append_images=images[1:])
43
1
import re def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' return [char.split() for char in re.split(R'''[^ a-z A-Z 0-9 \s]''' , str_ )] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = split_input(str_ ) return "".join( [''''''.join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' try: __UpperCamelCase :Union[str, Any] = split_input(SCREAMING_SNAKE_CASE ) if upper: __UpperCamelCase :Union[str, Any] = ''''''.join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: __UpperCamelCase :Any = ''''''.join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' return to_simple_case(SCREAMING_SNAKE_CASE ) def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' try: __UpperCamelCase :Tuple = to_simple_case(SCREAMING_SNAKE_CASE ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' return to_complex_case(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , '''_''' ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' return to_complex_case(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , '''-''' ) if __name__ == "__main__": __import__('''doctest''').testmod()
43
import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging __lowercase = logging.get_logger(__name__) def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = R'''\w+[.]\d+''' __UpperCamelCase :List[str] = re.findall(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for pat in pats: __UpperCamelCase :int = key.replace(SCREAMING_SNAKE_CASE , '''_'''.join(pat.split('''.''' ) ) ) return key def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = pt_tuple_key[:-1] + ('''scale''',) if ( any('''norm''' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): __UpperCamelCase :str = pt_tuple_key[:-1] + ('''scale''',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: __UpperCamelCase :Any = pt_tuple_key[:-1] + ('''scale''',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: __UpperCamelCase :str = pt_tuple_key[:-1] + ('''embedding''',) return renamed_pt_tuple_key, pt_tensor # conv layer __UpperCamelCase :List[str] = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: __UpperCamelCase :List[Any] = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer __UpperCamelCase :List[str] = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight": __UpperCamelCase :Any = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight __UpperCamelCase :int = pt_tuple_key[:-1] + ('''weight''',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias __UpperCamelCase :int = pt_tuple_key[:-1] + ('''bias''',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=42 ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params __UpperCamelCase :str = flax_model.init_weights(PRNGKey(SCREAMING_SNAKE_CASE ) ) __UpperCamelCase :int = flatten_dict(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): __UpperCamelCase :List[Any] = rename_key(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = tuple(renamed_pt_key.split('''.''' ) ) # Correctly rename weight parameters __UpperCamelCase , __UpperCamelCase :Any = rename_key_and_reshape_tensor(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ f"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown __UpperCamelCase :str = jnp.asarray(SCREAMING_SNAKE_CASE ) return unflatten_dict(SCREAMING_SNAKE_CASE )
43
1
import copy import tempfile import unittest from transformers import MaMaaaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.utils import cached_property from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaTokenizer from transformers.models.mam_aaa.modeling_mam_aaa import MaMaaaDecoder, MaMaaaEncoder def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , ): '''simple docstring''' if attention_mask is None: __UpperCamelCase :List[str] = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: __UpperCamelCase :Tuple = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: __UpperCamelCase :Dict = torch.ones(config.encoder_layers , config.encoder_attention_heads , device=SCREAMING_SNAKE_CASE ) if decoder_head_mask is None: __UpperCamelCase :int = torch.ones(config.decoder_layers , config.decoder_attention_heads , device=SCREAMING_SNAKE_CASE ) if cross_attn_head_mask is None: __UpperCamelCase :Union[str, Any] = torch.ones(config.decoder_layers , config.decoder_attention_heads , device=SCREAMING_SNAKE_CASE ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } class lowerCamelCase_ : '''simple docstring''' def __init__( self , __lowercase , __lowercase=13 , __lowercase=7 , __lowercase=True , __lowercase=False , __lowercase=99 , __lowercase=16 , __lowercase=2 , __lowercase=4 , __lowercase=4 , __lowercase="relu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=20 , __lowercase=2 , __lowercase=1 , __lowercase=0 , ) -> Tuple: __UpperCamelCase :int = parent __UpperCamelCase :Optional[int] = batch_size __UpperCamelCase :int = seq_length __UpperCamelCase :Tuple = is_training __UpperCamelCase :Optional[int] = use_labels __UpperCamelCase :Any = vocab_size __UpperCamelCase :Tuple = hidden_size __UpperCamelCase :int = num_hidden_layers __UpperCamelCase :str = num_attention_heads __UpperCamelCase :Any = intermediate_size __UpperCamelCase :Any = hidden_act __UpperCamelCase :Union[str, Any] = hidden_dropout_prob __UpperCamelCase :Any = attention_probs_dropout_prob __UpperCamelCase :List[str] = encoder_layerdrop __UpperCamelCase :List[Any] = decoder_layerdrop __UpperCamelCase :Tuple = max_position_embeddings __UpperCamelCase :Tuple = eos_token_id __UpperCamelCase :Optional[Any] = pad_token_id __UpperCamelCase :Any = bos_token_id def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __UpperCamelCase :Any = self.eos_token_id # Eos Token __UpperCamelCase :Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) # we need to clamp the input ids here to avoid having pad token in between # this is because for M2M100 the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input __UpperCamelCase :Union[str, Any] = input_ids.clamp(self.pad_token_id + 1) __UpperCamelCase :List[str] = decoder_input_ids.clamp(self.pad_token_id + 1) __UpperCamelCase :Any = self.get_config() __UpperCamelCase :List[Any] = prepare_mam_aaa_inputs_dict(__lowercase , __lowercase , __lowercase) return config, inputs_dict def UpperCamelCase__ ( self) -> List[Any]: return MaMaaaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , encoder_layerdrop=self.encoder_layerdrop , decoder_layerdrop=self.decoder_layerdrop , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , ) def UpperCamelCase__ ( self) -> Any: __UpperCamelCase , __UpperCamelCase :List[str] = self.prepare_config_and_inputs() return config, inputs_dict def UpperCamelCase__ ( self , __lowercase , __lowercase) -> Optional[Any]: __UpperCamelCase :Tuple = MaMaaaModel(config=__lowercase).get_decoder().to(__lowercase).eval() __UpperCamelCase :str = inputs_dict['''input_ids'''] __UpperCamelCase :Union[str, Any] = inputs_dict['''attention_mask'''] __UpperCamelCase :List[str] = inputs_dict['''head_mask'''] # first forward pass __UpperCamelCase :Dict = model(__lowercase , attention_mask=__lowercase , head_mask=__lowercase , use_cache=__lowercase) __UpperCamelCase , __UpperCamelCase :Optional[Any] = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids __UpperCamelCase :List[str] = ids_tensor((self.batch_size, 3) , config.vocab_size) __UpperCamelCase :Union[str, Any] = ids_tensor((self.batch_size, 3) , 2) # append to next input_ids and __UpperCamelCase :List[str] = torch.cat([input_ids, next_tokens] , dim=-1) __UpperCamelCase :Optional[Any] = torch.cat([attention_mask, next_attn_mask] , dim=-1) __UpperCamelCase :Union[str, Any] = model(__lowercase , attention_mask=__lowercase)['''last_hidden_state'''] __UpperCamelCase :str = model(__lowercase , attention_mask=__lowercase , past_key_values=__lowercase)[ '''last_hidden_state''' ] # select random slice __UpperCamelCase :List[str] = ids_tensor((1,) , output_from_past.shape[-1]).item() __UpperCamelCase :Optional[int] = output_from_no_past[:, -3:, random_slice_idx].detach() __UpperCamelCase :Optional[Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1]) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__lowercase , __lowercase , atol=1E-2)) def UpperCamelCase__ ( self , __lowercase , __lowercase) -> int: __UpperCamelCase :Dict = MaMaaaModel(config=__lowercase).to(__lowercase).eval() __UpperCamelCase :Tuple = model(**__lowercase) __UpperCamelCase :List[Any] = outputs.encoder_last_hidden_state __UpperCamelCase :List[str] = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: __UpperCamelCase :List[str] = model.get_encoder() encoder.save_pretrained(__lowercase) __UpperCamelCase :Union[str, Any] = MaMaaaEncoder.from_pretrained(__lowercase).to(__lowercase) __UpperCamelCase :Tuple = encoder(inputs_dict['''input_ids'''] , attention_mask=inputs_dict['''attention_mask'''])[ 0 ] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1E-3) with tempfile.TemporaryDirectory() as tmpdirname: __UpperCamelCase :Any = model.get_decoder() decoder.save_pretrained(__lowercase) __UpperCamelCase :str = MaMaaaDecoder.from_pretrained(__lowercase).to(__lowercase) __UpperCamelCase :Any = decoder( input_ids=inputs_dict['''decoder_input_ids'''] , attention_mask=inputs_dict['''decoder_attention_mask'''] , encoder_hidden_states=__lowercase , encoder_attention_mask=inputs_dict['''attention_mask'''] , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1E-3) @require_torch class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : Optional[Any] = ( ( MaMaaaModel, MaMaaaForConditionalGeneration, ) if is_torch_available() else () ) a__ : str = (MaMaaaForConditionalGeneration,) if is_torch_available() else () a__ : str = ( { """conversational""": MaMaaaForConditionalGeneration, """feature-extraction""": MaMaaaModel, """summarization""": MaMaaaForConditionalGeneration, """text2text-generation""": MaMaaaForConditionalGeneration, """translation""": MaMaaaForConditionalGeneration, } if is_torch_available() else {} ) a__ : int = True a__ : int = True a__ : Union[str, Any] = False a__ : int = False def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase) -> Union[str, Any]: if pipeline_test_casse_name == "TranslationPipelineTests": # Get `ValueError: Translation requires a `src_lang` and a `tgt_lang` for this model`. # `M2M100Config` was never used in pipeline tests: cannot create a simple tokenizer. return True return False def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Optional[int] = MaMaaaModelTester(self) __UpperCamelCase :Optional[int] = ConfigTester(self , config_class=__lowercase) def UpperCamelCase__ ( self) -> str: self.config_tester.run_common_tests() def UpperCamelCase__ ( self) -> int: __UpperCamelCase , __UpperCamelCase :Optional[Any] = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: __UpperCamelCase :str = model_class(__lowercase) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__lowercase) __UpperCamelCase , __UpperCamelCase :Dict = model_class.from_pretrained(__lowercase , output_loading_info=__lowercase) self.assertEqual(info['''missing_keys'''] , []) def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(*__lowercase) def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :Any = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*__lowercase) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase , __UpperCamelCase :Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in (MaMaaaModel, MaMaaaForConditionalGeneration): __UpperCamelCase :Optional[Any] = model_class(__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :Union[str, Any] = copy.deepcopy(self._prepare_for_class(__lowercase , __lowercase)) if not self.is_encoder_decoder: __UpperCamelCase :Dict = inputs['''input_ids'''] del inputs["input_ids"] else: __UpperCamelCase :str = inputs['''input_ids'''] __UpperCamelCase :Dict = inputs.get('''decoder_input_ids''' , __lowercase) del inputs["input_ids"] inputs.pop('''decoder_input_ids''' , __lowercase) __UpperCamelCase :Union[str, Any] = model.get_input_embeddings() if not self.is_encoder_decoder: __UpperCamelCase :List[Any] = wte(__lowercase) else: __UpperCamelCase :int = wte(__lowercase) __UpperCamelCase :List[str] = wte(__lowercase) with torch.no_grad(): model(**__lowercase)[0] def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase , __UpperCamelCase :int = self.model_tester.prepare_config_and_inputs() __UpperCamelCase :Dict = input_dict['''input_ids'''] __UpperCamelCase :Any = input_ids.ne(1).to(__lowercase) __UpperCamelCase :Optional[int] = MaMaaaForConditionalGeneration(__lowercase).eval().to(__lowercase) if torch_device == "cuda": model.half() model.generate(__lowercase , attention_mask=__lowercase) model.generate(num_beams=4 , do_sample=__lowercase , early_stopping=__lowercase , num_return_sequences=3) def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' return torch.tensor(SCREAMING_SNAKE_CASE , dtype=torch.long , device=SCREAMING_SNAKE_CASE ) __lowercase = 1e-4 @require_torch @require_sentencepiece @require_tokenizers @slow class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase__ ( self) -> int: return MaMaaaTokenizer.from_pretrained('''facebook/m2m100_418M''') def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :Union[str, Any] = MaMaaaModel.from_pretrained('''facebook/m2m100_418M''').to(__lowercase) __UpperCamelCase :Dict = _long_tensor([[128_028, 98, 12, 30_527, 2_732, 159, 7_755, 61_904, 39_144, 38, 2]]) __UpperCamelCase :int = _long_tensor([[2, 128_028, 98, 12, 30_527, 2_732, 159, 7_755, 61_904, 39_144, 38]]) __UpperCamelCase :int = prepare_mam_aaa_inputs_dict(model.config , __lowercase , __lowercase) with torch.no_grad(): __UpperCamelCase :List[Any] = model(**__lowercase)[0] __UpperCamelCase :Optional[Any] = torch.Size((1, 11, 1_024)) self.assertEqual(output.shape , __lowercase) # change to expected output here __UpperCamelCase :Tuple = torch.tensor( [[-0.77_80, -0.16_76, 0.10_38], [-6.75_56, -1.39_92, 0.05_67], [-7.53_83, -0.59_20, -0.27_79]] , device=__lowercase) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowercase , atol=__lowercase)) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Optional[int] = MaMaaaForConditionalGeneration.from_pretrained('''facebook/m2m100_418M''').to(__lowercase) # change to intended input __UpperCamelCase :List[Any] = _long_tensor([[128_028, 98, 12, 30_527, 2_732, 159, 7_755, 61_904, 39_144, 38, 2]]) __UpperCamelCase :str = _long_tensor([[2, 128_028, 98, 12, 30_527, 2_732, 159, 7_755, 61_904, 39_144, 38]]) __UpperCamelCase :List[Any] = prepare_mam_aaa_inputs_dict(model.config , __lowercase , __lowercase) with torch.no_grad(): __UpperCamelCase :Optional[int] = model(**__lowercase)[0] __UpperCamelCase :List[Any] = torch.Size((1, 11, model.config.vocab_size)) self.assertEqual(output.shape , __lowercase) # change to expected output here __UpperCamelCase :int = torch.tensor( [[-1.04_48, -1.04_11, 3.79_92], [-3.21_91, -3.23_86, -1.34_51], [-3.62_10, -3.59_93, 0.49_25]] , device=__lowercase) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowercase , atol=__lowercase)) def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :Tuple = MaMaaaForConditionalGeneration.from_pretrained('''facebook/m2m100_418M''').to(__lowercase) __UpperCamelCase :Optional[int] = MaMaaaTokenizer.from_pretrained('''facebook/m2m100_418M''' , src_lang='''fr''' , tgt_lang='''en''') __UpperCamelCase :Optional[int] = [ '''L\'affaire NSA souligne l\'absence totale de débat sur le renseignement''', '''Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.''', '''Lorsque François Hollande téléphone à Barack Obama ou quand le ministre des affaires étrangères Laurent''' ''' Fabius convoque l\'ambassadeur des Etats-Unis, ils réagissent à une vraie découverte, qui est celle de''' ''' l\'ampleur de la surveillance américaine sur l\'ensemble des communications en France.''', ] # The below article tests that we don't add any hypotheses outside of the top n_beams __UpperCamelCase :List[str] = tokenizer(__lowercase , padding=__lowercase , return_tensors='''pt''') __UpperCamelCase :Dict = model.generate( input_ids=dct['''input_ids'''].to(__lowercase) , attention_mask=dct['''attention_mask'''].to(__lowercase) , num_beams=5 , forced_bos_token_id=tokenizer.get_lang_id('''en''') , ) __UpperCamelCase :Any = [ '''The NSA case highlights the total absence of intelligence debate''', '''I think there are two levels of response from the French government.''', '''When François Hollande calls Barack Obama or when Foreign Minister Laurent Fabius calls the U.S.''' ''' Ambassador, they respond to a real discovery, which is that of the scale of U.S. surveillance on all''' ''' communications in France.''', ] __UpperCamelCase :str = tokenizer.batch_decode( hypotheses_batch.tolist() , clean_up_tokenization_spaces=__lowercase , skip_special_tokens=__lowercase) assert generated == expected_en
43
import argparse import torch from ...utils import logging from . import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert logging.set_verbosity_info() def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = AlbertConfig.from_json_file(SCREAMING_SNAKE_CASE ) print(f"""Building PyTorch model from configuration: {config}""" ) __UpperCamelCase :List[str] = AlbertForPreTraining(SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint load_tf_weights_in_albert(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--albert_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained ALBERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __lowercase = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.albert_config_file, args.pytorch_dump_path)
43
1
def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): raise ValueError('''multiplicative_persistence() only accepts integral values''' ) if num < 0: raise ValueError('''multiplicative_persistence() does not accept negative values''' ) __UpperCamelCase :int = 0 __UpperCamelCase :List[Any] = str(SCREAMING_SNAKE_CASE ) while len(SCREAMING_SNAKE_CASE ) != 1: __UpperCamelCase :List[Any] = [int(SCREAMING_SNAKE_CASE ) for i in num_string] __UpperCamelCase :Tuple = 1 for i in range(0 , len(SCREAMING_SNAKE_CASE ) ): total *= numbers[i] __UpperCamelCase :Any = str(SCREAMING_SNAKE_CASE ) steps += 1 return steps def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): raise ValueError('''additive_persistence() only accepts integral values''' ) if num < 0: raise ValueError('''additive_persistence() does not accept negative values''' ) __UpperCamelCase :Optional[int] = 0 __UpperCamelCase :Optional[int] = str(SCREAMING_SNAKE_CASE ) while len(SCREAMING_SNAKE_CASE ) != 1: __UpperCamelCase :Optional[int] = [int(SCREAMING_SNAKE_CASE ) for i in num_string] __UpperCamelCase :Tuple = 0 for i in range(0 , len(SCREAMING_SNAKE_CASE ) ): total += numbers[i] __UpperCamelCase :Any = str(SCREAMING_SNAKE_CASE ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
43
import math import qiskit def lowerCamelCase ( SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1 ): '''simple docstring''' if ( isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) or isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) or isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ): raise TypeError('''inputs must be integers.''' ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError('''inputs must be positive.''' ) if ( (math.floor(SCREAMING_SNAKE_CASE ) != input_a) or (math.floor(SCREAMING_SNAKE_CASE ) != input_a) or (math.floor(SCREAMING_SNAKE_CASE ) != carry_in) ): raise ValueError('''inputs must be exact integers.''' ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError('''inputs must be less or equal to 2.''' ) # build registers __UpperCamelCase :List[str] = qiskit.QuantumRegister(4 , '''qr''' ) __UpperCamelCase :str = qiskit.ClassicalRegister(2 , '''cr''' ) # list the entries __UpperCamelCase :Tuple = [input_a, input_a, carry_in] __UpperCamelCase :Optional[int] = qiskit.QuantumCircuit(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(SCREAMING_SNAKE_CASE ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(SCREAMING_SNAKE_CASE ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(SCREAMING_SNAKE_CASE ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , SCREAMING_SNAKE_CASE ) # measure the last two qbits __UpperCamelCase :Optional[Any] = qiskit.Aer.get_backend('''aer_simulator''' ) __UpperCamelCase :Tuple = qiskit.execute(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , shots=1_000 ) return job.result().get_counts(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(F'Total sum count for state is: {quantum_full_adder(1, 1, 1)}')
43
1
from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutputWithNoAttention, TFBaseModelOutputWithPoolingAndNoAttention, TFSequenceClassifierOutput, ) from ...modeling_tf_utils import TFPreTrainedModel, TFSequenceClassificationLoss, keras_serializable, unpack_inputs from ...tf_utils import shape_list from ...utils import logging from .configuration_regnet import RegNetConfig __lowercase = logging.get_logger(__name__) # General docstring __lowercase = '''RegNetConfig''' # Base docstring __lowercase = '''facebook/regnet-y-040''' __lowercase = [1, 1088, 7, 7] # Image classification docstring __lowercase = '''facebook/regnet-y-040''' __lowercase = '''tabby, tabby cat''' __lowercase = [ '''facebook/regnet-y-040''', # See all regnet models at https://huggingface.co/models?filter=regnet ] class lowerCamelCase_ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self , __lowercase , __lowercase = 3 , __lowercase = 1 , __lowercase = 1 , __lowercase = "relu" , **__lowercase , ) -> Dict: super().__init__(**__lowercase) # The padding and conv has been verified in # https://colab.research.google.com/gist/sayakpaul/854bc10eeaf21c9ee2119e0b9f3841a7/scratchpad.ipynb __UpperCamelCase :Optional[int] = tf.keras.layers.ZeroPaddingaD(padding=kernel_size // 2) __UpperCamelCase :Any = tf.keras.layers.ConvaD( filters=__lowercase , kernel_size=__lowercase , strides=__lowercase , padding='''VALID''' , groups=__lowercase , use_bias=__lowercase , name='''convolution''' , ) __UpperCamelCase :Tuple = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='''normalization''') __UpperCamelCase :List[str] = ACTaFN[activation] if activation is not None else tf.identity def UpperCamelCase__ ( self , __lowercase) -> List[Any]: __UpperCamelCase :Optional[int] = self.convolution(self.padding(__lowercase)) __UpperCamelCase :Optional[int] = self.normalization(__lowercase) __UpperCamelCase :Tuple = self.activation(__lowercase) return hidden_state class lowerCamelCase_ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self , __lowercase , **__lowercase) -> int: super().__init__(**__lowercase) __UpperCamelCase :Tuple = config.num_channels __UpperCamelCase :Optional[int] = TFRegNetConvLayer( out_channels=config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act , name='''embedder''' , ) def UpperCamelCase__ ( self , __lowercase) -> Dict: __UpperCamelCase :Dict = shape_list(__lowercase)[1] if tf.executing_eagerly() and num_channels != self.num_channels: raise ValueError( '''Make sure that the channel dimension of the pixel values match with the one set in the configuration.''') # When running on CPU, `tf.keras.layers.Conv2D` doesn't support `NCHW` format. # So change the input format from `NCHW` to `NHWC`. # shape = (batch_size, in_height, in_width, in_channels=num_channels) __UpperCamelCase :Dict = tf.transpose(__lowercase , perm=(0, 2, 3, 1)) __UpperCamelCase :Optional[int] = self.embedder(__lowercase) return hidden_state class lowerCamelCase_ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self , __lowercase , __lowercase = 2 , **__lowercase) -> Optional[Any]: super().__init__(**__lowercase) __UpperCamelCase :Union[str, Any] = tf.keras.layers.ConvaD( filters=__lowercase , kernel_size=1 , strides=__lowercase , use_bias=__lowercase , name='''convolution''') __UpperCamelCase :int = tf.keras.layers.BatchNormalization(epsilon=1E-5 , momentum=0.9 , name='''normalization''') def UpperCamelCase__ ( self , __lowercase , __lowercase = False) -> tf.Tensor: return self.normalization(self.convolution(__lowercase) , training=__lowercase) class lowerCamelCase_ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self , __lowercase , __lowercase , **__lowercase) -> Union[str, Any]: super().__init__(**__lowercase) __UpperCamelCase :int = tf.keras.layers.GlobalAveragePoolingaD(keepdims=__lowercase , name='''pooler''') __UpperCamelCase :List[Any] = [ tf.keras.layers.ConvaD(filters=__lowercase , kernel_size=1 , activation='''relu''' , name='''attention.0'''), tf.keras.layers.ConvaD(filters=__lowercase , kernel_size=1 , activation='''sigmoid''' , name='''attention.2'''), ] def UpperCamelCase__ ( self , __lowercase) -> Any: # [batch_size, h, w, num_channels] -> [batch_size, 1, 1, num_channels] __UpperCamelCase :List[str] = self.pooler(__lowercase) for layer_module in self.attention: __UpperCamelCase :List[str] = layer_module(__lowercase) __UpperCamelCase :int = hidden_state * pooled return hidden_state class lowerCamelCase_ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self , __lowercase , __lowercase , __lowercase , __lowercase = 1 , **__lowercase) -> Tuple: super().__init__(**__lowercase) __UpperCamelCase :str = in_channels != out_channels or stride != 1 __UpperCamelCase :Any = max(1 , out_channels // config.groups_width) __UpperCamelCase :Optional[Any] = ( TFRegNetShortCut(__lowercase , stride=__lowercase , name='''shortcut''') if should_apply_shortcut else tf.keras.layers.Activation('''linear''' , name='''shortcut''') ) # `self.layers` instead of `self.layer` because that is a reserved argument. __UpperCamelCase :Optional[int] = [ TFRegNetConvLayer(__lowercase , kernel_size=1 , activation=config.hidden_act , name='''layer.0'''), TFRegNetConvLayer( __lowercase , stride=__lowercase , groups=__lowercase , activation=config.hidden_act , name='''layer.1'''), TFRegNetConvLayer(__lowercase , kernel_size=1 , activation=__lowercase , name='''layer.2'''), ] __UpperCamelCase :Dict = ACTaFN[config.hidden_act] def UpperCamelCase__ ( self , __lowercase) -> Any: __UpperCamelCase :Optional[int] = hidden_state for layer_module in self.layers: __UpperCamelCase :Any = layer_module(__lowercase) __UpperCamelCase :Tuple = self.shortcut(__lowercase) hidden_state += residual __UpperCamelCase :Optional[Any] = self.activation(__lowercase) return hidden_state class lowerCamelCase_ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self , __lowercase , __lowercase , __lowercase , __lowercase = 1 , **__lowercase) -> Union[str, Any]: super().__init__(**__lowercase) __UpperCamelCase :int = in_channels != out_channels or stride != 1 __UpperCamelCase :Optional[int] = max(1 , out_channels // config.groups_width) __UpperCamelCase :Dict = ( TFRegNetShortCut(__lowercase , stride=__lowercase , name='''shortcut''') if should_apply_shortcut else tf.keras.layers.Activation('''linear''' , name='''shortcut''') ) __UpperCamelCase :Dict = [ TFRegNetConvLayer(__lowercase , kernel_size=1 , activation=config.hidden_act , name='''layer.0'''), TFRegNetConvLayer( __lowercase , stride=__lowercase , groups=__lowercase , activation=config.hidden_act , name='''layer.1'''), TFRegNetSELayer(__lowercase , reduced_channels=int(round(in_channels / 4)) , name='''layer.2'''), TFRegNetConvLayer(__lowercase , kernel_size=1 , activation=__lowercase , name='''layer.3'''), ] __UpperCamelCase :Dict = ACTaFN[config.hidden_act] def UpperCamelCase__ ( self , __lowercase) -> Optional[int]: __UpperCamelCase :str = hidden_state for layer_module in self.layers: __UpperCamelCase :Optional[Any] = layer_module(__lowercase) __UpperCamelCase :List[Any] = self.shortcut(__lowercase) hidden_state += residual __UpperCamelCase :List[str] = self.activation(__lowercase) return hidden_state class lowerCamelCase_ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self , __lowercase , __lowercase , __lowercase , __lowercase = 2 , __lowercase = 2 , **__lowercase) -> int: super().__init__(**__lowercase) __UpperCamelCase :List[Any] = TFRegNetXLayer if config.layer_type == '''x''' else TFRegNetYLayer __UpperCamelCase :List[str] = [ # downsampling is done in the first layer with stride of 2 layer(__lowercase , __lowercase , __lowercase , stride=__lowercase , name='''layers.0'''), *[layer(__lowercase , __lowercase , __lowercase , name=f"""layers.{i+1}""") for i in range(depth - 1)], ] def UpperCamelCase__ ( self , __lowercase) -> Optional[Any]: for layer_module in self.layers: __UpperCamelCase :Tuple = layer_module(__lowercase) return hidden_state class lowerCamelCase_ ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self , __lowercase , **__lowercase) -> str: super().__init__(**__lowercase) __UpperCamelCase :Any = [] # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( TFRegNetStage( __lowercase , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , name='''stages.0''' , )) __UpperCamelCase :str = zip(config.hidden_sizes , config.hidden_sizes[1:]) for i, ((in_channels, out_channels), depth) in enumerate(zip(__lowercase , config.depths[1:])): self.stages.append(TFRegNetStage(__lowercase , __lowercase , __lowercase , depth=__lowercase , name=f"""stages.{i+1}""")) def UpperCamelCase__ ( self , __lowercase , __lowercase = False , __lowercase = True) -> TFBaseModelOutputWithNoAttention: __UpperCamelCase :Optional[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: __UpperCamelCase :Tuple = hidden_states + (hidden_state,) __UpperCamelCase :Optional[Any] = stage_module(__lowercase) if output_hidden_states: __UpperCamelCase :Tuple = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None) return TFBaseModelOutputWithNoAttention(last_hidden_state=__lowercase , hidden_states=__lowercase) @keras_serializable class lowerCamelCase_ ( tf.keras.layers.Layer ): '''simple docstring''' a__ : Optional[Any] = RegNetConfig def __init__( self , __lowercase , **__lowercase) -> Union[str, Any]: super().__init__(**__lowercase) __UpperCamelCase :List[str] = config __UpperCamelCase :List[Any] = TFRegNetEmbeddings(__lowercase , name='''embedder''') __UpperCamelCase :Any = TFRegNetEncoder(__lowercase , name='''encoder''') __UpperCamelCase :Optional[int] = tf.keras.layers.GlobalAveragePoolingaD(keepdims=__lowercase , name='''pooler''') @unpack_inputs def UpperCamelCase__ ( self , __lowercase , __lowercase = None , __lowercase = None , __lowercase = False , ) -> TFBaseModelOutputWithPoolingAndNoAttention: __UpperCamelCase :Optional[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __UpperCamelCase :Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict __UpperCamelCase :Union[str, Any] = self.embedder(__lowercase , training=__lowercase) __UpperCamelCase :Any = self.encoder( __lowercase , output_hidden_states=__lowercase , return_dict=__lowercase , training=__lowercase) __UpperCamelCase :List[str] = encoder_outputs[0] __UpperCamelCase :List[str] = self.pooler(__lowercase) # Change to NCHW output format have uniformity in the modules __UpperCamelCase :str = tf.transpose(__lowercase , perm=(0, 3, 1, 2)) __UpperCamelCase :List[Any] = tf.transpose(__lowercase , perm=(0, 3, 1, 2)) # Change the other hidden state outputs to NCHW as well if output_hidden_states: __UpperCamelCase :List[str] = tuple([tf.transpose(__lowercase , perm=(0, 3, 1, 2)) for h in encoder_outputs[1]]) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=__lowercase , pooler_output=__lowercase , hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states , ) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : int = RegNetConfig a__ : List[str] = """regnet""" a__ : Optional[int] = """pixel_values""" @property def UpperCamelCase__ ( self) -> str: return {"pixel_values": tf.TensorSpec(shape=(None, self.config.num_channels, 224, 224) , dtype=tf.floataa)} __lowercase = r''' Parameters: This model is a Tensorflow [tf.keras.layers.Layer](https://www.tensorflow.org/api_docs/python/tf/keras/layers/Layer) sub-class. Use it as a regular Tensorflow Module and refer to the Tensorflow documentation for all matter related to general usage and behavior. config ([`RegNetConfig`]): Model configuration class with all the parameters of the model. Initializing with a config file does not load the weights associated with the model, only the configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights. ''' __lowercase = r''' Args: pixel_values (`tf.Tensor` of shape `(batch_size, num_channels, height, width)`): Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ConveNextImageProcessor.__call__`] for details. output_hidden_states (`bool`, *optional*): Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for more detail. return_dict (`bool`, *optional*): Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. ''' @add_start_docstrings( """The bare RegNet model outputting raw features without any specific head on top.""" , UpperCAmelCase_ , ) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' def __init__( self , __lowercase , *__lowercase , **__lowercase) -> List[Any]: super().__init__(__lowercase , *__lowercase , **__lowercase) __UpperCamelCase :Tuple = TFRegNetMainLayer(__lowercase , name='''regnet''') @unpack_inputs @add_start_docstrings_to_model_forward(__lowercase) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=__lowercase , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def UpperCamelCase__ ( self , __lowercase , __lowercase = None , __lowercase = None , __lowercase=False , ) -> Union[TFBaseModelOutputWithPoolingAndNoAttention, Tuple[tf.Tensor]]: __UpperCamelCase :int = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __UpperCamelCase :Optional[Any] = return_dict if return_dict is not None else self.config.use_return_dict __UpperCamelCase :List[Any] = self.regnet( pixel_values=__lowercase , output_hidden_states=__lowercase , return_dict=__lowercase , training=__lowercase , ) if not return_dict: return (outputs[0],) + outputs[1:] return TFBaseModelOutputWithPoolingAndNoAttention( last_hidden_state=outputs.last_hidden_state , pooler_output=outputs.pooler_output , hidden_states=outputs.hidden_states , ) @add_start_docstrings( """ RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for ImageNet. """ , UpperCAmelCase_ , ) class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ ): '''simple docstring''' def __init__( self , __lowercase , *__lowercase , **__lowercase) -> int: super().__init__(__lowercase , *__lowercase , **__lowercase) __UpperCamelCase :Optional[Any] = config.num_labels __UpperCamelCase :str = TFRegNetMainLayer(__lowercase , name='''regnet''') # classification head __UpperCamelCase :List[Any] = [ tf.keras.layers.Flatten(), tf.keras.layers.Dense(config.num_labels , name='''classifier.1''') if config.num_labels > 0 else tf.identity, ] @unpack_inputs @add_start_docstrings_to_model_forward(__lowercase) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=__lowercase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def UpperCamelCase__ ( self , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase=False , ) -> Union[TFSequenceClassifierOutput, Tuple[tf.Tensor]]: __UpperCamelCase :Optional[int] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __UpperCamelCase :int = return_dict if return_dict is not None else self.config.use_return_dict __UpperCamelCase :Union[str, Any] = self.regnet( __lowercase , output_hidden_states=__lowercase , return_dict=__lowercase , training=__lowercase) __UpperCamelCase :List[str] = outputs.pooler_output if return_dict else outputs[1] __UpperCamelCase :str = self.classifier[0](__lowercase) __UpperCamelCase :List[Any] = self.classifier[1](__lowercase) __UpperCamelCase :Tuple = None if labels is None else self.hf_compute_loss(labels=__lowercase , logits=__lowercase) if not return_dict: __UpperCamelCase :Tuple = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return TFSequenceClassifierOutput(loss=__lowercase , logits=__lowercase , hidden_states=outputs.hidden_states)
43
import random def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = a[left_index] __UpperCamelCase :Any = left_index + 1 for j in range(left_index + 1 , SCREAMING_SNAKE_CASE ): if a[j] < pivot: __UpperCamelCase , __UpperCamelCase :str = a[i], a[j] i += 1 __UpperCamelCase , __UpperCamelCase :Optional[int] = a[i - 1], a[left_index] return i - 1 def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' if left < right: __UpperCamelCase :int = random.randint(SCREAMING_SNAKE_CASE , right - 1 ) __UpperCamelCase , __UpperCamelCase :List[str] = ( a[left], a[pivot], ) # switches the pivot with the left most bound __UpperCamelCase :Dict = partition(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) quick_sort_random( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # recursive quicksort to the left of the pivot point quick_sort_random( SCREAMING_SNAKE_CASE , pivot_index + 1 , SCREAMING_SNAKE_CASE ) # recursive quicksort to the right of the pivot point def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = input('''Enter numbers separated by a comma:\n''' ).strip() __UpperCamelCase :Union[str, Any] = [int(SCREAMING_SNAKE_CASE ) for item in user_input.split(''',''' )] quick_sort_random(SCREAMING_SNAKE_CASE , 0 , len(SCREAMING_SNAKE_CASE ) ) print(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
43
1
import gc import unittest import numpy as np import torch from diffusers import StableDiffusionKDiffusionPipeline from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() @slow @require_torch_gpu class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Union[str, Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Dict = StableDiffusionKDiffusionPipeline.from_pretrained('''CompVis/stable-diffusion-v1-4''') __UpperCamelCase :List[Any] = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) sd_pipe.set_scheduler('''sample_euler''') __UpperCamelCase :Tuple = '''A painting of a squirrel eating a burger''' __UpperCamelCase :Union[str, Any] = torch.manual_seed(0) __UpperCamelCase :List[str] = sd_pipe([prompt] , generator=__lowercase , guidance_scale=9.0 , num_inference_steps=20 , output_type='''np''') __UpperCamelCase :str = output.images __UpperCamelCase :List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) __UpperCamelCase :Optional[int] = np.array([0.04_47, 0.04_92, 0.04_68, 0.04_08, 0.03_83, 0.04_08, 0.03_54, 0.03_80, 0.03_39]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Union[str, Any] = StableDiffusionKDiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-1-base''') __UpperCamelCase :Optional[Any] = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) sd_pipe.set_scheduler('''sample_euler''') __UpperCamelCase :Tuple = '''A painting of a squirrel eating a burger''' __UpperCamelCase :List[str] = torch.manual_seed(0) __UpperCamelCase :List[Any] = sd_pipe([prompt] , generator=__lowercase , guidance_scale=9.0 , num_inference_steps=20 , output_type='''np''') __UpperCamelCase :int = output.images __UpperCamelCase :Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) __UpperCamelCase :str = np.array([0.12_37, 0.13_20, 0.14_38, 0.13_59, 0.13_90, 0.11_32, 0.12_77, 0.11_75, 0.11_12]) assert np.abs(image_slice.flatten() - expected_slice).max() < 5E-1 def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Union[str, Any] = StableDiffusionKDiffusionPipeline.from_pretrained('''stabilityai/stable-diffusion-2-1-base''') __UpperCamelCase :str = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) sd_pipe.set_scheduler('''sample_dpmpp_2m''') __UpperCamelCase :Optional[int] = '''A painting of a squirrel eating a burger''' __UpperCamelCase :str = torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = sd_pipe( [prompt] , generator=__lowercase , guidance_scale=7.5 , num_inference_steps=15 , output_type='''np''' , use_karras_sigmas=__lowercase , ) __UpperCamelCase :List[str] = output.images __UpperCamelCase :Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) __UpperCamelCase :Tuple = np.array( [0.11_38_16_89, 0.12_11_29_21, 0.1_38_94_57, 0.12_54_96_06, 0.1_24_49_64, 0.10_83_15_17, 0.11_56_28_66, 0.10_86_78_16, 0.10_49_90_48]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2
43
def lowerCamelCase ( SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1_000 ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = 1 __UpperCamelCase :Any = 0 for divide_by_number in range(SCREAMING_SNAKE_CASE , digit + 1 ): __UpperCamelCase :list[int] = [] __UpperCamelCase :Optional[int] = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = len(SCREAMING_SNAKE_CASE ) __UpperCamelCase :int = divide_by_number else: has_been_divided.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
43
1
def lowerCamelCase ( SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1_000 ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = 1 __UpperCamelCase :Any = 0 for divide_by_number in range(SCREAMING_SNAKE_CASE , digit + 1 ): __UpperCamelCase :list[int] = [] __UpperCamelCase :Optional[int] = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = len(SCREAMING_SNAKE_CASE ) __UpperCamelCase :int = divide_by_number else: has_been_divided.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
43
import argparse import json from tqdm import tqdm def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--src_path''' , type=SCREAMING_SNAKE_CASE , default='''biencoder-nq-dev.json''' , help='''Path to raw DPR training data''' , ) parser.add_argument( '''--evaluation_set''' , type=SCREAMING_SNAKE_CASE , help='''where to store parsed evaluation_set file''' , ) parser.add_argument( '''--gold_data_path''' , type=SCREAMING_SNAKE_CASE , help='''where to store parsed gold_data_path file''' , ) __UpperCamelCase :str = parser.parse_args() with open(args.src_path , '''r''' ) as src_file, open(args.evaluation_set , '''w''' ) as eval_file, open( args.gold_data_path , '''w''' ) as gold_file: __UpperCamelCase :List[str] = json.load(SCREAMING_SNAKE_CASE ) for dpr_record in tqdm(SCREAMING_SNAKE_CASE ): __UpperCamelCase :List[str] = dpr_record['''question'''] __UpperCamelCase :Tuple = [context['''title'''] for context in dpr_record['''positive_ctxs''']] eval_file.write(question + '''\n''' ) gold_file.write('''\t'''.join(SCREAMING_SNAKE_CASE ) + '''\n''' ) if __name__ == "__main__": main()
43
1
import unittest from transformers import DonutProcessor __lowercase = '''naver-clova-ix/donut-base''' class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :Union[str, Any] = DonutProcessor.from_pretrained(__lowercase) def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :Any = { '''name''': '''John Doe''', '''age''': '''99''', '''city''': '''Atlanta''', '''state''': '''GA''', '''zip''': '''30301''', '''phone''': '''123-4567''', '''nicknames''': [{'''nickname''': '''Johnny'''}, {'''nickname''': '''JD'''}], } __UpperCamelCase :List[str] = ( '''<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>''' '''<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>''' '''<s_nicknames><s_nickname>Johnny</s_nickname>''' '''<sep/><s_nickname>JD</s_nickname></s_nicknames>''' ) __UpperCamelCase :Union[str, Any] = self.processor.tokenajson(__lowercase) self.assertDictEqual(__lowercase , __lowercase)
43
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. __lowercase = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. __lowercase = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. __lowercase = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Any = len([g for position, g in enumerate(SCREAMING_SNAKE_CASE ) if g == main_target[position]] ) return (item, float(SCREAMING_SNAKE_CASE )) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = random.randint(0 , len(SCREAMING_SNAKE_CASE ) - 1 ) __UpperCamelCase :Tuple = parent_a[:random_slice] + parent_a[random_slice:] __UpperCamelCase :Union[str, Any] = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :int = list(SCREAMING_SNAKE_CASE ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: __UpperCamelCase :str = random.choice(SCREAMING_SNAKE_CASE ) return "".join(SCREAMING_SNAKE_CASE ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ): '''simple docstring''' __UpperCamelCase :int = [] # Generate more children proportionally to the fitness score. __UpperCamelCase :int = int(parent_a[1] * 100 ) + 1 __UpperCamelCase :List[str] = 10 if child_n >= 10 else child_n for _ in range(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = population_score[random.randint(0 , SCREAMING_SNAKE_CASE )][0] __UpperCamelCase , __UpperCamelCase :Any = crossover(parent_a[0] , SCREAMING_SNAKE_CASE ) # Append new string to the population list. pop.append(mutate(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) pop.append(mutate(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) return pop def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = True ): '''simple docstring''' if N_POPULATION < N_SELECTED: __UpperCamelCase :List[Any] = f"""{N_POPULATION} must be bigger than {N_SELECTED}""" raise ValueError(SCREAMING_SNAKE_CASE ) # Verify that the target contains no genes besides the ones inside genes variable. __UpperCamelCase :List[str] = sorted({c for c in target if c not in genes} ) if not_in_genes_list: __UpperCamelCase :Optional[int] = f"""{not_in_genes_list} is not in genes list, evolution cannot converge""" raise ValueError(SCREAMING_SNAKE_CASE ) # Generate random starting population. __UpperCamelCase :int = [] for _ in range(SCREAMING_SNAKE_CASE ): population.append(''''''.join([random.choice(SCREAMING_SNAKE_CASE ) for i in range(len(SCREAMING_SNAKE_CASE ) )] ) ) # Just some logs to know what the algorithms is doing. __UpperCamelCase , __UpperCamelCase :List[Any] = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(SCREAMING_SNAKE_CASE ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. __UpperCamelCase :Tuple = [evaluate(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for item in population] # Check if there is a matching evolution. __UpperCamelCase :Tuple = sorted(SCREAMING_SNAKE_CASE , key=lambda SCREAMING_SNAKE_CASE : x[1] , reverse=SCREAMING_SNAKE_CASE ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( f"""\nGeneration: {generation}""" f"""\nTotal Population:{total_population}""" f"""\nBest score: {population_score[0][1]}""" f"""\nBest string: {population_score[0][0]}""" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. __UpperCamelCase :str = population[: int(N_POPULATION / 3 )] population.clear() population.extend(SCREAMING_SNAKE_CASE ) # Normalize population score to be between 0 and 1. __UpperCamelCase :Union[str, Any] = [ (item, score / len(SCREAMING_SNAKE_CASE )) for item, score in population_score ] # This is selection for i in range(SCREAMING_SNAKE_CASE ): population.extend(select(population_score[int(SCREAMING_SNAKE_CASE )] , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(SCREAMING_SNAKE_CASE ) > N_POPULATION: break if __name__ == "__main__": __lowercase = ( '''This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!''' ) __lowercase = list( ''' ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm''' '''nopqrstuvwxyz.,;!?+-*#@^\'èéòà€ù=)(&%$£/\\''' ) __lowercase , __lowercase , __lowercase = basic(target_str, genes_list) print( F'\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}' )
43
1
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[int] = ["""image_processor""", """tokenizer"""] a__ : Optional[int] = """LayoutLMv2ImageProcessor""" a__ : Any = ("""LayoutXLMTokenizer""", """LayoutXLMTokenizerFast""") def __init__( self , __lowercase=None , __lowercase=None , **__lowercase) -> Any: if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , __lowercase , ) __UpperCamelCase :int = kwargs.pop('''feature_extractor''') __UpperCamelCase :Tuple = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''') if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''') super().__init__(__lowercase , __lowercase) def __call__( self , __lowercase , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = True , __lowercase = False , __lowercase = None , __lowercase = None , __lowercase = 0 , __lowercase = None , __lowercase = None , __lowercase = None , __lowercase = False , __lowercase = False , __lowercase = False , __lowercase = False , __lowercase = True , __lowercase = None , **__lowercase , ) -> BatchEncoding: # verify input if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( '''You cannot provide bounding boxes ''' '''if you initialized the image processor with apply_ocr set to True.''') if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( '''You cannot provide word labels if you initialized the image processor with apply_ocr set to True.''') if return_overflowing_tokens is True and return_offsets_mapping is False: raise ValueError('''You cannot return overflowing tokens without returning the offsets mapping.''') # first, apply the image processor __UpperCamelCase :int = self.image_processor(images=__lowercase , return_tensors=__lowercase) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(__lowercase , __lowercase): __UpperCamelCase :Dict = [text] # add batch dimension (as the image processor always adds a batch dimension) __UpperCamelCase :Optional[int] = features['''words'''] __UpperCamelCase :List[Any] = self.tokenizer( text=text if text is not None else features['''words'''] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['''boxes'''] , word_labels=__lowercase , add_special_tokens=__lowercase , padding=__lowercase , truncation=__lowercase , max_length=__lowercase , stride=__lowercase , pad_to_multiple_of=__lowercase , return_token_type_ids=__lowercase , return_attention_mask=__lowercase , return_overflowing_tokens=__lowercase , return_special_tokens_mask=__lowercase , return_offsets_mapping=__lowercase , return_length=__lowercase , verbose=__lowercase , return_tensors=__lowercase , **__lowercase , ) # add pixel values __UpperCamelCase :Dict = features.pop('''pixel_values''') if return_overflowing_tokens is True: __UpperCamelCase :Optional[Any] = self.get_overflowing_images(__lowercase , encoded_inputs['''overflow_to_sample_mapping''']) __UpperCamelCase :int = images return encoded_inputs def UpperCamelCase__ ( self , __lowercase , __lowercase) -> Tuple: # in case there's an overflow, ensure each `input_ids` sample is mapped to its corresponding image __UpperCamelCase :int = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx]) if len(__lowercase) != len(__lowercase): raise ValueError( '''Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got''' f""" {len(__lowercase)} and {len(__lowercase)}""") return images_with_overflow def UpperCamelCase__ ( self , *__lowercase , **__lowercase) -> List[Any]: return self.tokenizer.batch_decode(*__lowercase , **__lowercase) def UpperCamelCase__ ( self , *__lowercase , **__lowercase) -> List[str]: return self.tokenizer.decode(*__lowercase , **__lowercase) @property def UpperCamelCase__ ( self) -> int: return ["input_ids", "bbox", "attention_mask", "image"] @property def UpperCamelCase__ ( self) -> List[Any]: warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , __lowercase , ) return self.image_processor_class @property def UpperCamelCase__ ( self) -> str: warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , __lowercase , ) return self.image_processor
43
import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler __lowercase = 16 __lowercase = 32 def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 16 , SCREAMING_SNAKE_CASE = "bert-base-cased" ): '''simple docstring''' __UpperCamelCase :List[str] = AutoTokenizer.from_pretrained(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(SCREAMING_SNAKE_CASE ): # max_length=None => use the model max length (it's actually the default) __UpperCamelCase :int = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset __UpperCamelCase :Tuple = datasets.map( SCREAMING_SNAKE_CASE , batched=SCREAMING_SNAKE_CASE , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , load_from_cache_file=SCREAMING_SNAKE_CASE ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __UpperCamelCase :List[str] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(SCREAMING_SNAKE_CASE ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(SCREAMING_SNAKE_CASE , padding='''max_length''' , max_length=128 , return_tensors='''pt''' ) return tokenizer.pad(SCREAMING_SNAKE_CASE , padding='''longest''' , return_tensors='''pt''' ) # Instantiate dataloaders. __UpperCamelCase :Union[str, Any] = DataLoader( tokenized_datasets['''train'''] , shuffle=SCREAMING_SNAKE_CASE , collate_fn=SCREAMING_SNAKE_CASE , batch_size=SCREAMING_SNAKE_CASE ) __UpperCamelCase :Dict = DataLoader( tokenized_datasets['''validation'''] , shuffle=SCREAMING_SNAKE_CASE , collate_fn=SCREAMING_SNAKE_CASE , batch_size=SCREAMING_SNAKE_CASE ) return train_dataloader, eval_dataloader def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __UpperCamelCase :int = config['''lr'''] __UpperCamelCase :str = int(config['''num_epochs'''] ) __UpperCamelCase :Any = int(config['''seed'''] ) __UpperCamelCase :Dict = int(config['''batch_size'''] ) __UpperCamelCase :Optional[Any] = args.model_name_or_path set_seed(SCREAMING_SNAKE_CASE ) __UpperCamelCase , __UpperCamelCase :Dict = get_dataloaders(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __UpperCamelCase :Any = AutoModelForSequenceClassification.from_pretrained(SCREAMING_SNAKE_CASE , return_dict=SCREAMING_SNAKE_CASE ) # Instantiate optimizer __UpperCamelCase :List[str] = ( AdamW if accelerator.state.deepspeed_plugin is None or '''optimizer''' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) __UpperCamelCase :Optional[Any] = optimizer_cls(params=model.parameters() , lr=SCREAMING_SNAKE_CASE ) if accelerator.state.deepspeed_plugin is not None: __UpperCamelCase :Dict = accelerator.state.deepspeed_plugin.deepspeed_config[ '''gradient_accumulation_steps''' ] else: __UpperCamelCase :Dict = 1 __UpperCamelCase :Tuple = (len(SCREAMING_SNAKE_CASE ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): __UpperCamelCase :str = get_linear_schedule_with_warmup( optimizer=SCREAMING_SNAKE_CASE , num_warmup_steps=0 , num_training_steps=SCREAMING_SNAKE_CASE , ) else: __UpperCamelCase :Dict = DummyScheduler(SCREAMING_SNAKE_CASE , total_num_steps=SCREAMING_SNAKE_CASE , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :int = accelerator.prepare( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # We need to keep track of how many total steps we have iterated over __UpperCamelCase :List[Any] = 0 # We also need to keep track of the stating epoch so files are named properly __UpperCamelCase :Dict = 0 # Now we train the model __UpperCamelCase :Any = evaluate.load('''glue''' , '''mrpc''' ) __UpperCamelCase :Union[str, Any] = 0 __UpperCamelCase :Optional[int] = {} for epoch in range(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): model.train() for step, batch in enumerate(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = model(**SCREAMING_SNAKE_CASE ) __UpperCamelCase :Tuple = outputs.loss __UpperCamelCase :str = loss / gradient_accumulation_steps accelerator.backward(SCREAMING_SNAKE_CASE ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() __UpperCamelCase :Any = 0 for step, batch in enumerate(SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __UpperCamelCase :Any = model(**SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times __UpperCamelCase , __UpperCamelCase :List[Any] = accelerator.gather( (predictions, batch['''labels''']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(SCREAMING_SNAKE_CASE ) - 1: __UpperCamelCase :List[str] = predictions[: len(eval_dataloader.dataset ) - samples_seen] __UpperCamelCase :Optional[int] = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=SCREAMING_SNAKE_CASE , references=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :Dict = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , SCREAMING_SNAKE_CASE ) __UpperCamelCase :str = eval_metric['''accuracy'''] if best_performance < eval_metric["accuracy"]: __UpperCamelCase :int = eval_metric['''accuracy'''] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f"""Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}""" accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , '''all_results.json''' ) , '''w''' ) as f: json.dump(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = argparse.ArgumentParser(description='''Simple example of training script tracking peak GPU memory usage.''' ) parser.add_argument( '''--model_name_or_path''' , type=SCREAMING_SNAKE_CASE , default='''bert-base-cased''' , help='''Path to pretrained model or model identifier from huggingface.co/models.''' , required=SCREAMING_SNAKE_CASE , ) parser.add_argument( '''--output_dir''' , type=SCREAMING_SNAKE_CASE , default='''.''' , help='''Optional save directory where all checkpoint folders will be stored. Default is the current working directory.''' , ) parser.add_argument( '''--performance_lower_bound''' , type=SCREAMING_SNAKE_CASE , default=SCREAMING_SNAKE_CASE , help='''Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.''' , ) parser.add_argument( '''--num_epochs''' , type=SCREAMING_SNAKE_CASE , default=3 , help='''Number of train epochs.''' , ) __UpperCamelCase :List[str] = parser.parse_args() __UpperCamelCase :Tuple = {'''lr''': 2e-5, '''num_epochs''': args.num_epochs, '''seed''': 42, '''batch_size''': 16} training_function(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
43
1
from __future__ import annotations from math import pi def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' if (inductance, frequency, reactance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if inductance < 0: raise ValueError('''Inductance cannot be negative''' ) if frequency < 0: raise ValueError('''Frequency cannot be negative''' ) if reactance < 0: raise ValueError('''Inductive reactance cannot be negative''' ) if inductance == 0: return {"inductance": reactance / (2 * pi * frequency)} elif frequency == 0: return {"frequency": reactance / (2 * pi * inductance)} elif reactance == 0: return {"reactance": 2 * pi * frequency * inductance} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
43
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __lowercase = logging.get_logger(__name__) __lowercase = { '''SenseTime/deformable-detr''': '''https://huggingface.co/sensetime/deformable-detr/resolve/main/config.json''', # See all Deformable DETR models at https://huggingface.co/models?filter=deformable-detr } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : List[str] = """deformable_detr""" a__ : Union[str, Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , __lowercase=True , __lowercase=None , __lowercase=3 , __lowercase=300 , __lowercase=1_024 , __lowercase=6 , __lowercase=1_024 , __lowercase=8 , __lowercase=6 , __lowercase=1_024 , __lowercase=8 , __lowercase=0.0 , __lowercase=True , __lowercase="relu" , __lowercase=256 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.02 , __lowercase=1.0 , __lowercase=True , __lowercase=False , __lowercase="sine" , __lowercase="resnet50" , __lowercase=True , __lowercase=False , __lowercase=4 , __lowercase=4 , __lowercase=4 , __lowercase=False , __lowercase=300 , __lowercase=False , __lowercase=1 , __lowercase=5 , __lowercase=2 , __lowercase=1 , __lowercase=1 , __lowercase=5 , __lowercase=2 , __lowercase=0.1 , __lowercase=0.25 , __lowercase=False , **__lowercase , ) -> int: if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''') if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''') __UpperCamelCase :str = CONFIG_MAPPING['''resnet'''](out_features=['''stage4''']) elif isinstance(__lowercase , __lowercase): __UpperCamelCase :str = backbone_config.get('''model_type''') __UpperCamelCase :Tuple = CONFIG_MAPPING[backbone_model_type] __UpperCamelCase :Any = config_class.from_dict(__lowercase) __UpperCamelCase :int = use_timm_backbone __UpperCamelCase :Dict = backbone_config __UpperCamelCase :Any = num_channels __UpperCamelCase :Optional[int] = num_queries __UpperCamelCase :Any = max_position_embeddings __UpperCamelCase :str = d_model __UpperCamelCase :Tuple = encoder_ffn_dim __UpperCamelCase :Union[str, Any] = encoder_layers __UpperCamelCase :List[Any] = encoder_attention_heads __UpperCamelCase :Any = decoder_ffn_dim __UpperCamelCase :List[str] = decoder_layers __UpperCamelCase :int = decoder_attention_heads __UpperCamelCase :str = dropout __UpperCamelCase :Any = attention_dropout __UpperCamelCase :int = activation_dropout __UpperCamelCase :List[Any] = activation_function __UpperCamelCase :List[Any] = init_std __UpperCamelCase :List[Any] = init_xavier_std __UpperCamelCase :int = encoder_layerdrop __UpperCamelCase :str = auxiliary_loss __UpperCamelCase :Optional[Any] = position_embedding_type __UpperCamelCase :Union[str, Any] = backbone __UpperCamelCase :Any = use_pretrained_backbone __UpperCamelCase :str = dilation # deformable attributes __UpperCamelCase :Optional[Any] = num_feature_levels __UpperCamelCase :str = encoder_n_points __UpperCamelCase :int = decoder_n_points __UpperCamelCase :Union[str, Any] = two_stage __UpperCamelCase :Optional[Any] = two_stage_num_proposals __UpperCamelCase :Dict = with_box_refine if two_stage is True and with_box_refine is False: raise ValueError('''If two_stage is True, with_box_refine must be True.''') # Hungarian matcher __UpperCamelCase :Optional[int] = class_cost __UpperCamelCase :List[Any] = bbox_cost __UpperCamelCase :str = giou_cost # Loss coefficients __UpperCamelCase :Tuple = mask_loss_coefficient __UpperCamelCase :Tuple = dice_loss_coefficient __UpperCamelCase :int = bbox_loss_coefficient __UpperCamelCase :Any = giou_loss_coefficient __UpperCamelCase :Dict = eos_coefficient __UpperCamelCase :Optional[Any] = focal_alpha __UpperCamelCase :Optional[Any] = disable_custom_kernels super().__init__(is_encoder_decoder=__lowercase , **__lowercase) @property def UpperCamelCase__ ( self) -> int: return self.encoder_attention_heads @property def UpperCamelCase__ ( self) -> int: return self.d_model def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Dict = copy.deepcopy(self.__dict__) if self.backbone_config is not None: __UpperCamelCase :Tuple = self.backbone_config.to_dict() __UpperCamelCase :List[Any] = self.__class__.model_type return output
43
1
import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> int: __UpperCamelCase :int = '''ZinengTang/tvlt-base''' __UpperCamelCase :str = tempfile.mkdtemp() def UpperCamelCase__ ( self , **__lowercase) -> List[str]: return TvltImageProcessor.from_pretrained(self.checkpoint , **__lowercase) def UpperCamelCase__ ( self , **__lowercase) -> List[Any]: return TvltFeatureExtractor.from_pretrained(self.checkpoint , **__lowercase) def UpperCamelCase__ ( self) -> int: shutil.rmtree(self.tmpdirname) def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :int = self.get_image_processor() __UpperCamelCase :Dict = self.get_feature_extractor() __UpperCamelCase :Optional[Any] = TvltProcessor(image_processor=__lowercase , feature_extractor=__lowercase) processor.save_pretrained(self.tmpdirname) __UpperCamelCase :Tuple = TvltProcessor.from_pretrained(self.tmpdirname) self.assertIsInstance(processor.feature_extractor , __lowercase) self.assertIsInstance(processor.image_processor , __lowercase) def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :Union[str, Any] = self.get_image_processor() __UpperCamelCase :List[Any] = self.get_feature_extractor() __UpperCamelCase :Optional[Any] = TvltProcessor(image_processor=__lowercase , feature_extractor=__lowercase) __UpperCamelCase :Tuple = np.ones([12_000]) __UpperCamelCase :Union[str, Any] = feature_extractor(__lowercase , return_tensors='''np''') __UpperCamelCase :Tuple = processor(audio=__lowercase , return_tensors='''np''') for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1E-2) def UpperCamelCase__ ( self) -> int: __UpperCamelCase :str = self.get_image_processor() __UpperCamelCase :int = self.get_feature_extractor() __UpperCamelCase :Optional[Any] = TvltProcessor(image_processor=__lowercase , feature_extractor=__lowercase) __UpperCamelCase :Dict = np.ones([3, 224, 224]) __UpperCamelCase :Optional[Any] = image_processor(__lowercase , return_tensors='''np''') __UpperCamelCase :Dict = processor(images=__lowercase , return_tensors='''np''') for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1E-2) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Any = self.get_image_processor() __UpperCamelCase :Tuple = self.get_feature_extractor() __UpperCamelCase :str = TvltProcessor(image_processor=__lowercase , feature_extractor=__lowercase) __UpperCamelCase :Optional[int] = np.ones([12_000]) __UpperCamelCase :Optional[Any] = np.ones([3, 224, 224]) __UpperCamelCase :Optional[Any] = processor(audio=__lowercase , images=__lowercase) self.assertListEqual(list(inputs.keys()) , ['''audio_values''', '''audio_mask''', '''pixel_values''', '''pixel_mask''']) # test if it raises when no input is passed with pytest.raises(__lowercase): processor() def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :List[str] = self.get_image_processor() __UpperCamelCase :Optional[Any] = self.get_feature_extractor() __UpperCamelCase :int = TvltProcessor(image_processor=__lowercase , feature_extractor=__lowercase) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='''`processor` and `image_processor`+`feature_extractor` model input names do not match''' , )
43
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[Any] = """facebook/bart-large-mnli""" a__ : int = ( """This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which """ """should be the text to classify, and `labels`, which should be the list of labels to use for classification. """ """It returns the most likely label in the list of provided `labels` for the input text.""" ) a__ : Optional[Any] = """text_classifier""" a__ : Any = AutoTokenizer a__ : str = AutoModelForSequenceClassification a__ : str = ["""text""", ["""text"""]] a__ : Optional[int] = ["""text"""] def UpperCamelCase__ ( self) -> Union[str, Any]: super().setup() __UpperCamelCase :int = self.model.config __UpperCamelCase :Optional[Any] = -1 for idx, label in config.idalabel.items(): if label.lower().startswith('''entail'''): __UpperCamelCase :List[Any] = int(__lowercase) if self.entailment_id == -1: raise ValueError('''Could not determine the entailment ID from the model config, please pass it at init.''') def UpperCamelCase__ ( self , __lowercase , __lowercase) -> Union[str, Any]: __UpperCamelCase :Any = labels return self.pre_processor( [text] * len(__lowercase) , [f"""This example is {label}""" for label in labels] , return_tensors='''pt''' , padding='''max_length''' , ) def UpperCamelCase__ ( self , __lowercase) -> Optional[Any]: __UpperCamelCase :List[Any] = outputs.logits __UpperCamelCase :Any = torch.argmax(logits[:, 2]).item() return self._labels[label_id]
43
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def __init__( self , __lowercase , __lowercase=7 , __lowercase=3 , __lowercase=18 , __lowercase=30 , __lowercase=400 , __lowercase=True , __lowercase=None , __lowercase=True , __lowercase=None , __lowercase=True , ) -> Dict: __UpperCamelCase :Union[str, Any] = size if size is not None else {'''shortest_edge''': 20} __UpperCamelCase :List[str] = crop_size if crop_size is not None else {'''height''': 18, '''width''': 18} __UpperCamelCase :int = parent __UpperCamelCase :Union[str, Any] = batch_size __UpperCamelCase :Tuple = num_channels __UpperCamelCase :Tuple = image_size __UpperCamelCase :List[Any] = min_resolution __UpperCamelCase :Tuple = max_resolution __UpperCamelCase :Tuple = do_resize __UpperCamelCase :Any = size __UpperCamelCase :Optional[int] = do_center_crop __UpperCamelCase :int = crop_size __UpperCamelCase :Optional[int] = do_flip_channel_order def UpperCamelCase__ ( self) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_flip_channel_order": self.do_flip_channel_order, } @require_torch @require_vision class lowerCamelCase_ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : Optional[int] = MobileViTImageProcessor if is_vision_available() else None def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :int = MobileViTImageProcessingTester(self) @property def UpperCamelCase__ ( self) -> List[Any]: return self.image_processor_tester.prepare_image_processor_dict() def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Union[str, Any] = self.image_processing_class(**self.image_processor_dict) self.assertTrue(hasattr(__lowercase , '''do_resize''')) self.assertTrue(hasattr(__lowercase , '''size''')) self.assertTrue(hasattr(__lowercase , '''do_center_crop''')) self.assertTrue(hasattr(__lowercase , '''center_crop''')) self.assertTrue(hasattr(__lowercase , '''do_flip_channel_order''')) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Optional[int] = self.image_processing_class.from_dict(self.image_processor_dict) self.assertEqual(image_processor.size , {'''shortest_edge''': 20}) self.assertEqual(image_processor.crop_size , {'''height''': 18, '''width''': 18}) __UpperCamelCase :Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84) self.assertEqual(image_processor.size , {'''shortest_edge''': 42}) self.assertEqual(image_processor.crop_size , {'''height''': 84, '''width''': 84}) def UpperCamelCase__ ( self) -> Optional[Any]: pass def UpperCamelCase__ ( self) -> List[Any]: # Initialize image_processing __UpperCamelCase :List[str] = self.image_processing_class(**self.image_processor_dict) # create random PIL images __UpperCamelCase :List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase) for image in image_inputs: self.assertIsInstance(__lowercase , Image.Image) # Test not batched input __UpperCamelCase :int = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched __UpperCamelCase :str = image_processing(__lowercase , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCamelCase__ ( self) -> Dict: # Initialize image_processing __UpperCamelCase :Union[str, Any] = self.image_processing_class(**self.image_processor_dict) # create random numpy tensors __UpperCamelCase :List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , numpify=__lowercase) for image in image_inputs: self.assertIsInstance(__lowercase , np.ndarray) # Test not batched input __UpperCamelCase :List[Any] = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched __UpperCamelCase :Optional[int] = image_processing(__lowercase , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) def UpperCamelCase__ ( self) -> List[str]: # Initialize image_processing __UpperCamelCase :List[Any] = self.image_processing_class(**self.image_processor_dict) # create random PyTorch tensors __UpperCamelCase :Union[str, Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=__lowercase , torchify=__lowercase) for image in image_inputs: self.assertIsInstance(__lowercase , torch.Tensor) # Test not batched input __UpperCamelCase :Optional[Any] = image_processing(image_inputs[0] , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , ) # Test batched __UpperCamelCase :Optional[int] = image_processing(__lowercase , return_tensors='''pt''').pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) , )
43
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : int = StableUnCLIPImgaImgPipeline a__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS a__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a__ : Optional[Any] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess a__ : int = frozenset([] ) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Tuple = 32 __UpperCamelCase :Optional[int] = embedder_hidden_size # image encoding components __UpperCamelCase :Union[str, Any] = CLIPImageProcessor(crop_size=32 , size=32) torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=__lowercase , projection_dim=__lowercase , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , )) # regular denoising components torch.manual_seed(0) __UpperCamelCase :str = StableUnCLIPImageNormalizer(embedding_dim=__lowercase) __UpperCamelCase :Optional[int] = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''') torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') torch.manual_seed(0) __UpperCamelCase :Dict = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__lowercase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , )) torch.manual_seed(0) __UpperCamelCase :List[Any] = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=__lowercase , layers_per_block=1 , upcast_attention=__lowercase , use_linear_projection=__lowercase , ) torch.manual_seed(0) __UpperCamelCase :Tuple = DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.0_00_85 , beta_end=0.0_12 , prediction_type='''v_prediction''' , set_alpha_to_one=__lowercase , steps_offset=1 , ) torch.manual_seed(0) __UpperCamelCase :List[str] = AutoencoderKL() __UpperCamelCase :Tuple = { # image encoding components '''feature_extractor''': feature_extractor, '''image_encoder''': image_encoder.eval(), # image noising components '''image_normalizer''': image_normalizer.eval(), '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder.eval(), '''unet''': unet.eval(), '''scheduler''': scheduler, '''vae''': vae.eval(), } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0 , __lowercase=True) -> str: if str(__lowercase).startswith('''mps'''): __UpperCamelCase :Union[str, Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :int = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :int = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowercase)).to(__lowercase) if pil_image: __UpperCamelCase :List[Any] = input_image * 0.5 + 0.5 __UpperCamelCase :Optional[Any] = input_image.clamp(0 , 1) __UpperCamelCase :int = input_image.cpu().permute(0 , 2 , 3 , 1).float().numpy() __UpperCamelCase :Optional[Any] = DiffusionPipeline.numpy_to_pil(__lowercase)[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :Dict = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Tuple = self.get_dummy_components() __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline(**__lowercase) __UpperCamelCase :Optional[Any] = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :List[Any] = self.get_dummy_inputs(__lowercase) inputs.update({'''image_embeds''': None}) __UpperCamelCase :Any = sd_pipe(**__lowercase).images __UpperCamelCase :List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCamelCase :List[Any] = np.array([0.38_72, 0.72_24, 0.56_01, 0.47_41, 0.68_72, 0.58_14, 0.46_36, 0.38_67, 0.50_78]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_attention_slicing_forward_pass(test_max_difference=__lowercase) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=__lowercase) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase__ ( self) -> Union[str, Any]: self._test_xformers_attention_forwardGenerator_pass(test_max_difference=__lowercase) @slow @require_torch_gpu class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Union[str, Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :int = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-l-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Dict = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :Dict = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Optional[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Optional[int] = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :List[Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) __UpperCamelCase :Union[str, Any] = pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :Optional[Any] = pipe( __lowercase , '''anime turtle''' , num_inference_steps=2 , output_type='''np''' , ) __UpperCamelCase :int = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
43
1
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed __lowercase = '''true''' def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=82 , SCREAMING_SNAKE_CASE=16 ): '''simple docstring''' set_seed(42 ) __UpperCamelCase :List[str] = RegressionModel() __UpperCamelCase :List[Any] = deepcopy(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = RegressionDataset(length=SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = DataLoader(SCREAMING_SNAKE_CASE , batch_size=SCREAMING_SNAKE_CASE ) model.to(accelerator.device ) __UpperCamelCase , __UpperCamelCase :Optional[Any] = accelerator.prepare(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return model, ddp_model, dataloader def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=False ): '''simple docstring''' __UpperCamelCase :List[str] = AutoTokenizer.from_pretrained('''hf-internal-testing/mrpc-bert-base-cased''' ) __UpperCamelCase :Optional[int] = load_dataset('''glue''' , '''mrpc''' , split='''validation''' ) def tokenize_function(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE ) return outputs with accelerator.main_process_first(): __UpperCamelCase :Any = dataset.map( SCREAMING_SNAKE_CASE , batched=SCREAMING_SNAKE_CASE , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) __UpperCamelCase :Union[str, Any] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(SCREAMING_SNAKE_CASE ): if use_longest: return tokenizer.pad(SCREAMING_SNAKE_CASE , padding='''longest''' , return_tensors='''pt''' ) return tokenizer.pad(SCREAMING_SNAKE_CASE , padding='''max_length''' , max_length=128 , return_tensors='''pt''' ) return DataLoader(SCREAMING_SNAKE_CASE , shuffle=SCREAMING_SNAKE_CASE , collate_fn=SCREAMING_SNAKE_CASE , batch_size=16 ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :str = Accelerator(dispatch_batches=SCREAMING_SNAKE_CASE , split_batches=SCREAMING_SNAKE_CASE ) __UpperCamelCase :int = get_dataloader(SCREAMING_SNAKE_CASE , not dispatch_batches ) __UpperCamelCase :List[str] = AutoModelForSequenceClassification.from_pretrained( '''hf-internal-testing/mrpc-bert-base-cased''' , return_dict=SCREAMING_SNAKE_CASE ) __UpperCamelCase , __UpperCamelCase :List[Any] = accelerator.prepare(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[str] = [] for batch in dataloader: __UpperCamelCase , __UpperCamelCase :Tuple = batch.values() with torch.no_grad(): __UpperCamelCase :Any = model(SCREAMING_SNAKE_CASE ) __UpperCamelCase , __UpperCamelCase :Any = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) __UpperCamelCase , __UpperCamelCase :List[Any] = [], [] for logit, targ in logits_and_targets: logits.append(SCREAMING_SNAKE_CASE ) targs.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase , __UpperCamelCase :str = torch.cat(SCREAMING_SNAKE_CASE ), torch.cat(SCREAMING_SNAKE_CASE ) return logits, targs def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=82 , SCREAMING_SNAKE_CASE=False , SCREAMING_SNAKE_CASE=False , SCREAMING_SNAKE_CASE=16 ): '''simple docstring''' __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Tuple = get_basic_setup(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) __UpperCamelCase , __UpperCamelCase :List[Any] = generate_predictions(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) assert ( len(SCREAMING_SNAKE_CASE ) == num_samples ), f"""Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(SCREAMING_SNAKE_CASE )}""" def lowerCamelCase ( SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = False ): '''simple docstring''' __UpperCamelCase :Dict = evaluate.load('''glue''' , '''mrpc''' ) __UpperCamelCase , __UpperCamelCase :Dict = get_mrpc_setup(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # First do baseline __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Optional[int] = setup['''no'''] model.to(SCREAMING_SNAKE_CASE ) model.eval() for batch in dataloader: batch.to(SCREAMING_SNAKE_CASE ) with torch.inference_mode(): __UpperCamelCase :Optional[Any] = model(**SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[str] = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=SCREAMING_SNAKE_CASE , references=batch['''labels'''] ) __UpperCamelCase :Tuple = metric.compute() # Then do distributed __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Any = setup['''ddp'''] model.eval() for batch in dataloader: with torch.inference_mode(): __UpperCamelCase :Optional[Any] = model(**SCREAMING_SNAKE_CASE ) __UpperCamelCase :Union[str, Any] = outputs.logits.argmax(dim=-1 ) __UpperCamelCase :Optional[Any] = batch['''labels'''] __UpperCamelCase , __UpperCamelCase :Optional[int] = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=SCREAMING_SNAKE_CASE , references=SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), f"""Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n""" def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :str = Accelerator(split_batches=SCREAMING_SNAKE_CASE , dispatch_batches=SCREAMING_SNAKE_CASE ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print('''**Testing gather_for_metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(f"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`""" ) test_mrpc(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test torch metrics**''' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: __UpperCamelCase :Union[str, Any] = Accelerator(split_batches=SCREAMING_SNAKE_CASE , dispatch_batches=SCREAMING_SNAKE_CASE ) if accelerator.is_local_main_process: print(f"""With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99""" ) test_torch_metrics(SCREAMING_SNAKE_CASE , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('''**Test last batch is not dropped when perfectly divisible**''' ) __UpperCamelCase :Union[str, Any] = Accelerator() test_torch_metrics(SCREAMING_SNAKE_CASE , 512 ) accelerator.state._reset_state() def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' main() if __name__ == "__main__": main()
43
import numpy as np def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 1e-12 , SCREAMING_SNAKE_CASE = 100 , ): '''simple docstring''' assert np.shape(SCREAMING_SNAKE_CASE )[0] == np.shape(SCREAMING_SNAKE_CASE )[1] # Ensure proper dimensionality. assert np.shape(SCREAMING_SNAKE_CASE )[0] == np.shape(SCREAMING_SNAKE_CASE )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(SCREAMING_SNAKE_CASE ) == np.iscomplexobj(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = np.iscomplexobj(SCREAMING_SNAKE_CASE ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(SCREAMING_SNAKE_CASE , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. __UpperCamelCase :str = False __UpperCamelCase :int = 0 __UpperCamelCase :Optional[Any] = 0 __UpperCamelCase :Union[str, Any] = 1e12 while not convergence: # Multiple matrix by the vector. __UpperCamelCase :List[str] = np.dot(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Normalize the resulting output vector. __UpperCamelCase :Tuple = w / np.linalg.norm(SCREAMING_SNAKE_CASE ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) __UpperCamelCase :int = vector.conj().T if is_complex else vector.T __UpperCamelCase :Optional[int] = np.dot(SCREAMING_SNAKE_CASE , np.dot(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) # Check convergence. __UpperCamelCase :Optional[Any] = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: __UpperCamelCase :Dict = True __UpperCamelCase :List[Any] = lambda_ if is_complex: __UpperCamelCase :Tuple = np.real(lambda_ ) return lambda_, vector def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :int = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) __UpperCamelCase :Optional[Any] = np.array([41, 4, 20] ) __UpperCamelCase :Any = real_input_matrix.astype(np.complexaaa ) __UpperCamelCase :Dict = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T __UpperCamelCase :Optional[int] = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": __UpperCamelCase :Any = real_input_matrix __UpperCamelCase :int = real_vector elif problem_type == "complex": __UpperCamelCase :Tuple = complex_input_matrix __UpperCamelCase :Optional[Any] = complex_vector # Our implementation. __UpperCamelCase , __UpperCamelCase :Dict = power_iteration(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). __UpperCamelCase , __UpperCamelCase :List[Any] = np.linalg.eigh(SCREAMING_SNAKE_CASE ) # Last eigenvalue is the maximum one. __UpperCamelCase :List[Any] = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. __UpperCamelCase :str = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(SCREAMING_SNAKE_CASE ) - np.abs(SCREAMING_SNAKE_CASE ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
43
1
import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow __lowercase = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ '''text-classification''', '''language-modeling''', '''summarization''', '''token-classification''', '''question-answering''', ] ] sys.path.extend(SRC_DIRS) if SRC_DIRS is not None: import run_clm_flax import run_flax_glue import run_flax_ner import run_mlm_flax import run_qa import run_summarization_flax import run_ta_mlm_flax logging.basicConfig(level=logging.DEBUG) __lowercase = logging.getLogger() def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :List[str] = argparse.ArgumentParser() parser.add_argument('''-f''' ) __UpperCamelCase :Dict = parser.parse_args() return args.f def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE="eval" ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = os.path.join(SCREAMING_SNAKE_CASE , f"""{split}_results.json""" ) if os.path.exists(SCREAMING_SNAKE_CASE ): with open(SCREAMING_SNAKE_CASE , '''r''' ) as f: return json.load(SCREAMING_SNAKE_CASE ) raise ValueError(f"""can't find {path}""" ) __lowercase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :List[str] = self.get_auto_remove_tmp_dir() __UpperCamelCase :List[str] = f""" run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --eval_steps=2 --warmup_steps=2 --seed=42 --max_seq_length=128 """.split() with patch.object(__lowercase , '''argv''' , __lowercase): run_flax_glue.main() __UpperCamelCase :Optional[Any] = get_results(__lowercase) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.75) @slow def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Dict = self.get_auto_remove_tmp_dir() __UpperCamelCase :Optional[int] = f""" run_clm_flax.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --block_size 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(__lowercase , '''argv''' , __lowercase): run_clm_flax.main() __UpperCamelCase :Tuple = get_results(__lowercase) self.assertLess(result['''eval_perplexity'''] , 100) @slow def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :int = self.get_auto_remove_tmp_dir() __UpperCamelCase :List[str] = f""" run_summarization.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --test_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=8 --do_train --do_eval --do_predict --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --predict_with_generate """.split() with patch.object(__lowercase , '''argv''' , __lowercase): run_summarization_flax.main() __UpperCamelCase :Tuple = get_results(__lowercase , split='''test''') self.assertGreaterEqual(result['''test_rouge1'''] , 10) self.assertGreaterEqual(result['''test_rouge2'''] , 2) self.assertGreaterEqual(result['''test_rougeL'''] , 7) self.assertGreaterEqual(result['''test_rougeLsum'''] , 7) @slow def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :List[Any] = self.get_auto_remove_tmp_dir() __UpperCamelCase :List[str] = f""" run_mlm.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --overwrite_output_dir --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --logging_steps 2 --eval_steps 2 --do_train --do_eval --num_train_epochs=1 """.split() with patch.object(__lowercase , '''argv''' , __lowercase): run_mlm_flax.main() __UpperCamelCase :Union[str, Any] = get_results(__lowercase) self.assertLess(result['''eval_perplexity'''] , 42) @slow def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :Union[str, Any] = self.get_auto_remove_tmp_dir() __UpperCamelCase :Optional[int] = f""" run_t5_mlm_flax.py --model_name_or_path t5-small --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(__lowercase , '''argv''' , __lowercase): run_ta_mlm_flax.main() __UpperCamelCase :Dict = get_results(__lowercase) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.42) @slow def UpperCamelCase__ ( self) -> List[Any]: # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu __UpperCamelCase :Any = 7 if get_gpu_count() > 1 else 2 __UpperCamelCase :List[str] = self.get_auto_remove_tmp_dir() __UpperCamelCase :Union[str, Any] = f""" run_flax_ner.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --overwrite_output_dir --do_train --do_eval --warmup_steps=2 --learning_rate=2e-4 --logging_steps 2 --eval_steps 2 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 """.split() with patch.object(__lowercase , '''argv''' , __lowercase): run_flax_ner.main() __UpperCamelCase :Tuple = get_results(__lowercase) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.75) self.assertGreaterEqual(result['''eval_f1'''] , 0.3) @slow def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Dict = self.get_auto_remove_tmp_dir() __UpperCamelCase :Dict = f""" run_qa.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=2 --do_train --do_eval --logging_steps 2 --eval_steps 2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 """.split() with patch.object(__lowercase , '''argv''' , __lowercase): run_qa.main() __UpperCamelCase :Any = get_results(__lowercase) self.assertGreaterEqual(result['''eval_f1'''] , 30) self.assertGreaterEqual(result['''eval_exact'''] , 30)
43
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __lowercase = logging.get_logger(__name__) __lowercase = {'''tokenizer_file''': '''tokenizer.json'''} __lowercase = { '''tokenizer_file''': { '''bigscience/tokenizer''': '''https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json''', '''bigscience/bloom-560m''': '''https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json''', '''bigscience/bloom-1b1''': '''https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json''', '''bigscience/bloom-1b7''': '''https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json''', '''bigscience/bloom-3b''': '''https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json''', '''bigscience/bloom-7b1''': '''https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json''', '''bigscience/bloom''': '''https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json''', }, } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : int = VOCAB_FILES_NAMES a__ : Tuple = PRETRAINED_VOCAB_FILES_MAP a__ : List[str] = ["""input_ids""", """attention_mask"""] a__ : int = None def __init__( self , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase="<unk>" , __lowercase="<s>" , __lowercase="</s>" , __lowercase="<pad>" , __lowercase=False , __lowercase=False , **__lowercase , ) -> List[str]: super().__init__( __lowercase , __lowercase , tokenizer_file=__lowercase , unk_token=__lowercase , bos_token=__lowercase , eos_token=__lowercase , pad_token=__lowercase , add_prefix_space=__lowercase , clean_up_tokenization_spaces=__lowercase , **__lowercase , ) __UpperCamelCase :int = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get('''add_prefix_space''' , __lowercase) != add_prefix_space: __UpperCamelCase :Any = getattr(__lowercase , pre_tok_state.pop('''type''')) __UpperCamelCase :str = add_prefix_space __UpperCamelCase :List[str] = pre_tok_class(**__lowercase) __UpperCamelCase :Tuple = add_prefix_space def UpperCamelCase__ ( self , *__lowercase , **__lowercase) -> BatchEncoding: __UpperCamelCase :Tuple = kwargs.get('''is_split_into_words''' , __lowercase) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with""" ''' pretokenized inputs.''') return super()._batch_encode_plus(*__lowercase , **__lowercase) def UpperCamelCase__ ( self , *__lowercase , **__lowercase) -> BatchEncoding: __UpperCamelCase :List[str] = kwargs.get('''is_split_into_words''' , __lowercase) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with""" ''' pretokenized inputs.''') return super()._encode_plus(*__lowercase , **__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase = None) -> Tuple[str]: __UpperCamelCase :Optional[Any] = self._tokenizer.model.save(__lowercase , name=__lowercase) return tuple(__lowercase) def UpperCamelCase__ ( self , __lowercase) -> List[int]: __UpperCamelCase :str = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(__lowercase , add_special_tokens=__lowercase) + [self.eos_token_id]) if len(__lowercase) > self.model_max_length: __UpperCamelCase :Any = input_ids[-self.model_max_length :] return input_ids
43
1
def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Any = 1 for i in range(1 , num + 1 ): fact *= i return fact def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = 0 while number > 0: __UpperCamelCase :Optional[Any] = number % 10 sum_of_digits += last_digit __UpperCamelCase :str = number // 10 # Removing the last_digit from the given number return sum_of_digits def lowerCamelCase ( SCREAMING_SNAKE_CASE = 100 ): '''simple docstring''' __UpperCamelCase :Dict = factorial(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Dict = split_and_add(SCREAMING_SNAKE_CASE ) return result if __name__ == "__main__": print(solution(int(input('''Enter the Number: ''').strip())))
43
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = {'''ctrl''': '''https://huggingface.co/ctrl/resolve/main/config.json'''} class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : str = """ctrl""" a__ : Dict = ["""past_key_values"""] a__ : Tuple = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , __lowercase=246_534 , __lowercase=256 , __lowercase=1_280 , __lowercase=8_192 , __lowercase=48 , __lowercase=16 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=1E-6 , __lowercase=0.02 , __lowercase=True , **__lowercase , ) -> List[Any]: __UpperCamelCase :List[str] = vocab_size __UpperCamelCase :Optional[Any] = n_positions __UpperCamelCase :Dict = n_embd __UpperCamelCase :Dict = n_layer __UpperCamelCase :List[Any] = n_head __UpperCamelCase :int = dff __UpperCamelCase :Union[str, Any] = resid_pdrop __UpperCamelCase :Optional[int] = embd_pdrop __UpperCamelCase :List[Any] = layer_norm_epsilon __UpperCamelCase :Dict = initializer_range __UpperCamelCase :Any = use_cache super().__init__(**__lowercase)
43
1
def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' if not isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): raise TypeError('''Input value must be an \'int\' type''' ) __UpperCamelCase :List[str] = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
43
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class lowerCamelCase_ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : str = TextToVideoSDPipeline a__ : Union[str, Any] = TEXT_TO_IMAGE_PARAMS a__ : Tuple = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. a__ : int = frozenset( [ """num_inference_steps""", """generator""", """latents""", """return_dict""", """callback""", """callback_steps""", ] ) def UpperCamelCase__ ( self) -> Optional[Any]: torch.manual_seed(0) __UpperCamelCase :str = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''DownBlock3D''') , up_block_types=('''UpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''') , cross_attention_dim=32 , attention_head_dim=4 , ) __UpperCamelCase :Optional[int] = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=__lowercase , set_alpha_to_one=__lowercase , ) torch.manual_seed(0) __UpperCamelCase :Optional[int] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0) __UpperCamelCase :Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act='''gelu''' , projection_dim=512 , ) __UpperCamelCase :Optional[Any] = CLIPTextModel(__lowercase) __UpperCamelCase :Optional[int] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') __UpperCamelCase :Union[str, Any] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0) -> Optional[int]: if str(__lowercase).startswith('''mps'''): __UpperCamelCase :List[Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :Tuple = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :int = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Optional[int] = self.get_dummy_components() __UpperCamelCase :Dict = TextToVideoSDPipeline(**__lowercase) __UpperCamelCase :Any = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :Optional[Any] = self.get_dummy_inputs(__lowercase) __UpperCamelCase :int = '''np''' __UpperCamelCase :List[str] = sd_pipe(**__lowercase).frames __UpperCamelCase :Optional[Any] = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) __UpperCamelCase :str = np.array([1_58.0, 1_60.0, 1_53.0, 1_25.0, 1_00.0, 1_21.0, 1_11.0, 93.0, 1_13.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase__ ( self) -> Tuple: self._test_attention_slicing_forward_pass(test_mean_pixel_difference=__lowercase , expected_max_diff=3E-3) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase__ ( self) -> Optional[int]: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__lowercase , expected_max_diff=1E-2) @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCamelCase__ ( self) -> Union[str, Any]: pass @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCamelCase__ ( self) -> Dict: pass @unittest.skip(reason='''`num_images_per_prompt` argument is not supported for this pipeline.''') def UpperCamelCase__ ( self) -> str: pass def UpperCamelCase__ ( self) -> List[str]: return super().test_progress_bar() @slow @skip_mps class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Union[str, Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy''') __UpperCamelCase :List[str] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''') __UpperCamelCase :Union[str, Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) __UpperCamelCase :str = pipe.to('''cuda''') __UpperCamelCase :Optional[Any] = '''Spiderman is surfing''' __UpperCamelCase :Union[str, Any] = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :List[Any] = pipe(__lowercase , generator=__lowercase , num_inference_steps=25 , output_type='''pt''').frames __UpperCamelCase :Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video).mean() < 5E-2 def UpperCamelCase__ ( self) -> int: __UpperCamelCase :str = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy''') __UpperCamelCase :Union[str, Any] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''') __UpperCamelCase :str = pipe.to('''cuda''') __UpperCamelCase :Union[str, Any] = '''Spiderman is surfing''' __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :List[Any] = pipe(__lowercase , generator=__lowercase , num_inference_steps=2 , output_type='''pt''').frames __UpperCamelCase :Optional[Any] = video_frames.cpu().numpy() assert np.abs(expected_video - video).mean() < 5E-2
43
1
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger __lowercase = get_logger(__name__) class lowerCamelCase_ ( enum.Enum ): '''simple docstring''' a__ : Union[str, Any] = """all_checks""" a__ : List[Any] = """basic_checks""" a__ : List[Any] = """no_checks""" class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None ): '''simple docstring''' if expected_checksums is None: logger.info('''Unable to verify checksums.''' ) return if len(set(SCREAMING_SNAKE_CASE ) - set(SCREAMING_SNAKE_CASE ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(SCREAMING_SNAKE_CASE ) - set(SCREAMING_SNAKE_CASE ) ) ) if len(set(SCREAMING_SNAKE_CASE ) - set(SCREAMING_SNAKE_CASE ) ) > 0: raise UnexpectedDownloadedFile(str(set(SCREAMING_SNAKE_CASE ) - set(SCREAMING_SNAKE_CASE ) ) ) __UpperCamelCase :int = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] __UpperCamelCase :Any = ''' for ''' + verification_name if verification_name is not None else '''''' if len(SCREAMING_SNAKE_CASE ) > 0: raise NonMatchingChecksumError( f"""Checksums didn't match{for_verification_name}:\n""" f"""{bad_urls}\n""" '''Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error''' ) logger.info('''All the checksums matched successfully''' + for_verification_name ) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' if expected_splits is None: logger.info('''Unable to verify splits sizes.''' ) return if len(set(SCREAMING_SNAKE_CASE ) - set(SCREAMING_SNAKE_CASE ) ) > 0: raise ExpectedMoreSplits(str(set(SCREAMING_SNAKE_CASE ) - set(SCREAMING_SNAKE_CASE ) ) ) if len(set(SCREAMING_SNAKE_CASE ) - set(SCREAMING_SNAKE_CASE ) ) > 0: raise UnexpectedSplits(str(set(SCREAMING_SNAKE_CASE ) - set(SCREAMING_SNAKE_CASE ) ) ) __UpperCamelCase :Union[str, Any] = [ {'''expected''': expected_splits[name], '''recorded''': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(SCREAMING_SNAKE_CASE ) > 0: raise NonMatchingSplitsSizesError(str(SCREAMING_SNAKE_CASE ) ) logger.info('''All the splits matched successfully.''' ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = True ): '''simple docstring''' if record_checksum: __UpperCamelCase :List[Any] = shaaaa() with open(SCREAMING_SNAKE_CASE , '''rb''' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , B'''''' ): m.update(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = m.hexdigest() else: __UpperCamelCase :Optional[int] = None return {"num_bytes": os.path.getsize(SCREAMING_SNAKE_CASE ), "checksum": checksum} def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
43
def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = [0 for i in range(len(SCREAMING_SNAKE_CASE ) )] # initialize interval's left pointer and right pointer __UpperCamelCase , __UpperCamelCase :str = 0, 0 for i in range(1 , len(SCREAMING_SNAKE_CASE ) ): # case when current index is inside the interval if i <= right_pointer: __UpperCamelCase :Union[str, Any] = min(right_pointer - i + 1 , z_result[i - left_pointer] ) __UpperCamelCase :Tuple = min_edge while go_next(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: __UpperCamelCase , __UpperCamelCase :Union[str, Any] = i, i + z_result[i] - 1 return z_result def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' return i + z_result[i] < len(SCREAMING_SNAKE_CASE ) and s[z_result[i]] == s[i + z_result[i]] def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string __UpperCamelCase :Tuple = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(SCREAMING_SNAKE_CASE ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
43
1
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import datasets import numpy as np import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, EvalPrediction, HfArgumentParser, PreTrainedTokenizer, TFAutoModelForSequenceClassification, TFTrainer, TFTrainingArguments, ) from transformers.utils import logging as hf_logging hf_logging.set_verbosity_info() hf_logging.enable_default_handler() hf_logging.enable_explicit_format() def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , ): '''simple docstring''' __UpperCamelCase :List[str] = {} if train_file is not None: __UpperCamelCase :Optional[Any] = [train_file] if eval_file is not None: __UpperCamelCase :Any = [eval_file] if test_file is not None: __UpperCamelCase :List[str] = [test_file] __UpperCamelCase :int = datasets.load_dataset('''csv''' , data_files=SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[Any] = list(ds[list(files.keys() )[0]].features.keys() ) __UpperCamelCase :int = features_name.pop(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Union[str, Any] = list(set(ds[list(files.keys() )[0]][label_name] ) ) __UpperCamelCase :List[Any] = {label: i for i, label in enumerate(SCREAMING_SNAKE_CASE )} __UpperCamelCase :str = tokenizer.model_input_names __UpperCamelCase :Optional[int] = {} if len(SCREAMING_SNAKE_CASE ) == 1: for k in files.keys(): __UpperCamelCase :str = ds[k].map( lambda SCREAMING_SNAKE_CASE : tokenizer.batch_encode_plus( example[features_name[0]] , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE , padding='''max_length''' ) , batched=SCREAMING_SNAKE_CASE , ) elif len(SCREAMING_SNAKE_CASE ) == 2: for k in files.keys(): __UpperCamelCase :int = ds[k].map( lambda SCREAMING_SNAKE_CASE : tokenizer.batch_encode_plus( (example[features_name[0]], example[features_name[1]]) , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE , padding='''max_length''' , ) , batched=SCREAMING_SNAKE_CASE , ) def gen_train(): for ex in transformed_ds[datasets.Split.TRAIN]: __UpperCamelCase :List[Any] = {k: v for k, v in ex.items() if k in input_names} __UpperCamelCase :Tuple = labelaid[ex[label_name]] yield (d, label) def gen_val(): for ex in transformed_ds[datasets.Split.VALIDATION]: __UpperCamelCase :List[str] = {k: v for k, v in ex.items() if k in input_names} __UpperCamelCase :Dict = labelaid[ex[label_name]] yield (d, label) def gen_test(): for ex in transformed_ds[datasets.Split.TEST]: __UpperCamelCase :Dict = {k: v for k, v in ex.items() if k in input_names} __UpperCamelCase :Optional[int] = labelaid[ex[label_name]] yield (d, label) __UpperCamelCase :Optional[int] = ( tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TRAIN in transformed_ds else None ) if train_ds is not None: __UpperCamelCase :List[Any] = train_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TRAIN] ) ) ) __UpperCamelCase :str = ( tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.VALIDATION in transformed_ds else None ) if val_ds is not None: __UpperCamelCase :List[Any] = val_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.VALIDATION] ) ) ) __UpperCamelCase :int = ( tf.data.Dataset.from_generator( SCREAMING_SNAKE_CASE , ({k: tf.intaa for k in input_names}, tf.intaa) , ({k: tf.TensorShape([None] ) for k in input_names}, tf.TensorShape([] )) , ) if datasets.Split.TEST in transformed_ds else None ) if test_ds is not None: __UpperCamelCase :Tuple = test_ds.apply(tf.data.experimental.assert_cardinality(len(ds[datasets.Split.TEST] ) ) ) return train_ds, val_ds, test_ds, labelaid __lowercase = logging.getLogger(__name__) @dataclass class lowerCamelCase_ : '''simple docstring''' a__ : int = field(metadata={"""help""": """Which column contains the label"""} ) a__ : str = field(default=UpperCAmelCase_ , metadata={"""help""": """The path of the training file"""} ) a__ : Optional[str] = field(default=UpperCAmelCase_ , metadata={"""help""": """The path of the development file"""} ) a__ : Optional[str] = field(default=UpperCAmelCase_ , metadata={"""help""": """The path of the test file"""} ) a__ : int = field( default=1_2_8 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) a__ : bool = field( default=UpperCAmelCase_ , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) @dataclass class lowerCamelCase_ : '''simple docstring''' a__ : str = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) a__ : Optional[str] = field( default=UpperCAmelCase_ , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) a__ : Optional[str] = field( default=UpperCAmelCase_ , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) a__ : bool = field(default=UpperCAmelCase_ , metadata={"""help""": """Set this flag to use fast tokenization."""} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. a__ : Optional[str] = field( default=UpperCAmelCase_ , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :List[Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TFTrainingArguments) ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :str = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ''' --overwrite_output_dir to overcome.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , level=logging.INFO , ) logger.info( f"""n_replicas: {training_args.n_replicas}, distributed training: {bool(training_args.n_replicas > 1 )}, """ f"""16-bits training: {training_args.fpaa}""" ) logger.info(f"""Training/evaluation parameters {training_args}""" ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. __UpperCamelCase :List[Any] = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Optional[int] = get_tfds( train_file=data_args.train_file , eval_file=data_args.dev_file , test_file=data_args.test_file , tokenizer=SCREAMING_SNAKE_CASE , label_column_id=data_args.label_column_id , max_seq_length=data_args.max_seq_length , ) __UpperCamelCase :Dict = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=len(SCREAMING_SNAKE_CASE ) , labelaid=SCREAMING_SNAKE_CASE , idalabel={id: label for label, id in labelaid.items()} , finetuning_task='''text-classification''' , cache_dir=model_args.cache_dir , ) with training_args.strategy.scope(): __UpperCamelCase :str = TFAutoModelForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_pt=bool('''.bin''' in model_args.model_name_or_path ) , config=SCREAMING_SNAKE_CASE , cache_dir=model_args.cache_dir , ) def compute_metrics(SCREAMING_SNAKE_CASE ) -> Dict: __UpperCamelCase :Dict = np.argmax(p.predictions , axis=1 ) return {"acc": (preds == p.label_ids).mean()} # Initialize our Trainer __UpperCamelCase :Optional[Any] = TFTrainer( model=SCREAMING_SNAKE_CASE , args=SCREAMING_SNAKE_CASE , train_dataset=SCREAMING_SNAKE_CASE , eval_dataset=SCREAMING_SNAKE_CASE , compute_metrics=SCREAMING_SNAKE_CASE , ) # Training if training_args.do_train: trainer.train() trainer.save_model() tokenizer.save_pretrained(training_args.output_dir ) # Evaluation __UpperCamelCase :int = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) __UpperCamelCase :List[str] = trainer.evaluate() __UpperCamelCase :Optional[Any] = os.path.join(training_args.output_dir , '''eval_results.txt''' ) with open(SCREAMING_SNAKE_CASE , '''w''' ) as writer: logger.info('''***** Eval results *****''' ) for key, value in result.items(): logger.info(f""" {key} = {value}""" ) writer.write(f"""{key} = {value}\n""" ) results.update(SCREAMING_SNAKE_CASE ) return results if __name__ == "__main__": main()
43
import math from typing import Any, Callable, List, Optional, Tuple, Union import numpy as np import torch from ...models import TaFilmDecoder from ...schedulers import DDPMScheduler from ...utils import is_onnx_available, logging, randn_tensor if is_onnx_available(): from ..onnx_utils import OnnxRuntimeModel from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline from .continous_encoder import SpectrogramContEncoder from .notes_encoder import SpectrogramNotesEncoder __lowercase = logging.get_logger(__name__) # pylint: disable=invalid-name __lowercase = 256 class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Tuple = ["""melgan"""] def __init__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> None: super().__init__() # From MELGAN __UpperCamelCase :int = math.log(1E-5) # Matches MelGAN training. __UpperCamelCase :int = 4.0 # Largest value for most examples __UpperCamelCase :str = 128 self.register_modules( notes_encoder=__lowercase , continuous_encoder=__lowercase , decoder=__lowercase , scheduler=__lowercase , melgan=__lowercase , ) def UpperCamelCase__ ( self , __lowercase , __lowercase=(-1.0, 1.0) , __lowercase=False) -> Dict: __UpperCamelCase , __UpperCamelCase :str = output_range if clip: __UpperCamelCase :Union[str, Any] = torch.clip(__lowercase , self.min_value , self.max_value) # Scale to [0, 1]. __UpperCamelCase :Union[str, Any] = (features - self.min_value) / (self.max_value - self.min_value) # Scale to [min_out, max_out]. return zero_one * (max_out - min_out) + min_out def UpperCamelCase__ ( self , __lowercase , __lowercase=(-1.0, 1.0) , __lowercase=False) -> Optional[int]: __UpperCamelCase , __UpperCamelCase :int = input_range __UpperCamelCase :Optional[int] = torch.clip(__lowercase , __lowercase , __lowercase) if clip else outputs # Scale to [0, 1]. __UpperCamelCase :List[str] = (outputs - min_out) / (max_out - min_out) # Scale to [self.min_value, self.max_value]. return zero_one * (self.max_value - self.min_value) + self.min_value def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase) -> List[Any]: __UpperCamelCase :List[str] = input_tokens > 0 __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self.notes_encoder( encoder_input_tokens=__lowercase , encoder_inputs_mask=__lowercase) __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self.continuous_encoder( encoder_inputs=__lowercase , encoder_inputs_mask=__lowercase) return [(tokens_encoded, tokens_mask), (continuous_encoded, continuous_mask)] def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase) -> str: __UpperCamelCase :Optional[int] = noise_time if not torch.is_tensor(__lowercase): __UpperCamelCase :str = torch.tensor([timesteps] , dtype=torch.long , device=input_tokens.device) elif torch.is_tensor(__lowercase) and len(timesteps.shape) == 0: __UpperCamelCase :Dict = timesteps[None].to(input_tokens.device) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML __UpperCamelCase :List[str] = timesteps * torch.ones(input_tokens.shape[0] , dtype=timesteps.dtype , device=timesteps.device) __UpperCamelCase :Tuple = self.decoder( encodings_and_masks=__lowercase , decoder_input_tokens=__lowercase , decoder_noise_time=__lowercase) return logits @torch.no_grad() def __call__( self , __lowercase , __lowercase = None , __lowercase = 100 , __lowercase = True , __lowercase = "numpy" , __lowercase = None , __lowercase = 1 , ) -> Union[AudioPipelineOutput, Tuple]: if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__lowercase , __lowercase) or callback_steps <= 0) ): raise ValueError( f"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" f""" {type(__lowercase)}.""") __UpperCamelCase :Union[str, Any] = np.zeros([1, TARGET_FEATURE_LENGTH, self.n_dims] , dtype=np.floataa) __UpperCamelCase :Union[str, Any] = np.zeros([1, 0, self.n_dims] , np.floataa) __UpperCamelCase :Union[str, Any] = torch.ones((1, TARGET_FEATURE_LENGTH) , dtype=__lowercase , device=self.device) for i, encoder_input_tokens in enumerate(__lowercase): if i == 0: __UpperCamelCase :int = torch.from_numpy(pred_mel[:1].copy()).to( device=self.device , dtype=self.decoder.dtype) # The first chunk has no previous context. __UpperCamelCase :int = torch.zeros((1, TARGET_FEATURE_LENGTH) , dtype=__lowercase , device=self.device) else: # The full song pipeline does not feed in a context feature, so the mask # will be all 0s after the feature converter. Because we know we're # feeding in a full context chunk from the previous prediction, set it # to all 1s. __UpperCamelCase :Tuple = ones __UpperCamelCase :Optional[Any] = self.scale_features( __lowercase , output_range=[-1.0, 1.0] , clip=__lowercase) __UpperCamelCase :int = self.encode( input_tokens=torch.IntTensor([encoder_input_tokens]).to(device=self.device) , continuous_inputs=__lowercase , continuous_mask=__lowercase , ) # Sample encoder_continuous_inputs shaped gaussian noise to begin loop __UpperCamelCase :int = randn_tensor( shape=encoder_continuous_inputs.shape , generator=__lowercase , device=self.device , dtype=self.decoder.dtype , ) # set step values self.scheduler.set_timesteps(__lowercase) # Denoising diffusion loop for j, t in enumerate(self.progress_bar(self.scheduler.timesteps)): __UpperCamelCase :Optional[int] = self.decode( encodings_and_masks=__lowercase , input_tokens=__lowercase , noise_time=t / self.scheduler.config.num_train_timesteps , ) # Compute previous output: x_t -> x_t-1 __UpperCamelCase :int = self.scheduler.step(__lowercase , __lowercase , __lowercase , generator=__lowercase).prev_sample __UpperCamelCase :Tuple = self.scale_to_features(__lowercase , input_range=[-1.0, 1.0]) __UpperCamelCase :List[Any] = mel[:1] __UpperCamelCase :Optional[Any] = mel.cpu().float().numpy() __UpperCamelCase :Any = np.concatenate([full_pred_mel, pred_mel[:1]] , axis=1) # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__lowercase , __lowercase) logger.info('''Generated segment''' , __lowercase) if output_type == "numpy" and not is_onnx_available(): raise ValueError( '''Cannot return output in \'np\' format if ONNX is not available. Make sure to have ONNX installed or set \'output_type\' to \'mel\'.''') elif output_type == "numpy" and self.melgan is None: raise ValueError( '''Cannot return output in \'np\' format if melgan component is not defined. Make sure to define `self.melgan` or set \'output_type\' to \'mel\'.''') if output_type == "numpy": __UpperCamelCase :Optional[Any] = self.melgan(input_features=full_pred_mel.astype(np.floataa)) else: __UpperCamelCase :List[str] = full_pred_mel if not return_dict: return (output,) return AudioPipelineOutput(audios=__lowercase)
43
1
import inspect import unittest from transformers import YolosConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import YolosForObjectDetection, YolosModel from transformers.models.yolos.modeling_yolos import YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowerCamelCase_ : '''simple docstring''' def __init__( self , __lowercase , __lowercase=13 , __lowercase=[30, 30] , __lowercase=2 , __lowercase=3 , __lowercase=True , __lowercase=True , __lowercase=32 , __lowercase=5 , __lowercase=4 , __lowercase=37 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=10 , __lowercase=0.02 , __lowercase=3 , __lowercase=None , __lowercase=8 , __lowercase=10 , ) -> List[str]: __UpperCamelCase :Optional[int] = parent __UpperCamelCase :List[str] = batch_size __UpperCamelCase :Tuple = image_size __UpperCamelCase :int = patch_size __UpperCamelCase :Union[str, Any] = num_channels __UpperCamelCase :Any = is_training __UpperCamelCase :Union[str, Any] = use_labels __UpperCamelCase :List[Any] = hidden_size __UpperCamelCase :Tuple = num_hidden_layers __UpperCamelCase :Tuple = num_attention_heads __UpperCamelCase :Union[str, Any] = intermediate_size __UpperCamelCase :Union[str, Any] = hidden_act __UpperCamelCase :int = hidden_dropout_prob __UpperCamelCase :Dict = attention_probs_dropout_prob __UpperCamelCase :Optional[int] = type_sequence_label_size __UpperCamelCase :List[str] = initializer_range __UpperCamelCase :Dict = num_labels __UpperCamelCase :Tuple = scope __UpperCamelCase :Optional[int] = n_targets __UpperCamelCase :Any = num_detection_tokens # we set the expected sequence length (which is used in several tests) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) + num_detection_tokens __UpperCamelCase :str = (image_size[1] // patch_size) * (image_size[0] // patch_size) __UpperCamelCase :Dict = num_patches + 1 + self.num_detection_tokens def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size[0], self.image_size[1]]) __UpperCamelCase :Tuple = None if self.use_labels: # labels is a list of Dict (each Dict being the labels for a given example in the batch) __UpperCamelCase :List[Any] = [] for i in range(self.batch_size): __UpperCamelCase :Optional[Any] = {} __UpperCamelCase :str = torch.randint( high=self.num_labels , size=(self.n_targets,) , device=__lowercase) __UpperCamelCase :Optional[Any] = torch.rand(self.n_targets , 4 , device=__lowercase) labels.append(__lowercase) __UpperCamelCase :Union[str, Any] = self.get_config() return config, pixel_values, labels def UpperCamelCase__ ( self) -> List[Any]: return YolosConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowercase , initializer_range=self.initializer_range , num_detection_tokens=self.num_detection_tokens , num_labels=self.num_labels , ) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase) -> Optional[Any]: __UpperCamelCase :Optional[Any] = YolosModel(config=__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :Union[str, Any] = model(__lowercase) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.expected_seq_len, self.hidden_size)) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase) -> int: __UpperCamelCase :str = YolosForObjectDetection(__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :str = model(pixel_values=__lowercase) __UpperCamelCase :Optional[int] = model(__lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1)) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4)) __UpperCamelCase :Dict = model(pixel_values=__lowercase , labels=__lowercase) self.parent.assertEqual(result.loss.shape , ()) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1)) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4)) def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :int = self.prepare_config_and_inputs() __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Union[str, Any] = config_and_inputs __UpperCamelCase :Union[str, Any] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : Optional[int] = (YolosModel, YolosForObjectDetection) if is_torch_available() else () a__ : Tuple = ( {"""feature-extraction""": YolosModel, """object-detection""": YolosForObjectDetection} if is_torch_available() else {} ) a__ : Optional[Any] = False a__ : Optional[Any] = False a__ : Union[str, Any] = False a__ : List[str] = False def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase=False) -> Tuple: __UpperCamelCase :Optional[Any] = super()._prepare_for_class(__lowercase , __lowercase , return_labels=__lowercase) if return_labels: if model_class.__name__ == "YolosForObjectDetection": __UpperCamelCase :Dict = [] for i in range(self.model_tester.batch_size): __UpperCamelCase :Union[str, Any] = {} __UpperCamelCase :Tuple = torch.ones( size=(self.model_tester.n_targets,) , device=__lowercase , dtype=torch.long) __UpperCamelCase :str = torch.ones( self.model_tester.n_targets , 4 , device=__lowercase , dtype=torch.float) labels.append(__lowercase) __UpperCamelCase :Union[str, Any] = labels return inputs_dict def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :Any = YolosModelTester(self) __UpperCamelCase :List[Any] = ConfigTester(self , config_class=__lowercase , has_text_modality=__lowercase , hidden_size=37) def UpperCamelCase__ ( self) -> Tuple: self.config_tester.run_common_tests() def UpperCamelCase__ ( self) -> List[str]: # YOLOS does not use inputs_embeds pass def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase , __UpperCamelCase :Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCamelCase :Tuple = model_class(__lowercase) self.assertIsInstance(model.get_input_embeddings() , (nn.Module)) __UpperCamelCase :Tuple = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowercase , nn.Linear)) def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase , __UpperCamelCase :List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCamelCase :str = model_class(__lowercase) __UpperCamelCase :Dict = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCamelCase :int = [*signature.parameters.keys()] __UpperCamelCase :Tuple = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __lowercase) def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase) def UpperCamelCase__ ( self) -> str: __UpperCamelCase , __UpperCamelCase :int = self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase :str = True # in YOLOS, the seq_len is different __UpperCamelCase :List[str] = self.model_tester.expected_seq_len for model_class in self.all_model_classes: __UpperCamelCase :Dict = True __UpperCamelCase :List[Any] = False __UpperCamelCase :Optional[Any] = True __UpperCamelCase :List[Any] = model_class(__lowercase) model.to(__lowercase) model.eval() with torch.no_grad(): __UpperCamelCase :Optional[Any] = model(**self._prepare_for_class(__lowercase , __lowercase)) __UpperCamelCase :List[str] = outputs.attentions self.assertEqual(len(__lowercase) , self.model_tester.num_hidden_layers) # check that output_attentions also work using config del inputs_dict["output_attentions"] __UpperCamelCase :Optional[Any] = True __UpperCamelCase :List[str] = model_class(__lowercase) model.to(__lowercase) model.eval() with torch.no_grad(): __UpperCamelCase :Optional[Any] = model(**self._prepare_for_class(__lowercase , __lowercase)) __UpperCamelCase :List[str] = outputs.attentions self.assertEqual(len(__lowercase) , self.model_tester.num_hidden_layers) self.assertListEqual( list(attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) __UpperCamelCase :int = len(__lowercase) # Check attention is always last and order is fine __UpperCamelCase :Any = True __UpperCamelCase :List[Any] = True __UpperCamelCase :Dict = model_class(__lowercase) model.to(__lowercase) model.eval() with torch.no_grad(): __UpperCamelCase :int = model(**self._prepare_for_class(__lowercase , __lowercase)) __UpperCamelCase :List[Any] = 1 self.assertEqual(out_len + added_hidden_states , len(__lowercase)) __UpperCamelCase :Union[str, Any] = outputs.attentions self.assertEqual(len(__lowercase) , self.model_tester.num_hidden_layers) self.assertListEqual( list(self_attentions[0].shape[-3:]) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) def UpperCamelCase__ ( self) -> Optional[Any]: def check_hidden_states_output(__lowercase , __lowercase , __lowercase): __UpperCamelCase :List[Any] = model_class(__lowercase) model.to(__lowercase) model.eval() with torch.no_grad(): __UpperCamelCase :Optional[Any] = model(**self._prepare_for_class(__lowercase , __lowercase)) __UpperCamelCase :Any = outputs.hidden_states __UpperCamelCase :Any = getattr( self.model_tester , '''expected_num_hidden_layers''' , self.model_tester.num_hidden_layers + 1) self.assertEqual(len(__lowercase) , __lowercase) # YOLOS has a different seq_length __UpperCamelCase :List[str] = self.model_tester.expected_seq_len self.assertListEqual( list(hidden_states[0].shape[-2:]) , [seq_length, self.model_tester.hidden_size] , ) __UpperCamelCase , __UpperCamelCase :Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCamelCase :Tuple = True check_hidden_states_output(__lowercase , __lowercase , __lowercase) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCamelCase :Any = True check_hidden_states_output(__lowercase , __lowercase , __lowercase) def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_object_detection(*__lowercase) @slow def UpperCamelCase__ ( self) -> Any: for model_name in YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase :Any = YolosModel.from_pretrained(__lowercase) self.assertIsNotNone(__lowercase) def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase__ ( self) -> List[str]: return AutoImageProcessor.from_pretrained('''hustvl/yolos-small''') if is_vision_available() else None @slow def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :List[str] = YolosForObjectDetection.from_pretrained('''hustvl/yolos-small''').to(__lowercase) __UpperCamelCase :Dict = self.default_image_processor __UpperCamelCase :Optional[int] = prepare_img() __UpperCamelCase :Any = image_processor(images=__lowercase , return_tensors='''pt''').to(__lowercase) # forward pass with torch.no_grad(): __UpperCamelCase :Optional[int] = model(inputs.pixel_values) # verify outputs __UpperCamelCase :str = torch.Size((1, 100, 92)) self.assertEqual(outputs.logits.shape , __lowercase) __UpperCamelCase :int = torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]] , device=__lowercase , ) __UpperCamelCase :Any = torch.tensor( [[0.25_59, 0.54_55, 0.47_06], [0.29_89, 0.72_79, 0.18_75], [0.77_32, 0.40_17, 0.44_62]] , device=__lowercase) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , __lowercase , atol=1E-4)) self.assertTrue(torch.allclose(outputs.pred_boxes[0, :3, :3] , __lowercase , atol=1E-4)) # verify postprocessing __UpperCamelCase :Tuple = image_processor.post_process_object_detection( __lowercase , threshold=0.3 , target_sizes=[image.size[::-1]])[0] __UpperCamelCase :Optional[int] = torch.tensor([0.99_94, 0.97_90, 0.99_64, 0.99_72, 0.98_61]).to(__lowercase) __UpperCamelCase :Optional[int] = [75, 75, 17, 63, 17] __UpperCamelCase :int = torch.tensor([3_35.06_09, 79.38_48, 3_75.42_16, 1_87.24_95]).to(__lowercase) self.assertEqual(len(results['''scores''']) , 5) self.assertTrue(torch.allclose(results['''scores'''] , __lowercase , atol=1E-4)) self.assertSequenceEqual(results['''labels'''].tolist() , __lowercase) self.assertTrue(torch.allclose(results['''boxes'''][0, :] , __lowercase))
43
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( HubertConfig, HubertForCTC, HubertModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __lowercase = logging.get_logger(__name__) __lowercase = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' for attribute in key.split('''.''' ): __UpperCamelCase :str = getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if weight_type is not None: __UpperCamelCase :Any = getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).shape else: __UpperCamelCase :Union[str, Any] = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __UpperCamelCase :str = value elif weight_type == "weight_g": __UpperCamelCase :List[str] = value elif weight_type == "weight_v": __UpperCamelCase :str = value elif weight_type == "bias": __UpperCamelCase :Union[str, Any] = value else: __UpperCamelCase :str = value logger.info(f"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = [] __UpperCamelCase :int = fairseq_model.state_dict() __UpperCamelCase :List[Any] = hf_model.hubert.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): __UpperCamelCase :List[Any] = False if "conv_layers" in name: load_conv_layer( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , hf_model.config.feat_extract_norm == '''group''' , ) __UpperCamelCase :List[str] = True else: for key, mapped_key in MAPPING.items(): __UpperCamelCase :Dict = '''hubert.''' + mapped_key if (is_finetuned and mapped_key != '''lm_head''') else mapped_key if key in name or (key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0] and not is_finetuned): __UpperCamelCase :Optional[Any] = True if "*" in mapped_key: __UpperCamelCase :List[str] = name.split(SCREAMING_SNAKE_CASE )[0].split('''.''' )[-2] __UpperCamelCase :Optional[int] = mapped_key.replace('''*''' , SCREAMING_SNAKE_CASE ) if "weight_g" in name: __UpperCamelCase :int = '''weight_g''' elif "weight_v" in name: __UpperCamelCase :List[Any] = '''weight_v''' elif "weight" in name: __UpperCamelCase :Dict = '''weight''' elif "bias" in name: __UpperCamelCase :Dict = '''bias''' else: __UpperCamelCase :Dict = None set_recursively(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(SCREAMING_SNAKE_CASE ) logger.warning(f"""Unused weights: {unused_weights}""" ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = full_name.split('''conv_layers.''' )[-1] __UpperCamelCase :Optional[int] = name.split('''.''' ) __UpperCamelCase :str = int(items[0] ) __UpperCamelCase :List[Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __UpperCamelCase :Dict = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __UpperCamelCase :Any = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) __UpperCamelCase :int = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) __UpperCamelCase :Union[str, Any] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(SCREAMING_SNAKE_CASE ) @torch.no_grad() def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=True ): '''simple docstring''' if config_path is not None: __UpperCamelCase :Tuple = HubertConfig.from_pretrained(SCREAMING_SNAKE_CASE ) else: __UpperCamelCase :Optional[int] = HubertConfig() if is_finetuned: if dict_path: __UpperCamelCase :Optional[int] = Dictionary.load(SCREAMING_SNAKE_CASE ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __UpperCamelCase :Optional[int] = target_dict.pad_index __UpperCamelCase :Dict = target_dict.bos_index __UpperCamelCase :str = target_dict.eos_index __UpperCamelCase :Dict = len(target_dict.symbols ) __UpperCamelCase :List[Any] = os.path.join(SCREAMING_SNAKE_CASE , '''vocab.json''' ) if not os.path.isdir(SCREAMING_SNAKE_CASE ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(SCREAMING_SNAKE_CASE ) ) return os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) with open(SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(target_dict.indices , SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = WavaVecaCTCTokenizer( SCREAMING_SNAKE_CASE , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :Union[str, Any] = True if config.feat_extract_norm == '''layer''' else False __UpperCamelCase :Any = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0 , do_normalize=SCREAMING_SNAKE_CASE , return_attention_mask=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :Any = WavaVecaProcessor(feature_extractor=SCREAMING_SNAKE_CASE , tokenizer=SCREAMING_SNAKE_CASE ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[str] = HubertForCTC(SCREAMING_SNAKE_CASE ) else: __UpperCamelCase :str = HubertModel(SCREAMING_SNAKE_CASE ) if is_finetuned: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :int = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Any = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) __UpperCamelCase :Dict = model[0].eval() recursively_load_weights(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) hf_wavavec.save_pretrained(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) __lowercase = parser.parse_args() convert_hubert_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
43
1
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class lowerCamelCase_ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : str = TextToVideoSDPipeline a__ : Union[str, Any] = TEXT_TO_IMAGE_PARAMS a__ : Tuple = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. a__ : int = frozenset( [ """num_inference_steps""", """generator""", """latents""", """return_dict""", """callback""", """callback_steps""", ] ) def UpperCamelCase__ ( self) -> Optional[Any]: torch.manual_seed(0) __UpperCamelCase :str = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''DownBlock3D''') , up_block_types=('''UpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''') , cross_attention_dim=32 , attention_head_dim=4 , ) __UpperCamelCase :Optional[int] = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=__lowercase , set_alpha_to_one=__lowercase , ) torch.manual_seed(0) __UpperCamelCase :Optional[int] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0) __UpperCamelCase :Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act='''gelu''' , projection_dim=512 , ) __UpperCamelCase :Optional[Any] = CLIPTextModel(__lowercase) __UpperCamelCase :Optional[int] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') __UpperCamelCase :Union[str, Any] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0) -> Optional[int]: if str(__lowercase).startswith('''mps'''): __UpperCamelCase :List[Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :Tuple = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :int = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Optional[int] = self.get_dummy_components() __UpperCamelCase :Dict = TextToVideoSDPipeline(**__lowercase) __UpperCamelCase :Any = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :Optional[Any] = self.get_dummy_inputs(__lowercase) __UpperCamelCase :int = '''np''' __UpperCamelCase :List[str] = sd_pipe(**__lowercase).frames __UpperCamelCase :Optional[Any] = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) __UpperCamelCase :str = np.array([1_58.0, 1_60.0, 1_53.0, 1_25.0, 1_00.0, 1_21.0, 1_11.0, 93.0, 1_13.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase__ ( self) -> Tuple: self._test_attention_slicing_forward_pass(test_mean_pixel_difference=__lowercase , expected_max_diff=3E-3) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase__ ( self) -> Optional[int]: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__lowercase , expected_max_diff=1E-2) @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCamelCase__ ( self) -> Union[str, Any]: pass @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCamelCase__ ( self) -> Dict: pass @unittest.skip(reason='''`num_images_per_prompt` argument is not supported for this pipeline.''') def UpperCamelCase__ ( self) -> str: pass def UpperCamelCase__ ( self) -> List[str]: return super().test_progress_bar() @slow @skip_mps class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Union[str, Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy''') __UpperCamelCase :List[str] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''') __UpperCamelCase :Union[str, Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) __UpperCamelCase :str = pipe.to('''cuda''') __UpperCamelCase :Optional[Any] = '''Spiderman is surfing''' __UpperCamelCase :Union[str, Any] = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :List[Any] = pipe(__lowercase , generator=__lowercase , num_inference_steps=25 , output_type='''pt''').frames __UpperCamelCase :Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video).mean() < 5E-2 def UpperCamelCase__ ( self) -> int: __UpperCamelCase :str = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy''') __UpperCamelCase :Union[str, Any] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''') __UpperCamelCase :str = pipe.to('''cuda''') __UpperCamelCase :Union[str, Any] = '''Spiderman is surfing''' __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :List[Any] = pipe(__lowercase , generator=__lowercase , num_inference_steps=2 , output_type='''pt''').frames __UpperCamelCase :Optional[Any] = video_frames.cpu().numpy() assert np.abs(expected_video - video).mean() < 5E-2
43
import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters __lowercase = (720, 1280) # Height, Width __lowercase = (0.4, 0.6) # if height or width lower than this scale, drop it. __lowercase = 1 / 100 __lowercase = '''''' __lowercase = '''''' __lowercase = '''''' __lowercase = 250 def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase , __UpperCamelCase :List[Any] = get_dataset(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for index in range(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = random.sample(range(len(SCREAMING_SNAKE_CASE ) ) , 4 ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :str = update_image_and_anno( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , filter_scale=SCREAMING_SNAKE_CASE , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __UpperCamelCase :List[Any] = random_chars(32 ) __UpperCamelCase :List[str] = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __UpperCamelCase :Tuple = f"""{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}""" cva.imwrite(f"""{file_root}.jpg""" , SCREAMING_SNAKE_CASE , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(f"""Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}""" ) __UpperCamelCase :Optional[Any] = [] for anno in new_annos: __UpperCamelCase :int = anno[3] - anno[1] __UpperCamelCase :Optional[int] = anno[4] - anno[2] __UpperCamelCase :int = anno[1] + width / 2 __UpperCamelCase :List[str] = anno[2] + height / 2 __UpperCamelCase :str = f"""{anno[0]} {x_center} {y_center} {width} {height}""" annos_list.append(SCREAMING_SNAKE_CASE ) with open(f"""{file_root}.txt""" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :str = [] __UpperCamelCase :str = [] for label_file in glob.glob(os.path.join(SCREAMING_SNAKE_CASE , '''*.txt''' ) ): __UpperCamelCase :Any = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(SCREAMING_SNAKE_CASE ) as in_file: __UpperCamelCase :str = in_file.readlines() __UpperCamelCase :Optional[int] = os.path.join(SCREAMING_SNAKE_CASE , f"""{label_name}.jpg""" ) __UpperCamelCase :int = [] for obj_list in obj_lists: __UpperCamelCase :Optional[int] = obj_list.rstrip('''\n''' ).split(''' ''' ) __UpperCamelCase :Any = float(obj[1] ) - float(obj[3] ) / 2 __UpperCamelCase :List[str] = float(obj[2] ) - float(obj[4] ) / 2 __UpperCamelCase :Dict = float(obj[1] ) + float(obj[3] ) / 2 __UpperCamelCase :List[str] = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(SCREAMING_SNAKE_CASE ) labels.append(SCREAMING_SNAKE_CASE ) return img_paths, labels def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 0.0 , ): '''simple docstring''' __UpperCamelCase :List[str] = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __UpperCamelCase :List[Any] = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __UpperCamelCase :int = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __UpperCamelCase :Optional[int] = int(scale_x * output_size[1] ) __UpperCamelCase :Any = int(scale_y * output_size[0] ) __UpperCamelCase :List[str] = [] __UpperCamelCase :Dict = [] for i, index in enumerate(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Any = all_img_list[index] path_list.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = all_annos[index] __UpperCamelCase :Union[str, Any] = cva.imread(SCREAMING_SNAKE_CASE ) if i == 0: # top-left __UpperCamelCase :str = cva.resize(SCREAMING_SNAKE_CASE , (divid_point_x, divid_point_y) ) __UpperCamelCase :Union[str, Any] = img for bbox in img_annos: __UpperCamelCase :Union[str, Any] = bbox[1] * scale_x __UpperCamelCase :Optional[Any] = bbox[2] * scale_y __UpperCamelCase :int = bbox[3] * scale_x __UpperCamelCase :Union[str, Any] = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __UpperCamelCase :str = cva.resize(SCREAMING_SNAKE_CASE , (output_size[1] - divid_point_x, divid_point_y) ) __UpperCamelCase :List[str] = img for bbox in img_annos: __UpperCamelCase :str = scale_x + bbox[1] * (1 - scale_x) __UpperCamelCase :Dict = bbox[2] * scale_y __UpperCamelCase :Optional[Any] = scale_x + bbox[3] * (1 - scale_x) __UpperCamelCase :List[Any] = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __UpperCamelCase :str = cva.resize(SCREAMING_SNAKE_CASE , (divid_point_x, output_size[0] - divid_point_y) ) __UpperCamelCase :Optional[int] = img for bbox in img_annos: __UpperCamelCase :Tuple = bbox[1] * scale_x __UpperCamelCase :Optional[Any] = scale_y + bbox[2] * (1 - scale_y) __UpperCamelCase :Tuple = bbox[3] * scale_x __UpperCamelCase :Dict = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __UpperCamelCase :Optional[int] = cva.resize( SCREAMING_SNAKE_CASE , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __UpperCamelCase :Optional[int] = img for bbox in img_annos: __UpperCamelCase :Optional[Any] = scale_x + bbox[1] * (1 - scale_x) __UpperCamelCase :Optional[int] = scale_y + bbox[2] * (1 - scale_y) __UpperCamelCase :Optional[Any] = scale_x + bbox[3] * (1 - scale_x) __UpperCamelCase :int = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __UpperCamelCase :List[Any] = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' assert number_char > 1, "The number of character should greater than 1" __UpperCamelCase :Optional[Any] = ascii_lowercase + digits return "".join(random.choice(SCREAMING_SNAKE_CASE ) for _ in range(SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": main() print('''DONE ✅''')
43
1
import random def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = a[left_index] __UpperCamelCase :Any = left_index + 1 for j in range(left_index + 1 , SCREAMING_SNAKE_CASE ): if a[j] < pivot: __UpperCamelCase , __UpperCamelCase :str = a[i], a[j] i += 1 __UpperCamelCase , __UpperCamelCase :Optional[int] = a[i - 1], a[left_index] return i - 1 def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' if left < right: __UpperCamelCase :int = random.randint(SCREAMING_SNAKE_CASE , right - 1 ) __UpperCamelCase , __UpperCamelCase :List[str] = ( a[left], a[pivot], ) # switches the pivot with the left most bound __UpperCamelCase :Dict = partition(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) quick_sort_random( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # recursive quicksort to the left of the pivot point quick_sort_random( SCREAMING_SNAKE_CASE , pivot_index + 1 , SCREAMING_SNAKE_CASE ) # recursive quicksort to the right of the pivot point def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = input('''Enter numbers separated by a comma:\n''' ).strip() __UpperCamelCase :Union[str, Any] = [int(SCREAMING_SNAKE_CASE ) for item in user_input.split(''',''' )] quick_sort_random(SCREAMING_SNAKE_CASE , 0 , len(SCREAMING_SNAKE_CASE ) ) print(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
43
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = { '''facebook/wav2vec2-base-960h''': '''https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/config.json''', # See all Wav2Vec2 models at https://huggingface.co/models?filter=wav2vec2 } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Union[str, Any] = """wav2vec2""" def __init__( self , __lowercase=32 , __lowercase=768 , __lowercase=12 , __lowercase=12 , __lowercase=3_072 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.02 , __lowercase=1E-5 , __lowercase="group" , __lowercase="gelu" , __lowercase=(512, 512, 512, 512, 512, 512, 512) , __lowercase=(5, 2, 2, 2, 2, 2, 2) , __lowercase=(10, 3, 3, 3, 3, 2, 2) , __lowercase=False , __lowercase=128 , __lowercase=16 , __lowercase=False , __lowercase=True , __lowercase=0.05 , __lowercase=10 , __lowercase=2 , __lowercase=0.0 , __lowercase=10 , __lowercase=0 , __lowercase=320 , __lowercase=2 , __lowercase=0.1 , __lowercase=100 , __lowercase=256 , __lowercase=256 , __lowercase=0.1 , __lowercase="sum" , __lowercase=False , __lowercase=False , __lowercase=256 , __lowercase=(512, 512, 512, 512, 1_500) , __lowercase=(5, 3, 3, 1, 1) , __lowercase=(1, 2, 3, 1, 1) , __lowercase=512 , __lowercase=0 , __lowercase=1 , __lowercase=2 , __lowercase=False , __lowercase=3 , __lowercase=2 , __lowercase=3 , __lowercase=None , __lowercase=None , **__lowercase , ) -> int: super().__init__(**__lowercase , pad_token_id=__lowercase , bos_token_id=__lowercase , eos_token_id=__lowercase) __UpperCamelCase :Any = hidden_size __UpperCamelCase :int = feat_extract_norm __UpperCamelCase :Tuple = feat_extract_activation __UpperCamelCase :Union[str, Any] = list(__lowercase) __UpperCamelCase :List[Any] = list(__lowercase) __UpperCamelCase :int = list(__lowercase) __UpperCamelCase :List[Any] = conv_bias __UpperCamelCase :Optional[int] = num_conv_pos_embeddings __UpperCamelCase :Dict = num_conv_pos_embedding_groups __UpperCamelCase :Any = len(self.conv_dim) __UpperCamelCase :List[str] = num_hidden_layers __UpperCamelCase :int = intermediate_size __UpperCamelCase :str = hidden_act __UpperCamelCase :Any = num_attention_heads __UpperCamelCase :int = hidden_dropout __UpperCamelCase :Tuple = attention_dropout __UpperCamelCase :List[str] = activation_dropout __UpperCamelCase :Optional[Any] = feat_proj_dropout __UpperCamelCase :Any = final_dropout __UpperCamelCase :Any = layerdrop __UpperCamelCase :str = layer_norm_eps __UpperCamelCase :Optional[Any] = initializer_range __UpperCamelCase :List[str] = vocab_size __UpperCamelCase :str = do_stable_layer_norm __UpperCamelCase :Union[str, Any] = use_weighted_layer_sum if ( (len(self.conv_stride) != self.num_feat_extract_layers) or (len(self.conv_kernel) != self.num_feat_extract_layers) or (len(self.conv_dim) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' f""" {len(self.conv_dim)}`, `len(config.conv_stride) = {len(self.conv_stride)}`,""" f""" `len(config.conv_kernel) = {len(self.conv_kernel)}`.""") # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __UpperCamelCase :List[Any] = apply_spec_augment __UpperCamelCase :Tuple = mask_time_prob __UpperCamelCase :int = mask_time_length __UpperCamelCase :Dict = mask_time_min_masks __UpperCamelCase :str = mask_feature_prob __UpperCamelCase :List[str] = mask_feature_length __UpperCamelCase :Union[str, Any] = mask_feature_min_masks # parameters for pretraining with codevector quantized representations __UpperCamelCase :Optional[Any] = num_codevectors_per_group __UpperCamelCase :List[Any] = num_codevector_groups __UpperCamelCase :Tuple = contrastive_logits_temperature __UpperCamelCase :Optional[int] = feat_quantizer_dropout __UpperCamelCase :Optional[int] = num_negatives __UpperCamelCase :List[Any] = codevector_dim __UpperCamelCase :str = proj_codevector_dim __UpperCamelCase :List[str] = diversity_loss_weight # ctc loss __UpperCamelCase :Tuple = ctc_loss_reduction __UpperCamelCase :Tuple = ctc_zero_infinity # adapter __UpperCamelCase :List[str] = add_adapter __UpperCamelCase :Tuple = adapter_kernel_size __UpperCamelCase :str = adapter_stride __UpperCamelCase :Tuple = num_adapter_layers __UpperCamelCase :Tuple = output_hidden_size or hidden_size __UpperCamelCase :Optional[Any] = adapter_attn_dim # SequenceClassification-specific parameter. Feel free to ignore for other classes. __UpperCamelCase :Optional[Any] = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. __UpperCamelCase :Optional[int] = list(__lowercase) __UpperCamelCase :List[Any] = list(__lowercase) __UpperCamelCase :List[Any] = list(__lowercase) __UpperCamelCase :str = xvector_output_dim @property def UpperCamelCase__ ( self) -> List[str]: return functools.reduce(operator.mul , self.conv_stride , 1)
43
1
import tempfile import unittest from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from transformers.testing_utils import ( is_torch_available, require_optimum, require_torch, slow, ) if is_torch_available(): import torch @require_torch @require_optimum @slow class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> str: __UpperCamelCase :List[str] = '''hf-internal-testing/tiny-random-t5''' __UpperCamelCase :str = AutoTokenizer.from_pretrained(__lowercase) __UpperCamelCase :Dict = AutoModelForSeqaSeqLM.from_pretrained(__lowercase) __UpperCamelCase :List[Any] = tokenizer('''This is me''' , return_tensors='''pt''') __UpperCamelCase :Dict = model.to_bettertransformer() self.assertTrue(any('''BetterTransformer''' in mod.__class__.__name__ for _, mod in model.named_modules())) __UpperCamelCase :Union[str, Any] = model.generate(**__lowercase) __UpperCamelCase :Any = model.reverse_bettertransformer() self.assertFalse(any('''BetterTransformer''' in mod.__class__.__name__ for _, mod in model.named_modules())) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__lowercase) __UpperCamelCase :List[Any] = AutoModelForSeqaSeqLM.from_pretrained(__lowercase) self.assertFalse( any('''BetterTransformer''' in mod.__class__.__name__ for _, mod in model_reloaded.named_modules())) __UpperCamelCase :str = model_reloaded.generate(**__lowercase) self.assertTrue(torch.allclose(__lowercase , __lowercase)) def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :Union[str, Any] = '''hf-internal-testing/tiny-random-t5''' __UpperCamelCase :List[str] = AutoModelForSeqaSeqLM.from_pretrained(__lowercase) __UpperCamelCase :Tuple = model.to_bettertransformer() with tempfile.TemporaryDirectory() as tmpdirname: with self.assertRaises(__lowercase): model.save_pretrained(__lowercase) __UpperCamelCase :Tuple = model.reverse_bettertransformer() model.save_pretrained(__lowercase)
43
from typing import List, Optional, Union import numpy as np import PIL.Image from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, PILImageResampling, get_image_size, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __lowercase = logging.get_logger(__name__) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[Any] = ["""pixel_values"""] def __init__( self , __lowercase = True , __lowercase = 32 , __lowercase=PILImageResampling.BILINEAR , __lowercase = True , **__lowercase , ) -> None: __UpperCamelCase :Optional[int] = do_resize __UpperCamelCase :Any = do_rescale __UpperCamelCase :str = size_divisor __UpperCamelCase :Dict = resample super().__init__(**__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase = None , **__lowercase) -> np.ndarray: __UpperCamelCase , __UpperCamelCase :int = get_image_size(__lowercase) # Rounds the height and width down to the closest multiple of size_divisor __UpperCamelCase :List[Any] = height // size_divisor * size_divisor __UpperCamelCase :List[str] = width // size_divisor * size_divisor __UpperCamelCase :str = resize(__lowercase , (new_h, new_w) , resample=__lowercase , data_format=__lowercase , **__lowercase) return image def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase = None , **__lowercase) -> np.ndarray: return rescale(image=__lowercase , scale=__lowercase , data_format=__lowercase , **__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase = None , __lowercase = None , __lowercase=None , __lowercase = None , __lowercase = None , __lowercase = ChannelDimension.FIRST , **__lowercase , ) -> BatchFeature: __UpperCamelCase :Union[str, Any] = do_resize if do_resize is not None else self.do_resize __UpperCamelCase :Tuple = do_rescale if do_rescale is not None else self.do_rescale __UpperCamelCase :List[str] = size_divisor if size_divisor is not None else self.size_divisor __UpperCamelCase :List[Any] = resample if resample is not None else self.resample if do_resize and size_divisor is None: raise ValueError('''size_divisor is required for resizing''') __UpperCamelCase :List[Any] = make_list_of_images(__lowercase) if not valid_images(__lowercase): raise ValueError('''Invalid image(s)''') # All transformations expect numpy arrays. __UpperCamelCase :Optional[Any] = [to_numpy_array(__lowercase) for img in images] if do_resize: __UpperCamelCase :List[str] = [self.resize(__lowercase , size_divisor=__lowercase , resample=__lowercase) for image in images] if do_rescale: __UpperCamelCase :Dict = [self.rescale(__lowercase , scale=1 / 255) for image in images] __UpperCamelCase :str = [to_channel_dimension_format(__lowercase , __lowercase) for image in images] __UpperCamelCase :int = {'''pixel_values''': images} return BatchFeature(data=__lowercase , tensor_type=__lowercase)
43
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_tf_available, is_torch_available, ) __lowercase = { '''configuration_speech_to_text''': ['''SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Speech2TextConfig'''], '''processing_speech_to_text''': ['''Speech2TextProcessor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = ['''Speech2TextTokenizer'''] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = ['''Speech2TextFeatureExtractor'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFSpeech2TextForConditionalGeneration''', '''TFSpeech2TextModel''', '''TFSpeech2TextPreTrainedModel''', ] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Speech2TextForConditionalGeneration''', '''Speech2TextModel''', '''Speech2TextPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_speech_to_text import SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, SpeechaTextConfig from .processing_speech_to_text import SpeechaTextProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speech_to_text import SpeechaTextTokenizer try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_speech_to_text import SpeechaTextFeatureExtractor try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_speech_to_text import ( TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, TFSpeechaTextForConditionalGeneration, TFSpeechaTextModel, TFSpeechaTextPreTrainedModel, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_to_text import ( SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechaTextForConditionalGeneration, SpeechaTextModel, SpeechaTextPreTrainedModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
43
from __future__ import annotations from PIL import Image # Define glider example __lowercase = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example __lowercase = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :int = [] for i in range(len(SCREAMING_SNAKE_CASE ) ): __UpperCamelCase :Dict = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours __UpperCamelCase :List[str] = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(SCREAMING_SNAKE_CASE ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(SCREAMING_SNAKE_CASE ) - 1: neighbour_count += cells[i + 1][j] if i < len(SCREAMING_SNAKE_CASE ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. __UpperCamelCase :List[str] = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(SCREAMING_SNAKE_CASE ) return next_generation def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = [] for _ in range(SCREAMING_SNAKE_CASE ): # Create output image __UpperCamelCase :Dict = Image.new('''RGB''' , (len(cells[0] ), len(SCREAMING_SNAKE_CASE )) ) __UpperCamelCase :Any = img.load() # Save cells to image for x in range(len(SCREAMING_SNAKE_CASE ) ): for y in range(len(cells[0] ) ): __UpperCamelCase :Optional[Any] = 255 - cells[y][x] * 255 __UpperCamelCase :int = (colour, colour, colour) # Save image images.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = new_generation(SCREAMING_SNAKE_CASE ) return images if __name__ == "__main__": __lowercase = generate_images(GLIDER, 16) images[0].save('''out.gif''', save_all=True, append_images=images[1:])
43
1
import argparse import hashlib import os import urllib import warnings import torch from torch import nn from tqdm import tqdm from transformers import WhisperConfig, WhisperForConditionalGeneration __lowercase = { '''tiny.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d3dd57d32accea0b295c96e26691aa14d8822fac7d9d27d5dc00b4ca2826dd03/tiny.en.pt''', '''tiny''': '''https://openaipublic.azureedge.net/main/whisper/models/65147644a518d12f04e32d6f3b26facc3f8dd46e5390956a9424a650c0ce22b9/tiny.pt''', '''base.en''': '''https://openaipublic.azureedge.net/main/whisper/models/25a8566e1d0c1e2231d1c762132cd20e0f96a85d16145c3a00adf5d1ac670ead/base.en.pt''', '''base''': '''https://openaipublic.azureedge.net/main/whisper/models/ed3a0b6b1c0edf879ad9b11b1af5a0e6ab5db9205f891f668f8b0e6c6326e34e/base.pt''', '''small.en''': '''https://openaipublic.azureedge.net/main/whisper/models/f953ad0fd29cacd07d5a9eda5624af0f6bcf2258be67c92b79389873d91e0872/small.en.pt''', '''small''': '''https://openaipublic.azureedge.net/main/whisper/models/9ecf779972d90ba49c06d968637d720dd632c55bbf19d441fb42bf17a411e794/small.pt''', '''medium.en''': '''https://openaipublic.azureedge.net/main/whisper/models/d7440d1dc186f76616474e0ff0b3b6b879abc9d1a4926b7adfa41db2d497ab4f/medium.en.pt''', '''medium''': '''https://openaipublic.azureedge.net/main/whisper/models/345ae4da62f9b3d59415adc60127b97c714f32e89e936602e85993674d08dcb1/medium.pt''', '''large''': '''https://openaipublic.azureedge.net/main/whisper/models/e4b87e7e0bf463eb8e6956e646f1e277e901512310def2c24bf0e11bd3c28e9a/large.pt''', '''large-v2''': '''https://openaipublic.azureedge.net/main/whisper/models/81f7c96c852ee8fc832187b0132e569d6c3065a3252ed18e56effd0b6a73e524/large-v2.pt''', } def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[int] = ['''layers''', '''blocks'''] for k in ignore_keys: state_dict.pop(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) __lowercase = { '''blocks''': '''layers''', '''mlp.0''': '''fc1''', '''mlp.2''': '''fc2''', '''mlp_ln''': '''final_layer_norm''', '''.attn.query''': '''.self_attn.q_proj''', '''.attn.key''': '''.self_attn.k_proj''', '''.attn.value''': '''.self_attn.v_proj''', '''.attn_ln''': '''.self_attn_layer_norm''', '''.attn.out''': '''.self_attn.out_proj''', '''.cross_attn.query''': '''.encoder_attn.q_proj''', '''.cross_attn.key''': '''.encoder_attn.k_proj''', '''.cross_attn.value''': '''.encoder_attn.v_proj''', '''.cross_attn_ln''': '''.encoder_attn_layer_norm''', '''.cross_attn.out''': '''.encoder_attn.out_proj''', '''decoder.ln.''': '''decoder.layer_norm.''', '''encoder.ln.''': '''encoder.layer_norm.''', '''token_embedding''': '''embed_tokens''', '''encoder.positional_embedding''': '''encoder.embed_positions.weight''', '''decoder.positional_embedding''': '''decoder.embed_positions.weight''', '''ln_post''': '''layer_norm''', } def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :str = list(s_dict.keys() ) for key in keys: __UpperCamelCase :Union[str, Any] = key for k, v in WHISPER_MAPPING.items(): if k in key: __UpperCamelCase :List[Any] = new_key.replace(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) print(f"""{key} -> {new_key}""" ) __UpperCamelCase :List[str] = s_dict.pop(SCREAMING_SNAKE_CASE ) return s_dict def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase , __UpperCamelCase :List[Any] = emb.weight.shape __UpperCamelCase :Tuple = nn.Linear(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , bias=SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = emb.weight.data return lin_layer def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) __UpperCamelCase :int = os.path.basename(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[str] = url.split('''/''' )[-2] __UpperCamelCase :Dict = os.path.join(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if os.path.exists(SCREAMING_SNAKE_CASE ) and not os.path.isfile(SCREAMING_SNAKE_CASE ): raise RuntimeError(f"""{download_target} exists and is not a regular file""" ) if os.path.isfile(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = open(SCREAMING_SNAKE_CASE , '''rb''' ).read() if hashlib.shaaaa(SCREAMING_SNAKE_CASE ).hexdigest() == expected_shaaaa: return model_bytes else: warnings.warn(f"""{download_target} exists, but the SHA256 checksum does not match; re-downloading the file""" ) with urllib.request.urlopen(SCREAMING_SNAKE_CASE ) as source, open(SCREAMING_SNAKE_CASE , '''wb''' ) as output: with tqdm( total=int(source.info().get('''Content-Length''' ) ) , ncols=80 , unit='''iB''' , unit_scale=SCREAMING_SNAKE_CASE , unit_divisor=1_024 ) as loop: while True: __UpperCamelCase :str = source.read(8_192 ) if not buffer: break output.write(SCREAMING_SNAKE_CASE ) loop.update(len(SCREAMING_SNAKE_CASE ) ) __UpperCamelCase :Union[str, Any] = open(SCREAMING_SNAKE_CASE , '''rb''' ).read() if hashlib.shaaaa(SCREAMING_SNAKE_CASE ).hexdigest() != expected_shaaaa: raise RuntimeError( '''Model has been downloaded but the SHA256 checksum does not not match. Please retry loading the model.''' ) return model_bytes def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' if ".pt" not in checkpoint_path: __UpperCamelCase :Dict = _download(_MODELS[checkpoint_path] ) else: __UpperCamelCase :Optional[Any] = torch.load(SCREAMING_SNAKE_CASE , map_location='''cpu''' ) __UpperCamelCase :Dict = original_checkpoint['''dims'''] __UpperCamelCase :Optional[Any] = original_checkpoint['''model_state_dict'''] __UpperCamelCase :Optional[int] = state_dict['''decoder.token_embedding.weight'''] remove_ignore_keys_(SCREAMING_SNAKE_CASE ) rename_keys(SCREAMING_SNAKE_CASE ) __UpperCamelCase :int = True __UpperCamelCase :Union[str, Any] = state_dict['''decoder.layers.0.fc1.weight'''].shape[0] __UpperCamelCase :List[Any] = WhisperConfig( vocab_size=dimensions['''n_vocab'''] , encoder_ffn_dim=SCREAMING_SNAKE_CASE , decoder_ffn_dim=SCREAMING_SNAKE_CASE , num_mel_bins=dimensions['''n_mels'''] , d_model=dimensions['''n_audio_state'''] , max_target_positions=dimensions['''n_text_ctx'''] , encoder_layers=dimensions['''n_audio_layer'''] , encoder_attention_heads=dimensions['''n_audio_head'''] , decoder_layers=dimensions['''n_text_layer'''] , decoder_attention_heads=dimensions['''n_text_state'''] , max_source_positions=dimensions['''n_audio_ctx'''] , ) __UpperCamelCase :Union[str, Any] = WhisperForConditionalGeneration(SCREAMING_SNAKE_CASE ) __UpperCamelCase , __UpperCamelCase :Dict = model.model.load_state_dict(SCREAMING_SNAKE_CASE , strict=SCREAMING_SNAKE_CASE ) if len(SCREAMING_SNAKE_CASE ) > 0 and not set(SCREAMING_SNAKE_CASE ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( '''Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,''' f""" but all the following weights are missing {missing}""" ) if tie_embeds: __UpperCamelCase :int = make_linear_from_emb(model.model.decoder.embed_tokens ) else: __UpperCamelCase :List[Any] = proj_out_weights model.save_pretrained(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() # # Required parameters parser.add_argument('''--checkpoint_path''', type=str, help='''Patht to the downloaded checkpoints''') parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') __lowercase = parser.parse_args() convert_openai_whisper_to_tfms(args.checkpoint_path, args.pytorch_dump_folder_path)
43
import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging __lowercase = logging.get_logger(__name__) def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = R'''\w+[.]\d+''' __UpperCamelCase :List[str] = re.findall(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for pat in pats: __UpperCamelCase :int = key.replace(SCREAMING_SNAKE_CASE , '''_'''.join(pat.split('''.''' ) ) ) return key def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = pt_tuple_key[:-1] + ('''scale''',) if ( any('''norm''' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): __UpperCamelCase :str = pt_tuple_key[:-1] + ('''scale''',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: __UpperCamelCase :Any = pt_tuple_key[:-1] + ('''scale''',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: __UpperCamelCase :str = pt_tuple_key[:-1] + ('''embedding''',) return renamed_pt_tuple_key, pt_tensor # conv layer __UpperCamelCase :List[str] = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: __UpperCamelCase :List[Any] = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer __UpperCamelCase :List[str] = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight": __UpperCamelCase :Any = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight __UpperCamelCase :int = pt_tuple_key[:-1] + ('''weight''',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias __UpperCamelCase :int = pt_tuple_key[:-1] + ('''bias''',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=42 ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params __UpperCamelCase :str = flax_model.init_weights(PRNGKey(SCREAMING_SNAKE_CASE ) ) __UpperCamelCase :int = flatten_dict(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): __UpperCamelCase :List[Any] = rename_key(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = tuple(renamed_pt_key.split('''.''' ) ) # Correctly rename weight parameters __UpperCamelCase , __UpperCamelCase :Any = rename_key_and_reshape_tensor(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ f"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown __UpperCamelCase :str = jnp.asarray(SCREAMING_SNAKE_CASE ) return unflatten_dict(SCREAMING_SNAKE_CASE )
43
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowercase = { '''configuration_xmod''': [ '''XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XmodConfig''', '''XmodOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''XMOD_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XmodForCausalLM''', '''XmodForMaskedLM''', '''XmodForMultipleChoice''', '''XmodForQuestionAnswering''', '''XmodForSequenceClassification''', '''XmodForTokenClassification''', '''XmodModel''', '''XmodPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
43
import argparse import torch from ...utils import logging from . import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert logging.set_verbosity_info() def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = AlbertConfig.from_json_file(SCREAMING_SNAKE_CASE ) print(f"""Building PyTorch model from configuration: {config}""" ) __UpperCamelCase :List[str] = AlbertForPreTraining(SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint load_tf_weights_in_albert(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--albert_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained ALBERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __lowercase = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.albert_config_file, args.pytorch_dump_path)
43
1
from .imports import is_tqdm_available if is_tqdm_available(): from tqdm.auto import tqdm as _tqdm from ..state import PartialState def lowerCamelCase ( SCREAMING_SNAKE_CASE = True , *SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ): '''simple docstring''' if not is_tqdm_available(): raise ImportError('''Accelerate\'s `tqdm` module requires `tqdm` to be installed. Please run `pip install tqdm`.''' ) __UpperCamelCase :int = False if main_process_only: __UpperCamelCase :Union[str, Any] = PartialState().local_process_index == 0 return _tqdm(*SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE , disable=SCREAMING_SNAKE_CASE )
43
import math import qiskit def lowerCamelCase ( SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1 ): '''simple docstring''' if ( isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) or isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) or isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ): raise TypeError('''inputs must be integers.''' ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError('''inputs must be positive.''' ) if ( (math.floor(SCREAMING_SNAKE_CASE ) != input_a) or (math.floor(SCREAMING_SNAKE_CASE ) != input_a) or (math.floor(SCREAMING_SNAKE_CASE ) != carry_in) ): raise ValueError('''inputs must be exact integers.''' ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError('''inputs must be less or equal to 2.''' ) # build registers __UpperCamelCase :List[str] = qiskit.QuantumRegister(4 , '''qr''' ) __UpperCamelCase :str = qiskit.ClassicalRegister(2 , '''cr''' ) # list the entries __UpperCamelCase :Tuple = [input_a, input_a, carry_in] __UpperCamelCase :Optional[int] = qiskit.QuantumCircuit(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(SCREAMING_SNAKE_CASE ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(SCREAMING_SNAKE_CASE ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(SCREAMING_SNAKE_CASE ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , SCREAMING_SNAKE_CASE ) # measure the last two qbits __UpperCamelCase :Optional[Any] = qiskit.Aer.get_backend('''aer_simulator''' ) __UpperCamelCase :Tuple = qiskit.execute(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , shots=1_000 ) return job.result().get_counts(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(F'Total sum count for state is: {quantum_full_adder(1, 1, 1)}')
43
1
import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration __lowercase = 5_0000 __lowercase = 5000 __lowercase , __lowercase = os.path.split(__file__) __lowercase = os.path.join(RESULTS_BASEPATH, '''results''', RESULTS_FILENAME.replace('''.py''', '''.json''')) @get_duration def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' for i in range(SCREAMING_SNAKE_CASE ): __UpperCamelCase :int = dataset[i] @get_duration def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' for i in range(0 , len(SCREAMING_SNAKE_CASE ) , SCREAMING_SNAKE_CASE ): __UpperCamelCase :str = dataset[i : i + batch_size] @get_duration def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' with dataset.formatted_as(type=SCREAMING_SNAKE_CASE ): for i in range(SCREAMING_SNAKE_CASE ): __UpperCamelCase :List[str] = dataset[i] @get_duration def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' with dataset.formatted_as(type=SCREAMING_SNAKE_CASE ): for i in range(0 , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): __UpperCamelCase :List[Any] = dataset[i : i + batch_size] def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Optional[Any] = {'''num examples''': SPEED_TEST_N_EXAMPLES} __UpperCamelCase :Optional[Any] = [ (read, {'''length''': SMALL_TEST}), (read, {'''length''': SPEED_TEST_N_EXAMPLES}), (read_batch, {'''length''': SPEED_TEST_N_EXAMPLES, '''batch_size''': 10}), (read_batch, {'''length''': SPEED_TEST_N_EXAMPLES, '''batch_size''': 100}), (read_batch, {'''length''': SPEED_TEST_N_EXAMPLES, '''batch_size''': 1_000}), (read_formatted, {'''type''': '''numpy''', '''length''': SMALL_TEST}), (read_formatted, {'''type''': '''pandas''', '''length''': SMALL_TEST}), (read_formatted, {'''type''': '''torch''', '''length''': SMALL_TEST}), (read_formatted, {'''type''': '''tensorflow''', '''length''': SMALL_TEST}), (read_formatted_batch, {'''type''': '''numpy''', '''length''': SMALL_TEST, '''batch_size''': 10}), (read_formatted_batch, {'''type''': '''numpy''', '''length''': SMALL_TEST, '''batch_size''': 1_000}), ] __UpperCamelCase :List[str] = [ (read, {'''length''': SMALL_TEST}), (read, {'''length''': SPEED_TEST_N_EXAMPLES}), (read_batch, {'''length''': SPEED_TEST_N_EXAMPLES, '''batch_size''': 10}), (read_batch, {'''length''': SPEED_TEST_N_EXAMPLES, '''batch_size''': 100}), (read_batch, {'''length''': SPEED_TEST_N_EXAMPLES, '''batch_size''': 1_000}), (read_formatted, {'''type''': '''numpy''', '''length''': SMALL_TEST}), (read_formatted_batch, {'''type''': '''numpy''', '''length''': SMALL_TEST, '''batch_size''': 10}), (read_formatted_batch, {'''type''': '''numpy''', '''length''': SMALL_TEST, '''batch_size''': 1_000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print('''generating dataset''' ) __UpperCamelCase :Any = datasets.Features( {'''list''': datasets.Sequence(datasets.Value('''float32''' ) ), '''numbers''': datasets.Value('''float32''' )} ) __UpperCamelCase :int = generate_example_dataset( os.path.join(SCREAMING_SNAKE_CASE , '''dataset.arrow''' ) , SCREAMING_SNAKE_CASE , num_examples=SCREAMING_SNAKE_CASE , seq_shapes={'''list''': (100,)} , ) print('''first set of iterations''' ) for func, kwargs in functions: print(func.__name__ , str(SCREAMING_SNAKE_CASE ) ) __UpperCamelCase :Union[str, Any] = func(SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) print('''shuffling dataset''' ) __UpperCamelCase :Union[str, Any] = dataset.shuffle() print('''Second set of iterations (after shuffling''' ) for func, kwargs in functions_shuffled: print('''shuffled ''' , func.__name__ , str(SCREAMING_SNAKE_CASE ) ) __UpperCamelCase :List[Any] = func( SCREAMING_SNAKE_CASE , **SCREAMING_SNAKE_CASE ) with open(SCREAMING_SNAKE_CASE , '''wb''' ) as f: f.write(json.dumps(SCREAMING_SNAKE_CASE ).encode('''utf-8''' ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
43
import random def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = a[left_index] __UpperCamelCase :Any = left_index + 1 for j in range(left_index + 1 , SCREAMING_SNAKE_CASE ): if a[j] < pivot: __UpperCamelCase , __UpperCamelCase :str = a[i], a[j] i += 1 __UpperCamelCase , __UpperCamelCase :Optional[int] = a[i - 1], a[left_index] return i - 1 def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' if left < right: __UpperCamelCase :int = random.randint(SCREAMING_SNAKE_CASE , right - 1 ) __UpperCamelCase , __UpperCamelCase :List[str] = ( a[left], a[pivot], ) # switches the pivot with the left most bound __UpperCamelCase :Dict = partition(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) quick_sort_random( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # recursive quicksort to the left of the pivot point quick_sort_random( SCREAMING_SNAKE_CASE , pivot_index + 1 , SCREAMING_SNAKE_CASE ) # recursive quicksort to the right of the pivot point def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = input('''Enter numbers separated by a comma:\n''' ).strip() __UpperCamelCase :Union[str, Any] = [int(SCREAMING_SNAKE_CASE ) for item in user_input.split(''',''' )] quick_sort_random(SCREAMING_SNAKE_CASE , 0 , len(SCREAMING_SNAKE_CASE ) ) print(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
43
1
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging __lowercase = logging.get_logger(__name__) __lowercase = { '''Helsinki-NLP/opus-mt-en-de''': '''https://huggingface.co/Helsinki-NLP/opus-mt-en-de/resolve/main/config.json''', # See all Marian models at https://huggingface.co/models?filter=marian } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[Any] = """marian""" a__ : str = ["""past_key_values"""] a__ : str = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , __lowercase=58_101 , __lowercase=None , __lowercase=1_024 , __lowercase=12 , __lowercase=4_096 , __lowercase=16 , __lowercase=12 , __lowercase=4_096 , __lowercase=16 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=True , __lowercase=True , __lowercase="gelu" , __lowercase=1_024 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.02 , __lowercase=58_100 , __lowercase=False , __lowercase=58_100 , __lowercase=0 , __lowercase=0 , __lowercase=True , **__lowercase , ) -> Union[str, Any]: __UpperCamelCase :List[Any] = vocab_size __UpperCamelCase :Dict = decoder_vocab_size or vocab_size __UpperCamelCase :Union[str, Any] = max_position_embeddings __UpperCamelCase :List[Any] = d_model __UpperCamelCase :Optional[int] = encoder_ffn_dim __UpperCamelCase :Tuple = encoder_layers __UpperCamelCase :Union[str, Any] = encoder_attention_heads __UpperCamelCase :Dict = decoder_ffn_dim __UpperCamelCase :Union[str, Any] = decoder_layers __UpperCamelCase :Union[str, Any] = decoder_attention_heads __UpperCamelCase :int = dropout __UpperCamelCase :List[str] = attention_dropout __UpperCamelCase :List[Any] = activation_dropout __UpperCamelCase :Tuple = activation_function __UpperCamelCase :Union[str, Any] = init_std __UpperCamelCase :List[str] = encoder_layerdrop __UpperCamelCase :List[Any] = decoder_layerdrop __UpperCamelCase :str = use_cache __UpperCamelCase :Dict = encoder_layers __UpperCamelCase :Dict = scale_embedding # scale factor will be sqrt(d_model) if True __UpperCamelCase :Union[str, Any] = share_encoder_decoder_embeddings super().__init__( pad_token_id=__lowercase , eos_token_id=__lowercase , is_encoder_decoder=__lowercase , decoder_start_token_id=__lowercase , forced_eos_token_id=__lowercase , **__lowercase , ) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.inputs def UpperCamelCase__ ( self) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: __UpperCamelCase :Dict = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ]) if self.use_past: __UpperCamelCase :int = {0: '''batch'''} __UpperCamelCase :Optional[int] = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} else: __UpperCamelCase :Optional[Any] = {0: '''batch''', 1: '''decoder_sequence'''} __UpperCamelCase :List[str] = {0: '''batch''', 1: '''decoder_sequence'''} if self.use_past: self.fill_with_past_key_values_(__lowercase , direction='''inputs''') elif self.task == "causal-lm": # TODO: figure this case out. __UpperCamelCase :List[Any] = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ]) if self.use_past: __UpperCamelCase , __UpperCamelCase :List[str] = self.num_layers for i in range(__lowercase): __UpperCamelCase :int = {0: '''batch''', 2: '''past_sequence + sequence'''} __UpperCamelCase :List[Any] = {0: '''batch''', 2: '''past_sequence + sequence'''} else: __UpperCamelCase :int = OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''attention_mask''', {0: '''batch''', 1: '''encoder_sequence'''}), ('''decoder_input_ids''', {0: '''batch''', 1: '''decoder_sequence'''}), ('''decoder_attention_mask''', {0: '''batch''', 1: '''decoder_sequence'''}), ]) return common_inputs @property # Copied from transformers.models.bart.configuration_bart.BartOnnxConfig.outputs def UpperCamelCase__ ( self) -> Mapping[str, Mapping[int, str]]: if self.task in ["default", "seq2seq-lm"]: __UpperCamelCase :List[str] = super().outputs else: __UpperCamelCase :Any = super(__lowercase , self).outputs if self.use_past: __UpperCamelCase , __UpperCamelCase :Dict = self.num_layers for i in range(__lowercase): __UpperCamelCase :List[Any] = {0: '''batch''', 2: '''past_sequence + sequence'''} __UpperCamelCase :Tuple = {0: '''batch''', 2: '''past_sequence + sequence'''} return common_outputs def UpperCamelCase__ ( self , __lowercase , __lowercase = -1 , __lowercase = -1 , __lowercase = False , __lowercase = None , ) -> Mapping[str, Any]: __UpperCamelCase :Dict = self._generate_dummy_inputs_for_encoder_and_decoder( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase) # Generate decoder inputs __UpperCamelCase :str = seq_length if not self.use_past else 1 __UpperCamelCase :List[str] = self._generate_dummy_inputs_for_encoder_and_decoder( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase) __UpperCamelCase :List[Any] = {f"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} __UpperCamelCase :Union[str, Any] = dict(**__lowercase , **__lowercase) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''') else: import torch __UpperCamelCase , __UpperCamelCase :List[str] = common_inputs['''input_ids'''].shape __UpperCamelCase :Optional[Any] = common_inputs['''decoder_input_ids'''].shape[1] __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self.num_attention_heads __UpperCamelCase :Optional[int] = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) __UpperCamelCase :Optional[int] = decoder_seq_length + 3 __UpperCamelCase :Tuple = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) __UpperCamelCase :Optional[Any] = torch.cat( [common_inputs['''decoder_attention_mask'''], torch.ones(__lowercase , __lowercase)] , dim=1) __UpperCamelCase :Union[str, Any] = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self.num_layers __UpperCamelCase :Dict = min(__lowercase , __lowercase) __UpperCamelCase :Dict = max(__lowercase , __lowercase) - min_num_layers __UpperCamelCase :str = '''encoder''' if num_encoder_layers > num_decoder_layers else '''decoder''' for _ in range(__lowercase): common_inputs["past_key_values"].append( ( torch.zeros(__lowercase), torch.zeros(__lowercase), torch.zeros(__lowercase), torch.zeros(__lowercase), )) # TODO: test this. __UpperCamelCase :Optional[int] = encoder_shape if remaining_side_name == '''encoder''' else decoder_shape for _ in range(__lowercase , __lowercase): common_inputs["past_key_values"].append((torch.zeros(__lowercase), torch.zeros(__lowercase))) return common_inputs def UpperCamelCase__ ( self , __lowercase , __lowercase = -1 , __lowercase = -1 , __lowercase = False , __lowercase = None , ) -> Mapping[str, Any]: __UpperCamelCase :int = self._generate_dummy_inputs_for_encoder_and_decoder( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase) if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''') else: import torch __UpperCamelCase , __UpperCamelCase :Dict = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values __UpperCamelCase :Tuple = seqlen + 2 __UpperCamelCase , __UpperCamelCase :List[str] = self.num_layers __UpperCamelCase , __UpperCamelCase :Tuple = self.num_attention_heads __UpperCamelCase :Optional[Any] = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) __UpperCamelCase :Optional[Any] = common_inputs['''attention_mask'''].dtype __UpperCamelCase :int = torch.cat( [common_inputs['''attention_mask'''], torch.ones(__lowercase , __lowercase , dtype=__lowercase)] , dim=1) __UpperCamelCase :Union[str, Any] = [ (torch.zeros(__lowercase), torch.zeros(__lowercase)) for _ in range(__lowercase) ] return common_inputs def UpperCamelCase__ ( self , __lowercase , __lowercase = -1 , __lowercase = -1 , __lowercase = False , __lowercase = None , ) -> Mapping[str, Any]: # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX __UpperCamelCase :Optional[Any] = compute_effective_axis_dimension( __lowercase , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX __UpperCamelCase :Optional[int] = tokenizer.num_special_tokens_to_add(__lowercase) __UpperCamelCase :List[Any] = compute_effective_axis_dimension( __lowercase , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__lowercase) # Generate dummy inputs according to compute batch and sequence __UpperCamelCase :Any = [''' '''.join([tokenizer.unk_token]) * seq_length] * batch_size __UpperCamelCase :Optional[Any] = dict(tokenizer(__lowercase , return_tensors=__lowercase)) return common_inputs def UpperCamelCase__ ( self , __lowercase , __lowercase = -1 , __lowercase = -1 , __lowercase = False , __lowercase = None , ) -> Mapping[str, Any]: if self.task in ["default", "seq2seq-lm"]: __UpperCamelCase :Optional[int] = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __lowercase , batch_size=__lowercase , seq_length=__lowercase , is_pair=__lowercase , framework=__lowercase) else: __UpperCamelCase :Dict = self._generate_dummy_inputs_for_causal_lm( __lowercase , batch_size=__lowercase , seq_length=__lowercase , is_pair=__lowercase , framework=__lowercase) return common_inputs def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase) -> List[str]: if self.task in ["default", "seq2seq-lm"]: __UpperCamelCase :Optional[int] = super()._flatten_past_key_values_(__lowercase , __lowercase , __lowercase , __lowercase) else: __UpperCamelCase :List[str] = super(__lowercase , self)._flatten_past_key_values_( __lowercase , __lowercase , __lowercase , __lowercase) @property def UpperCamelCase__ ( self) -> float: return 1E-4
43
def lowerCamelCase ( SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1_000 ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = 1 __UpperCamelCase :Any = 0 for divide_by_number in range(SCREAMING_SNAKE_CASE , digit + 1 ): __UpperCamelCase :list[int] = [] __UpperCamelCase :Optional[int] = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = len(SCREAMING_SNAKE_CASE ) __UpperCamelCase :int = divide_by_number else: has_been_divided.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
43
1
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __lowercase = logging.get_logger(__name__) __lowercase = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''ctc_proj''', '''mask_emb''': '''masked_spec_embed''', } __lowercase = [ '''ctc_proj''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' for attribute in key.split('''.''' ): if is_finetuned: if attribute in ["quantizer", "project_q", "project_hid"]: # those layers are only relevant for pretraining and should be dropped return if attribute == "ctc_proj": # we should rename `ctc_proj` to `lm_head` for fine-tuned phoneme models __UpperCamelCase :str = '''lm_head''' __UpperCamelCase :Any = getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if weight_type is not None: __UpperCamelCase :int = getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).shape else: __UpperCamelCase :Any = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __UpperCamelCase :Optional[Any] = value elif weight_type == "weight_g": __UpperCamelCase :Any = value elif weight_type == "weight_v": __UpperCamelCase :Any = value elif weight_type == "bias": __UpperCamelCase :Any = value else: __UpperCamelCase :Optional[Any] = value logger.info(f"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Any = [] __UpperCamelCase :Tuple = fairseq_model.state_dict() __UpperCamelCase :Tuple = hf_model.unispeech.feature_extractor for name, value in fairseq_dict.items(): __UpperCamelCase :Optional[Any] = False if "conv_layers" in name: load_conv_layer( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , hf_model.config.feat_extract_norm == '''group''' , ) __UpperCamelCase :List[Any] = True else: for key, mapped_key in MAPPING.items(): __UpperCamelCase :Optional[int] = '''unispeech.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: __UpperCamelCase :Optional[int] = True if "*" in mapped_key: __UpperCamelCase :str = name.split(SCREAMING_SNAKE_CASE )[0].split('''.''' )[-2] __UpperCamelCase :Tuple = mapped_key.replace('''*''' , SCREAMING_SNAKE_CASE ) if "weight_g" in name: __UpperCamelCase :Union[str, Any] = '''weight_g''' elif "weight_v" in name: __UpperCamelCase :Optional[int] = '''weight_v''' elif "bias" in name: __UpperCamelCase :str = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj __UpperCamelCase :int = '''weight''' else: __UpperCamelCase :str = None set_recursively(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(SCREAMING_SNAKE_CASE ) logger.warning(f"""Unused weights: {unused_weights}""" ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[int] = full_name.split('''conv_layers.''' )[-1] __UpperCamelCase :Optional[int] = name.split('''.''' ) __UpperCamelCase :int = int(items[0] ) __UpperCamelCase :List[str] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __UpperCamelCase :Union[str, Any] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __UpperCamelCase :int = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) __UpperCamelCase :str = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) __UpperCamelCase :List[str] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(SCREAMING_SNAKE_CASE ) @torch.no_grad() def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=True ): '''simple docstring''' if config_path is not None: __UpperCamelCase :str = UniSpeechConfig.from_pretrained(SCREAMING_SNAKE_CASE ) else: __UpperCamelCase :Optional[Any] = UniSpeechConfig() if is_finetuned: if dict_path: __UpperCamelCase :Tuple = Dictionary.load_from_json(SCREAMING_SNAKE_CASE ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __UpperCamelCase :Tuple = target_dict.pad_index __UpperCamelCase :Any = target_dict.bos_index __UpperCamelCase :List[str] = target_dict.eos_index __UpperCamelCase :int = len(target_dict.symbols ) __UpperCamelCase :Optional[Any] = os.path.join(SCREAMING_SNAKE_CASE , '''vocab.json''' ) if not os.path.isdir(SCREAMING_SNAKE_CASE ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(SCREAMING_SNAKE_CASE ) ) return os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[str] = target_dict.indices # fairseq has the <pad> and <s> switched __UpperCamelCase :int = 42 __UpperCamelCase :str = 43 with open(SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = WavaVecaPhonemeCTCTokenizer( SCREAMING_SNAKE_CASE , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :int = True if config.feat_extract_norm == '''layer''' else False __UpperCamelCase :List[Any] = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0 , do_normalize=SCREAMING_SNAKE_CASE , return_attention_mask=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :List[str] = WavaVecaProcessor(feature_extractor=SCREAMING_SNAKE_CASE , tokenizer=SCREAMING_SNAKE_CASE ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) __UpperCamelCase :str = UniSpeechForCTC(SCREAMING_SNAKE_CASE ) else: __UpperCamelCase :List[Any] = UniSpeechForPreTraining(SCREAMING_SNAKE_CASE ) if is_finetuned: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :int = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] ), '''w2v_path''': checkpoint_path} ) else: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Dict = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) __UpperCamelCase :Optional[int] = model[0].eval() recursively_load_weights(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) hf_unispeech.save_pretrained(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) __lowercase = parser.parse_args() convert_unispeech_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
43
import argparse import json from tqdm import tqdm def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--src_path''' , type=SCREAMING_SNAKE_CASE , default='''biencoder-nq-dev.json''' , help='''Path to raw DPR training data''' , ) parser.add_argument( '''--evaluation_set''' , type=SCREAMING_SNAKE_CASE , help='''where to store parsed evaluation_set file''' , ) parser.add_argument( '''--gold_data_path''' , type=SCREAMING_SNAKE_CASE , help='''where to store parsed gold_data_path file''' , ) __UpperCamelCase :str = parser.parse_args() with open(args.src_path , '''r''' ) as src_file, open(args.evaluation_set , '''w''' ) as eval_file, open( args.gold_data_path , '''w''' ) as gold_file: __UpperCamelCase :List[str] = json.load(SCREAMING_SNAKE_CASE ) for dpr_record in tqdm(SCREAMING_SNAKE_CASE ): __UpperCamelCase :List[str] = dpr_record['''question'''] __UpperCamelCase :Tuple = [context['''title'''] for context in dpr_record['''positive_ctxs''']] eval_file.write(question + '''\n''' ) gold_file.write('''\t'''.join(SCREAMING_SNAKE_CASE ) + '''\n''' ) if __name__ == "__main__": main()
43
1
import os from collections import namedtuple import pytest from datasets import ClassLabel, Features, Sequence, Value from datasets.commands.test import TestCommand from datasets.info import DatasetInfo, DatasetInfosDict __lowercase = namedtuple( '''_TestCommandArgs''', [ '''dataset''', '''name''', '''cache_dir''', '''data_dir''', '''all_configs''', '''save_infos''', '''ignore_verifications''', '''force_redownload''', '''clear_cache''', ], defaults=[None, None, None, False, False, False, False, False], ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' return (abs(source - target ) / target) < 0.01 @pytest.mark.integration def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = _TestCommandArgs(dataset=SCREAMING_SNAKE_CASE , all_configs=SCREAMING_SNAKE_CASE , save_infos=SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[str] = TestCommand(*SCREAMING_SNAKE_CASE ) test_command.run() __UpperCamelCase :int = os.path.join(SCREAMING_SNAKE_CASE , '''README.md''' ) assert os.path.exists(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[str] = DatasetInfosDict.from_directory(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Union[str, Any] = DatasetInfosDict( { '''default''': DatasetInfo( features=Features( { '''tokens''': Sequence(Value('''string''' ) ), '''ner_tags''': Sequence( ClassLabel(names=['''O''', '''B-PER''', '''I-PER''', '''B-ORG''', '''I-ORG''', '''B-LOC''', '''I-LOC'''] ) ), '''langs''': Sequence(Value('''string''' ) ), '''spans''': Sequence(Value('''string''' ) ), } ) , splits=[ { '''name''': '''train''', '''num_bytes''': 2_351_563, '''num_examples''': 10_000, }, { '''name''': '''validation''', '''num_bytes''': 238_418, '''num_examples''': 1_000, }, ] , download_size=3_940_680 , dataset_size=2_589_981 , ) } ) assert dataset_infos.keys() == expected_dataset_infos.keys() for key in DatasetInfo._INCLUDED_INFO_IN_YAML: __UpperCamelCase , __UpperCamelCase :Optional[int] = getattr(dataset_infos['''default'''] , SCREAMING_SNAKE_CASE ), getattr(expected_dataset_infos['''default'''] , SCREAMING_SNAKE_CASE ) if key == "num_bytes": assert is_apercent_close(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) elif key == "splits": assert list(SCREAMING_SNAKE_CASE ) == list(SCREAMING_SNAKE_CASE ) for split in result: assert result[split].name == expected[split].name assert result[split].num_examples == expected[split].num_examples assert is_apercent_close(result[split].num_bytes , expected[split].num_bytes ) else: result == expected
43
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. __lowercase = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. __lowercase = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. __lowercase = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Any = len([g for position, g in enumerate(SCREAMING_SNAKE_CASE ) if g == main_target[position]] ) return (item, float(SCREAMING_SNAKE_CASE )) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = random.randint(0 , len(SCREAMING_SNAKE_CASE ) - 1 ) __UpperCamelCase :Tuple = parent_a[:random_slice] + parent_a[random_slice:] __UpperCamelCase :Union[str, Any] = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :int = list(SCREAMING_SNAKE_CASE ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: __UpperCamelCase :str = random.choice(SCREAMING_SNAKE_CASE ) return "".join(SCREAMING_SNAKE_CASE ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ): '''simple docstring''' __UpperCamelCase :int = [] # Generate more children proportionally to the fitness score. __UpperCamelCase :int = int(parent_a[1] * 100 ) + 1 __UpperCamelCase :List[str] = 10 if child_n >= 10 else child_n for _ in range(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = population_score[random.randint(0 , SCREAMING_SNAKE_CASE )][0] __UpperCamelCase , __UpperCamelCase :Any = crossover(parent_a[0] , SCREAMING_SNAKE_CASE ) # Append new string to the population list. pop.append(mutate(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) pop.append(mutate(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) return pop def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = True ): '''simple docstring''' if N_POPULATION < N_SELECTED: __UpperCamelCase :List[Any] = f"""{N_POPULATION} must be bigger than {N_SELECTED}""" raise ValueError(SCREAMING_SNAKE_CASE ) # Verify that the target contains no genes besides the ones inside genes variable. __UpperCamelCase :List[str] = sorted({c for c in target if c not in genes} ) if not_in_genes_list: __UpperCamelCase :Optional[int] = f"""{not_in_genes_list} is not in genes list, evolution cannot converge""" raise ValueError(SCREAMING_SNAKE_CASE ) # Generate random starting population. __UpperCamelCase :int = [] for _ in range(SCREAMING_SNAKE_CASE ): population.append(''''''.join([random.choice(SCREAMING_SNAKE_CASE ) for i in range(len(SCREAMING_SNAKE_CASE ) )] ) ) # Just some logs to know what the algorithms is doing. __UpperCamelCase , __UpperCamelCase :List[Any] = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(SCREAMING_SNAKE_CASE ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. __UpperCamelCase :Tuple = [evaluate(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for item in population] # Check if there is a matching evolution. __UpperCamelCase :Tuple = sorted(SCREAMING_SNAKE_CASE , key=lambda SCREAMING_SNAKE_CASE : x[1] , reverse=SCREAMING_SNAKE_CASE ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( f"""\nGeneration: {generation}""" f"""\nTotal Population:{total_population}""" f"""\nBest score: {population_score[0][1]}""" f"""\nBest string: {population_score[0][0]}""" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. __UpperCamelCase :str = population[: int(N_POPULATION / 3 )] population.clear() population.extend(SCREAMING_SNAKE_CASE ) # Normalize population score to be between 0 and 1. __UpperCamelCase :Union[str, Any] = [ (item, score / len(SCREAMING_SNAKE_CASE )) for item, score in population_score ] # This is selection for i in range(SCREAMING_SNAKE_CASE ): population.extend(select(population_score[int(SCREAMING_SNAKE_CASE )] , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(SCREAMING_SNAKE_CASE ) > N_POPULATION: break if __name__ == "__main__": __lowercase = ( '''This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!''' ) __lowercase = list( ''' ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm''' '''nopqrstuvwxyz.,;!?+-*#@^\'èéòà€ù=)(&%$£/\\''' ) __lowercase , __lowercase , __lowercase = basic(target_str, genes_list) print( F'\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}' )
43
1
import math import qiskit def lowerCamelCase ( SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1 ): '''simple docstring''' if ( isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) or isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) or isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ): raise TypeError('''inputs must be integers.''' ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError('''inputs must be positive.''' ) if ( (math.floor(SCREAMING_SNAKE_CASE ) != input_a) or (math.floor(SCREAMING_SNAKE_CASE ) != input_a) or (math.floor(SCREAMING_SNAKE_CASE ) != carry_in) ): raise ValueError('''inputs must be exact integers.''' ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError('''inputs must be less or equal to 2.''' ) # build registers __UpperCamelCase :List[str] = qiskit.QuantumRegister(4 , '''qr''' ) __UpperCamelCase :str = qiskit.ClassicalRegister(2 , '''cr''' ) # list the entries __UpperCamelCase :Tuple = [input_a, input_a, carry_in] __UpperCamelCase :Optional[int] = qiskit.QuantumCircuit(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(SCREAMING_SNAKE_CASE ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(SCREAMING_SNAKE_CASE ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(SCREAMING_SNAKE_CASE ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , SCREAMING_SNAKE_CASE ) # measure the last two qbits __UpperCamelCase :Optional[Any] = qiskit.Aer.get_backend('''aer_simulator''' ) __UpperCamelCase :Tuple = qiskit.execute(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , shots=1_000 ) return job.result().get_counts(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(F'Total sum count for state is: {quantum_full_adder(1, 1, 1)}')
43
import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler __lowercase = 16 __lowercase = 32 def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 16 , SCREAMING_SNAKE_CASE = "bert-base-cased" ): '''simple docstring''' __UpperCamelCase :List[str] = AutoTokenizer.from_pretrained(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(SCREAMING_SNAKE_CASE ): # max_length=None => use the model max length (it's actually the default) __UpperCamelCase :int = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset __UpperCamelCase :Tuple = datasets.map( SCREAMING_SNAKE_CASE , batched=SCREAMING_SNAKE_CASE , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , load_from_cache_file=SCREAMING_SNAKE_CASE ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __UpperCamelCase :List[str] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(SCREAMING_SNAKE_CASE ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(SCREAMING_SNAKE_CASE , padding='''max_length''' , max_length=128 , return_tensors='''pt''' ) return tokenizer.pad(SCREAMING_SNAKE_CASE , padding='''longest''' , return_tensors='''pt''' ) # Instantiate dataloaders. __UpperCamelCase :Union[str, Any] = DataLoader( tokenized_datasets['''train'''] , shuffle=SCREAMING_SNAKE_CASE , collate_fn=SCREAMING_SNAKE_CASE , batch_size=SCREAMING_SNAKE_CASE ) __UpperCamelCase :Dict = DataLoader( tokenized_datasets['''validation'''] , shuffle=SCREAMING_SNAKE_CASE , collate_fn=SCREAMING_SNAKE_CASE , batch_size=SCREAMING_SNAKE_CASE ) return train_dataloader, eval_dataloader def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __UpperCamelCase :int = config['''lr'''] __UpperCamelCase :str = int(config['''num_epochs'''] ) __UpperCamelCase :Any = int(config['''seed'''] ) __UpperCamelCase :Dict = int(config['''batch_size'''] ) __UpperCamelCase :Optional[Any] = args.model_name_or_path set_seed(SCREAMING_SNAKE_CASE ) __UpperCamelCase , __UpperCamelCase :Dict = get_dataloaders(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __UpperCamelCase :Any = AutoModelForSequenceClassification.from_pretrained(SCREAMING_SNAKE_CASE , return_dict=SCREAMING_SNAKE_CASE ) # Instantiate optimizer __UpperCamelCase :List[str] = ( AdamW if accelerator.state.deepspeed_plugin is None or '''optimizer''' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) __UpperCamelCase :Optional[Any] = optimizer_cls(params=model.parameters() , lr=SCREAMING_SNAKE_CASE ) if accelerator.state.deepspeed_plugin is not None: __UpperCamelCase :Dict = accelerator.state.deepspeed_plugin.deepspeed_config[ '''gradient_accumulation_steps''' ] else: __UpperCamelCase :Dict = 1 __UpperCamelCase :Tuple = (len(SCREAMING_SNAKE_CASE ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): __UpperCamelCase :str = get_linear_schedule_with_warmup( optimizer=SCREAMING_SNAKE_CASE , num_warmup_steps=0 , num_training_steps=SCREAMING_SNAKE_CASE , ) else: __UpperCamelCase :Dict = DummyScheduler(SCREAMING_SNAKE_CASE , total_num_steps=SCREAMING_SNAKE_CASE , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :int = accelerator.prepare( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # We need to keep track of how many total steps we have iterated over __UpperCamelCase :List[Any] = 0 # We also need to keep track of the stating epoch so files are named properly __UpperCamelCase :Dict = 0 # Now we train the model __UpperCamelCase :Any = evaluate.load('''glue''' , '''mrpc''' ) __UpperCamelCase :Union[str, Any] = 0 __UpperCamelCase :Optional[int] = {} for epoch in range(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): model.train() for step, batch in enumerate(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = model(**SCREAMING_SNAKE_CASE ) __UpperCamelCase :Tuple = outputs.loss __UpperCamelCase :str = loss / gradient_accumulation_steps accelerator.backward(SCREAMING_SNAKE_CASE ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() __UpperCamelCase :Any = 0 for step, batch in enumerate(SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __UpperCamelCase :Any = model(**SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times __UpperCamelCase , __UpperCamelCase :List[Any] = accelerator.gather( (predictions, batch['''labels''']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(SCREAMING_SNAKE_CASE ) - 1: __UpperCamelCase :List[str] = predictions[: len(eval_dataloader.dataset ) - samples_seen] __UpperCamelCase :Optional[int] = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=SCREAMING_SNAKE_CASE , references=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :Dict = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , SCREAMING_SNAKE_CASE ) __UpperCamelCase :str = eval_metric['''accuracy'''] if best_performance < eval_metric["accuracy"]: __UpperCamelCase :int = eval_metric['''accuracy'''] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f"""Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}""" accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , '''all_results.json''' ) , '''w''' ) as f: json.dump(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = argparse.ArgumentParser(description='''Simple example of training script tracking peak GPU memory usage.''' ) parser.add_argument( '''--model_name_or_path''' , type=SCREAMING_SNAKE_CASE , default='''bert-base-cased''' , help='''Path to pretrained model or model identifier from huggingface.co/models.''' , required=SCREAMING_SNAKE_CASE , ) parser.add_argument( '''--output_dir''' , type=SCREAMING_SNAKE_CASE , default='''.''' , help='''Optional save directory where all checkpoint folders will be stored. Default is the current working directory.''' , ) parser.add_argument( '''--performance_lower_bound''' , type=SCREAMING_SNAKE_CASE , default=SCREAMING_SNAKE_CASE , help='''Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.''' , ) parser.add_argument( '''--num_epochs''' , type=SCREAMING_SNAKE_CASE , default=3 , help='''Number of train epochs.''' , ) __UpperCamelCase :List[str] = parser.parse_args() __UpperCamelCase :Tuple = {'''lr''': 2e-5, '''num_epochs''': args.num_epochs, '''seed''': 42, '''batch_size''': 16} training_function(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
43
1
from scipy.stats import spearmanr import datasets __lowercase = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' __lowercase = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' __lowercase = r'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCamelCase_ ( datasets.Metric ): '''simple docstring''' def UpperCamelCase__ ( self) -> Optional[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''float'''), '''references''': datasets.Value('''float'''), }) , reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] , ) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase=False) -> List[str]: __UpperCamelCase :Optional[Any] = spearmanr(__lowercase , __lowercase) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
43
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __lowercase = logging.get_logger(__name__) __lowercase = { '''SenseTime/deformable-detr''': '''https://huggingface.co/sensetime/deformable-detr/resolve/main/config.json''', # See all Deformable DETR models at https://huggingface.co/models?filter=deformable-detr } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : List[str] = """deformable_detr""" a__ : Union[str, Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , __lowercase=True , __lowercase=None , __lowercase=3 , __lowercase=300 , __lowercase=1_024 , __lowercase=6 , __lowercase=1_024 , __lowercase=8 , __lowercase=6 , __lowercase=1_024 , __lowercase=8 , __lowercase=0.0 , __lowercase=True , __lowercase="relu" , __lowercase=256 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.02 , __lowercase=1.0 , __lowercase=True , __lowercase=False , __lowercase="sine" , __lowercase="resnet50" , __lowercase=True , __lowercase=False , __lowercase=4 , __lowercase=4 , __lowercase=4 , __lowercase=False , __lowercase=300 , __lowercase=False , __lowercase=1 , __lowercase=5 , __lowercase=2 , __lowercase=1 , __lowercase=1 , __lowercase=5 , __lowercase=2 , __lowercase=0.1 , __lowercase=0.25 , __lowercase=False , **__lowercase , ) -> int: if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''') if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''') __UpperCamelCase :str = CONFIG_MAPPING['''resnet'''](out_features=['''stage4''']) elif isinstance(__lowercase , __lowercase): __UpperCamelCase :str = backbone_config.get('''model_type''') __UpperCamelCase :Tuple = CONFIG_MAPPING[backbone_model_type] __UpperCamelCase :Any = config_class.from_dict(__lowercase) __UpperCamelCase :int = use_timm_backbone __UpperCamelCase :Dict = backbone_config __UpperCamelCase :Any = num_channels __UpperCamelCase :Optional[int] = num_queries __UpperCamelCase :Any = max_position_embeddings __UpperCamelCase :str = d_model __UpperCamelCase :Tuple = encoder_ffn_dim __UpperCamelCase :Union[str, Any] = encoder_layers __UpperCamelCase :List[Any] = encoder_attention_heads __UpperCamelCase :Any = decoder_ffn_dim __UpperCamelCase :List[str] = decoder_layers __UpperCamelCase :int = decoder_attention_heads __UpperCamelCase :str = dropout __UpperCamelCase :Any = attention_dropout __UpperCamelCase :int = activation_dropout __UpperCamelCase :List[Any] = activation_function __UpperCamelCase :List[Any] = init_std __UpperCamelCase :List[Any] = init_xavier_std __UpperCamelCase :int = encoder_layerdrop __UpperCamelCase :str = auxiliary_loss __UpperCamelCase :Optional[Any] = position_embedding_type __UpperCamelCase :Union[str, Any] = backbone __UpperCamelCase :Any = use_pretrained_backbone __UpperCamelCase :str = dilation # deformable attributes __UpperCamelCase :Optional[Any] = num_feature_levels __UpperCamelCase :str = encoder_n_points __UpperCamelCase :int = decoder_n_points __UpperCamelCase :Union[str, Any] = two_stage __UpperCamelCase :Optional[Any] = two_stage_num_proposals __UpperCamelCase :Dict = with_box_refine if two_stage is True and with_box_refine is False: raise ValueError('''If two_stage is True, with_box_refine must be True.''') # Hungarian matcher __UpperCamelCase :Optional[int] = class_cost __UpperCamelCase :List[Any] = bbox_cost __UpperCamelCase :str = giou_cost # Loss coefficients __UpperCamelCase :Tuple = mask_loss_coefficient __UpperCamelCase :Tuple = dice_loss_coefficient __UpperCamelCase :int = bbox_loss_coefficient __UpperCamelCase :Any = giou_loss_coefficient __UpperCamelCase :Dict = eos_coefficient __UpperCamelCase :Optional[Any] = focal_alpha __UpperCamelCase :Optional[Any] = disable_custom_kernels super().__init__(is_encoder_decoder=__lowercase , **__lowercase) @property def UpperCamelCase__ ( self) -> int: return self.encoder_attention_heads @property def UpperCamelCase__ ( self) -> int: return self.d_model def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Dict = copy.deepcopy(self.__dict__) if self.backbone_config is not None: __UpperCamelCase :Tuple = self.backbone_config.to_dict() __UpperCamelCase :List[Any] = self.__class__.model_type return output
43
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = { '''google/realm-cc-news-pretrained-embedder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json''' ), '''google/realm-cc-news-pretrained-encoder''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json''' ), '''google/realm-cc-news-pretrained-scorer''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json''' ), '''google/realm-cc-news-pretrained-openqa''': ( '''https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json''' ), '''google/realm-orqa-nq-openqa''': '''https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json''', '''google/realm-orqa-nq-reader''': '''https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json''', '''google/realm-orqa-wq-openqa''': '''https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json''', '''google/realm-orqa-wq-reader''': '''https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json''', # See all REALM models at https://huggingface.co/models?filter=realm } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Any = """realm""" def __init__( self , __lowercase=30_522 , __lowercase=768 , __lowercase=128 , __lowercase=12 , __lowercase=12 , __lowercase=8 , __lowercase=3_072 , __lowercase="gelu_new" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=512 , __lowercase=2 , __lowercase=0.02 , __lowercase=1E-1_2 , __lowercase=256 , __lowercase=10 , __lowercase=1E-3 , __lowercase=5 , __lowercase=320 , __lowercase=13_353_718 , __lowercase=5_000 , __lowercase=1 , __lowercase=0 , __lowercase=2 , **__lowercase , ) -> Tuple: super().__init__(pad_token_id=__lowercase , bos_token_id=__lowercase , eos_token_id=__lowercase , **__lowercase) # Common config __UpperCamelCase :Union[str, Any] = vocab_size __UpperCamelCase :int = max_position_embeddings __UpperCamelCase :List[Any] = hidden_size __UpperCamelCase :List[str] = retriever_proj_size __UpperCamelCase :int = num_hidden_layers __UpperCamelCase :str = num_attention_heads __UpperCamelCase :List[str] = num_candidates __UpperCamelCase :List[Any] = intermediate_size __UpperCamelCase :List[str] = hidden_act __UpperCamelCase :List[Any] = hidden_dropout_prob __UpperCamelCase :int = attention_probs_dropout_prob __UpperCamelCase :List[Any] = initializer_range __UpperCamelCase :Dict = type_vocab_size __UpperCamelCase :List[str] = layer_norm_eps # Reader config __UpperCamelCase :List[Any] = span_hidden_size __UpperCamelCase :int = max_span_width __UpperCamelCase :List[str] = reader_layer_norm_eps __UpperCamelCase :Optional[int] = reader_beam_size __UpperCamelCase :Optional[Any] = reader_seq_len # Retrieval config __UpperCamelCase :str = num_block_records __UpperCamelCase :Any = searcher_beam_size
43
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[Any] = """facebook/bart-large-mnli""" a__ : int = ( """This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which """ """should be the text to classify, and `labels`, which should be the list of labels to use for classification. """ """It returns the most likely label in the list of provided `labels` for the input text.""" ) a__ : Optional[Any] = """text_classifier""" a__ : Any = AutoTokenizer a__ : str = AutoModelForSequenceClassification a__ : str = ["""text""", ["""text"""]] a__ : Optional[int] = ["""text"""] def UpperCamelCase__ ( self) -> Union[str, Any]: super().setup() __UpperCamelCase :int = self.model.config __UpperCamelCase :Optional[Any] = -1 for idx, label in config.idalabel.items(): if label.lower().startswith('''entail'''): __UpperCamelCase :List[Any] = int(__lowercase) if self.entailment_id == -1: raise ValueError('''Could not determine the entailment ID from the model config, please pass it at init.''') def UpperCamelCase__ ( self , __lowercase , __lowercase) -> Union[str, Any]: __UpperCamelCase :Any = labels return self.pre_processor( [text] * len(__lowercase) , [f"""This example is {label}""" for label in labels] , return_tensors='''pt''' , padding='''max_length''' , ) def UpperCamelCase__ ( self , __lowercase) -> Optional[Any]: __UpperCamelCase :List[Any] = outputs.logits __UpperCamelCase :Any = torch.argmax(logits[:, 2]).item() return self._labels[label_id]
43
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, TensorType __lowercase = logging.get_logger(__name__) __lowercase = { '''openai/imagegpt-small''': '''''', '''openai/imagegpt-medium''': '''''', '''openai/imagegpt-large''': '''''', } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Tuple = """imagegpt""" a__ : int = ["""past_key_values"""] a__ : List[str] = { """hidden_size""": """n_embd""", """max_position_embeddings""": """n_positions""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , __lowercase=512 + 1 , __lowercase=32 * 32 , __lowercase=512 , __lowercase=24 , __lowercase=8 , __lowercase=None , __lowercase="quick_gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=1E-5 , __lowercase=0.02 , __lowercase=True , __lowercase=True , __lowercase=False , __lowercase=False , __lowercase=False , **__lowercase , ) -> Optional[Any]: __UpperCamelCase :Optional[int] = vocab_size __UpperCamelCase :Any = n_positions __UpperCamelCase :int = n_embd __UpperCamelCase :Optional[int] = n_layer __UpperCamelCase :Optional[int] = n_head __UpperCamelCase :Optional[Any] = n_inner __UpperCamelCase :str = activation_function __UpperCamelCase :Tuple = resid_pdrop __UpperCamelCase :Dict = embd_pdrop __UpperCamelCase :Optional[Any] = attn_pdrop __UpperCamelCase :Optional[Any] = layer_norm_epsilon __UpperCamelCase :Tuple = initializer_range __UpperCamelCase :Union[str, Any] = scale_attn_weights __UpperCamelCase :Tuple = use_cache __UpperCamelCase :Optional[int] = scale_attn_by_inverse_layer_idx __UpperCamelCase :Tuple = reorder_and_upcast_attn __UpperCamelCase :int = tie_word_embeddings super().__init__(tie_word_embeddings=__lowercase , **__lowercase) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' @property def UpperCamelCase__ ( self) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('''input_ids''', {0: '''batch''', 1: '''sequence'''}), ]) def UpperCamelCase__ ( self , __lowercase , __lowercase = 1 , __lowercase = -1 , __lowercase = False , __lowercase = None , __lowercase = 3 , __lowercase = 32 , __lowercase = 32 , ) -> Mapping[str, Any]: __UpperCamelCase :Optional[int] = self._generate_dummy_images(__lowercase , __lowercase , __lowercase , __lowercase) __UpperCamelCase :List[str] = dict(preprocessor(images=__lowercase , return_tensors=__lowercase)) return inputs
43
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : int = StableUnCLIPImgaImgPipeline a__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS a__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a__ : Optional[Any] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess a__ : int = frozenset([] ) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Tuple = 32 __UpperCamelCase :Optional[int] = embedder_hidden_size # image encoding components __UpperCamelCase :Union[str, Any] = CLIPImageProcessor(crop_size=32 , size=32) torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=__lowercase , projection_dim=__lowercase , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , )) # regular denoising components torch.manual_seed(0) __UpperCamelCase :str = StableUnCLIPImageNormalizer(embedding_dim=__lowercase) __UpperCamelCase :Optional[int] = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''') torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') torch.manual_seed(0) __UpperCamelCase :Dict = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__lowercase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , )) torch.manual_seed(0) __UpperCamelCase :List[Any] = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=__lowercase , layers_per_block=1 , upcast_attention=__lowercase , use_linear_projection=__lowercase , ) torch.manual_seed(0) __UpperCamelCase :Tuple = DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.0_00_85 , beta_end=0.0_12 , prediction_type='''v_prediction''' , set_alpha_to_one=__lowercase , steps_offset=1 , ) torch.manual_seed(0) __UpperCamelCase :List[str] = AutoencoderKL() __UpperCamelCase :Tuple = { # image encoding components '''feature_extractor''': feature_extractor, '''image_encoder''': image_encoder.eval(), # image noising components '''image_normalizer''': image_normalizer.eval(), '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder.eval(), '''unet''': unet.eval(), '''scheduler''': scheduler, '''vae''': vae.eval(), } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0 , __lowercase=True) -> str: if str(__lowercase).startswith('''mps'''): __UpperCamelCase :Union[str, Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :int = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :int = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowercase)).to(__lowercase) if pil_image: __UpperCamelCase :List[Any] = input_image * 0.5 + 0.5 __UpperCamelCase :Optional[Any] = input_image.clamp(0 , 1) __UpperCamelCase :int = input_image.cpu().permute(0 , 2 , 3 , 1).float().numpy() __UpperCamelCase :Optional[Any] = DiffusionPipeline.numpy_to_pil(__lowercase)[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :Dict = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Tuple = self.get_dummy_components() __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline(**__lowercase) __UpperCamelCase :Optional[Any] = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :List[Any] = self.get_dummy_inputs(__lowercase) inputs.update({'''image_embeds''': None}) __UpperCamelCase :Any = sd_pipe(**__lowercase).images __UpperCamelCase :List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCamelCase :List[Any] = np.array([0.38_72, 0.72_24, 0.56_01, 0.47_41, 0.68_72, 0.58_14, 0.46_36, 0.38_67, 0.50_78]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_attention_slicing_forward_pass(test_max_difference=__lowercase) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=__lowercase) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase__ ( self) -> Union[str, Any]: self._test_xformers_attention_forwardGenerator_pass(test_max_difference=__lowercase) @slow @require_torch_gpu class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Union[str, Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :int = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-l-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Dict = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :Dict = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Optional[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Optional[int] = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :List[Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) __UpperCamelCase :Union[str, Any] = pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :Optional[Any] = pipe( __lowercase , '''anime turtle''' , num_inference_steps=2 , output_type='''np''' , ) __UpperCamelCase :int = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
43
1
import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, require_tf, slow if is_tf_available(): import tensorflow as tf if is_tensorflow_text_available(): from transformers.models.bert import TFBertTokenizer __lowercase = ['''bert-base-uncased''', '''bert-base-cased'''] __lowercase = '''hf-internal-testing/tiny-bert-tf-only''' if is_tf_available(): class lowerCamelCase_ ( tf.keras.Model ): '''simple docstring''' def __init__( self , __lowercase) -> List[Any]: super().__init__() __UpperCamelCase :Dict = tokenizer __UpperCamelCase :Any = AutoConfig.from_pretrained(__lowercase) __UpperCamelCase :Optional[Any] = TFAutoModel.from_config(__lowercase) def UpperCamelCase__ ( self , __lowercase) -> List[str]: __UpperCamelCase :str = self.tokenizer(__lowercase) __UpperCamelCase :Union[str, Any] = self.bert(**__lowercase) return out["pooler_output"] @require_tf @require_tensorflow_text class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Dict: super().setUp() __UpperCamelCase :int = [ BertTokenizer.from_pretrained(__lowercase) for checkpoint in (TOKENIZER_CHECKPOINTS * 2) ] # repeat for when fast_bert_tokenizer=false __UpperCamelCase :List[str] = [TFBertTokenizer.from_pretrained(__lowercase) for checkpoint in TOKENIZER_CHECKPOINTS] + [ TFBertTokenizer.from_pretrained(__lowercase , use_fast_bert_tokenizer=__lowercase) for checkpoint in TOKENIZER_CHECKPOINTS ] assert len(self.tokenizers) == len(self.tf_tokenizers) __UpperCamelCase :List[Any] = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] __UpperCamelCase :Any = list(zip(self.test_sentences , self.test_sentences[::-1])) def UpperCamelCase__ ( self) -> Union[str, Any]: for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers): for test_inputs in (self.test_sentences, self.paired_sentences): __UpperCamelCase :Tuple = tokenizer(__lowercase , return_tensors='''tf''' , padding='''longest''') __UpperCamelCase :Optional[int] = tf_tokenizer(__lowercase) for key in python_outputs.keys(): self.assertTrue(tf.reduce_all(python_outputs[key].shape == tf_outputs[key].shape)) self.assertTrue(tf.reduce_all(tf.cast(python_outputs[key] , tf.intaa) == tf_outputs[key])) @slow def UpperCamelCase__ ( self) -> Optional[int]: for tf_tokenizer in self.tf_tokenizers: __UpperCamelCase :Optional[Any] = tf_tokenizer(self.paired_sentences) __UpperCamelCase :Union[str, Any] = tf_tokenizer( text=[sentence[0] for sentence in self.paired_sentences] , text_pair=[sentence[1] for sentence in self.paired_sentences] , ) for key in merged_outputs.keys(): self.assertTrue(tf.reduce_all(tf.cast(merged_outputs[key] , tf.intaa) == separated_outputs[key])) @slow def UpperCamelCase__ ( self) -> str: for tf_tokenizer in self.tf_tokenizers: __UpperCamelCase :int = tf.function(__lowercase) for test_inputs in (self.test_sentences, self.paired_sentences): __UpperCamelCase :Any = tf.constant(__lowercase) __UpperCamelCase :List[str] = compiled_tokenizer(__lowercase) __UpperCamelCase :str = tf_tokenizer(__lowercase) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key])) @slow def UpperCamelCase__ ( self) -> List[Any]: for tf_tokenizer in self.tf_tokenizers: __UpperCamelCase :Union[str, Any] = ModelToSave(tokenizer=__lowercase) __UpperCamelCase :Dict = tf.convert_to_tensor(self.test_sentences) __UpperCamelCase :int = model(__lowercase) # Build model with some sample inputs with TemporaryDirectory() as tempdir: __UpperCamelCase :int = Path(__lowercase) / '''saved.model''' model.save(__lowercase) __UpperCamelCase :List[Any] = tf.keras.models.load_model(__lowercase) __UpperCamelCase :int = loaded_model(__lowercase) # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertLessEqual(tf.reduce_max(tf.abs(out - loaded_output)) , 1E-5)
43
import numpy as np def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 1e-12 , SCREAMING_SNAKE_CASE = 100 , ): '''simple docstring''' assert np.shape(SCREAMING_SNAKE_CASE )[0] == np.shape(SCREAMING_SNAKE_CASE )[1] # Ensure proper dimensionality. assert np.shape(SCREAMING_SNAKE_CASE )[0] == np.shape(SCREAMING_SNAKE_CASE )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(SCREAMING_SNAKE_CASE ) == np.iscomplexobj(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = np.iscomplexobj(SCREAMING_SNAKE_CASE ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(SCREAMING_SNAKE_CASE , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. __UpperCamelCase :str = False __UpperCamelCase :int = 0 __UpperCamelCase :Optional[Any] = 0 __UpperCamelCase :Union[str, Any] = 1e12 while not convergence: # Multiple matrix by the vector. __UpperCamelCase :List[str] = np.dot(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Normalize the resulting output vector. __UpperCamelCase :Tuple = w / np.linalg.norm(SCREAMING_SNAKE_CASE ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) __UpperCamelCase :int = vector.conj().T if is_complex else vector.T __UpperCamelCase :Optional[int] = np.dot(SCREAMING_SNAKE_CASE , np.dot(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) # Check convergence. __UpperCamelCase :Optional[Any] = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: __UpperCamelCase :Dict = True __UpperCamelCase :List[Any] = lambda_ if is_complex: __UpperCamelCase :Tuple = np.real(lambda_ ) return lambda_, vector def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :int = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) __UpperCamelCase :Optional[Any] = np.array([41, 4, 20] ) __UpperCamelCase :Any = real_input_matrix.astype(np.complexaaa ) __UpperCamelCase :Dict = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T __UpperCamelCase :Optional[int] = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": __UpperCamelCase :Any = real_input_matrix __UpperCamelCase :int = real_vector elif problem_type == "complex": __UpperCamelCase :Tuple = complex_input_matrix __UpperCamelCase :Optional[Any] = complex_vector # Our implementation. __UpperCamelCase , __UpperCamelCase :Dict = power_iteration(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). __UpperCamelCase , __UpperCamelCase :List[Any] = np.linalg.eigh(SCREAMING_SNAKE_CASE ) # Last eigenvalue is the maximum one. __UpperCamelCase :List[Any] = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. __UpperCamelCase :str = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(SCREAMING_SNAKE_CASE ) - np.abs(SCREAMING_SNAKE_CASE ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
43
1
import re from filelock import FileLock try: import nltk __lowercase = True except (ImportError, ModuleNotFoundError): __lowercase = False if NLTK_AVAILABLE: with FileLock('''.lock''') as lock: nltk.download('''punkt''', quiet=True) def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' re.sub('''<n>''' , '''''' , SCREAMING_SNAKE_CASE ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(SCREAMING_SNAKE_CASE ) )
43
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __lowercase = logging.get_logger(__name__) __lowercase = {'''tokenizer_file''': '''tokenizer.json'''} __lowercase = { '''tokenizer_file''': { '''bigscience/tokenizer''': '''https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json''', '''bigscience/bloom-560m''': '''https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json''', '''bigscience/bloom-1b1''': '''https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json''', '''bigscience/bloom-1b7''': '''https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json''', '''bigscience/bloom-3b''': '''https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json''', '''bigscience/bloom-7b1''': '''https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json''', '''bigscience/bloom''': '''https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json''', }, } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : int = VOCAB_FILES_NAMES a__ : Tuple = PRETRAINED_VOCAB_FILES_MAP a__ : List[str] = ["""input_ids""", """attention_mask"""] a__ : int = None def __init__( self , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase="<unk>" , __lowercase="<s>" , __lowercase="</s>" , __lowercase="<pad>" , __lowercase=False , __lowercase=False , **__lowercase , ) -> List[str]: super().__init__( __lowercase , __lowercase , tokenizer_file=__lowercase , unk_token=__lowercase , bos_token=__lowercase , eos_token=__lowercase , pad_token=__lowercase , add_prefix_space=__lowercase , clean_up_tokenization_spaces=__lowercase , **__lowercase , ) __UpperCamelCase :int = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get('''add_prefix_space''' , __lowercase) != add_prefix_space: __UpperCamelCase :Any = getattr(__lowercase , pre_tok_state.pop('''type''')) __UpperCamelCase :str = add_prefix_space __UpperCamelCase :List[str] = pre_tok_class(**__lowercase) __UpperCamelCase :Tuple = add_prefix_space def UpperCamelCase__ ( self , *__lowercase , **__lowercase) -> BatchEncoding: __UpperCamelCase :Tuple = kwargs.get('''is_split_into_words''' , __lowercase) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with""" ''' pretokenized inputs.''') return super()._batch_encode_plus(*__lowercase , **__lowercase) def UpperCamelCase__ ( self , *__lowercase , **__lowercase) -> BatchEncoding: __UpperCamelCase :List[str] = kwargs.get('''is_split_into_words''' , __lowercase) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with""" ''' pretokenized inputs.''') return super()._encode_plus(*__lowercase , **__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase = None) -> Tuple[str]: __UpperCamelCase :Optional[Any] = self._tokenizer.model.save(__lowercase , name=__lowercase) return tuple(__lowercase) def UpperCamelCase__ ( self , __lowercase) -> List[int]: __UpperCamelCase :str = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(__lowercase , add_special_tokens=__lowercase) + [self.eos_token_id]) if len(__lowercase) > self.model_max_length: __UpperCamelCase :Any = input_ids[-self.model_max_length :] return input_ids
43
1
import os import textwrap import pyarrow as pa import pytest from datasets import ClassLabel, Features, Image from datasets.packaged_modules.csv.csv import Csv from ..utils import require_pil @pytest.fixture def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Dict = tmp_path / '''file.csv''' __UpperCamelCase :Any = textwrap.dedent( '''\ header1,header2 1,2 10,20 ''' ) with open(SCREAMING_SNAKE_CASE , '''w''' ) as f: f.write(SCREAMING_SNAKE_CASE ) return str(SCREAMING_SNAKE_CASE ) @pytest.fixture def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[str] = tmp_path / '''malformed_file.csv''' __UpperCamelCase :Tuple = textwrap.dedent( '''\ header1,header2 1,2 10,20, ''' ) with open(SCREAMING_SNAKE_CASE , '''w''' ) as f: f.write(SCREAMING_SNAKE_CASE ) return str(SCREAMING_SNAKE_CASE ) @pytest.fixture def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = tmp_path / '''csv_with_image.csv''' __UpperCamelCase :str = textwrap.dedent( f"""\ image {image_file} """ ) with open(SCREAMING_SNAKE_CASE , '''w''' ) as f: f.write(SCREAMING_SNAKE_CASE ) return str(SCREAMING_SNAKE_CASE ) @pytest.fixture def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = tmp_path / '''csv_with_label.csv''' __UpperCamelCase :Tuple = textwrap.dedent( '''\ label good bad good ''' ) with open(SCREAMING_SNAKE_CASE , '''w''' ) as f: f.write(SCREAMING_SNAKE_CASE ) return str(SCREAMING_SNAKE_CASE ) @pytest.fixture def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = tmp_path / '''csv_with_int_list.csv''' __UpperCamelCase :Optional[Any] = textwrap.dedent( '''\ int_list 1 2 3 4 5 6 7 8 9 ''' ) with open(SCREAMING_SNAKE_CASE , '''w''' ) as f: f.write(SCREAMING_SNAKE_CASE ) return str(SCREAMING_SNAKE_CASE ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = Csv() __UpperCamelCase :Optional[int] = csv._generate_tables([[csv_file, malformed_csv_file]] ) with pytest.raises(SCREAMING_SNAKE_CASE , match='''Error tokenizing data''' ): for _ in generator: pass assert any( record.levelname == '''ERROR''' and '''Failed to read file''' in record.message and os.path.basename(SCREAMING_SNAKE_CASE ) in record.message for record in caplog.records ) @require_pil def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' with open(SCREAMING_SNAKE_CASE , encoding='''utf-8''' ) as f: __UpperCamelCase :List[Any] = f.read().splitlines()[1] __UpperCamelCase :List[Any] = Csv(encoding='''utf-8''' , features=Features({'''image''': Image()} ) ) __UpperCamelCase :Optional[int] = csv._generate_tables([[csv_file_with_image]] ) __UpperCamelCase :Optional[Any] = pa.concat_tables([table for _, table in generator] ) assert pa_table.schema.field('''image''' ).type == Image()() __UpperCamelCase :Tuple = pa_table.to_pydict()['''image'''] assert generated_content == [{"path": image_file, "bytes": None}] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' with open(SCREAMING_SNAKE_CASE , encoding='''utf-8''' ) as f: __UpperCamelCase :List[Any] = f.read().splitlines()[1:] __UpperCamelCase :Optional[int] = Csv(encoding='''utf-8''' , features=Features({'''label''': ClassLabel(names=['''good''', '''bad'''] )} ) ) __UpperCamelCase :List[Any] = csv._generate_tables([[csv_file_with_label]] ) __UpperCamelCase :List[str] = pa.concat_tables([table for _, table in generator] ) assert pa_table.schema.field('''label''' ).type == ClassLabel(names=['''good''', '''bad'''] )() __UpperCamelCase :List[str] = pa_table.to_pydict()['''label'''] assert generated_content == [ClassLabel(names=['''good''', '''bad'''] ).straint(SCREAMING_SNAKE_CASE ) for label in labels] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :int = Csv(encoding='''utf-8''' , sep=''',''' , converters={'''int_list''': lambda SCREAMING_SNAKE_CASE : [int(SCREAMING_SNAKE_CASE ) for i in x.split()]} ) __UpperCamelCase :Union[str, Any] = csv._generate_tables([[csv_file_with_int_list]] ) __UpperCamelCase :Optional[int] = pa.concat_tables([table for _, table in generator] ) assert pa.types.is_list(pa_table.schema.field('''int_list''' ).type ) __UpperCamelCase :str = pa_table.to_pydict()['''int_list'''] assert generated_content == [[1, 2, 3], [4, 5, 6], [7, 8, 9]]
43
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = {'''ctrl''': '''https://huggingface.co/ctrl/resolve/main/config.json'''} class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : str = """ctrl""" a__ : Dict = ["""past_key_values"""] a__ : Tuple = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , __lowercase=246_534 , __lowercase=256 , __lowercase=1_280 , __lowercase=8_192 , __lowercase=48 , __lowercase=16 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=1E-6 , __lowercase=0.02 , __lowercase=True , **__lowercase , ) -> List[Any]: __UpperCamelCase :List[str] = vocab_size __UpperCamelCase :Optional[Any] = n_positions __UpperCamelCase :Dict = n_embd __UpperCamelCase :Dict = n_layer __UpperCamelCase :List[Any] = n_head __UpperCamelCase :int = dff __UpperCamelCase :Union[str, Any] = resid_pdrop __UpperCamelCase :Optional[int] = embd_pdrop __UpperCamelCase :List[Any] = layer_norm_epsilon __UpperCamelCase :Dict = initializer_range __UpperCamelCase :Any = use_cache super().__init__(**__lowercase)
43
1
import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class lowerCamelCase_ ( nn.Module ): '''simple docstring''' def __init__( self) -> Any: super().__init__() __UpperCamelCase :Optional[int] = nn.Linear(3 , 4) __UpperCamelCase :List[Any] = nn.BatchNormad(4) __UpperCamelCase :List[Any] = nn.Linear(4 , 5) def UpperCamelCase__ ( self , __lowercase) -> Dict: return self.lineara(self.batchnorm(self.lineara(__lowercase))) class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :Dict = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(__lowercase , model.state_dict()) __UpperCamelCase :str = os.path.join(__lowercase , '''index.json''') self.assertTrue(os.path.isfile(__lowercase)) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: __UpperCamelCase :Tuple = os.path.join(__lowercase , f"""{key}.dat""") self.assertTrue(os.path.isfile(__lowercase)) # TODO: add tests on the fact weights are properly loaded def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :List[str] = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: __UpperCamelCase :List[Any] = torch.randn(2 , 3 , dtype=__lowercase) with TemporaryDirectory() as tmp_dir: __UpperCamelCase :Tuple = offload_weight(__lowercase , '''weight''' , __lowercase , {}) __UpperCamelCase :Optional[int] = os.path.join(__lowercase , '''weight.dat''') self.assertTrue(os.path.isfile(__lowercase)) self.assertDictEqual(__lowercase , {'''weight''': {'''shape''': [2, 3], '''dtype''': str(__lowercase).split('''.''')[1]}}) __UpperCamelCase :List[str] = load_offloaded_weight(__lowercase , index['''weight''']) self.assertTrue(torch.equal(__lowercase , __lowercase)) def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :Optional[Any] = ModelForTest() __UpperCamelCase :Any = model.state_dict() __UpperCamelCase :str = {k: v for k, v in state_dict.items() if '''linear2''' not in k} __UpperCamelCase :str = {k: v for k, v in state_dict.items() if '''linear2''' in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(__lowercase , __lowercase) __UpperCamelCase :Optional[int] = OffloadedWeightsLoader(state_dict=__lowercase , save_folder=__lowercase) # Every key is there with the right value self.assertEqual(sorted(__lowercase) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(__lowercase , weight_map[key])) __UpperCamelCase :Union[str, Any] = {k: v for k, v in state_dict.items() if '''weight''' in k} __UpperCamelCase :Optional[Any] = {k: v for k, v in state_dict.items() if '''weight''' not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(__lowercase , __lowercase) __UpperCamelCase :List[str] = OffloadedWeightsLoader(state_dict=__lowercase , save_folder=__lowercase) # Every key is there with the right value self.assertEqual(sorted(__lowercase) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(__lowercase , weight_map[key])) with TemporaryDirectory() as tmp_dir: offload_state_dict(__lowercase , __lowercase) # Duplicates are removed __UpperCamelCase :Tuple = OffloadedWeightsLoader(state_dict=__lowercase , save_folder=__lowercase) # Every key is there with the right value self.assertEqual(sorted(__lowercase) , sorted(state_dict.keys())) for key, param in state_dict.items(): self.assertTrue(torch.allclose(__lowercase , weight_map[key])) def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :List[Any] = {'''a.1''': 0, '''a.10''': 1, '''a.2''': 2} __UpperCamelCase :Optional[Any] = extract_submodules_state_dict(__lowercase , ['''a.1''', '''a.2''']) self.assertDictEqual(__lowercase , {'''a.1''': 0, '''a.2''': 2}) __UpperCamelCase :List[Any] = {'''a.1.a''': 0, '''a.10.a''': 1, '''a.2.a''': 2} __UpperCamelCase :Optional[int] = extract_submodules_state_dict(__lowercase , ['''a.1''', '''a.2''']) self.assertDictEqual(__lowercase , {'''a.1.a''': 0, '''a.2.a''': 2})
43
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class lowerCamelCase_ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : str = TextToVideoSDPipeline a__ : Union[str, Any] = TEXT_TO_IMAGE_PARAMS a__ : Tuple = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. a__ : int = frozenset( [ """num_inference_steps""", """generator""", """latents""", """return_dict""", """callback""", """callback_steps""", ] ) def UpperCamelCase__ ( self) -> Optional[Any]: torch.manual_seed(0) __UpperCamelCase :str = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''DownBlock3D''') , up_block_types=('''UpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''') , cross_attention_dim=32 , attention_head_dim=4 , ) __UpperCamelCase :Optional[int] = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=__lowercase , set_alpha_to_one=__lowercase , ) torch.manual_seed(0) __UpperCamelCase :Optional[int] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0) __UpperCamelCase :Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act='''gelu''' , projection_dim=512 , ) __UpperCamelCase :Optional[Any] = CLIPTextModel(__lowercase) __UpperCamelCase :Optional[int] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') __UpperCamelCase :Union[str, Any] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0) -> Optional[int]: if str(__lowercase).startswith('''mps'''): __UpperCamelCase :List[Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :Tuple = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :int = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Optional[int] = self.get_dummy_components() __UpperCamelCase :Dict = TextToVideoSDPipeline(**__lowercase) __UpperCamelCase :Any = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :Optional[Any] = self.get_dummy_inputs(__lowercase) __UpperCamelCase :int = '''np''' __UpperCamelCase :List[str] = sd_pipe(**__lowercase).frames __UpperCamelCase :Optional[Any] = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) __UpperCamelCase :str = np.array([1_58.0, 1_60.0, 1_53.0, 1_25.0, 1_00.0, 1_21.0, 1_11.0, 93.0, 1_13.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase__ ( self) -> Tuple: self._test_attention_slicing_forward_pass(test_mean_pixel_difference=__lowercase , expected_max_diff=3E-3) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase__ ( self) -> Optional[int]: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__lowercase , expected_max_diff=1E-2) @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCamelCase__ ( self) -> Union[str, Any]: pass @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCamelCase__ ( self) -> Dict: pass @unittest.skip(reason='''`num_images_per_prompt` argument is not supported for this pipeline.''') def UpperCamelCase__ ( self) -> str: pass def UpperCamelCase__ ( self) -> List[str]: return super().test_progress_bar() @slow @skip_mps class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Union[str, Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy''') __UpperCamelCase :List[str] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''') __UpperCamelCase :Union[str, Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) __UpperCamelCase :str = pipe.to('''cuda''') __UpperCamelCase :Optional[Any] = '''Spiderman is surfing''' __UpperCamelCase :Union[str, Any] = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :List[Any] = pipe(__lowercase , generator=__lowercase , num_inference_steps=25 , output_type='''pt''').frames __UpperCamelCase :Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video).mean() < 5E-2 def UpperCamelCase__ ( self) -> int: __UpperCamelCase :str = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy''') __UpperCamelCase :Union[str, Any] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''') __UpperCamelCase :str = pipe.to('''cuda''') __UpperCamelCase :Union[str, Any] = '''Spiderman is surfing''' __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :List[Any] = pipe(__lowercase , generator=__lowercase , num_inference_steps=2 , output_type='''pt''').frames __UpperCamelCase :Optional[Any] = video_frames.cpu().numpy() assert np.abs(expected_video - video).mean() < 5E-2
43
1
import os import random import sys from . import cryptomath_module as cryptomath from . import rabin_miller __lowercase = 3 def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' print('''Generating primitive root of p''' ) while True: __UpperCamelCase :Tuple = random.randrange(3 , SCREAMING_SNAKE_CASE ) if pow(SCREAMING_SNAKE_CASE , 2 , SCREAMING_SNAKE_CASE ) == 1: continue if pow(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) == 1: continue return g def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' print('''Generating prime p...''' ) __UpperCamelCase :List[Any] = rabin_miller.generate_large_prime(SCREAMING_SNAKE_CASE ) # select large prime number. __UpperCamelCase :Optional[Any] = primitive_root(SCREAMING_SNAKE_CASE ) # one primitive root on modulo p. __UpperCamelCase :Optional[int] = random.randrange(3 , SCREAMING_SNAKE_CASE ) # private_key -> have to be greater than 2 for safety. __UpperCamelCase :List[Any] = cryptomath.find_mod_inverse(pow(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) , SCREAMING_SNAKE_CASE ) __UpperCamelCase :int = (key_size, e_a, e_a, p) __UpperCamelCase :str = (key_size, d) return public_key, private_key def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' if os.path.exists(f"""{name}_pubkey.txt""" ) or os.path.exists(f"""{name}_privkey.txt""" ): print('''\nWARNING:''' ) print( f"""\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n""" '''Use a different name or delete these files and re-run this program.''' ) sys.exit() __UpperCamelCase , __UpperCamelCase :str = generate_key(SCREAMING_SNAKE_CASE ) print(f"""\nWriting public key to file {name}_pubkey.txt...""" ) with open(f"""{name}_pubkey.txt""" , '''w''' ) as fo: fo.write(f"""{public_key[0]},{public_key[1]},{public_key[2]},{public_key[3]}""" ) print(f"""Writing private key to file {name}_privkey.txt...""" ) with open(f"""{name}_privkey.txt""" , '''w''' ) as fo: fo.write(f"""{private_key[0]},{private_key[1]}""" ) def lowerCamelCase ( ): '''simple docstring''' print('''Making key files...''' ) make_key_files('''elgamal''' , 2_048 ) print('''Key files generation successful''' ) if __name__ == "__main__": main()
43
def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = [0 for i in range(len(SCREAMING_SNAKE_CASE ) )] # initialize interval's left pointer and right pointer __UpperCamelCase , __UpperCamelCase :str = 0, 0 for i in range(1 , len(SCREAMING_SNAKE_CASE ) ): # case when current index is inside the interval if i <= right_pointer: __UpperCamelCase :Union[str, Any] = min(right_pointer - i + 1 , z_result[i - left_pointer] ) __UpperCamelCase :Tuple = min_edge while go_next(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: __UpperCamelCase , __UpperCamelCase :Union[str, Any] = i, i + z_result[i] - 1 return z_result def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' return i + z_result[i] < len(SCREAMING_SNAKE_CASE ) and s[z_result[i]] == s[i + z_result[i]] def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string __UpperCamelCase :Tuple = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(SCREAMING_SNAKE_CASE ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
43
1
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : Tuple = StableDiffusionInstructPixaPixPipeline a__ : Any = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""height""", """width""", """cross_attention_kwargs"""} a__ : Any = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS a__ : Any = IMAGE_TO_IMAGE_IMAGE_PARAMS a__ : Any = IMAGE_TO_IMAGE_IMAGE_PARAMS def UpperCamelCase__ ( self) -> Optional[int]: torch.manual_seed(0) __UpperCamelCase :Optional[Any] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=8 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) __UpperCamelCase :Dict = PNDMScheduler(skip_prk_steps=__lowercase) torch.manual_seed(0) __UpperCamelCase :Tuple = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) __UpperCamelCase :List[Any] = CLIPTextModel(__lowercase) __UpperCamelCase :str = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') __UpperCamelCase :List[str] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0) -> Union[str, Any]: __UpperCamelCase :Optional[Any] = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowercase)).to(__lowercase) __UpperCamelCase :Union[str, Any] = image.cpu().permute(0 , 2 , 3 , 1)[0] __UpperCamelCase :Tuple = Image.fromarray(np.uinta(__lowercase)).convert('''RGB''') if str(__lowercase).startswith('''mps'''): __UpperCamelCase :List[Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :Any = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''image_guidance_scale''': 1, '''output_type''': '''numpy''', } return inputs def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :List[str] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Dict = self.get_dummy_components() __UpperCamelCase :List[Any] = StableDiffusionInstructPixaPixPipeline(**__lowercase) __UpperCamelCase :List[str] = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :List[str] = self.get_dummy_inputs(__lowercase) __UpperCamelCase :int = sd_pipe(**__lowercase).images __UpperCamelCase :int = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCamelCase :List[Any] = np.array([0.75_26, 0.37_50, 0.45_47, 0.61_17, 0.58_66, 0.50_16, 0.43_27, 0.56_42, 0.48_15]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :int = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Dict = self.get_dummy_components() __UpperCamelCase :int = StableDiffusionInstructPixaPixPipeline(**__lowercase) __UpperCamelCase :List[Any] = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :Any = self.get_dummy_inputs(__lowercase) __UpperCamelCase :int = '''french fries''' __UpperCamelCase :Tuple = sd_pipe(**__lowercase , negative_prompt=__lowercase) __UpperCamelCase :Dict = output.images __UpperCamelCase :Any = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCamelCase :List[Any] = np.array([0.75_11, 0.36_42, 0.45_53, 0.62_36, 0.57_97, 0.50_13, 0.43_43, 0.56_11, 0.48_31]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :int = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :str = self.get_dummy_components() __UpperCamelCase :str = StableDiffusionInstructPixaPixPipeline(**__lowercase) __UpperCamelCase :List[Any] = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :List[Any] = self.get_dummy_inputs(__lowercase) __UpperCamelCase :Tuple = [inputs['''prompt''']] * 2 __UpperCamelCase :Dict = np.array(inputs['''image''']).astype(np.floataa) / 2_55.0 __UpperCamelCase :Tuple = torch.from_numpy(__lowercase).unsqueeze(0).to(__lowercase) __UpperCamelCase :int = image / 2 + 0.5 __UpperCamelCase :List[str] = image.permute(0 , 3 , 1 , 2) __UpperCamelCase :Dict = image.repeat(2 , 1 , 1 , 1) __UpperCamelCase :Tuple = sd_pipe(**__lowercase).images __UpperCamelCase :Optional[int] = image[-1, -3:, -3:, -1] assert image.shape == (2, 32, 32, 3) __UpperCamelCase :List[str] = np.array([0.58_12, 0.57_48, 0.52_22, 0.59_08, 0.56_95, 0.71_74, 0.68_04, 0.55_23, 0.55_79]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :List[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Union[str, Any] = self.get_dummy_components() __UpperCamelCase :Union[str, Any] = EulerAncestralDiscreteScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''') __UpperCamelCase :Any = StableDiffusionInstructPixaPixPipeline(**__lowercase) __UpperCamelCase :Union[str, Any] = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :Optional[Any] = self.get_dummy_inputs(__lowercase) __UpperCamelCase :int = sd_pipe(**__lowercase).images __UpperCamelCase :List[Any] = image[0, -3:, -3:, -1] __UpperCamelCase :str = [round(__lowercase , 4) for x in image_slice.flatten().tolist()] print(''','''.join([str(__lowercase) for x in slice])) assert image.shape == (1, 32, 32, 3) __UpperCamelCase :str = np.array([0.74_17, 0.38_42, 0.47_32, 0.57_76, 0.58_91, 0.51_39, 0.40_52, 0.56_73, 0.49_86]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> Union[str, Any]: super().test_inference_batch_single_identical(expected_max_diff=3E-3) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :int = self.get_dummy_components() __UpperCamelCase :Any = StableDiffusionInstructPixaPixPipeline(**__lowercase) __UpperCamelCase :Union[str, Any] = VaeImageProcessor(do_resize=__lowercase , do_normalize=__lowercase) __UpperCamelCase :Dict = pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :Any = pipe(**self.get_dummy_inputs_by_type(__lowercase , input_image_type='''pt'''))[0] __UpperCamelCase :int = components['''vae'''] __UpperCamelCase :Dict = self.get_dummy_inputs_by_type(__lowercase , input_image_type='''pt''') for image_param in self.image_latents_params: if image_param in inputs.keys(): __UpperCamelCase :str = vae.encode(inputs[image_param]).latent_dist.mode() __UpperCamelCase :Tuple = pipe(**__lowercase)[0] __UpperCamelCase :Optional[Any] = np.abs(out - out_latents_inputs).max() self.assertLess(__lowercase , 1E-4 , '''passing latents as image input generate different result from passing image''') @slow @require_torch_gpu class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Any: super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self , __lowercase=0) -> List[Any]: __UpperCamelCase :Optional[int] = torch.manual_seed(__lowercase) __UpperCamelCase :Optional[Any] = load_image( '''https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg''') __UpperCamelCase :Any = { '''prompt''': '''turn him into a cyborg''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''image_guidance_scale''': 1.0, '''output_type''': '''numpy''', } return inputs def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Any = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' , safety_checker=__lowercase) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) pipe.enable_attention_slicing() __UpperCamelCase :Union[str, Any] = self.get_inputs() __UpperCamelCase :Dict = pipe(**__lowercase).images __UpperCamelCase :Union[str, Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __UpperCamelCase :Any = np.array([0.59_02, 0.60_15, 0.60_27, 0.59_83, 0.60_92, 0.60_61, 0.57_65, 0.57_85, 0.55_55]) assert np.abs(expected_slice - image_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :int = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' , safety_checker=__lowercase) __UpperCamelCase :Dict = LMSDiscreteScheduler.from_config(pipe.scheduler.config) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) pipe.enable_attention_slicing() __UpperCamelCase :Any = self.get_inputs() __UpperCamelCase :Union[str, Any] = pipe(**__lowercase).images __UpperCamelCase :int = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __UpperCamelCase :List[str] = np.array([0.65_78, 0.68_17, 0.69_72, 0.67_61, 0.68_56, 0.69_16, 0.64_28, 0.65_16, 0.63_01]) assert np.abs(expected_slice - image_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' , safety_checker=__lowercase) __UpperCamelCase :List[Any] = DDIMScheduler.from_config(pipe.scheduler.config) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) pipe.enable_attention_slicing() __UpperCamelCase :Any = self.get_inputs() __UpperCamelCase :Dict = pipe(**__lowercase).images __UpperCamelCase :List[Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __UpperCamelCase :List[str] = np.array([0.38_28, 0.38_34, 0.38_18, 0.37_92, 0.38_65, 0.37_52, 0.37_92, 0.38_47, 0.37_53]) assert np.abs(expected_slice - image_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :Dict = 0 def callback_fn(__lowercase , __lowercase , __lowercase) -> None: __UpperCamelCase :Any = True nonlocal number_of_steps number_of_steps += 1 if step == 1: __UpperCamelCase :str = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) __UpperCamelCase :Optional[int] = latents[0, -3:, -3:, -1] __UpperCamelCase :Tuple = np.array([-0.24_63, -0.46_44, -0.97_56, 1.51_76, 1.44_14, 0.78_66, 0.98_97, 0.85_21, 0.79_83]) assert np.abs(latents_slice.flatten() - expected_slice).max() < 5E-2 elif step == 2: __UpperCamelCase :Optional[Any] = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) __UpperCamelCase :str = latents[0, -3:, -3:, -1] __UpperCamelCase :int = np.array([-0.26_44, -0.46_26, -0.96_53, 1.51_76, 1.45_51, 0.76_86, 0.98_05, 0.84_52, 0.81_15]) assert np.abs(latents_slice.flatten() - expected_slice).max() < 5E-2 __UpperCamelCase :Tuple = False __UpperCamelCase :Optional[Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' , safety_checker=__lowercase , torch_dtype=torch.floataa) __UpperCamelCase :List[str] = pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) pipe.enable_attention_slicing() __UpperCamelCase :Dict = self.get_inputs() pipe(**__lowercase , callback=__lowercase , callback_steps=1) assert callback_fn.has_been_called assert number_of_steps == 3 def UpperCamelCase__ ( self) -> str: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __UpperCamelCase :Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( '''timbrooks/instruct-pix2pix''' , safety_checker=__lowercase , torch_dtype=torch.floataa) __UpperCamelCase :Tuple = pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) pipe.enable_attention_slicing(1) pipe.enable_sequential_cpu_offload() __UpperCamelCase :Any = self.get_inputs() __UpperCamelCase :Tuple = pipe(**__lowercase) __UpperCamelCase :Union[str, Any] = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 10**9 def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Any = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 __UpperCamelCase :str = inputs['''image'''].resize((504, 504)) __UpperCamelCase :str = '''timbrooks/instruct-pix2pix''' __UpperCamelCase :str = StableDiffusionInstructPixaPixPipeline.from_pretrained( __lowercase , safety_checker=__lowercase , ) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) pipe.enable_attention_slicing() __UpperCamelCase :int = pipe(**__lowercase) __UpperCamelCase :Tuple = output.images[0] __UpperCamelCase :Dict = image[255:258, 383:386, -1] assert image.shape == (504, 504, 3) __UpperCamelCase :List[str] = np.array([0.27_26, 0.25_29, 0.26_64, 0.26_55, 0.26_41, 0.26_42, 0.25_91, 0.26_49, 0.25_90]) assert np.abs(image_slice.flatten() - expected_slice).max() < 5E-3
43
import math from typing import Any, Callable, List, Optional, Tuple, Union import numpy as np import torch from ...models import TaFilmDecoder from ...schedulers import DDPMScheduler from ...utils import is_onnx_available, logging, randn_tensor if is_onnx_available(): from ..onnx_utils import OnnxRuntimeModel from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline from .continous_encoder import SpectrogramContEncoder from .notes_encoder import SpectrogramNotesEncoder __lowercase = logging.get_logger(__name__) # pylint: disable=invalid-name __lowercase = 256 class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Tuple = ["""melgan"""] def __init__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> None: super().__init__() # From MELGAN __UpperCamelCase :int = math.log(1E-5) # Matches MelGAN training. __UpperCamelCase :int = 4.0 # Largest value for most examples __UpperCamelCase :str = 128 self.register_modules( notes_encoder=__lowercase , continuous_encoder=__lowercase , decoder=__lowercase , scheduler=__lowercase , melgan=__lowercase , ) def UpperCamelCase__ ( self , __lowercase , __lowercase=(-1.0, 1.0) , __lowercase=False) -> Dict: __UpperCamelCase , __UpperCamelCase :str = output_range if clip: __UpperCamelCase :Union[str, Any] = torch.clip(__lowercase , self.min_value , self.max_value) # Scale to [0, 1]. __UpperCamelCase :Union[str, Any] = (features - self.min_value) / (self.max_value - self.min_value) # Scale to [min_out, max_out]. return zero_one * (max_out - min_out) + min_out def UpperCamelCase__ ( self , __lowercase , __lowercase=(-1.0, 1.0) , __lowercase=False) -> Optional[int]: __UpperCamelCase , __UpperCamelCase :int = input_range __UpperCamelCase :Optional[int] = torch.clip(__lowercase , __lowercase , __lowercase) if clip else outputs # Scale to [0, 1]. __UpperCamelCase :List[str] = (outputs - min_out) / (max_out - min_out) # Scale to [self.min_value, self.max_value]. return zero_one * (self.max_value - self.min_value) + self.min_value def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase) -> List[Any]: __UpperCamelCase :List[str] = input_tokens > 0 __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self.notes_encoder( encoder_input_tokens=__lowercase , encoder_inputs_mask=__lowercase) __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self.continuous_encoder( encoder_inputs=__lowercase , encoder_inputs_mask=__lowercase) return [(tokens_encoded, tokens_mask), (continuous_encoded, continuous_mask)] def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase) -> str: __UpperCamelCase :Optional[int] = noise_time if not torch.is_tensor(__lowercase): __UpperCamelCase :str = torch.tensor([timesteps] , dtype=torch.long , device=input_tokens.device) elif torch.is_tensor(__lowercase) and len(timesteps.shape) == 0: __UpperCamelCase :Dict = timesteps[None].to(input_tokens.device) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML __UpperCamelCase :List[str] = timesteps * torch.ones(input_tokens.shape[0] , dtype=timesteps.dtype , device=timesteps.device) __UpperCamelCase :Tuple = self.decoder( encodings_and_masks=__lowercase , decoder_input_tokens=__lowercase , decoder_noise_time=__lowercase) return logits @torch.no_grad() def __call__( self , __lowercase , __lowercase = None , __lowercase = 100 , __lowercase = True , __lowercase = "numpy" , __lowercase = None , __lowercase = 1 , ) -> Union[AudioPipelineOutput, Tuple]: if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__lowercase , __lowercase) or callback_steps <= 0) ): raise ValueError( f"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" f""" {type(__lowercase)}.""") __UpperCamelCase :Union[str, Any] = np.zeros([1, TARGET_FEATURE_LENGTH, self.n_dims] , dtype=np.floataa) __UpperCamelCase :Union[str, Any] = np.zeros([1, 0, self.n_dims] , np.floataa) __UpperCamelCase :Union[str, Any] = torch.ones((1, TARGET_FEATURE_LENGTH) , dtype=__lowercase , device=self.device) for i, encoder_input_tokens in enumerate(__lowercase): if i == 0: __UpperCamelCase :int = torch.from_numpy(pred_mel[:1].copy()).to( device=self.device , dtype=self.decoder.dtype) # The first chunk has no previous context. __UpperCamelCase :int = torch.zeros((1, TARGET_FEATURE_LENGTH) , dtype=__lowercase , device=self.device) else: # The full song pipeline does not feed in a context feature, so the mask # will be all 0s after the feature converter. Because we know we're # feeding in a full context chunk from the previous prediction, set it # to all 1s. __UpperCamelCase :Tuple = ones __UpperCamelCase :Optional[Any] = self.scale_features( __lowercase , output_range=[-1.0, 1.0] , clip=__lowercase) __UpperCamelCase :int = self.encode( input_tokens=torch.IntTensor([encoder_input_tokens]).to(device=self.device) , continuous_inputs=__lowercase , continuous_mask=__lowercase , ) # Sample encoder_continuous_inputs shaped gaussian noise to begin loop __UpperCamelCase :int = randn_tensor( shape=encoder_continuous_inputs.shape , generator=__lowercase , device=self.device , dtype=self.decoder.dtype , ) # set step values self.scheduler.set_timesteps(__lowercase) # Denoising diffusion loop for j, t in enumerate(self.progress_bar(self.scheduler.timesteps)): __UpperCamelCase :Optional[int] = self.decode( encodings_and_masks=__lowercase , input_tokens=__lowercase , noise_time=t / self.scheduler.config.num_train_timesteps , ) # Compute previous output: x_t -> x_t-1 __UpperCamelCase :int = self.scheduler.step(__lowercase , __lowercase , __lowercase , generator=__lowercase).prev_sample __UpperCamelCase :Tuple = self.scale_to_features(__lowercase , input_range=[-1.0, 1.0]) __UpperCamelCase :List[Any] = mel[:1] __UpperCamelCase :Optional[Any] = mel.cpu().float().numpy() __UpperCamelCase :Any = np.concatenate([full_pred_mel, pred_mel[:1]] , axis=1) # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__lowercase , __lowercase) logger.info('''Generated segment''' , __lowercase) if output_type == "numpy" and not is_onnx_available(): raise ValueError( '''Cannot return output in \'np\' format if ONNX is not available. Make sure to have ONNX installed or set \'output_type\' to \'mel\'.''') elif output_type == "numpy" and self.melgan is None: raise ValueError( '''Cannot return output in \'np\' format if melgan component is not defined. Make sure to define `self.melgan` or set \'output_type\' to \'mel\'.''') if output_type == "numpy": __UpperCamelCase :Optional[Any] = self.melgan(input_features=full_pred_mel.astype(np.floataa)) else: __UpperCamelCase :List[str] = full_pred_mel if not return_dict: return (output,) return AudioPipelineOutput(audios=__lowercase)
43
1
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : int = StableUnCLIPImgaImgPipeline a__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS a__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a__ : Optional[Any] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess a__ : int = frozenset([] ) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Tuple = 32 __UpperCamelCase :Optional[int] = embedder_hidden_size # image encoding components __UpperCamelCase :Union[str, Any] = CLIPImageProcessor(crop_size=32 , size=32) torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=__lowercase , projection_dim=__lowercase , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , )) # regular denoising components torch.manual_seed(0) __UpperCamelCase :str = StableUnCLIPImageNormalizer(embedding_dim=__lowercase) __UpperCamelCase :Optional[int] = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''') torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') torch.manual_seed(0) __UpperCamelCase :Dict = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__lowercase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , )) torch.manual_seed(0) __UpperCamelCase :List[Any] = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=__lowercase , layers_per_block=1 , upcast_attention=__lowercase , use_linear_projection=__lowercase , ) torch.manual_seed(0) __UpperCamelCase :Tuple = DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.0_00_85 , beta_end=0.0_12 , prediction_type='''v_prediction''' , set_alpha_to_one=__lowercase , steps_offset=1 , ) torch.manual_seed(0) __UpperCamelCase :List[str] = AutoencoderKL() __UpperCamelCase :Tuple = { # image encoding components '''feature_extractor''': feature_extractor, '''image_encoder''': image_encoder.eval(), # image noising components '''image_normalizer''': image_normalizer.eval(), '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder.eval(), '''unet''': unet.eval(), '''scheduler''': scheduler, '''vae''': vae.eval(), } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0 , __lowercase=True) -> str: if str(__lowercase).startswith('''mps'''): __UpperCamelCase :Union[str, Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :int = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :int = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowercase)).to(__lowercase) if pil_image: __UpperCamelCase :List[Any] = input_image * 0.5 + 0.5 __UpperCamelCase :Optional[Any] = input_image.clamp(0 , 1) __UpperCamelCase :int = input_image.cpu().permute(0 , 2 , 3 , 1).float().numpy() __UpperCamelCase :Optional[Any] = DiffusionPipeline.numpy_to_pil(__lowercase)[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :Dict = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Tuple = self.get_dummy_components() __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline(**__lowercase) __UpperCamelCase :Optional[Any] = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :List[Any] = self.get_dummy_inputs(__lowercase) inputs.update({'''image_embeds''': None}) __UpperCamelCase :Any = sd_pipe(**__lowercase).images __UpperCamelCase :List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCamelCase :List[Any] = np.array([0.38_72, 0.72_24, 0.56_01, 0.47_41, 0.68_72, 0.58_14, 0.46_36, 0.38_67, 0.50_78]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_attention_slicing_forward_pass(test_max_difference=__lowercase) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=__lowercase) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase__ ( self) -> Union[str, Any]: self._test_xformers_attention_forwardGenerator_pass(test_max_difference=__lowercase) @slow @require_torch_gpu class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Union[str, Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :int = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-l-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Dict = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :Dict = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Optional[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Optional[int] = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :List[Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) __UpperCamelCase :Union[str, Any] = pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :Optional[Any] = pipe( __lowercase , '''anime turtle''' , num_inference_steps=2 , output_type='''np''' , ) __UpperCamelCase :int = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
43
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( HubertConfig, HubertForCTC, HubertModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __lowercase = logging.get_logger(__name__) __lowercase = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' for attribute in key.split('''.''' ): __UpperCamelCase :str = getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if weight_type is not None: __UpperCamelCase :Any = getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).shape else: __UpperCamelCase :Union[str, Any] = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __UpperCamelCase :str = value elif weight_type == "weight_g": __UpperCamelCase :List[str] = value elif weight_type == "weight_v": __UpperCamelCase :str = value elif weight_type == "bias": __UpperCamelCase :Union[str, Any] = value else: __UpperCamelCase :str = value logger.info(f"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = [] __UpperCamelCase :int = fairseq_model.state_dict() __UpperCamelCase :List[Any] = hf_model.hubert.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): __UpperCamelCase :List[Any] = False if "conv_layers" in name: load_conv_layer( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , hf_model.config.feat_extract_norm == '''group''' , ) __UpperCamelCase :List[str] = True else: for key, mapped_key in MAPPING.items(): __UpperCamelCase :Dict = '''hubert.''' + mapped_key if (is_finetuned and mapped_key != '''lm_head''') else mapped_key if key in name or (key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0] and not is_finetuned): __UpperCamelCase :Optional[Any] = True if "*" in mapped_key: __UpperCamelCase :List[str] = name.split(SCREAMING_SNAKE_CASE )[0].split('''.''' )[-2] __UpperCamelCase :Optional[int] = mapped_key.replace('''*''' , SCREAMING_SNAKE_CASE ) if "weight_g" in name: __UpperCamelCase :int = '''weight_g''' elif "weight_v" in name: __UpperCamelCase :List[Any] = '''weight_v''' elif "weight" in name: __UpperCamelCase :Dict = '''weight''' elif "bias" in name: __UpperCamelCase :Dict = '''bias''' else: __UpperCamelCase :Dict = None set_recursively(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(SCREAMING_SNAKE_CASE ) logger.warning(f"""Unused weights: {unused_weights}""" ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = full_name.split('''conv_layers.''' )[-1] __UpperCamelCase :Optional[int] = name.split('''.''' ) __UpperCamelCase :str = int(items[0] ) __UpperCamelCase :List[Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __UpperCamelCase :Dict = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __UpperCamelCase :Any = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) __UpperCamelCase :int = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) __UpperCamelCase :Union[str, Any] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(SCREAMING_SNAKE_CASE ) @torch.no_grad() def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=True ): '''simple docstring''' if config_path is not None: __UpperCamelCase :Tuple = HubertConfig.from_pretrained(SCREAMING_SNAKE_CASE ) else: __UpperCamelCase :Optional[int] = HubertConfig() if is_finetuned: if dict_path: __UpperCamelCase :Optional[int] = Dictionary.load(SCREAMING_SNAKE_CASE ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __UpperCamelCase :Optional[int] = target_dict.pad_index __UpperCamelCase :Dict = target_dict.bos_index __UpperCamelCase :str = target_dict.eos_index __UpperCamelCase :Dict = len(target_dict.symbols ) __UpperCamelCase :List[Any] = os.path.join(SCREAMING_SNAKE_CASE , '''vocab.json''' ) if not os.path.isdir(SCREAMING_SNAKE_CASE ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(SCREAMING_SNAKE_CASE ) ) return os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) with open(SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(target_dict.indices , SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = WavaVecaCTCTokenizer( SCREAMING_SNAKE_CASE , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :Union[str, Any] = True if config.feat_extract_norm == '''layer''' else False __UpperCamelCase :Any = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0 , do_normalize=SCREAMING_SNAKE_CASE , return_attention_mask=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :Any = WavaVecaProcessor(feature_extractor=SCREAMING_SNAKE_CASE , tokenizer=SCREAMING_SNAKE_CASE ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[str] = HubertForCTC(SCREAMING_SNAKE_CASE ) else: __UpperCamelCase :str = HubertModel(SCREAMING_SNAKE_CASE ) if is_finetuned: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :int = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Any = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) __UpperCamelCase :Dict = model[0].eval() recursively_load_weights(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) hf_wavavec.save_pretrained(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) __lowercase = parser.parse_args() convert_hubert_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
43
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __lowercase = {'''configuration_glpn''': ['''GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GLPNConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = ['''GLPNFeatureExtractor'''] __lowercase = ['''GLPNImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''GLPN_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GLPNForDepthEstimation''', '''GLPNLayer''', '''GLPNModel''', '''GLPNPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_glpn import GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP, GLPNConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_glpn import GLPNFeatureExtractor from .image_processing_glpn import GLPNImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_glpn import ( GLPN_PRETRAINED_MODEL_ARCHIVE_LIST, GLPNForDepthEstimation, GLPNLayer, GLPNModel, GLPNPreTrainedModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
43
import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters __lowercase = (720, 1280) # Height, Width __lowercase = (0.4, 0.6) # if height or width lower than this scale, drop it. __lowercase = 1 / 100 __lowercase = '''''' __lowercase = '''''' __lowercase = '''''' __lowercase = 250 def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase , __UpperCamelCase :List[Any] = get_dataset(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for index in range(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = random.sample(range(len(SCREAMING_SNAKE_CASE ) ) , 4 ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :str = update_image_and_anno( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , filter_scale=SCREAMING_SNAKE_CASE , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __UpperCamelCase :List[Any] = random_chars(32 ) __UpperCamelCase :List[str] = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __UpperCamelCase :Tuple = f"""{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}""" cva.imwrite(f"""{file_root}.jpg""" , SCREAMING_SNAKE_CASE , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(f"""Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}""" ) __UpperCamelCase :Optional[Any] = [] for anno in new_annos: __UpperCamelCase :int = anno[3] - anno[1] __UpperCamelCase :Optional[int] = anno[4] - anno[2] __UpperCamelCase :int = anno[1] + width / 2 __UpperCamelCase :List[str] = anno[2] + height / 2 __UpperCamelCase :str = f"""{anno[0]} {x_center} {y_center} {width} {height}""" annos_list.append(SCREAMING_SNAKE_CASE ) with open(f"""{file_root}.txt""" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :str = [] __UpperCamelCase :str = [] for label_file in glob.glob(os.path.join(SCREAMING_SNAKE_CASE , '''*.txt''' ) ): __UpperCamelCase :Any = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(SCREAMING_SNAKE_CASE ) as in_file: __UpperCamelCase :str = in_file.readlines() __UpperCamelCase :Optional[int] = os.path.join(SCREAMING_SNAKE_CASE , f"""{label_name}.jpg""" ) __UpperCamelCase :int = [] for obj_list in obj_lists: __UpperCamelCase :Optional[int] = obj_list.rstrip('''\n''' ).split(''' ''' ) __UpperCamelCase :Any = float(obj[1] ) - float(obj[3] ) / 2 __UpperCamelCase :List[str] = float(obj[2] ) - float(obj[4] ) / 2 __UpperCamelCase :Dict = float(obj[1] ) + float(obj[3] ) / 2 __UpperCamelCase :List[str] = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(SCREAMING_SNAKE_CASE ) labels.append(SCREAMING_SNAKE_CASE ) return img_paths, labels def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 0.0 , ): '''simple docstring''' __UpperCamelCase :List[str] = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __UpperCamelCase :List[Any] = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __UpperCamelCase :int = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __UpperCamelCase :Optional[int] = int(scale_x * output_size[1] ) __UpperCamelCase :Any = int(scale_y * output_size[0] ) __UpperCamelCase :List[str] = [] __UpperCamelCase :Dict = [] for i, index in enumerate(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Any = all_img_list[index] path_list.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = all_annos[index] __UpperCamelCase :Union[str, Any] = cva.imread(SCREAMING_SNAKE_CASE ) if i == 0: # top-left __UpperCamelCase :str = cva.resize(SCREAMING_SNAKE_CASE , (divid_point_x, divid_point_y) ) __UpperCamelCase :Union[str, Any] = img for bbox in img_annos: __UpperCamelCase :Union[str, Any] = bbox[1] * scale_x __UpperCamelCase :Optional[Any] = bbox[2] * scale_y __UpperCamelCase :int = bbox[3] * scale_x __UpperCamelCase :Union[str, Any] = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __UpperCamelCase :str = cva.resize(SCREAMING_SNAKE_CASE , (output_size[1] - divid_point_x, divid_point_y) ) __UpperCamelCase :List[str] = img for bbox in img_annos: __UpperCamelCase :str = scale_x + bbox[1] * (1 - scale_x) __UpperCamelCase :Dict = bbox[2] * scale_y __UpperCamelCase :Optional[Any] = scale_x + bbox[3] * (1 - scale_x) __UpperCamelCase :List[Any] = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __UpperCamelCase :str = cva.resize(SCREAMING_SNAKE_CASE , (divid_point_x, output_size[0] - divid_point_y) ) __UpperCamelCase :Optional[int] = img for bbox in img_annos: __UpperCamelCase :Tuple = bbox[1] * scale_x __UpperCamelCase :Optional[Any] = scale_y + bbox[2] * (1 - scale_y) __UpperCamelCase :Tuple = bbox[3] * scale_x __UpperCamelCase :Dict = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __UpperCamelCase :Optional[int] = cva.resize( SCREAMING_SNAKE_CASE , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __UpperCamelCase :Optional[int] = img for bbox in img_annos: __UpperCamelCase :Optional[Any] = scale_x + bbox[1] * (1 - scale_x) __UpperCamelCase :Optional[int] = scale_y + bbox[2] * (1 - scale_y) __UpperCamelCase :Optional[Any] = scale_x + bbox[3] * (1 - scale_x) __UpperCamelCase :int = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __UpperCamelCase :List[Any] = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' assert number_char > 1, "The number of character should greater than 1" __UpperCamelCase :Optional[Any] = ascii_lowercase + digits return "".join(random.choice(SCREAMING_SNAKE_CASE ) for _ in range(SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": main() print('''DONE ✅''')
43
1
__lowercase = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(10_0000)] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 100_000] number //= 100_000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution __lowercase = [None] * 1000_0000 __lowercase = True __lowercase = False def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore __UpperCamelCase :Optional[Any] = chain(next_number(SCREAMING_SNAKE_CASE ) ) __UpperCamelCase :Union[str, Any] = number_chain while number < 10_000_000: __UpperCamelCase :List[Any] = number_chain number *= 10 return number_chain def lowerCamelCase ( SCREAMING_SNAKE_CASE = 10_000_000 ): '''simple docstring''' for i in range(1 , SCREAMING_SNAKE_CASE ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod() print(F'{solution() = }')
43
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = { '''facebook/wav2vec2-base-960h''': '''https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/config.json''', # See all Wav2Vec2 models at https://huggingface.co/models?filter=wav2vec2 } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Union[str, Any] = """wav2vec2""" def __init__( self , __lowercase=32 , __lowercase=768 , __lowercase=12 , __lowercase=12 , __lowercase=3_072 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.02 , __lowercase=1E-5 , __lowercase="group" , __lowercase="gelu" , __lowercase=(512, 512, 512, 512, 512, 512, 512) , __lowercase=(5, 2, 2, 2, 2, 2, 2) , __lowercase=(10, 3, 3, 3, 3, 2, 2) , __lowercase=False , __lowercase=128 , __lowercase=16 , __lowercase=False , __lowercase=True , __lowercase=0.05 , __lowercase=10 , __lowercase=2 , __lowercase=0.0 , __lowercase=10 , __lowercase=0 , __lowercase=320 , __lowercase=2 , __lowercase=0.1 , __lowercase=100 , __lowercase=256 , __lowercase=256 , __lowercase=0.1 , __lowercase="sum" , __lowercase=False , __lowercase=False , __lowercase=256 , __lowercase=(512, 512, 512, 512, 1_500) , __lowercase=(5, 3, 3, 1, 1) , __lowercase=(1, 2, 3, 1, 1) , __lowercase=512 , __lowercase=0 , __lowercase=1 , __lowercase=2 , __lowercase=False , __lowercase=3 , __lowercase=2 , __lowercase=3 , __lowercase=None , __lowercase=None , **__lowercase , ) -> int: super().__init__(**__lowercase , pad_token_id=__lowercase , bos_token_id=__lowercase , eos_token_id=__lowercase) __UpperCamelCase :Any = hidden_size __UpperCamelCase :int = feat_extract_norm __UpperCamelCase :Tuple = feat_extract_activation __UpperCamelCase :Union[str, Any] = list(__lowercase) __UpperCamelCase :List[Any] = list(__lowercase) __UpperCamelCase :int = list(__lowercase) __UpperCamelCase :List[Any] = conv_bias __UpperCamelCase :Optional[int] = num_conv_pos_embeddings __UpperCamelCase :Dict = num_conv_pos_embedding_groups __UpperCamelCase :Any = len(self.conv_dim) __UpperCamelCase :List[str] = num_hidden_layers __UpperCamelCase :int = intermediate_size __UpperCamelCase :str = hidden_act __UpperCamelCase :Any = num_attention_heads __UpperCamelCase :int = hidden_dropout __UpperCamelCase :Tuple = attention_dropout __UpperCamelCase :List[str] = activation_dropout __UpperCamelCase :Optional[Any] = feat_proj_dropout __UpperCamelCase :Any = final_dropout __UpperCamelCase :Any = layerdrop __UpperCamelCase :str = layer_norm_eps __UpperCamelCase :Optional[Any] = initializer_range __UpperCamelCase :List[str] = vocab_size __UpperCamelCase :str = do_stable_layer_norm __UpperCamelCase :Union[str, Any] = use_weighted_layer_sum if ( (len(self.conv_stride) != self.num_feat_extract_layers) or (len(self.conv_kernel) != self.num_feat_extract_layers) or (len(self.conv_dim) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' f""" {len(self.conv_dim)}`, `len(config.conv_stride) = {len(self.conv_stride)}`,""" f""" `len(config.conv_kernel) = {len(self.conv_kernel)}`.""") # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __UpperCamelCase :List[Any] = apply_spec_augment __UpperCamelCase :Tuple = mask_time_prob __UpperCamelCase :int = mask_time_length __UpperCamelCase :Dict = mask_time_min_masks __UpperCamelCase :str = mask_feature_prob __UpperCamelCase :List[str] = mask_feature_length __UpperCamelCase :Union[str, Any] = mask_feature_min_masks # parameters for pretraining with codevector quantized representations __UpperCamelCase :Optional[Any] = num_codevectors_per_group __UpperCamelCase :List[Any] = num_codevector_groups __UpperCamelCase :Tuple = contrastive_logits_temperature __UpperCamelCase :Optional[int] = feat_quantizer_dropout __UpperCamelCase :Optional[int] = num_negatives __UpperCamelCase :List[Any] = codevector_dim __UpperCamelCase :str = proj_codevector_dim __UpperCamelCase :List[str] = diversity_loss_weight # ctc loss __UpperCamelCase :Tuple = ctc_loss_reduction __UpperCamelCase :Tuple = ctc_zero_infinity # adapter __UpperCamelCase :List[str] = add_adapter __UpperCamelCase :Tuple = adapter_kernel_size __UpperCamelCase :str = adapter_stride __UpperCamelCase :Tuple = num_adapter_layers __UpperCamelCase :Tuple = output_hidden_size or hidden_size __UpperCamelCase :Optional[Any] = adapter_attn_dim # SequenceClassification-specific parameter. Feel free to ignore for other classes. __UpperCamelCase :Optional[Any] = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. __UpperCamelCase :Optional[int] = list(__lowercase) __UpperCamelCase :List[Any] = list(__lowercase) __UpperCamelCase :List[Any] = list(__lowercase) __UpperCamelCase :str = xvector_output_dim @property def UpperCamelCase__ ( self) -> List[str]: return functools.reduce(operator.mul , self.conv_stride , 1)
43
1
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[str] = filter(lambda SCREAMING_SNAKE_CASE : p.requires_grad , model.parameters() ) __UpperCamelCase :List[str] = sum([np.prod(p.size() ) for p in model_parameters] ) return params __lowercase = logging.getLogger(__name__) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' if metric == "rouge2": __UpperCamelCase :Any = '''{val_avg_rouge2:.4f}-{step_count}''' elif metric == "bleu": __UpperCamelCase :List[Any] = '''{val_avg_bleu:.4f}-{step_count}''' elif metric == "em": __UpperCamelCase :Union[str, Any] = '''{val_avg_em:.4f}-{step_count}''' else: raise NotImplementedError( f"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" ''' function.''' ) __UpperCamelCase :List[Any] = ModelCheckpoint( dirpath=SCREAMING_SNAKE_CASE , filename=SCREAMING_SNAKE_CASE , monitor=f"""val_{metric}""" , mode='''max''' , save_top_k=3 , every_n_epochs=1 , ) return checkpoint_callback def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' return EarlyStopping( monitor=f"""val_{metric}""" , mode='''min''' if '''loss''' in metric else '''max''' , patience=SCREAMING_SNAKE_CASE , verbose=SCREAMING_SNAKE_CASE , ) class lowerCamelCase_ ( pl.Callback ): '''simple docstring''' def UpperCamelCase__ ( self , __lowercase , __lowercase) -> Optional[Any]: __UpperCamelCase :List[str] = {f"""lr_group_{i}""": param['''lr'''] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups)} pl_module.logger.log_metrics(__lowercase) @rank_zero_only def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase=True) -> None: logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""") __UpperCamelCase :Dict = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ['''log''', '''progress_bar''', '''preds''']}) # Log results __UpperCamelCase :Tuple = Path(pl_module.hparams.output_dir) if type_path == "test": __UpperCamelCase :str = od / '''test_results.txt''' __UpperCamelCase :Optional[Any] = od / '''test_generations.txt''' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. __UpperCamelCase :List[Any] = od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" __UpperCamelCase :Optional[int] = od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=__lowercase) generations_file.parent.mkdir(exist_ok=__lowercase) with open(__lowercase , '''a+''') as writer: for key in sorted(__lowercase): if key in ["log", "progress_bar", "preds"]: continue __UpperCamelCase :Union[str, Any] = metrics[key] if isinstance(__lowercase , torch.Tensor): __UpperCamelCase :List[Any] = val.item() __UpperCamelCase :Optional[Any] = f"""{key}: {val:.6f}\n""" writer.write(__lowercase) if not save_generations: return if "preds" in metrics: __UpperCamelCase :Union[str, Any] = '''\n'''.join(metrics['''preds''']) generations_file.open('''w+''').write(__lowercase) @rank_zero_only def UpperCamelCase__ ( self , __lowercase , __lowercase) -> List[Any]: try: __UpperCamelCase :int = pl_module.model.model.num_parameters() except AttributeError: __UpperCamelCase :Tuple = pl_module.model.num_parameters() __UpperCamelCase :Optional[Any] = count_trainable_parameters(__lowercase) # mp stands for million parameters trainer.logger.log_metrics({'''n_params''': npars, '''mp''': npars / 1E6, '''grad_mp''': n_trainable_pars / 1E6}) @rank_zero_only def UpperCamelCase__ ( self , __lowercase , __lowercase) -> Optional[int]: save_json(pl_module.metrics , pl_module.metrics_save_path) return self._write_logs(__lowercase , __lowercase , '''test''') @rank_zero_only def UpperCamelCase__ ( self , __lowercase , __lowercase) -> Tuple: save_json(pl_module.metrics , pl_module.metrics_save_path) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
43
from typing import List, Optional, Union import numpy as np import PIL.Image from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, PILImageResampling, get_image_size, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __lowercase = logging.get_logger(__name__) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[Any] = ["""pixel_values"""] def __init__( self , __lowercase = True , __lowercase = 32 , __lowercase=PILImageResampling.BILINEAR , __lowercase = True , **__lowercase , ) -> None: __UpperCamelCase :Optional[int] = do_resize __UpperCamelCase :Any = do_rescale __UpperCamelCase :str = size_divisor __UpperCamelCase :Dict = resample super().__init__(**__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase = None , **__lowercase) -> np.ndarray: __UpperCamelCase , __UpperCamelCase :int = get_image_size(__lowercase) # Rounds the height and width down to the closest multiple of size_divisor __UpperCamelCase :List[Any] = height // size_divisor * size_divisor __UpperCamelCase :List[str] = width // size_divisor * size_divisor __UpperCamelCase :str = resize(__lowercase , (new_h, new_w) , resample=__lowercase , data_format=__lowercase , **__lowercase) return image def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase = None , **__lowercase) -> np.ndarray: return rescale(image=__lowercase , scale=__lowercase , data_format=__lowercase , **__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase = None , __lowercase = None , __lowercase=None , __lowercase = None , __lowercase = None , __lowercase = ChannelDimension.FIRST , **__lowercase , ) -> BatchFeature: __UpperCamelCase :Union[str, Any] = do_resize if do_resize is not None else self.do_resize __UpperCamelCase :Tuple = do_rescale if do_rescale is not None else self.do_rescale __UpperCamelCase :List[str] = size_divisor if size_divisor is not None else self.size_divisor __UpperCamelCase :List[Any] = resample if resample is not None else self.resample if do_resize and size_divisor is None: raise ValueError('''size_divisor is required for resizing''') __UpperCamelCase :List[Any] = make_list_of_images(__lowercase) if not valid_images(__lowercase): raise ValueError('''Invalid image(s)''') # All transformations expect numpy arrays. __UpperCamelCase :Optional[Any] = [to_numpy_array(__lowercase) for img in images] if do_resize: __UpperCamelCase :List[str] = [self.resize(__lowercase , size_divisor=__lowercase , resample=__lowercase) for image in images] if do_rescale: __UpperCamelCase :Dict = [self.rescale(__lowercase , scale=1 / 255) for image in images] __UpperCamelCase :str = [to_channel_dimension_format(__lowercase , __lowercase) for image in images] __UpperCamelCase :int = {'''pixel_values''': images} return BatchFeature(data=__lowercase , tensor_type=__lowercase)
43
1
import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler __lowercase = 16 __lowercase = 32 def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 16 , SCREAMING_SNAKE_CASE = "bert-base-cased" ): '''simple docstring''' __UpperCamelCase :List[str] = AutoTokenizer.from_pretrained(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(SCREAMING_SNAKE_CASE ): # max_length=None => use the model max length (it's actually the default) __UpperCamelCase :int = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset __UpperCamelCase :Tuple = datasets.map( SCREAMING_SNAKE_CASE , batched=SCREAMING_SNAKE_CASE , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , load_from_cache_file=SCREAMING_SNAKE_CASE ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __UpperCamelCase :List[str] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(SCREAMING_SNAKE_CASE ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(SCREAMING_SNAKE_CASE , padding='''max_length''' , max_length=128 , return_tensors='''pt''' ) return tokenizer.pad(SCREAMING_SNAKE_CASE , padding='''longest''' , return_tensors='''pt''' ) # Instantiate dataloaders. __UpperCamelCase :Union[str, Any] = DataLoader( tokenized_datasets['''train'''] , shuffle=SCREAMING_SNAKE_CASE , collate_fn=SCREAMING_SNAKE_CASE , batch_size=SCREAMING_SNAKE_CASE ) __UpperCamelCase :Dict = DataLoader( tokenized_datasets['''validation'''] , shuffle=SCREAMING_SNAKE_CASE , collate_fn=SCREAMING_SNAKE_CASE , batch_size=SCREAMING_SNAKE_CASE ) return train_dataloader, eval_dataloader def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __UpperCamelCase :int = config['''lr'''] __UpperCamelCase :str = int(config['''num_epochs'''] ) __UpperCamelCase :Any = int(config['''seed'''] ) __UpperCamelCase :Dict = int(config['''batch_size'''] ) __UpperCamelCase :Optional[Any] = args.model_name_or_path set_seed(SCREAMING_SNAKE_CASE ) __UpperCamelCase , __UpperCamelCase :Dict = get_dataloaders(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __UpperCamelCase :Any = AutoModelForSequenceClassification.from_pretrained(SCREAMING_SNAKE_CASE , return_dict=SCREAMING_SNAKE_CASE ) # Instantiate optimizer __UpperCamelCase :List[str] = ( AdamW if accelerator.state.deepspeed_plugin is None or '''optimizer''' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) __UpperCamelCase :Optional[Any] = optimizer_cls(params=model.parameters() , lr=SCREAMING_SNAKE_CASE ) if accelerator.state.deepspeed_plugin is not None: __UpperCamelCase :Dict = accelerator.state.deepspeed_plugin.deepspeed_config[ '''gradient_accumulation_steps''' ] else: __UpperCamelCase :Dict = 1 __UpperCamelCase :Tuple = (len(SCREAMING_SNAKE_CASE ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): __UpperCamelCase :str = get_linear_schedule_with_warmup( optimizer=SCREAMING_SNAKE_CASE , num_warmup_steps=0 , num_training_steps=SCREAMING_SNAKE_CASE , ) else: __UpperCamelCase :Dict = DummyScheduler(SCREAMING_SNAKE_CASE , total_num_steps=SCREAMING_SNAKE_CASE , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :int = accelerator.prepare( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # We need to keep track of how many total steps we have iterated over __UpperCamelCase :List[Any] = 0 # We also need to keep track of the stating epoch so files are named properly __UpperCamelCase :Dict = 0 # Now we train the model __UpperCamelCase :Any = evaluate.load('''glue''' , '''mrpc''' ) __UpperCamelCase :Union[str, Any] = 0 __UpperCamelCase :Optional[int] = {} for epoch in range(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): model.train() for step, batch in enumerate(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = model(**SCREAMING_SNAKE_CASE ) __UpperCamelCase :Tuple = outputs.loss __UpperCamelCase :str = loss / gradient_accumulation_steps accelerator.backward(SCREAMING_SNAKE_CASE ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() __UpperCamelCase :Any = 0 for step, batch in enumerate(SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __UpperCamelCase :Any = model(**SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times __UpperCamelCase , __UpperCamelCase :List[Any] = accelerator.gather( (predictions, batch['''labels''']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(SCREAMING_SNAKE_CASE ) - 1: __UpperCamelCase :List[str] = predictions[: len(eval_dataloader.dataset ) - samples_seen] __UpperCamelCase :Optional[int] = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=SCREAMING_SNAKE_CASE , references=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :Dict = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , SCREAMING_SNAKE_CASE ) __UpperCamelCase :str = eval_metric['''accuracy'''] if best_performance < eval_metric["accuracy"]: __UpperCamelCase :int = eval_metric['''accuracy'''] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f"""Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}""" accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , '''all_results.json''' ) , '''w''' ) as f: json.dump(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = argparse.ArgumentParser(description='''Simple example of training script tracking peak GPU memory usage.''' ) parser.add_argument( '''--model_name_or_path''' , type=SCREAMING_SNAKE_CASE , default='''bert-base-cased''' , help='''Path to pretrained model or model identifier from huggingface.co/models.''' , required=SCREAMING_SNAKE_CASE , ) parser.add_argument( '''--output_dir''' , type=SCREAMING_SNAKE_CASE , default='''.''' , help='''Optional save directory where all checkpoint folders will be stored. Default is the current working directory.''' , ) parser.add_argument( '''--performance_lower_bound''' , type=SCREAMING_SNAKE_CASE , default=SCREAMING_SNAKE_CASE , help='''Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.''' , ) parser.add_argument( '''--num_epochs''' , type=SCREAMING_SNAKE_CASE , default=3 , help='''Number of train epochs.''' , ) __UpperCamelCase :List[str] = parser.parse_args() __UpperCamelCase :Tuple = {'''lr''': 2e-5, '''num_epochs''': args.num_epochs, '''seed''': 42, '''batch_size''': 16} training_function(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
43
from __future__ import annotations from PIL import Image # Define glider example __lowercase = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example __lowercase = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :int = [] for i in range(len(SCREAMING_SNAKE_CASE ) ): __UpperCamelCase :Dict = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours __UpperCamelCase :List[str] = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(SCREAMING_SNAKE_CASE ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(SCREAMING_SNAKE_CASE ) - 1: neighbour_count += cells[i + 1][j] if i < len(SCREAMING_SNAKE_CASE ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. __UpperCamelCase :List[str] = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(SCREAMING_SNAKE_CASE ) return next_generation def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = [] for _ in range(SCREAMING_SNAKE_CASE ): # Create output image __UpperCamelCase :Dict = Image.new('''RGB''' , (len(cells[0] ), len(SCREAMING_SNAKE_CASE )) ) __UpperCamelCase :Any = img.load() # Save cells to image for x in range(len(SCREAMING_SNAKE_CASE ) ): for y in range(len(cells[0] ) ): __UpperCamelCase :Optional[Any] = 255 - cells[y][x] * 255 __UpperCamelCase :int = (colour, colour, colour) # Save image images.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = new_generation(SCREAMING_SNAKE_CASE ) return images if __name__ == "__main__": __lowercase = generate_images(GLIDER, 16) images[0].save('''out.gif''', save_all=True, append_images=images[1:])
43
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowercase = { '''configuration_electra''': ['''ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ElectraConfig''', '''ElectraOnnxConfig'''], '''tokenization_electra''': ['''ElectraTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = ['''ElectraTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ElectraForCausalLM''', '''ElectraForMaskedLM''', '''ElectraForMultipleChoice''', '''ElectraForPreTraining''', '''ElectraForQuestionAnswering''', '''ElectraForSequenceClassification''', '''ElectraForTokenClassification''', '''ElectraModel''', '''ElectraPreTrainedModel''', '''load_tf_weights_in_electra''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFElectraForMaskedLM''', '''TFElectraForMultipleChoice''', '''TFElectraForPreTraining''', '''TFElectraForQuestionAnswering''', '''TFElectraForSequenceClassification''', '''TFElectraForTokenClassification''', '''TFElectraModel''', '''TFElectraPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''FlaxElectraForCausalLM''', '''FlaxElectraForMaskedLM''', '''FlaxElectraForMultipleChoice''', '''FlaxElectraForPreTraining''', '''FlaxElectraForQuestionAnswering''', '''FlaxElectraForSequenceClassification''', '''FlaxElectraForTokenClassification''', '''FlaxElectraModel''', '''FlaxElectraPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_electra import ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ElectraConfig, ElectraOnnxConfig from .tokenization_electra import ElectraTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_electra_fast import ElectraTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_electra import ( ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, ElectraForCausalLM, ElectraForMaskedLM, ElectraForMultipleChoice, ElectraForPreTraining, ElectraForQuestionAnswering, ElectraForSequenceClassification, ElectraForTokenClassification, ElectraModel, ElectraPreTrainedModel, load_tf_weights_in_electra, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_electra import ( TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, TFElectraForMaskedLM, TFElectraForMultipleChoice, TFElectraForPreTraining, TFElectraForQuestionAnswering, TFElectraForSequenceClassification, TFElectraForTokenClassification, TFElectraModel, TFElectraPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_electra import ( FlaxElectraForCausalLM, FlaxElectraForMaskedLM, FlaxElectraForMultipleChoice, FlaxElectraForPreTraining, FlaxElectraForQuestionAnswering, FlaxElectraForSequenceClassification, FlaxElectraForTokenClassification, FlaxElectraModel, FlaxElectraPreTrainedModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
43
import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging __lowercase = logging.get_logger(__name__) def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = R'''\w+[.]\d+''' __UpperCamelCase :List[str] = re.findall(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for pat in pats: __UpperCamelCase :int = key.replace(SCREAMING_SNAKE_CASE , '''_'''.join(pat.split('''.''' ) ) ) return key def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = pt_tuple_key[:-1] + ('''scale''',) if ( any('''norm''' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): __UpperCamelCase :str = pt_tuple_key[:-1] + ('''scale''',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: __UpperCamelCase :Any = pt_tuple_key[:-1] + ('''scale''',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: __UpperCamelCase :str = pt_tuple_key[:-1] + ('''embedding''',) return renamed_pt_tuple_key, pt_tensor # conv layer __UpperCamelCase :List[str] = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: __UpperCamelCase :List[Any] = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer __UpperCamelCase :List[str] = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight": __UpperCamelCase :Any = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight __UpperCamelCase :int = pt_tuple_key[:-1] + ('''weight''',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias __UpperCamelCase :int = pt_tuple_key[:-1] + ('''bias''',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=42 ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params __UpperCamelCase :str = flax_model.init_weights(PRNGKey(SCREAMING_SNAKE_CASE ) ) __UpperCamelCase :int = flatten_dict(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): __UpperCamelCase :List[Any] = rename_key(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = tuple(renamed_pt_key.split('''.''' ) ) # Correctly rename weight parameters __UpperCamelCase , __UpperCamelCase :Any = rename_key_and_reshape_tensor(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ f"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown __UpperCamelCase :str = jnp.asarray(SCREAMING_SNAKE_CASE ) return unflatten_dict(SCREAMING_SNAKE_CASE )
43
1
from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_herbert import HerbertTokenizer __lowercase = logging.get_logger(__name__) __lowercase = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} __lowercase = { '''vocab_file''': { '''allegro/herbert-base-cased''': '''https://huggingface.co/allegro/herbert-base-cased/resolve/main/vocab.json''' }, '''merges_file''': { '''allegro/herbert-base-cased''': '''https://huggingface.co/allegro/herbert-base-cased/resolve/main/merges.txt''' }, } __lowercase = {'''allegro/herbert-base-cased''': 514} __lowercase = {} class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Dict = VOCAB_FILES_NAMES a__ : Tuple = PRETRAINED_VOCAB_FILES_MAP a__ : Optional[Any] = PRETRAINED_INIT_CONFIGURATION a__ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a__ : Optional[Any] = HerbertTokenizer def __init__( self , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase="<s>" , __lowercase="<unk>" , __lowercase="<pad>" , __lowercase="<mask>" , __lowercase="</s>" , **__lowercase , ) -> Optional[int]: super().__init__( __lowercase , __lowercase , tokenizer_file=__lowercase , cls_token=__lowercase , unk_token=__lowercase , pad_token=__lowercase , mask_token=__lowercase , sep_token=__lowercase , **__lowercase , ) def UpperCamelCase__ ( self , __lowercase , __lowercase = None) -> List[int]: __UpperCamelCase :Union[str, Any] = [self.cls_token_id] __UpperCamelCase :Optional[Any] = [self.sep_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase__ ( self , __lowercase , __lowercase = None , __lowercase = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowercase , token_ids_a=__lowercase , already_has_special_tokens=__lowercase) if token_ids_a is None: return [1] + ([0] * len(__lowercase)) + [1] return [1] + ([0] * len(__lowercase)) + [1] + ([0] * len(__lowercase)) + [1] def UpperCamelCase__ ( self , __lowercase , __lowercase = None) -> List[int]: __UpperCamelCase :List[Any] = [self.sep_token_id] __UpperCamelCase :Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def UpperCamelCase__ ( self , __lowercase , __lowercase = None) -> Tuple[str]: __UpperCamelCase :Optional[int] = self._tokenizer.model.save(__lowercase , name=__lowercase) return tuple(__lowercase)
43
import argparse import torch from ...utils import logging from . import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert logging.set_verbosity_info() def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = AlbertConfig.from_json_file(SCREAMING_SNAKE_CASE ) print(f"""Building PyTorch model from configuration: {config}""" ) __UpperCamelCase :List[str] = AlbertForPreTraining(SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint load_tf_weights_in_albert(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--albert_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained ALBERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __lowercase = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.albert_config_file, args.pytorch_dump_path)
43
1
from torch import nn def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(f"""Unsupported activation function: {act_fn}""" )
43
import math import qiskit def lowerCamelCase ( SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1 ): '''simple docstring''' if ( isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) or isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) or isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ): raise TypeError('''inputs must be integers.''' ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError('''inputs must be positive.''' ) if ( (math.floor(SCREAMING_SNAKE_CASE ) != input_a) or (math.floor(SCREAMING_SNAKE_CASE ) != input_a) or (math.floor(SCREAMING_SNAKE_CASE ) != carry_in) ): raise ValueError('''inputs must be exact integers.''' ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError('''inputs must be less or equal to 2.''' ) # build registers __UpperCamelCase :List[str] = qiskit.QuantumRegister(4 , '''qr''' ) __UpperCamelCase :str = qiskit.ClassicalRegister(2 , '''cr''' ) # list the entries __UpperCamelCase :Tuple = [input_a, input_a, carry_in] __UpperCamelCase :Optional[int] = qiskit.QuantumCircuit(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(SCREAMING_SNAKE_CASE ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(SCREAMING_SNAKE_CASE ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(SCREAMING_SNAKE_CASE ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , SCREAMING_SNAKE_CASE ) # measure the last two qbits __UpperCamelCase :Optional[Any] = qiskit.Aer.get_backend('''aer_simulator''' ) __UpperCamelCase :Tuple = qiskit.execute(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , shots=1_000 ) return job.result().get_counts(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(F'Total sum count for state is: {quantum_full_adder(1, 1, 1)}')
43
1
from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class lowerCamelCase_ : '''simple docstring''' a__ : List[str] a__ : Optional[str] = None # Automatically constructed a__ : ClassVar[str] = "dict" a__ : ClassVar[Any] = None a__ : str = field(default="""Translation""" , init=UpperCAmelCase_ , repr=UpperCAmelCase_ ) def __call__( self) -> List[Any]: return pa.struct({lang: pa.string() for lang in sorted(self.languages)}) def UpperCamelCase__ ( self) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value return {k: Value('''string''') for k in sorted(self.languages)} @dataclass class lowerCamelCase_ : '''simple docstring''' a__ : Optional[List] = None a__ : Optional[int] = None a__ : Optional[str] = None # Automatically constructed a__ : ClassVar[str] = "dict" a__ : ClassVar[Any] = None a__ : str = field(default="""TranslationVariableLanguages""" , init=UpperCAmelCase_ , repr=UpperCAmelCase_ ) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :int = sorted(set(self.languages)) if self.languages else None __UpperCamelCase :List[Any] = len(self.languages) if self.languages else None def __call__( self) -> Tuple: return pa.struct({'''language''': pa.list_(pa.string()), '''translation''': pa.list_(pa.string())}) def UpperCamelCase__ ( self , __lowercase) -> int: __UpperCamelCase :Any = set(self.languages) if self.languages and set(__lowercase) - lang_set: raise ValueError( f"""Some languages in example ({', '.join(sorted(set(__lowercase) - lang_set))}) are not in valid set ({', '.join(__lowercase)}).""") # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. __UpperCamelCase :List[Any] = [] for lang, text in translation_dict.items(): if isinstance(__lowercase , __lowercase): translation_tuples.append((lang, text)) else: translation_tuples.extend([(lang, el) for el in text]) # Ensure translations are in ascending order by language code. __UpperCamelCase , __UpperCamelCase :Dict = zip(*sorted(__lowercase)) return {"language": languages, "translation": translations} def UpperCamelCase__ ( self) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Sequence, Value return { "language": Sequence(Value('''string''')), "translation": Sequence(Value('''string''')), }
43
import random def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = a[left_index] __UpperCamelCase :Any = left_index + 1 for j in range(left_index + 1 , SCREAMING_SNAKE_CASE ): if a[j] < pivot: __UpperCamelCase , __UpperCamelCase :str = a[i], a[j] i += 1 __UpperCamelCase , __UpperCamelCase :Optional[int] = a[i - 1], a[left_index] return i - 1 def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' if left < right: __UpperCamelCase :int = random.randint(SCREAMING_SNAKE_CASE , right - 1 ) __UpperCamelCase , __UpperCamelCase :List[str] = ( a[left], a[pivot], ) # switches the pivot with the left most bound __UpperCamelCase :Dict = partition(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) quick_sort_random( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # recursive quicksort to the left of the pivot point quick_sort_random( SCREAMING_SNAKE_CASE , pivot_index + 1 , SCREAMING_SNAKE_CASE ) # recursive quicksort to the right of the pivot point def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = input('''Enter numbers separated by a comma:\n''' ).strip() __UpperCamelCase :Union[str, Any] = [int(SCREAMING_SNAKE_CASE ) for item in user_input.split(''',''' )] quick_sort_random(SCREAMING_SNAKE_CASE , 0 , len(SCREAMING_SNAKE_CASE ) ) print(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
43
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = {'''ctrl''': '''https://huggingface.co/ctrl/resolve/main/config.json'''} class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : str = """ctrl""" a__ : Dict = ["""past_key_values"""] a__ : Tuple = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , __lowercase=246_534 , __lowercase=256 , __lowercase=1_280 , __lowercase=8_192 , __lowercase=48 , __lowercase=16 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=1E-6 , __lowercase=0.02 , __lowercase=True , **__lowercase , ) -> List[Any]: __UpperCamelCase :List[str] = vocab_size __UpperCamelCase :Optional[Any] = n_positions __UpperCamelCase :Dict = n_embd __UpperCamelCase :Dict = n_layer __UpperCamelCase :List[Any] = n_head __UpperCamelCase :int = dff __UpperCamelCase :Union[str, Any] = resid_pdrop __UpperCamelCase :Optional[int] = embd_pdrop __UpperCamelCase :List[Any] = layer_norm_epsilon __UpperCamelCase :Dict = initializer_range __UpperCamelCase :Any = use_cache super().__init__(**__lowercase)
43
def lowerCamelCase ( SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1_000 ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = 1 __UpperCamelCase :Any = 0 for divide_by_number in range(SCREAMING_SNAKE_CASE , digit + 1 ): __UpperCamelCase :list[int] = [] __UpperCamelCase :Optional[int] = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = len(SCREAMING_SNAKE_CASE ) __UpperCamelCase :int = divide_by_number else: has_been_divided.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
43
1
__lowercase = [ (1000, '''M'''), (900, '''CM'''), (500, '''D'''), (400, '''CD'''), (100, '''C'''), (90, '''XC'''), (50, '''L'''), (40, '''XL'''), (10, '''X'''), (9, '''IX'''), (5, '''V'''), (4, '''IV'''), (1, '''I'''), ] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = {'''I''': 1, '''V''': 5, '''X''': 10, '''L''': 50, '''C''': 100, '''D''': 500, '''M''': 1_000} __UpperCamelCase :List[Any] = 0 __UpperCamelCase :str = 0 while place < len(SCREAMING_SNAKE_CASE ): if (place + 1 < len(SCREAMING_SNAKE_CASE )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = [] for arabic, roman in ROMAN: ((__UpperCamelCase) , (__UpperCamelCase)) :Tuple = divmod(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) result.append(roman * factor ) if number == 0: break return "".join(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": import doctest doctest.testmod()
43
import argparse import json from tqdm import tqdm def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--src_path''' , type=SCREAMING_SNAKE_CASE , default='''biencoder-nq-dev.json''' , help='''Path to raw DPR training data''' , ) parser.add_argument( '''--evaluation_set''' , type=SCREAMING_SNAKE_CASE , help='''where to store parsed evaluation_set file''' , ) parser.add_argument( '''--gold_data_path''' , type=SCREAMING_SNAKE_CASE , help='''where to store parsed gold_data_path file''' , ) __UpperCamelCase :str = parser.parse_args() with open(args.src_path , '''r''' ) as src_file, open(args.evaluation_set , '''w''' ) as eval_file, open( args.gold_data_path , '''w''' ) as gold_file: __UpperCamelCase :List[str] = json.load(SCREAMING_SNAKE_CASE ) for dpr_record in tqdm(SCREAMING_SNAKE_CASE ): __UpperCamelCase :List[str] = dpr_record['''question'''] __UpperCamelCase :Tuple = [context['''title'''] for context in dpr_record['''positive_ctxs''']] eval_file.write(question + '''\n''' ) gold_file.write('''\t'''.join(SCREAMING_SNAKE_CASE ) + '''\n''' ) if __name__ == "__main__": main()
43
1
import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, get_torch_dist_unique_port, require_apex, require_bitsandbytes, require_fairscale, require_torch, require_torch_gpu, require_torch_multi_gpu, require_torch_non_multi_gpu, slow, ) from transformers.trainer_callback import TrainerState from transformers.trainer_utils import set_seed __lowercase = os.path.abspath(os.path.dirname(__file__)) with ExtendSysPath(F'{bindir}/../../examples/pytorch/translation'): from run_translation import main # noqa set_seed(42) __lowercase = '''sshleifer/student_marian_en_ro_6_1''' __lowercase = '''sshleifer/tiny-mbart''' @require_torch class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' def UpperCamelCase__ ( self , __lowercase=False , __lowercase=None , __lowercase=True , __lowercase=True , __lowercase=True , __lowercase=True , ) -> Optional[int]: __UpperCamelCase :int = self.run_trainer( eval_steps=1 , max_len=12 , model_name=__lowercase , num_train_epochs=1 , distributed=__lowercase , extra_args_str=__lowercase , predict_with_generate=__lowercase , do_train=__lowercase , do_eval=__lowercase , do_predict=__lowercase , ) __UpperCamelCase :int = TrainerState.load_from_json(os.path.join(__lowercase , '''trainer_state.json''')).log_history if not do_eval: return __UpperCamelCase :Union[str, Any] = [log for log in logs if '''eval_loss''' in log.keys()] __UpperCamelCase :Optional[int] = eval_metrics[0] if predict_with_generate: assert "eval_bleu" in first_step_stats __UpperCamelCase :Optional[Any] = eval_metrics[-1] assert isinstance(last_step_stats['''eval_bleu'''] , __lowercase) assert not math.isnan(float(last_step_stats['''eval_loss'''])), "eval_loss must not be `nan`" @require_torch_non_multi_gpu def UpperCamelCase__ ( self) -> str: self.run_seqaseq_quick() @require_torch_multi_gpu def UpperCamelCase__ ( self) -> Any: self.run_seqaseq_quick(distributed=__lowercase) @require_torch_multi_gpu def UpperCamelCase__ ( self) -> Union[str, Any]: self.run_seqaseq_quick(distributed=__lowercase) @unittest.skip('''Requires an update of the env running those tests''') @require_torch_multi_gpu @require_fairscale def UpperCamelCase__ ( self) -> Union[str, Any]: self.run_seqaseq_quick(distributed=__lowercase , extra_args_str='''--sharded_ddp simple''') @unittest.skip('''Requires an update of the env running those tests''') @require_torch_multi_gpu @require_fairscale def UpperCamelCase__ ( self) -> List[Any]: self.run_seqaseq_quick(distributed=__lowercase , extra_args_str='''--sharded_ddp simple --fp16''') @unittest.skip('''Requires an update of the env running those tests''') @require_torch_multi_gpu @require_fairscale def UpperCamelCase__ ( self) -> Any: self.run_seqaseq_quick(distributed=__lowercase , extra_args_str='''--sharded_ddp zero_dp_2''' , predict_with_generate=__lowercase) @unittest.skip('''Requires an update of the env running those tests''') @require_torch_multi_gpu @require_fairscale def UpperCamelCase__ ( self) -> Dict: self.run_seqaseq_quick( distributed=__lowercase , extra_args_str='''--sharded_ddp zero_dp_2 --fp16''' , predict_with_generate=__lowercase) @require_apex @require_torch_gpu def UpperCamelCase__ ( self) -> Optional[int]: # XXX: apex breaks the trainer if it's run twice e.g. run_seq2seq.main() from the same # program and it breaks other tests that run from the same pytest worker, therefore until this is # sorted out it must be run only in an external program, that is distributed=True in this # test and only under one or more gpus - if we want cpu will need to make a special test # # specifically to the problem traced it to self.optimizer.step() - if it's run 2nd time via # 2nd main() call it botches the future eval. # self.run_seqaseq_quick(distributed=__lowercase , extra_args_str='''--fp16 --fp16_backend=apex''') # test 2nd time - was getting eval_loss': nan' # to reproduce the problem set distributed=False self.run_seqaseq_quick(distributed=__lowercase , extra_args_str='''--fp16 --fp16_backend=apex''') @parameterized.expand(['''base''', '''low''', '''high''', '''mixed''']) @require_torch_multi_gpu def UpperCamelCase__ ( self , __lowercase) -> str: # as each sub-test is slow-ish split into multiple sub-tests to avoid CI timeout __UpperCamelCase :List[str] = { # test with the default log_level - should be info and thus log info once '''base''': {'''extra_args_str''': '''''', '''n_matches''': 1}, # test with low log_level and log_level_replica - should be noisy on all processes # now the info string should appear twice on 2 processes '''low''': {'''extra_args_str''': '''--log_level debug --log_level_replica debug''', '''n_matches''': 2}, # test with high log_level and low log_level_replica # now the info string should appear once only on the replica '''high''': {'''extra_args_str''': '''--log_level error --log_level_replica debug''', '''n_matches''': 1}, # test with high log_level and log_level_replica - should be quiet on all processes '''mixed''': {'''extra_args_str''': '''--log_level error --log_level_replica error''', '''n_matches''': 0}, } __UpperCamelCase :Union[str, Any] = experiments[experiment_id] __UpperCamelCase :List[Any] = {'''distributed''': True, '''predict_with_generate''': False, '''do_eval''': False, '''do_predict''': False} __UpperCamelCase :Any = '''Running training''' with CaptureStderr() as cl: self.run_seqaseq_quick(**__lowercase , extra_args_str=data['''extra_args_str''']) __UpperCamelCase :str = len(re.findall(__lowercase , cl.err)) self.assertEqual(__lowercase , data['''n_matches''']) @slow def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :Optional[Any] = self.run_trainer( eval_steps=2 , max_len=128 , model_name=__lowercase , learning_rate=3E-4 , num_train_epochs=10 , distributed=__lowercase , ) # Check metrics __UpperCamelCase :int = TrainerState.load_from_json(os.path.join(__lowercase , '''trainer_state.json''')).log_history __UpperCamelCase :Dict = [log for log in logs if '''eval_loss''' in log.keys()] __UpperCamelCase :Optional[Any] = eval_metrics[0] __UpperCamelCase :List[Any] = eval_metrics[-1] assert first_step_stats["eval_loss"] > last_step_stats["eval_loss"], "model learned nothing" assert isinstance(last_step_stats['''eval_bleu'''] , __lowercase) # test if do_predict saves generations and metrics __UpperCamelCase :Any = os.listdir(__lowercase) __UpperCamelCase :List[Any] = {os.path.basename(__lowercase) for p in contents} assert "generated_predictions.txt" in contents assert "predict_results.json" in contents @slow @require_bitsandbytes def UpperCamelCase__ ( self) -> List[str]: from transformers.training_args import OptimizerNames def train_and_return_metrics(__lowercase) -> Tuple[int, float]: __UpperCamelCase :Optional[Any] = '''--skip_memory_metrics 0''' __UpperCamelCase :List[str] = self.run_trainer( max_len=128 , model_name=__lowercase , learning_rate=3E-4 , num_train_epochs=1 , optim=__lowercase , distributed=__lowercase , extra_args_str=__lowercase , do_eval=__lowercase , do_predict=__lowercase , n_gpus_to_use=1 , ) # Check metrics __UpperCamelCase :int = TrainerState.load_from_json(Path(__lowercase , '''trainer_state.json''')).log_history __UpperCamelCase :List[str] = int(logs[0]['''train_mem_gpu_peaked_delta'''] / 2**20) __UpperCamelCase :Union[str, Any] = int(logs[0]['''train_mem_gpu_alloc_delta'''] / 2**20) __UpperCamelCase :List[str] = logs[0]['''train_loss'''] return gpu_peak_mem_mb, gpu_alloc_mem_mb, loss __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Optional[Any] = train_and_return_metrics(OptimizerNames.ADAMW_TORCH.value) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Dict = train_and_return_metrics(OptimizerNames.ADAMW_BNB.value) __UpperCamelCase :List[Any] = gpu_alloc_mem_orig - gpu_alloc_mem_bnb __UpperCamelCase :List[Any] = gpu_peak_mem_orig + gpu_alloc_mem_orig __UpperCamelCase :int = gpu_peak_mem_bnb + gpu_alloc_mem_bnb __UpperCamelCase :Optional[int] = gpu_total_mem_orig - gpu_total_mem_bnb # sshleifer/student_marian_en_ro_6_1 has 54M parameter, 29M of which is `nn.Embedding` which # doesn't get quantized and remains in fp32. Therefore we only have 25M parameters quantized # in 2 bytes and the diff in optim memory usage is derived as so: # # - normal 25*8=~200MB (8 bytes per param) # - bnb 25*2= ~50MB (2 bytes per param) # # Thus we should expect ~150MB total memory saved. # # Peak memory should be the same - the total should be different by about that same margin # # After leaving a small margin to accommodate for differences between gpus let's check # that we have at least 120MB in savings __UpperCamelCase :str = 120 # uncomment the following if this test starts failing - requires py38 for a new print feature # gpu_peak_mem_diff = gpu_peak_mem_orig - gpu_peak_mem_bnb # print(f"{gpu_alloc_mem_orig=}MB {gpu_peak_mem_orig=}MB {gpu_alloc_mem_orig+gpu_peak_mem_orig=}MB") # print(f" {gpu_alloc_mem_bnb=}MB {gpu_peak_mem_bnb=}MB {gpu_alloc_mem_bnb+gpu_peak_mem_bnb=}MB") # print(f"{gpu_alloc_mem_diff=}MB") # print(f"{gpu_peak_mem_diff=}MB") # print(f"{gpu_total_mem_orig=}MB, {gpu_total_mem_bnb=}MB") # print(f"{gpu_total_mem_diff=}MB, {gpu_total_mem_diff=}MB") self.assertGreater( __lowercase , __lowercase , '''should use ~150MB less alloc gpu memory with BNB, compared to without it for this model but got''' f""" a difference of {gpu_alloc_mem_diff}MB, with gpu_alloc_mem_orig={gpu_alloc_mem_orig}MB and""" f""" gpu_alloc_mem_bnb={gpu_alloc_mem_bnb}MB""" , ) self.assertGreater( __lowercase , __lowercase , '''should use ~150MB less total gpu memory with BNB, compared to without it for this model but got''' f""" a difference of {gpu_total_mem_diff}MB, with gpu_total_mem_orig={gpu_total_mem_orig}MB and""" f""" gpu_total_mem_bnb={gpu_total_mem_bnb}MB""" , ) self.assertEqual( __lowercase , __lowercase , f"""loss should be the same, but got loss_orig={loss_orig}, loss_bnb={loss_bnb}""") def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase = 3E-3 , __lowercase = "adafactor" , __lowercase = False , __lowercase = None , __lowercase = 0 , __lowercase = True , __lowercase = True , __lowercase = True , __lowercase = True , __lowercase = None , ) -> Union[str, Any]: __UpperCamelCase :Any = self.test_file_dir / '''../fixtures/tests_samples/wmt_en_ro''' __UpperCamelCase :List[Any] = self.get_auto_remove_tmp_dir() __UpperCamelCase :Optional[Any] = f""" --model_name_or_path {model_name} --train_file {data_dir}/train.json --validation_file {data_dir}/val.json --test_file {data_dir}/test.json --output_dir {output_dir} --overwrite_output_dir --max_train_samples 8 --max_source_length {max_len} --max_target_length {max_len} --do_train --num_train_epochs {str(__lowercase)} --per_device_train_batch_size 4 --learning_rate {learning_rate} --warmup_steps 8 --logging_steps 0 --logging_strategy no --save_steps {str(__lowercase)} --group_by_length --label_smoothing_factor 0.1 --target_lang ro_RO --source_lang en_XX """.split() __UpperCamelCase :Dict = f""" --do_eval --per_device_eval_batch_size 4 --max_eval_samples 8 --val_max_target_length {max_len} --evaluation_strategy steps --eval_steps {str(__lowercase)} """.split() __UpperCamelCase :Tuple = ''' --do_predict '''.split() __UpperCamelCase :Optional[int] = [] if do_train: args += args_train if do_eval: args += args_eval if do_predict: args += args_predict if predict_with_generate: args += "--predict_with_generate".split() if do_train: if optim == "adafactor": args += "--adafactor".split() else: args += f"""--optim {optim}""".split() if extra_args_str is not None: args += extra_args_str.split() if distributed: if n_gpus_to_use is None: __UpperCamelCase :Optional[Any] = get_gpu_count() __UpperCamelCase :List[Any] = get_torch_dist_unique_port() __UpperCamelCase :Any = f""" -m torch.distributed.run --nproc_per_node={n_gpus_to_use} --master_port={master_port} {self.examples_dir_str}/pytorch/translation/run_translation.py """.split() __UpperCamelCase :str = [sys.executable] + distributed_args + args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(__lowercase , env=self.get_env()) else: __UpperCamelCase :int = ['''run_translation.py'''] + args with patch.object(__lowercase , '''argv''' , __lowercase): main() return output_dir
43
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. __lowercase = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. __lowercase = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. __lowercase = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Any = len([g for position, g in enumerate(SCREAMING_SNAKE_CASE ) if g == main_target[position]] ) return (item, float(SCREAMING_SNAKE_CASE )) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = random.randint(0 , len(SCREAMING_SNAKE_CASE ) - 1 ) __UpperCamelCase :Tuple = parent_a[:random_slice] + parent_a[random_slice:] __UpperCamelCase :Union[str, Any] = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :int = list(SCREAMING_SNAKE_CASE ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: __UpperCamelCase :str = random.choice(SCREAMING_SNAKE_CASE ) return "".join(SCREAMING_SNAKE_CASE ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , ): '''simple docstring''' __UpperCamelCase :int = [] # Generate more children proportionally to the fitness score. __UpperCamelCase :int = int(parent_a[1] * 100 ) + 1 __UpperCamelCase :List[str] = 10 if child_n >= 10 else child_n for _ in range(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = population_score[random.randint(0 , SCREAMING_SNAKE_CASE )][0] __UpperCamelCase , __UpperCamelCase :Any = crossover(parent_a[0] , SCREAMING_SNAKE_CASE ) # Append new string to the population list. pop.append(mutate(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) pop.append(mutate(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) return pop def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = True ): '''simple docstring''' if N_POPULATION < N_SELECTED: __UpperCamelCase :List[Any] = f"""{N_POPULATION} must be bigger than {N_SELECTED}""" raise ValueError(SCREAMING_SNAKE_CASE ) # Verify that the target contains no genes besides the ones inside genes variable. __UpperCamelCase :List[str] = sorted({c for c in target if c not in genes} ) if not_in_genes_list: __UpperCamelCase :Optional[int] = f"""{not_in_genes_list} is not in genes list, evolution cannot converge""" raise ValueError(SCREAMING_SNAKE_CASE ) # Generate random starting population. __UpperCamelCase :int = [] for _ in range(SCREAMING_SNAKE_CASE ): population.append(''''''.join([random.choice(SCREAMING_SNAKE_CASE ) for i in range(len(SCREAMING_SNAKE_CASE ) )] ) ) # Just some logs to know what the algorithms is doing. __UpperCamelCase , __UpperCamelCase :List[Any] = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(SCREAMING_SNAKE_CASE ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. __UpperCamelCase :Tuple = [evaluate(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for item in population] # Check if there is a matching evolution. __UpperCamelCase :Tuple = sorted(SCREAMING_SNAKE_CASE , key=lambda SCREAMING_SNAKE_CASE : x[1] , reverse=SCREAMING_SNAKE_CASE ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( f"""\nGeneration: {generation}""" f"""\nTotal Population:{total_population}""" f"""\nBest score: {population_score[0][1]}""" f"""\nBest string: {population_score[0][0]}""" ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. __UpperCamelCase :str = population[: int(N_POPULATION / 3 )] population.clear() population.extend(SCREAMING_SNAKE_CASE ) # Normalize population score to be between 0 and 1. __UpperCamelCase :Union[str, Any] = [ (item, score / len(SCREAMING_SNAKE_CASE )) for item, score in population_score ] # This is selection for i in range(SCREAMING_SNAKE_CASE ): population.extend(select(population_score[int(SCREAMING_SNAKE_CASE )] , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(SCREAMING_SNAKE_CASE ) > N_POPULATION: break if __name__ == "__main__": __lowercase = ( '''This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!''' ) __lowercase = list( ''' ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm''' '''nopqrstuvwxyz.,;!?+-*#@^\'èéòà€ù=)(&%$£/\\''' ) __lowercase , __lowercase , __lowercase = basic(target_str, genes_list) print( F'\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}' )
43
1
import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all image processors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...image_processing_utils import ImageProcessingMixin from ...utils import CONFIG_NAME, IMAGE_PROCESSOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) __lowercase = logging.get_logger(__name__) __lowercase = OrderedDict( [ ('''align''', '''EfficientNetImageProcessor'''), ('''beit''', '''BeitImageProcessor'''), ('''bit''', '''BitImageProcessor'''), ('''blip''', '''BlipImageProcessor'''), ('''blip-2''', '''BlipImageProcessor'''), ('''bridgetower''', '''BridgeTowerImageProcessor'''), ('''chinese_clip''', '''ChineseCLIPImageProcessor'''), ('''clip''', '''CLIPImageProcessor'''), ('''clipseg''', '''ViTImageProcessor'''), ('''conditional_detr''', '''ConditionalDetrImageProcessor'''), ('''convnext''', '''ConvNextImageProcessor'''), ('''convnextv2''', '''ConvNextImageProcessor'''), ('''cvt''', '''ConvNextImageProcessor'''), ('''data2vec-vision''', '''BeitImageProcessor'''), ('''deformable_detr''', '''DeformableDetrImageProcessor'''), ('''deit''', '''DeiTImageProcessor'''), ('''deta''', '''DetaImageProcessor'''), ('''detr''', '''DetrImageProcessor'''), ('''dinat''', '''ViTImageProcessor'''), ('''donut-swin''', '''DonutImageProcessor'''), ('''dpt''', '''DPTImageProcessor'''), ('''efficientformer''', '''EfficientFormerImageProcessor'''), ('''efficientnet''', '''EfficientNetImageProcessor'''), ('''flava''', '''FlavaImageProcessor'''), ('''focalnet''', '''BitImageProcessor'''), ('''git''', '''CLIPImageProcessor'''), ('''glpn''', '''GLPNImageProcessor'''), ('''groupvit''', '''CLIPImageProcessor'''), ('''imagegpt''', '''ImageGPTImageProcessor'''), ('''instructblip''', '''BlipImageProcessor'''), ('''layoutlmv2''', '''LayoutLMv2ImageProcessor'''), ('''layoutlmv3''', '''LayoutLMv3ImageProcessor'''), ('''levit''', '''LevitImageProcessor'''), ('''mask2former''', '''Mask2FormerImageProcessor'''), ('''maskformer''', '''MaskFormerImageProcessor'''), ('''mgp-str''', '''ViTImageProcessor'''), ('''mobilenet_v1''', '''MobileNetV1ImageProcessor'''), ('''mobilenet_v2''', '''MobileNetV2ImageProcessor'''), ('''mobilevit''', '''MobileViTImageProcessor'''), ('''mobilevit''', '''MobileViTImageProcessor'''), ('''mobilevitv2''', '''MobileViTImageProcessor'''), ('''nat''', '''ViTImageProcessor'''), ('''oneformer''', '''OneFormerImageProcessor'''), ('''owlvit''', '''OwlViTImageProcessor'''), ('''perceiver''', '''PerceiverImageProcessor'''), ('''pix2struct''', '''Pix2StructImageProcessor'''), ('''poolformer''', '''PoolFormerImageProcessor'''), ('''regnet''', '''ConvNextImageProcessor'''), ('''resnet''', '''ConvNextImageProcessor'''), ('''sam''', '''SamImageProcessor'''), ('''segformer''', '''SegformerImageProcessor'''), ('''swiftformer''', '''ViTImageProcessor'''), ('''swin''', '''ViTImageProcessor'''), ('''swin2sr''', '''Swin2SRImageProcessor'''), ('''swinv2''', '''ViTImageProcessor'''), ('''table-transformer''', '''DetrImageProcessor'''), ('''timesformer''', '''VideoMAEImageProcessor'''), ('''tvlt''', '''TvltImageProcessor'''), ('''upernet''', '''SegformerImageProcessor'''), ('''van''', '''ConvNextImageProcessor'''), ('''videomae''', '''VideoMAEImageProcessor'''), ('''vilt''', '''ViltImageProcessor'''), ('''vit''', '''ViTImageProcessor'''), ('''vit_hybrid''', '''ViTHybridImageProcessor'''), ('''vit_mae''', '''ViTImageProcessor'''), ('''vit_msn''', '''ViTImageProcessor'''), ('''xclip''', '''CLIPImageProcessor'''), ('''yolos''', '''YolosImageProcessor'''), ] ) __lowercase = _LazyAutoMapping(CONFIG_MAPPING_NAMES, IMAGE_PROCESSOR_MAPPING_NAMES) def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' for module_name, extractors in IMAGE_PROCESSOR_MAPPING_NAMES.items(): if class_name in extractors: __UpperCamelCase :Tuple = model_type_to_module_name(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Tuple = importlib.import_module(f""".{module_name}""" , '''transformers.models''' ) try: return getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) except AttributeError: continue for _, extractor in IMAGE_PROCESSOR_MAPPING._extra_content.items(): if getattr(SCREAMING_SNAKE_CASE , '''__name__''' , SCREAMING_SNAKE_CASE ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. __UpperCamelCase :List[str] = importlib.import_module('''transformers''' ) if hasattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): return getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return None def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = False , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = False , **SCREAMING_SNAKE_CASE , ): '''simple docstring''' __UpperCamelCase :Dict = get_file_from_repo( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , cache_dir=SCREAMING_SNAKE_CASE , force_download=SCREAMING_SNAKE_CASE , resume_download=SCREAMING_SNAKE_CASE , proxies=SCREAMING_SNAKE_CASE , use_auth_token=SCREAMING_SNAKE_CASE , revision=SCREAMING_SNAKE_CASE , local_files_only=SCREAMING_SNAKE_CASE , ) if resolved_config_file is None: logger.info( '''Could not locate the image processor configuration file, will try to use the model config instead.''' ) return {} with open(SCREAMING_SNAKE_CASE , encoding='''utf-8''' ) as reader: return json.load(SCREAMING_SNAKE_CASE ) class lowerCamelCase_ : '''simple docstring''' def __init__( self) -> Optional[Any]: raise EnvironmentError( '''AutoImageProcessor is designed to be instantiated ''' '''using the `AutoImageProcessor.from_pretrained(pretrained_model_name_or_path)` method.''') @classmethod @replace_list_option_in_docstrings(__lowercase) def UpperCamelCase__ ( cls , __lowercase , **__lowercase) -> Union[str, Any]: __UpperCamelCase :Dict = kwargs.pop('''config''' , __lowercase) __UpperCamelCase :List[str] = kwargs.pop('''trust_remote_code''' , __lowercase) __UpperCamelCase :Optional[int] = True __UpperCamelCase , __UpperCamelCase :List[str] = ImageProcessingMixin.get_image_processor_dict(__lowercase , **__lowercase) __UpperCamelCase :Optional[int] = config_dict.get('''image_processor_type''' , __lowercase) __UpperCamelCase :Any = None if "AutoImageProcessor" in config_dict.get('''auto_map''' , {}): __UpperCamelCase :int = config_dict['''auto_map''']['''AutoImageProcessor'''] # If we still don't have the image processor class, check if we're loading from a previous feature extractor config # and if so, infer the image processor class from there. if image_processor_class is None and image_processor_auto_map is None: __UpperCamelCase :Dict = config_dict.pop('''feature_extractor_type''' , __lowercase) if feature_extractor_class is not None: logger.warning( '''Could not find image processor class in the image processor config or the model config. Loading''' ''' based on pattern matching with the model\'s feature extractor configuration.''') __UpperCamelCase :str = feature_extractor_class.replace('''FeatureExtractor''' , '''ImageProcessor''') if "AutoFeatureExtractor" in config_dict.get('''auto_map''' , {}): __UpperCamelCase :Optional[int] = config_dict['''auto_map''']['''AutoFeatureExtractor'''] __UpperCamelCase :Union[str, Any] = feature_extractor_auto_map.replace('''FeatureExtractor''' , '''ImageProcessor''') logger.warning( '''Could not find image processor auto map in the image processor config or the model config.''' ''' Loading based on pattern matching with the model\'s feature extractor configuration.''') # If we don't find the image processor class in the image processor config, let's try the model config. if image_processor_class is None and image_processor_auto_map is None: if not isinstance(__lowercase , __lowercase): __UpperCamelCase :Union[str, Any] = AutoConfig.from_pretrained(__lowercase , **__lowercase) # It could be in `config.image_processor_type`` __UpperCamelCase :List[str] = getattr(__lowercase , '''image_processor_type''' , __lowercase) if hasattr(__lowercase , '''auto_map''') and "AutoImageProcessor" in config.auto_map: __UpperCamelCase :Union[str, Any] = config.auto_map['''AutoImageProcessor'''] if image_processor_class is not None: __UpperCamelCase :Tuple = image_processor_class_from_name(__lowercase) __UpperCamelCase :Union[str, Any] = image_processor_auto_map is not None __UpperCamelCase :str = image_processor_class is not None or type(__lowercase) in IMAGE_PROCESSOR_MAPPING __UpperCamelCase :Optional[Any] = resolve_trust_remote_code( __lowercase , __lowercase , __lowercase , __lowercase) if has_remote_code and trust_remote_code: __UpperCamelCase :int = get_class_from_dynamic_module( __lowercase , __lowercase , **__lowercase) __UpperCamelCase :Dict = kwargs.pop('''code_revision''' , __lowercase) if os.path.isdir(__lowercase): image_processor_class.register_for_auto_class() return image_processor_class.from_dict(__lowercase , **__lowercase) elif image_processor_class is not None: return image_processor_class.from_dict(__lowercase , **__lowercase) # Last try: we use the IMAGE_PROCESSOR_MAPPING. elif type(__lowercase) in IMAGE_PROCESSOR_MAPPING: __UpperCamelCase :int = IMAGE_PROCESSOR_MAPPING[type(__lowercase)] return image_processor_class.from_dict(__lowercase , **__lowercase) raise ValueError( f"""Unrecognized image processor in {pretrained_model_name_or_path}. Should have a """ f"""`image_processor_type` key in its {IMAGE_PROCESSOR_NAME} of {CONFIG_NAME}, or one of the following """ f"""`model_type` keys in its {CONFIG_NAME}: {', '.join(c for c in IMAGE_PROCESSOR_MAPPING_NAMES.keys())}""") @staticmethod def UpperCamelCase__ ( __lowercase , __lowercase) -> Union[str, Any]: IMAGE_PROCESSOR_MAPPING.register(__lowercase , __lowercase)
43
import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler __lowercase = 16 __lowercase = 32 def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 16 , SCREAMING_SNAKE_CASE = "bert-base-cased" ): '''simple docstring''' __UpperCamelCase :List[str] = AutoTokenizer.from_pretrained(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(SCREAMING_SNAKE_CASE ): # max_length=None => use the model max length (it's actually the default) __UpperCamelCase :int = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=SCREAMING_SNAKE_CASE , max_length=SCREAMING_SNAKE_CASE ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset __UpperCamelCase :Tuple = datasets.map( SCREAMING_SNAKE_CASE , batched=SCREAMING_SNAKE_CASE , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , load_from_cache_file=SCREAMING_SNAKE_CASE ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library __UpperCamelCase :List[str] = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(SCREAMING_SNAKE_CASE ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(SCREAMING_SNAKE_CASE , padding='''max_length''' , max_length=128 , return_tensors='''pt''' ) return tokenizer.pad(SCREAMING_SNAKE_CASE , padding='''longest''' , return_tensors='''pt''' ) # Instantiate dataloaders. __UpperCamelCase :Union[str, Any] = DataLoader( tokenized_datasets['''train'''] , shuffle=SCREAMING_SNAKE_CASE , collate_fn=SCREAMING_SNAKE_CASE , batch_size=SCREAMING_SNAKE_CASE ) __UpperCamelCase :Dict = DataLoader( tokenized_datasets['''validation'''] , shuffle=SCREAMING_SNAKE_CASE , collate_fn=SCREAMING_SNAKE_CASE , batch_size=SCREAMING_SNAKE_CASE ) return train_dataloader, eval_dataloader def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs __UpperCamelCase :int = config['''lr'''] __UpperCamelCase :str = int(config['''num_epochs'''] ) __UpperCamelCase :Any = int(config['''seed'''] ) __UpperCamelCase :Dict = int(config['''batch_size'''] ) __UpperCamelCase :Optional[Any] = args.model_name_or_path set_seed(SCREAMING_SNAKE_CASE ) __UpperCamelCase , __UpperCamelCase :Dict = get_dataloaders(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) __UpperCamelCase :Any = AutoModelForSequenceClassification.from_pretrained(SCREAMING_SNAKE_CASE , return_dict=SCREAMING_SNAKE_CASE ) # Instantiate optimizer __UpperCamelCase :List[str] = ( AdamW if accelerator.state.deepspeed_plugin is None or '''optimizer''' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) __UpperCamelCase :Optional[Any] = optimizer_cls(params=model.parameters() , lr=SCREAMING_SNAKE_CASE ) if accelerator.state.deepspeed_plugin is not None: __UpperCamelCase :Dict = accelerator.state.deepspeed_plugin.deepspeed_config[ '''gradient_accumulation_steps''' ] else: __UpperCamelCase :Dict = 1 __UpperCamelCase :Tuple = (len(SCREAMING_SNAKE_CASE ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): __UpperCamelCase :str = get_linear_schedule_with_warmup( optimizer=SCREAMING_SNAKE_CASE , num_warmup_steps=0 , num_training_steps=SCREAMING_SNAKE_CASE , ) else: __UpperCamelCase :Dict = DummyScheduler(SCREAMING_SNAKE_CASE , total_num_steps=SCREAMING_SNAKE_CASE , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :int = accelerator.prepare( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # We need to keep track of how many total steps we have iterated over __UpperCamelCase :List[Any] = 0 # We also need to keep track of the stating epoch so files are named properly __UpperCamelCase :Dict = 0 # Now we train the model __UpperCamelCase :Any = evaluate.load('''glue''' , '''mrpc''' ) __UpperCamelCase :Union[str, Any] = 0 __UpperCamelCase :Optional[int] = {} for epoch in range(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): model.train() for step, batch in enumerate(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = model(**SCREAMING_SNAKE_CASE ) __UpperCamelCase :Tuple = outputs.loss __UpperCamelCase :str = loss / gradient_accumulation_steps accelerator.backward(SCREAMING_SNAKE_CASE ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() __UpperCamelCase :Any = 0 for step, batch in enumerate(SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): __UpperCamelCase :Any = model(**SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times __UpperCamelCase , __UpperCamelCase :List[Any] = accelerator.gather( (predictions, batch['''labels''']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(SCREAMING_SNAKE_CASE ) - 1: __UpperCamelCase :List[str] = predictions[: len(eval_dataloader.dataset ) - samples_seen] __UpperCamelCase :Optional[int] = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=SCREAMING_SNAKE_CASE , references=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :Dict = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , SCREAMING_SNAKE_CASE ) __UpperCamelCase :str = eval_metric['''accuracy'''] if best_performance < eval_metric["accuracy"]: __UpperCamelCase :int = eval_metric['''accuracy'''] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f"""Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}""" accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , '''all_results.json''' ) , '''w''' ) as f: json.dump(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = argparse.ArgumentParser(description='''Simple example of training script tracking peak GPU memory usage.''' ) parser.add_argument( '''--model_name_or_path''' , type=SCREAMING_SNAKE_CASE , default='''bert-base-cased''' , help='''Path to pretrained model or model identifier from huggingface.co/models.''' , required=SCREAMING_SNAKE_CASE , ) parser.add_argument( '''--output_dir''' , type=SCREAMING_SNAKE_CASE , default='''.''' , help='''Optional save directory where all checkpoint folders will be stored. Default is the current working directory.''' , ) parser.add_argument( '''--performance_lower_bound''' , type=SCREAMING_SNAKE_CASE , default=SCREAMING_SNAKE_CASE , help='''Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.''' , ) parser.add_argument( '''--num_epochs''' , type=SCREAMING_SNAKE_CASE , default=3 , help='''Number of train epochs.''' , ) __UpperCamelCase :List[str] = parser.parse_args() __UpperCamelCase :Tuple = {'''lr''': 2e-5, '''num_epochs''': args.num_epochs, '''seed''': 42, '''batch_size''': 16} training_function(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
43
1
__lowercase = {str(digit): digit**5 for digit in range(10)} def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' return sum(DIGITS_FIFTH_POWER[digit] for digit in str(SCREAMING_SNAKE_CASE ) ) def lowerCamelCase ( ): '''simple docstring''' return sum( number for number in range(1_000 , 1_000_000 ) if number == digits_fifth_powers_sum(SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": print(solution())
43
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __lowercase = logging.get_logger(__name__) __lowercase = { '''SenseTime/deformable-detr''': '''https://huggingface.co/sensetime/deformable-detr/resolve/main/config.json''', # See all Deformable DETR models at https://huggingface.co/models?filter=deformable-detr } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : List[str] = """deformable_detr""" a__ : Union[str, Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , __lowercase=True , __lowercase=None , __lowercase=3 , __lowercase=300 , __lowercase=1_024 , __lowercase=6 , __lowercase=1_024 , __lowercase=8 , __lowercase=6 , __lowercase=1_024 , __lowercase=8 , __lowercase=0.0 , __lowercase=True , __lowercase="relu" , __lowercase=256 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.02 , __lowercase=1.0 , __lowercase=True , __lowercase=False , __lowercase="sine" , __lowercase="resnet50" , __lowercase=True , __lowercase=False , __lowercase=4 , __lowercase=4 , __lowercase=4 , __lowercase=False , __lowercase=300 , __lowercase=False , __lowercase=1 , __lowercase=5 , __lowercase=2 , __lowercase=1 , __lowercase=1 , __lowercase=5 , __lowercase=2 , __lowercase=0.1 , __lowercase=0.25 , __lowercase=False , **__lowercase , ) -> int: if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''') if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''') __UpperCamelCase :str = CONFIG_MAPPING['''resnet'''](out_features=['''stage4''']) elif isinstance(__lowercase , __lowercase): __UpperCamelCase :str = backbone_config.get('''model_type''') __UpperCamelCase :Tuple = CONFIG_MAPPING[backbone_model_type] __UpperCamelCase :Any = config_class.from_dict(__lowercase) __UpperCamelCase :int = use_timm_backbone __UpperCamelCase :Dict = backbone_config __UpperCamelCase :Any = num_channels __UpperCamelCase :Optional[int] = num_queries __UpperCamelCase :Any = max_position_embeddings __UpperCamelCase :str = d_model __UpperCamelCase :Tuple = encoder_ffn_dim __UpperCamelCase :Union[str, Any] = encoder_layers __UpperCamelCase :List[Any] = encoder_attention_heads __UpperCamelCase :Any = decoder_ffn_dim __UpperCamelCase :List[str] = decoder_layers __UpperCamelCase :int = decoder_attention_heads __UpperCamelCase :str = dropout __UpperCamelCase :Any = attention_dropout __UpperCamelCase :int = activation_dropout __UpperCamelCase :List[Any] = activation_function __UpperCamelCase :List[Any] = init_std __UpperCamelCase :List[Any] = init_xavier_std __UpperCamelCase :int = encoder_layerdrop __UpperCamelCase :str = auxiliary_loss __UpperCamelCase :Optional[Any] = position_embedding_type __UpperCamelCase :Union[str, Any] = backbone __UpperCamelCase :Any = use_pretrained_backbone __UpperCamelCase :str = dilation # deformable attributes __UpperCamelCase :Optional[Any] = num_feature_levels __UpperCamelCase :str = encoder_n_points __UpperCamelCase :int = decoder_n_points __UpperCamelCase :Union[str, Any] = two_stage __UpperCamelCase :Optional[Any] = two_stage_num_proposals __UpperCamelCase :Dict = with_box_refine if two_stage is True and with_box_refine is False: raise ValueError('''If two_stage is True, with_box_refine must be True.''') # Hungarian matcher __UpperCamelCase :Optional[int] = class_cost __UpperCamelCase :List[Any] = bbox_cost __UpperCamelCase :str = giou_cost # Loss coefficients __UpperCamelCase :Tuple = mask_loss_coefficient __UpperCamelCase :Tuple = dice_loss_coefficient __UpperCamelCase :int = bbox_loss_coefficient __UpperCamelCase :Any = giou_loss_coefficient __UpperCamelCase :Dict = eos_coefficient __UpperCamelCase :Optional[Any] = focal_alpha __UpperCamelCase :Optional[Any] = disable_custom_kernels super().__init__(is_encoder_decoder=__lowercase , **__lowercase) @property def UpperCamelCase__ ( self) -> int: return self.encoder_attention_heads @property def UpperCamelCase__ ( self) -> int: return self.d_model def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Dict = copy.deepcopy(self.__dict__) if self.backbone_config is not None: __UpperCamelCase :Tuple = self.backbone_config.to_dict() __UpperCamelCase :List[Any] = self.__class__.model_type return output
43
1
from __future__ import annotations from collections.abc import Callable from typing import Any, Generic, TypeVar __lowercase = TypeVar('''T''') class lowerCamelCase_ ( Generic[T] ): '''simple docstring''' def __init__( self , __lowercase , __lowercase) -> None: __UpperCamelCase :Any | T = None __UpperCamelCase :int = len(__lowercase) __UpperCamelCase :list[T] = [any_type for _ in range(self.N)] + arr __UpperCamelCase :int = fnc self.build() def UpperCamelCase__ ( self) -> None: for p in range(self.N - 1 , 0 , -1): __UpperCamelCase :List[str] = self.fn(self.st[p * 2] , self.st[p * 2 + 1]) def UpperCamelCase__ ( self , __lowercase , __lowercase) -> None: p += self.N __UpperCamelCase :int = v while p > 1: __UpperCamelCase :Optional[int] = p // 2 __UpperCamelCase :Optional[int] = self.fn(self.st[p * 2] , self.st[p * 2 + 1]) def UpperCamelCase__ ( self , __lowercase , __lowercase) -> T | None: # noqa: E741 __UpperCamelCase , __UpperCamelCase :Any = l + self.N, r + self.N __UpperCamelCase :T | None = None while l <= r: if l % 2 == 1: __UpperCamelCase :Optional[int] = self.st[l] if res is None else self.fn(__lowercase , self.st[l]) if r % 2 == 0: __UpperCamelCase :Dict = self.st[r] if res is None else self.fn(__lowercase , self.st[r]) __UpperCamelCase , __UpperCamelCase :Dict = (l + 1) // 2, (r - 1) // 2 return res if __name__ == "__main__": from functools import reduce __lowercase = [1, 10, -2, 9, -3, 8, 4, -7, 5, 6, 11, -12] __lowercase = { 0: 7, 1: 2, 2: 6, 3: -14, 4: 5, 5: 4, 6: 7, 7: -10, 8: 9, 9: 10, 10: 12, 11: 1, } __lowercase = SegmentTree(test_array, min) __lowercase = SegmentTree(test_array, max) __lowercase = SegmentTree(test_array, lambda a, b: a + b) def lowerCamelCase ( ): '''simple docstring''' for i in range(len(SCREAMING_SNAKE_CASE ) ): for j in range(SCREAMING_SNAKE_CASE , len(SCREAMING_SNAKE_CASE ) ): __UpperCamelCase :Union[str, Any] = reduce(SCREAMING_SNAKE_CASE , test_array[i : j + 1] ) __UpperCamelCase :Dict = reduce(SCREAMING_SNAKE_CASE , test_array[i : j + 1] ) __UpperCamelCase :str = reduce(lambda SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : a + b , test_array[i : j + 1] ) assert min_range == min_segment_tree.query(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) assert max_range == max_segment_tree.query(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) assert sum_range == sum_segment_tree.query(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) test_all_segments() for index, value in test_updates.items(): __lowercase = value min_segment_tree.update(index, value) max_segment_tree.update(index, value) sum_segment_tree.update(index, value) test_all_segments()
43
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.auto import AutoModelForSequenceClassification, AutoTokenizer from .base import PipelineTool class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[Any] = """facebook/bart-large-mnli""" a__ : int = ( """This is a tool that classifies an English text using provided labels. It takes two inputs: `text`, which """ """should be the text to classify, and `labels`, which should be the list of labels to use for classification. """ """It returns the most likely label in the list of provided `labels` for the input text.""" ) a__ : Optional[Any] = """text_classifier""" a__ : Any = AutoTokenizer a__ : str = AutoModelForSequenceClassification a__ : str = ["""text""", ["""text"""]] a__ : Optional[int] = ["""text"""] def UpperCamelCase__ ( self) -> Union[str, Any]: super().setup() __UpperCamelCase :int = self.model.config __UpperCamelCase :Optional[Any] = -1 for idx, label in config.idalabel.items(): if label.lower().startswith('''entail'''): __UpperCamelCase :List[Any] = int(__lowercase) if self.entailment_id == -1: raise ValueError('''Could not determine the entailment ID from the model config, please pass it at init.''') def UpperCamelCase__ ( self , __lowercase , __lowercase) -> Union[str, Any]: __UpperCamelCase :Any = labels return self.pre_processor( [text] * len(__lowercase) , [f"""This example is {label}""" for label in labels] , return_tensors='''pt''' , padding='''max_length''' , ) def UpperCamelCase__ ( self , __lowercase) -> Optional[Any]: __UpperCamelCase :List[Any] = outputs.logits __UpperCamelCase :Any = torch.argmax(logits[:, 2]).item() return self._labels[label_id]
43
1
def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' return x if y == 0 else greatest_common_divisor(SCREAMING_SNAKE_CASE , x % y ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' return (x * y) // greatest_common_divisor(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) def lowerCamelCase ( SCREAMING_SNAKE_CASE = 20 ): '''simple docstring''' __UpperCamelCase :Tuple = 1 for i in range(1 , n + 1 ): __UpperCamelCase :List[Any] = lcm(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) return g if __name__ == "__main__": print(F'{solution() = }')
43
import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModel, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import AutoencoderKL, DDIMScheduler, DDPMScheduler, StableUnCLIPImgaImgPipeline, UNetaDConditionModel from diffusers.pipelines.pipeline_utils import DiffusionPipeline from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import ( enable_full_determinism, floats_tensor, load_image, load_numpy, require_torch_gpu, skip_mps, slow, torch_device, ) from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : int = StableUnCLIPImgaImgPipeline a__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS a__ : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS a__ : Optional[Any] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess a__ : int = frozenset([] ) def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Tuple = 32 __UpperCamelCase :Optional[int] = embedder_hidden_size # image encoding components __UpperCamelCase :Union[str, Any] = CLIPImageProcessor(crop_size=32 , size=32) torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPVisionModelWithProjection( CLIPVisionConfig( hidden_size=__lowercase , projection_dim=__lowercase , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , )) # regular denoising components torch.manual_seed(0) __UpperCamelCase :str = StableUnCLIPImageNormalizer(embedding_dim=__lowercase) __UpperCamelCase :Optional[int] = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''') torch.manual_seed(0) __UpperCamelCase :Union[str, Any] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') torch.manual_seed(0) __UpperCamelCase :Dict = CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=__lowercase , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , )) torch.manual_seed(0) __UpperCamelCase :List[Any] = UNetaDConditionModel( sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(32, 64) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=__lowercase , layers_per_block=1 , upcast_attention=__lowercase , use_linear_projection=__lowercase , ) torch.manual_seed(0) __UpperCamelCase :Tuple = DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.0_00_85 , beta_end=0.0_12 , prediction_type='''v_prediction''' , set_alpha_to_one=__lowercase , steps_offset=1 , ) torch.manual_seed(0) __UpperCamelCase :List[str] = AutoencoderKL() __UpperCamelCase :Tuple = { # image encoding components '''feature_extractor''': feature_extractor, '''image_encoder''': image_encoder.eval(), # image noising components '''image_normalizer''': image_normalizer.eval(), '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder.eval(), '''unet''': unet.eval(), '''scheduler''': scheduler, '''vae''': vae.eval(), } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0 , __lowercase=True) -> str: if str(__lowercase).startswith('''mps'''): __UpperCamelCase :Union[str, Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :int = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :int = floats_tensor((1, 3, 32, 32) , rng=random.Random(__lowercase)).to(__lowercase) if pil_image: __UpperCamelCase :List[Any] = input_image * 0.5 + 0.5 __UpperCamelCase :Optional[Any] = input_image.clamp(0 , 1) __UpperCamelCase :int = input_image.cpu().permute(0 , 2 , 3 , 1).float().numpy() __UpperCamelCase :Optional[Any] = DiffusionPipeline.numpy_to_pil(__lowercase)[0] return { "prompt": "An anime racoon running a marathon", "image": input_image, "generator": generator, "num_inference_steps": 2, "output_type": "np", } @skip_mps def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :Dict = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Tuple = self.get_dummy_components() __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline(**__lowercase) __UpperCamelCase :Optional[Any] = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :List[Any] = self.get_dummy_inputs(__lowercase) inputs.update({'''image_embeds''': None}) __UpperCamelCase :Any = sd_pipe(**__lowercase).images __UpperCamelCase :List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCamelCase :List[Any] = np.array([0.38_72, 0.72_24, 0.56_01, 0.47_41, 0.68_72, 0.58_14, 0.46_36, 0.38_67, 0.50_78]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-3 def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_attention_slicing_forward_pass(test_max_difference=__lowercase) def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :Optional[Any] = torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=__lowercase) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase__ ( self) -> Union[str, Any]: self._test_xformers_attention_forwardGenerator_pass(test_max_difference=__lowercase) @slow @require_torch_gpu class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Union[str, Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :int = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-l-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Dict = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :Dict = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Optional[Any] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') __UpperCamelCase :Any = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_h_img2img_anime_turtle_fp16.npy''') __UpperCamelCase :Any = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :Optional[int] = pipe(__lowercase , '''anime turle''' , generator=__lowercase , output_type='''np''') __UpperCamelCase :List[Any] = output.images[0] assert image.shape == (768, 768, 3) assert_mean_pixel_difference(__lowercase , __lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Dict = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/turtle.png''') torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __UpperCamelCase :List[Any] = StableUnCLIPImgaImgPipeline.from_pretrained( '''fusing/stable-unclip-2-1-h-img2img''' , torch_dtype=torch.floataa) __UpperCamelCase :Union[str, Any] = pipe.to(__lowercase) pipe.set_progress_bar_config(disable=__lowercase) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __UpperCamelCase :Optional[Any] = pipe( __lowercase , '''anime turtle''' , num_inference_steps=2 , output_type='''np''' , ) __UpperCamelCase :int = torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 10**9
43
1
import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class lowerCamelCase_ : '''simple docstring''' def __init__( self , __lowercase , __lowercase=13 , __lowercase=7 , __lowercase=True , __lowercase=True , __lowercase=False , __lowercase=True , __lowercase=99 , __lowercase=32 , __lowercase=5 , __lowercase=4 , __lowercase=37 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=512 , __lowercase=16 , __lowercase=2 , __lowercase=0.02 , __lowercase=3 , __lowercase=4 , __lowercase=None , ) -> Dict: __UpperCamelCase :Union[str, Any] = parent __UpperCamelCase :List[str] = batch_size __UpperCamelCase :int = seq_length __UpperCamelCase :str = is_training __UpperCamelCase :Optional[Any] = use_input_mask __UpperCamelCase :List[Any] = use_token_type_ids __UpperCamelCase :Optional[Any] = use_labels __UpperCamelCase :List[Any] = vocab_size __UpperCamelCase :Optional[int] = hidden_size __UpperCamelCase :Tuple = num_hidden_layers __UpperCamelCase :Tuple = num_attention_heads __UpperCamelCase :Optional[int] = intermediate_size __UpperCamelCase :Union[str, Any] = hidden_act __UpperCamelCase :int = hidden_dropout_prob __UpperCamelCase :Optional[int] = attention_probs_dropout_prob __UpperCamelCase :Tuple = max_position_embeddings __UpperCamelCase :str = type_vocab_size __UpperCamelCase :List[Any] = type_sequence_label_size __UpperCamelCase :Optional[int] = initializer_range __UpperCamelCase :str = num_labels __UpperCamelCase :Any = num_choices __UpperCamelCase :Any = scope def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __UpperCamelCase :int = None if self.use_input_mask: __UpperCamelCase :Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length]) __UpperCamelCase :str = None if self.use_token_type_ids: __UpperCamelCase :Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size) __UpperCamelCase :Optional[Any] = None __UpperCamelCase :Dict = None __UpperCamelCase :Any = None if self.use_labels: __UpperCamelCase :List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size) __UpperCamelCase :Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels) __UpperCamelCase :Optional[int] = ids_tensor([self.batch_size] , self.num_choices) __UpperCamelCase :int = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase__ ( self) -> str: return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowercase , initializer_range=self.initializer_range , ) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase) -> Tuple: __UpperCamelCase :str = LlamaModel(config=__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :str = model(__lowercase , attention_mask=__lowercase) __UpperCamelCase :Any = model(__lowercase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> Tuple: __UpperCamelCase :Optional[Any] = True __UpperCamelCase :Dict = LlamaModel(__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :int = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , ) __UpperCamelCase :Any = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , ) __UpperCamelCase :Tuple = model(__lowercase , attention_mask=__lowercase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> Optional[Any]: __UpperCamelCase :Optional[int] = LlamaForCausalLM(config=__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :Any = model(__lowercase , attention_mask=__lowercase , labels=__lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> List[Any]: __UpperCamelCase :str = True __UpperCamelCase :Tuple = True __UpperCamelCase :List[Any] = LlamaForCausalLM(config=__lowercase) model.to(__lowercase) model.eval() # first forward pass __UpperCamelCase :Union[str, Any] = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , use_cache=__lowercase , ) __UpperCamelCase :Optional[Any] = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __UpperCamelCase :Optional[int] = ids_tensor((self.batch_size, 3) , config.vocab_size) __UpperCamelCase :int = ids_tensor((self.batch_size, 3) , vocab_size=2) # append to next input_ids and __UpperCamelCase :Any = torch.cat([input_ids, next_tokens] , dim=-1) __UpperCamelCase :Optional[int] = torch.cat([input_mask, next_mask] , dim=-1) __UpperCamelCase :Union[str, Any] = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , output_hidden_states=__lowercase , )['''hidden_states'''][0] __UpperCamelCase :Tuple = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , past_key_values=__lowercase , output_hidden_states=__lowercase , )['''hidden_states'''][0] # select random slice __UpperCamelCase :Any = ids_tensor((1,) , output_from_past.shape[-1]).item() __UpperCamelCase :Optional[Any] = output_from_no_past[:, -3:, random_slice_idx].detach() __UpperCamelCase :Union[str, Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1]) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__lowercase , __lowercase , atol=1E-3)) def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Union[str, Any] = self.prepare_config_and_inputs() ( ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ) :List[Any] = config_and_inputs __UpperCamelCase :Any = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : Union[str, Any] = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () a__ : Union[str, Any] = (LlamaForCausalLM,) if is_torch_available() else () a__ : List[Any] = ( { """feature-extraction""": LlamaModel, """text-classification""": LlamaForSequenceClassification, """text-generation""": LlamaForCausalLM, """zero-shot""": LlamaForSequenceClassification, } if is_torch_available() else {} ) a__ : Tuple = False a__ : Any = False def UpperCamelCase__ ( self) -> str: __UpperCamelCase :Dict = LlamaModelTester(self) __UpperCamelCase :Union[str, Any] = ConfigTester(self , config_class=__lowercase , hidden_size=37) def UpperCamelCase__ ( self) -> Dict: self.config_tester.run_common_tests() def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase) def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :List[str] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __UpperCamelCase :Any = type self.model_tester.create_and_check_model(*__lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase :Dict = 3 __UpperCamelCase :int = input_dict['''input_ids'''] __UpperCamelCase :Tuple = input_ids.ne(1).to(__lowercase) __UpperCamelCase :Any = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size) __UpperCamelCase :str = LlamaForSequenceClassification(__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :str = model(__lowercase , attention_mask=__lowercase , labels=__lowercase) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) def UpperCamelCase__ ( self) -> int: __UpperCamelCase , __UpperCamelCase :str = self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase :Optional[int] = 3 __UpperCamelCase :Optional[int] = '''single_label_classification''' __UpperCamelCase :List[str] = input_dict['''input_ids'''] __UpperCamelCase :str = input_ids.ne(1).to(__lowercase) __UpperCamelCase :Optional[int] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size) __UpperCamelCase :Tuple = LlamaForSequenceClassification(__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :List[Any] = model(__lowercase , attention_mask=__lowercase , labels=__lowercase) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase , __UpperCamelCase :Dict = self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase :int = 3 __UpperCamelCase :Dict = '''multi_label_classification''' __UpperCamelCase :Optional[int] = input_dict['''input_ids'''] __UpperCamelCase :int = input_ids.ne(1).to(__lowercase) __UpperCamelCase :Optional[int] = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size).to(torch.float) __UpperCamelCase :str = LlamaForSequenceClassification(__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :Any = model(__lowercase , attention_mask=__lowercase , labels=__lowercase) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels)) @unittest.skip('''LLaMA buffers include complex numbers, which breaks this test''') def UpperCamelCase__ ( self) -> Optional[Any]: pass @parameterized.expand([('''linear''',), ('''dynamic''',)]) def UpperCamelCase__ ( self , __lowercase) -> Optional[int]: __UpperCamelCase , __UpperCamelCase :List[str] = self.model_tester.prepare_config_and_inputs_for_common() __UpperCamelCase :Union[str, Any] = ids_tensor([1, 10] , config.vocab_size) __UpperCamelCase :List[Any] = ids_tensor([1, int(config.max_position_embeddings * 1.5)] , config.vocab_size) set_seed(42) # Fixed seed at init time so the two models get the same random weights __UpperCamelCase :str = LlamaModel(__lowercase) original_model.to(__lowercase) original_model.eval() __UpperCamelCase :Optional[Any] = original_model(__lowercase).last_hidden_state __UpperCamelCase :Optional[int] = original_model(__lowercase).last_hidden_state set_seed(42) # Fixed seed at init time so the two models get the same random weights __UpperCamelCase :List[str] = {'''type''': scaling_type, '''factor''': 10.0} __UpperCamelCase :int = LlamaModel(__lowercase) scaled_model.to(__lowercase) scaled_model.eval() __UpperCamelCase :List[str] = scaled_model(__lowercase).last_hidden_state __UpperCamelCase :List[Any] = scaled_model(__lowercase).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(__lowercase , __lowercase , atol=1E-5)) else: self.assertFalse(torch.allclose(__lowercase , __lowercase , atol=1E-5)) # The output should be different for long inputs self.assertFalse(torch.allclose(__lowercase , __lowercase , atol=1E-5)) @require_torch class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' @unittest.skip('''Logits are not exactly the same, once we fix the instabalities somehow, will update!''') @slow def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase :Dict = [1, 306, 4_658, 278, 6_593, 310, 2_834, 338] __UpperCamelCase :Optional[Any] = LlamaForCausalLM.from_pretrained('''meta-llama/Llama-2-7b-hf''' , device_map='''auto''') __UpperCamelCase :Tuple = model(torch.tensor([input_ids])) # Expected mean on dim = -1 __UpperCamelCase :Any = torch.tensor([[-6.65_50, -4.12_27, -4.98_59, -3.24_06, 0.82_62, -3.00_33, 1.29_64, -3.36_99]]) torch.testing.assert_close(out.mean(-1) , __lowercase , atol=1E-2 , rtol=1E-2) # slicing logits[0, 0, 0:30] # fmt: off __UpperCamelCase :Any = torch.tensor([-12.82_81, -7.44_53, -0.46_39, -8.06_25, -7.25_00, -8.00_00, -6.48_83, -7.76_95, -7.84_38, -7.03_12, -6.21_88, -7.13_28, -1.84_96, 1.99_61, -8.62_50, -6.72_27, -12.82_81, -6.94_92, -7.07_42, -7.78_52, -7.58_20, -7.90_62, -6.93_75, -7.98_05, -8.34_38, -8.15_62, -8.04_69, -7.62_50, -7.74_22, -7.33_98,]) # fmt: on torch.testing.assert_close(out[0, 0, :30] , __lowercase , atol=1E-5 , rtol=1E-5) @unittest.skip('''Logits are not exactly the same, once we fix the instabalities somehow, will update!''') @slow def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :str = [1, 306, 4_658, 278, 6_593, 310, 2_834, 338] __UpperCamelCase :Any = LlamaForCausalLM.from_pretrained('''meta-llama/Llama-2-13b-hf''' , device_map='''auto''') __UpperCamelCase :Optional[int] = model(torch.tensor(__lowercase)) # Expected mean on dim = -1 __UpperCamelCase :Any = torch.tensor([[-2.06_22, -1.27_94, -1.16_38, -0.97_88, -1.46_03, -1.02_38, -1.78_93, -1.44_11]]) torch.testing.assert_close(out.mean(-1) , __lowercase , atol=1E-2 , rtol=1E-2) # slicing logits[0, 0, 0:30] # fmt: off __UpperCamelCase :Union[str, Any] = torch.tensor([-8.14_06, -8.05_47, 2.74_61, -1.23_44, -0.14_48, -1.82_62, -1.00_20, -1.81_54, -1.68_95, -1.85_16, -2.35_74, -0.92_77, 3.75_98, 6.57_42, -1.29_98, -0.11_77, -8.14_06, -2.96_88, -2.91_99, -3.16_99, -3.52_54, -2.35_55, -2.79_88, -3.41_41, -2.82_62, -4.51_95, -3.33_79, -3.31_64, -2.78_32, -3.02_73]) # fmt: on torch.testing.assert_close(out[0, 0, :30] , __lowercase , atol=1E-5 , rtol=1E-5) @unittest.skip('''Logits are not exactly the same, once we fix the instabalities somehow, will update!''') @slow def UpperCamelCase__ ( self) -> int: __UpperCamelCase :Any = [1, 306, 4_658, 278, 6_593, 310, 2_834, 338] __UpperCamelCase :Optional[int] = LlamaForCausalLM.from_pretrained('''meta-llama/Llama-2-13b-chat-hf''' , device_map='''auto''') __UpperCamelCase :str = model(torch.tensor(__lowercase)) # Expected mean on dim = -1 __UpperCamelCase :List[Any] = torch.tensor([[-0.85_62, -1.85_20, -0.75_51, -0.41_62, -1.51_61, -1.20_38, -2.48_23, -2.32_54]]) torch.testing.assert_close(out.mean(-1) , __lowercase , atol=1E-2 , rtol=1E-2) # slicing logits[0, 0, 0:30] # fmt: off __UpperCamelCase :Union[str, Any] = torch.tensor([-2.22_27, 4.88_28, 0.90_23, -0.45_78, -0.78_71, -0.10_33, -0.62_21, -0.57_86, -0.78_03, -1.06_74, -1.29_20, -0.15_70, 0.80_08, 2.07_23, -0.94_97, 0.27_71, -2.22_27, -0.76_12, -1.43_46, -1.20_61, -1.64_26, -0.30_00, -0.71_39, -1.19_34, -1.86_91, -1.69_73, -1.59_47, -1.27_05, -0.35_23, -0.55_13]) # fmt: on torch.testing.assert_close(out.mean(-1) , __lowercase , atol=1E-2 , rtol=1E-2) @unittest.skip( '''Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test''') @slow def UpperCamelCase__ ( self) -> Tuple: __UpperCamelCase :Any = [1, 306, 4_658, 278, 6_593, 310, 2_834, 338] __UpperCamelCase :List[Any] = LlamaForCausalLM.from_pretrained('''meta-llama/Llama-2-70b-hf''' , device_map='''auto''') __UpperCamelCase :List[str] = model(torch.tensor(__lowercase)) __UpperCamelCase :Dict = torch.tensor( [[-4.23_27, -3.33_60, -4.66_65, -4.76_31, -1.81_80, -3.41_70, -1.42_11, -3.18_10]] , dtype=torch.floataa) torch.testing.assert_close(out.mean(-1) , __lowercase , atol=1E-2 , rtol=1E-2) # fmt: off __UpperCamelCase :str = torch.tensor([-9.49_22, -3.95_51, 1.79_98, -5.67_58, -5.10_55, -5.89_84, -4.83_20, -6.80_86, -6.53_91, -5.61_72, -5.58_20, -5.53_52, 1.78_81, 3.62_89, -6.51_17, -3.47_85, -9.50_00, -6.03_52, -6.81_25, -6.01_95, -6.68_36, -5.47_27, -6.28_12, -6.03_91, -7.33_98, -7.42_97, -7.48_44, -6.58_20, -5.87_89, -5.53_12]) # fmt: on torch.testing.assert_close(out[0, 0, :30] , __lowercase , atol=1E-5 , rtol=1E-5) @unittest.skip('''Model is curently gated''') @slow def UpperCamelCase__ ( self) -> str: __UpperCamelCase :List[Any] = '''Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the "princi''' __UpperCamelCase :Optional[int] = '''Simply put, the theory of relativity states that ''' __UpperCamelCase :Tuple = LlamaTokenizer.from_pretrained('''meta-llama/Llama-2-13b-chat-hf''') __UpperCamelCase :Dict = tokenizer.encode(__lowercase , return_tensors='''pt''') __UpperCamelCase :Tuple = LlamaForCausalLM.from_pretrained( '''meta-llama/Llama-2-13b-chat-hf''' , device_map='''sequential''' , use_safetensors=__lowercase) # greedy generation outputs __UpperCamelCase :List[Any] = model.generate(__lowercase , max_new_tokens=64 , top_p=__lowercase , temperature=1 , do_sample=__lowercase) __UpperCamelCase :str = tokenizer.decode(generated_ids[0] , skip_special_tokens=__lowercase) self.assertEqual(__lowercase , __lowercase)
43
import numpy as np def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 1e-12 , SCREAMING_SNAKE_CASE = 100 , ): '''simple docstring''' assert np.shape(SCREAMING_SNAKE_CASE )[0] == np.shape(SCREAMING_SNAKE_CASE )[1] # Ensure proper dimensionality. assert np.shape(SCREAMING_SNAKE_CASE )[0] == np.shape(SCREAMING_SNAKE_CASE )[0] # Ensure inputs are either both complex or both real assert np.iscomplexobj(SCREAMING_SNAKE_CASE ) == np.iscomplexobj(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = np.iscomplexobj(SCREAMING_SNAKE_CASE ) if is_complex: # Ensure complex input_matrix is Hermitian assert np.array_equal(SCREAMING_SNAKE_CASE , input_matrix.conj().T ) # Set convergence to False. Will define convergence when we exceed max_iterations # or when we have small changes from one iteration to next. __UpperCamelCase :str = False __UpperCamelCase :int = 0 __UpperCamelCase :Optional[Any] = 0 __UpperCamelCase :Union[str, Any] = 1e12 while not convergence: # Multiple matrix by the vector. __UpperCamelCase :List[str] = np.dot(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Normalize the resulting output vector. __UpperCamelCase :Tuple = w / np.linalg.norm(SCREAMING_SNAKE_CASE ) # Find rayleigh quotient # (faster than usual b/c we know vector is normalized already) __UpperCamelCase :int = vector.conj().T if is_complex else vector.T __UpperCamelCase :Optional[int] = np.dot(SCREAMING_SNAKE_CASE , np.dot(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) # Check convergence. __UpperCamelCase :Optional[Any] = np.abs(lambda_ - lambda_previous ) / lambda_ iterations += 1 if error <= error_tol or iterations >= max_iterations: __UpperCamelCase :Dict = True __UpperCamelCase :List[Any] = lambda_ if is_complex: __UpperCamelCase :Tuple = np.real(lambda_ ) return lambda_, vector def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :int = np.array([[41, 4, 20], [4, 26, 30], [20, 30, 50]] ) __UpperCamelCase :Optional[Any] = np.array([41, 4, 20] ) __UpperCamelCase :Any = real_input_matrix.astype(np.complexaaa ) __UpperCamelCase :Dict = np.triu(1j * complex_input_matrix , 1 ) complex_input_matrix += imag_matrix complex_input_matrix += -1 * imag_matrix.T __UpperCamelCase :Optional[int] = np.array([41, 4, 20] ).astype(np.complexaaa ) for problem_type in ["real", "complex"]: if problem_type == "real": __UpperCamelCase :Any = real_input_matrix __UpperCamelCase :int = real_vector elif problem_type == "complex": __UpperCamelCase :Tuple = complex_input_matrix __UpperCamelCase :Optional[Any] = complex_vector # Our implementation. __UpperCamelCase , __UpperCamelCase :Dict = power_iteration(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Numpy implementation. # Get eigenvalues and eigenvectors using built-in numpy # eigh (eigh used for symmetric or hermetian matrices). __UpperCamelCase , __UpperCamelCase :List[Any] = np.linalg.eigh(SCREAMING_SNAKE_CASE ) # Last eigenvalue is the maximum one. __UpperCamelCase :List[Any] = eigen_values[-1] # Last column in this matrix is eigenvector corresponding to largest eigenvalue. __UpperCamelCase :str = eigen_vectors[:, -1] # Check our implementation and numpy gives close answers. assert np.abs(eigen_value - eigen_value_max ) <= 1e-6 # Take absolute values element wise of each eigenvector. # as they are only unique to a minus sign. assert np.linalg.norm(np.abs(SCREAMING_SNAKE_CASE ) - np.abs(SCREAMING_SNAKE_CASE ) ) <= 1e-6 if __name__ == "__main__": import doctest doctest.testmod() test_power_iteration()
43
1
def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' if not all(x.isalpha() for x in string ): raise ValueError('''String must only contain alphabetic characters.''' ) __UpperCamelCase :int = sorted(string.lower() ) return len(SCREAMING_SNAKE_CASE ) == len(set(SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": __lowercase = input('''Enter a string ''').strip() __lowercase = is_isogram(input_str) print(F'{input_str} is {"an" if isogram else "not an"} isogram.')
43
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __lowercase = logging.get_logger(__name__) __lowercase = {'''tokenizer_file''': '''tokenizer.json'''} __lowercase = { '''tokenizer_file''': { '''bigscience/tokenizer''': '''https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json''', '''bigscience/bloom-560m''': '''https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json''', '''bigscience/bloom-1b1''': '''https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json''', '''bigscience/bloom-1b7''': '''https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json''', '''bigscience/bloom-3b''': '''https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json''', '''bigscience/bloom-7b1''': '''https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json''', '''bigscience/bloom''': '''https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json''', }, } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : int = VOCAB_FILES_NAMES a__ : Tuple = PRETRAINED_VOCAB_FILES_MAP a__ : List[str] = ["""input_ids""", """attention_mask"""] a__ : int = None def __init__( self , __lowercase=None , __lowercase=None , __lowercase=None , __lowercase="<unk>" , __lowercase="<s>" , __lowercase="</s>" , __lowercase="<pad>" , __lowercase=False , __lowercase=False , **__lowercase , ) -> List[str]: super().__init__( __lowercase , __lowercase , tokenizer_file=__lowercase , unk_token=__lowercase , bos_token=__lowercase , eos_token=__lowercase , pad_token=__lowercase , add_prefix_space=__lowercase , clean_up_tokenization_spaces=__lowercase , **__lowercase , ) __UpperCamelCase :int = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__()) if pre_tok_state.get('''add_prefix_space''' , __lowercase) != add_prefix_space: __UpperCamelCase :Any = getattr(__lowercase , pre_tok_state.pop('''type''')) __UpperCamelCase :str = add_prefix_space __UpperCamelCase :List[str] = pre_tok_class(**__lowercase) __UpperCamelCase :Tuple = add_prefix_space def UpperCamelCase__ ( self , *__lowercase , **__lowercase) -> BatchEncoding: __UpperCamelCase :Tuple = kwargs.get('''is_split_into_words''' , __lowercase) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with""" ''' pretokenized inputs.''') return super()._batch_encode_plus(*__lowercase , **__lowercase) def UpperCamelCase__ ( self , *__lowercase , **__lowercase) -> BatchEncoding: __UpperCamelCase :List[str] = kwargs.get('''is_split_into_words''' , __lowercase) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with""" ''' pretokenized inputs.''') return super()._encode_plus(*__lowercase , **__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase = None) -> Tuple[str]: __UpperCamelCase :Optional[Any] = self._tokenizer.model.save(__lowercase , name=__lowercase) return tuple(__lowercase) def UpperCamelCase__ ( self , __lowercase) -> List[int]: __UpperCamelCase :str = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(__lowercase , add_special_tokens=__lowercase) + [self.eos_token_id]) if len(__lowercase) > self.model_max_length: __UpperCamelCase :Any = input_ids[-self.model_max_length :] return input_ids
43
1
import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class lowerCamelCase_ : '''simple docstring''' def __init__( self , __lowercase , __lowercase=13 , __lowercase=7 , __lowercase=True , __lowercase=True , __lowercase=99 , __lowercase=32 , __lowercase=5 , __lowercase=4 , __lowercase=37 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=50 , __lowercase=0.02 , __lowercase=True , __lowercase=None , ) -> Dict: __UpperCamelCase :Union[str, Any] = parent __UpperCamelCase :List[Any] = batch_size __UpperCamelCase :Tuple = seq_length __UpperCamelCase :List[Any] = is_training __UpperCamelCase :str = use_input_mask __UpperCamelCase :str = vocab_size __UpperCamelCase :str = hidden_size __UpperCamelCase :Tuple = num_hidden_layers __UpperCamelCase :int = num_attention_heads __UpperCamelCase :Union[str, Any] = intermediate_size __UpperCamelCase :Tuple = hidden_act __UpperCamelCase :List[Any] = hidden_dropout_prob __UpperCamelCase :Any = attention_probs_dropout_prob __UpperCamelCase :Any = max_position_embeddings __UpperCamelCase :int = initializer_range __UpperCamelCase :Tuple = use_labels __UpperCamelCase :Any = scope def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __UpperCamelCase :Any = None if self.use_input_mask: __UpperCamelCase :Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length]) if self.use_labels: __UpperCamelCase :int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) __UpperCamelCase :int = self.get_config() return config, input_ids, input_mask, token_labels def UpperCamelCase__ ( self) -> Any: return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=__lowercase , initializer_range=self.initializer_range , ) def UpperCamelCase__ ( self) -> int: ( ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ) :Dict = self.prepare_config_and_inputs() __UpperCamelCase :Dict = True __UpperCamelCase :str = floats_tensor([self.batch_size, self.seq_length, self.hidden_size]) __UpperCamelCase :str = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , **__lowercase , ) -> Dict: __UpperCamelCase :Tuple = BertGenerationEncoder(config=__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :Tuple = model(__lowercase , attention_mask=__lowercase) __UpperCamelCase :Dict = model(__lowercase) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , **__lowercase , ) -> int: __UpperCamelCase :str = True __UpperCamelCase :Any = BertGenerationEncoder(config=__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :Union[str, Any] = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , ) __UpperCamelCase :Optional[Any] = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , **__lowercase , ) -> Optional[int]: __UpperCamelCase :Any = True __UpperCamelCase :Union[str, Any] = True __UpperCamelCase :Optional[int] = BertGenerationDecoder(config=__lowercase).to(__lowercase).eval() # first forward pass __UpperCamelCase :int = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , use_cache=__lowercase , ) __UpperCamelCase :List[Any] = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __UpperCamelCase :Optional[Any] = ids_tensor((self.batch_size, 3) , config.vocab_size) __UpperCamelCase :List[Any] = ids_tensor((self.batch_size, 3) , vocab_size=2) # append to next input_ids and __UpperCamelCase :Optional[int] = torch.cat([input_ids, next_tokens] , dim=-1) __UpperCamelCase :int = torch.cat([input_mask, next_mask] , dim=-1) __UpperCamelCase :Optional[Any] = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , output_hidden_states=__lowercase , )['''hidden_states'''][0] __UpperCamelCase :str = model( __lowercase , attention_mask=__lowercase , encoder_hidden_states=__lowercase , encoder_attention_mask=__lowercase , past_key_values=__lowercase , output_hidden_states=__lowercase , )['''hidden_states'''][0] # select random slice __UpperCamelCase :Union[str, Any] = ids_tensor((1,) , output_from_past.shape[-1]).item() __UpperCamelCase :Optional[int] = output_from_no_past[:, -3:, random_slice_idx].detach() __UpperCamelCase :Tuple = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1]) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__lowercase , __lowercase , atol=1E-3)) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase , *__lowercase , ) -> str: __UpperCamelCase :str = BertGenerationDecoder(__lowercase) model.to(__lowercase) model.eval() __UpperCamelCase :List[str] = model(__lowercase , attention_mask=__lowercase , labels=__lowercase) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size)) def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self.prepare_config_and_inputs() __UpperCamelCase :Union[str, Any] = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class lowerCamelCase_ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : List[str] = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () a__ : Union[str, Any] = (BertGenerationDecoder,) if is_torch_available() else () a__ : Optional[Any] = ( {"""feature-extraction""": BertGenerationEncoder, """text-generation""": BertGenerationDecoder} if is_torch_available() else {} ) def UpperCamelCase__ ( self) -> str: __UpperCamelCase :List[Any] = BertGenerationEncoderTester(self) __UpperCamelCase :Optional[Any] = ConfigTester(self , config_class=__lowercase , hidden_size=37) def UpperCamelCase__ ( self) -> int: self.config_tester.run_common_tests() def UpperCamelCase__ ( self) -> str: __UpperCamelCase :List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase) def UpperCamelCase__ ( self) -> List[str]: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Any = self.model_tester.prepare_config_and_inputs() __UpperCamelCase :Any = '''bert''' self.model_tester.create_and_check_model(__lowercase , __lowercase , __lowercase , __lowercase) def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :str = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*__lowercase) def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :int = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*__lowercase) def UpperCamelCase__ ( self) -> List[Any]: # This regression test was failing with PyTorch < 1.3 ( ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ( __UpperCamelCase ) , ) :int = self.model_tester.prepare_config_and_inputs_for_decoder() __UpperCamelCase :List[Any] = None self.model_tester.create_and_check_model_as_decoder( __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Dict = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*__lowercase) @slow def UpperCamelCase__ ( self) -> Union[str, Any]: __UpperCamelCase :Dict = BertGenerationEncoder.from_pretrained('''google/bert_for_seq_generation_L-24_bbc_encoder''') self.assertIsNotNone(__lowercase) @require_torch class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase__ ( self) -> Any: __UpperCamelCase :Any = BertGenerationEncoder.from_pretrained('''google/bert_for_seq_generation_L-24_bbc_encoder''') __UpperCamelCase :Optional[Any] = torch.tensor([[101, 7_592, 1_010, 2_026, 3_899, 2_003, 10_140, 102]]) with torch.no_grad(): __UpperCamelCase :Any = model(__lowercase)[0] __UpperCamelCase :Optional[Any] = torch.Size([1, 8, 1_024]) self.assertEqual(output.shape , __lowercase) __UpperCamelCase :str = torch.tensor( [[[0.17_75, 0.00_83, -0.03_21], [1.60_02, 0.12_87, 0.39_12], [2.14_73, 0.57_91, 0.60_66]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowercase , atol=1E-4)) @require_torch class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase__ ( self) -> List[Any]: __UpperCamelCase :List[str] = BertGenerationDecoder.from_pretrained('''google/bert_for_seq_generation_L-24_bbc_encoder''') __UpperCamelCase :Tuple = torch.tensor([[101, 7_592, 1_010, 2_026, 3_899, 2_003, 10_140, 102]]) with torch.no_grad(): __UpperCamelCase :int = model(__lowercase)[0] __UpperCamelCase :Optional[int] = torch.Size([1, 8, 50_358]) self.assertEqual(output.shape , __lowercase) __UpperCamelCase :Optional[Any] = torch.tensor( [[[-0.57_88, -2.59_94, -3.70_54], [0.04_38, 4.79_97, 1.87_95], [1.58_62, 6.64_09, 4.46_38]]]) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowercase , atol=1E-4))
43
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = {'''ctrl''': '''https://huggingface.co/ctrl/resolve/main/config.json'''} class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : str = """ctrl""" a__ : Dict = ["""past_key_values"""] a__ : Tuple = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , __lowercase=246_534 , __lowercase=256 , __lowercase=1_280 , __lowercase=8_192 , __lowercase=48 , __lowercase=16 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=1E-6 , __lowercase=0.02 , __lowercase=True , **__lowercase , ) -> List[Any]: __UpperCamelCase :List[str] = vocab_size __UpperCamelCase :Optional[Any] = n_positions __UpperCamelCase :Dict = n_embd __UpperCamelCase :Dict = n_layer __UpperCamelCase :List[Any] = n_head __UpperCamelCase :int = dff __UpperCamelCase :Union[str, Any] = resid_pdrop __UpperCamelCase :Optional[int] = embd_pdrop __UpperCamelCase :List[Any] = layer_norm_epsilon __UpperCamelCase :Dict = initializer_range __UpperCamelCase :Any = use_cache super().__init__(**__lowercase)
43
1
# Lint as: python3 # pylint: enable=line-too-long # pylint: disable=g-import-not-at-top,g-bad-import-order,wrong-import-position __lowercase = '''2.13.1''' import platform import pyarrow from packaging import version if version.parse(platform.python_version()) < version.parse('''3.7'''): raise ImportWarning( '''To use `datasets`, Python>=3.7 is required, and the current version of Python doesn\'t match this condition.''' ) if version.parse(pyarrow.__version__).major < 8: raise ImportWarning( '''To use `datasets`, the module `pyarrow>=8.0.0` is required, and the current version of `pyarrow` doesn\'t match this condition.\n''' '''If you are running this in a Google Colab, you should probably just restart the runtime to use the right version of `pyarrow`.''' ) del platform del pyarrow del version from .arrow_dataset import Dataset from .arrow_reader import ReadInstruction from .builder import ArrowBasedBuilder, BeamBasedBuilder, BuilderConfig, DatasetBuilder, GeneratorBasedBuilder from .combine import concatenate_datasets, interleave_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .download import * from .features import * from .fingerprint import disable_caching, enable_caching, is_caching_enabled, set_caching_enabled from .info import DatasetInfo, MetricInfo from .inspect import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, list_datasets, list_metrics, ) from .iterable_dataset import IterableDataset from .load import load_dataset, load_dataset_builder, load_from_disk, load_metric from .metric import Metric from .splits import ( NamedSplit, NamedSplitAll, Split, SplitBase, SplitDict, SplitGenerator, SplitInfo, SubSplitInfo, percent, ) from .tasks import * from .utils import * from .utils import logging # deprecated modules from datasets import arrow_dataset as _arrow_dataset # isort:skip from datasets import utils as _utils # isort:skip from datasets.utils import download_manager as _deprecated_download_manager # isort:skip __lowercase = concatenate_datasets __lowercase = DownloadConfig __lowercase = DownloadManager __lowercase = DownloadMode __lowercase = DownloadConfig __lowercase = DownloadMode __lowercase = DownloadManager del _arrow_dataset, _utils, _deprecated_download_manager
43
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class lowerCamelCase_ ( UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a__ : str = TextToVideoSDPipeline a__ : Union[str, Any] = TEXT_TO_IMAGE_PARAMS a__ : Tuple = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. a__ : int = frozenset( [ """num_inference_steps""", """generator""", """latents""", """return_dict""", """callback""", """callback_steps""", ] ) def UpperCamelCase__ ( self) -> Optional[Any]: torch.manual_seed(0) __UpperCamelCase :str = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''CrossAttnDownBlock3D''', '''DownBlock3D''') , up_block_types=('''UpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''', '''CrossAttnUpBlock3D''') , cross_attention_dim=32 , attention_head_dim=4 , ) __UpperCamelCase :Optional[int] = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=__lowercase , set_alpha_to_one=__lowercase , ) torch.manual_seed(0) __UpperCamelCase :Optional[int] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0) __UpperCamelCase :Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act='''gelu''' , projection_dim=512 , ) __UpperCamelCase :Optional[Any] = CLIPTextModel(__lowercase) __UpperCamelCase :Optional[int] = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') __UpperCamelCase :Union[str, Any] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def UpperCamelCase__ ( self , __lowercase , __lowercase=0) -> Optional[int]: if str(__lowercase).startswith('''mps'''): __UpperCamelCase :List[Any] = torch.manual_seed(__lowercase) else: __UpperCamelCase :Tuple = torch.Generator(device=__lowercase).manual_seed(__lowercase) __UpperCamelCase :Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def UpperCamelCase__ ( self) -> Optional[Any]: __UpperCamelCase :int = '''cpu''' # ensure determinism for the device-dependent torch.Generator __UpperCamelCase :Optional[int] = self.get_dummy_components() __UpperCamelCase :Dict = TextToVideoSDPipeline(**__lowercase) __UpperCamelCase :Any = sd_pipe.to(__lowercase) sd_pipe.set_progress_bar_config(disable=__lowercase) __UpperCamelCase :Optional[Any] = self.get_dummy_inputs(__lowercase) __UpperCamelCase :int = '''np''' __UpperCamelCase :List[str] = sd_pipe(**__lowercase).frames __UpperCamelCase :Optional[Any] = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) __UpperCamelCase :str = np.array([1_58.0, 1_60.0, 1_53.0, 1_25.0, 1_00.0, 1_21.0, 1_11.0, 93.0, 1_13.0]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 def UpperCamelCase__ ( self) -> Tuple: self._test_attention_slicing_forward_pass(test_mean_pixel_difference=__lowercase , expected_max_diff=3E-3) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase__ ( self) -> Optional[int]: self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__lowercase , expected_max_diff=1E-2) @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCamelCase__ ( self) -> Union[str, Any]: pass @unittest.skip(reason='''Batching needs to be properly figured out first for this pipeline.''') def UpperCamelCase__ ( self) -> Dict: pass @unittest.skip(reason='''`num_images_per_prompt` argument is not supported for this pipeline.''') def UpperCamelCase__ ( self) -> str: pass def UpperCamelCase__ ( self) -> List[str]: return super().test_progress_bar() @slow @skip_mps class lowerCamelCase_ ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self) -> Dict: __UpperCamelCase :Union[str, Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy''') __UpperCamelCase :List[str] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''') __UpperCamelCase :Union[str, Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) __UpperCamelCase :str = pipe.to('''cuda''') __UpperCamelCase :Optional[Any] = '''Spiderman is surfing''' __UpperCamelCase :Union[str, Any] = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :List[Any] = pipe(__lowercase , generator=__lowercase , num_inference_steps=25 , output_type='''pt''').frames __UpperCamelCase :Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video).mean() < 5E-2 def UpperCamelCase__ ( self) -> int: __UpperCamelCase :str = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy''') __UpperCamelCase :Union[str, Any] = TextToVideoSDPipeline.from_pretrained('''damo-vilab/text-to-video-ms-1.7b''') __UpperCamelCase :str = pipe.to('''cuda''') __UpperCamelCase :Union[str, Any] = '''Spiderman is surfing''' __UpperCamelCase :int = torch.Generator(device='''cpu''').manual_seed(0) __UpperCamelCase :List[Any] = pipe(__lowercase , generator=__lowercase , num_inference_steps=2 , output_type='''pt''').frames __UpperCamelCase :Optional[Any] = video_frames.cpu().numpy() assert np.abs(expected_video - video).mean() < 5E-2
43
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __lowercase = { '''configuration_transfo_xl''': ['''TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TransfoXLConfig'''], '''tokenization_transfo_xl''': ['''TransfoXLCorpus''', '''TransfoXLTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AdaptiveEmbedding''', '''TransfoXLForSequenceClassification''', '''TransfoXLLMHeadModel''', '''TransfoXLModel''', '''TransfoXLPreTrainedModel''', '''load_tf_weights_in_transfo_xl''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFAdaptiveEmbedding''', '''TFTransfoXLForSequenceClassification''', '''TFTransfoXLLMHeadModel''', '''TFTransfoXLMainLayer''', '''TFTransfoXLModel''', '''TFTransfoXLPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_transfo_xl import TRANSFO_XL_PRETRAINED_CONFIG_ARCHIVE_MAP, TransfoXLConfig from .tokenization_transfo_xl import TransfoXLCorpus, TransfoXLTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_transfo_xl import ( TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, AdaptiveEmbedding, TransfoXLForSequenceClassification, TransfoXLLMHeadModel, TransfoXLModel, TransfoXLPreTrainedModel, load_tf_weights_in_transfo_xl, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_transfo_xl import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFAdaptiveEmbedding, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLMainLayer, TFTransfoXLModel, TFTransfoXLPreTrainedModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
43
def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = [0 for i in range(len(SCREAMING_SNAKE_CASE ) )] # initialize interval's left pointer and right pointer __UpperCamelCase , __UpperCamelCase :str = 0, 0 for i in range(1 , len(SCREAMING_SNAKE_CASE ) ): # case when current index is inside the interval if i <= right_pointer: __UpperCamelCase :Union[str, Any] = min(right_pointer - i + 1 , z_result[i - left_pointer] ) __UpperCamelCase :Tuple = min_edge while go_next(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: __UpperCamelCase , __UpperCamelCase :Union[str, Any] = i, i + z_result[i] - 1 return z_result def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' return i + z_result[i] < len(SCREAMING_SNAKE_CASE ) and s[z_result[i]] == s[i + z_result[i]] def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string __UpperCamelCase :Tuple = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(SCREAMING_SNAKE_CASE ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
43
1
import dataclasses import re import string from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple import numpy as np from . import residue_constants __lowercase = Mapping[str, np.ndarray] __lowercase = Mapping[str, Any] # Is a nested dict. __lowercase = 0.0_1 @dataclasses.dataclass(frozen=UpperCAmelCase_ ) class lowerCamelCase_ : '''simple docstring''' a__ : np.ndarray # [num_res, num_atom_type, 3] # Amino-acid type for each residue represented as an integer between 0 and # 20, where 20 is 'X'. a__ : np.ndarray # [num_res] # Binary float mask to indicate presence of a particular atom. 1.0 if an atom # is present and 0.0 if not. This should be used for loss masking. a__ : np.ndarray # [num_res, num_atom_type] # Residue index as used in PDB. It is not necessarily continuous or 0-indexed. a__ : np.ndarray # [num_res] # B-factors, or temperature factors, of each residue (in sq. angstroms units), # representing the displacement of the residue from its ground truth mean # value. a__ : np.ndarray # [num_res, num_atom_type] # Chain indices for multi-chain predictions a__ : Optional[np.ndarray] = None # Optional remark about the protein. Included as a comment in output PDB # files a__ : Optional[str] = None # Templates used to generate this protein (prediction-only) a__ : Optional[Sequence[str]] = None # Chain corresponding to each parent a__ : Optional[Sequence[int]] = None def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[int] = R'''(\[[A-Z]+\]\n)''' __UpperCamelCase :List[str] = [tag.strip() for tag in re.split(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if len(SCREAMING_SNAKE_CASE ) > 0] __UpperCamelCase :Iterator[Tuple[str, List[str]]] = zip(tags[0::2] , [l.split('''\n''' ) for l in tags[1::2]] ) __UpperCamelCase :List[str] = ["N", "CA", "C"] __UpperCamelCase :int = None __UpperCamelCase :Optional[int] = None __UpperCamelCase :Dict = None for g in groups: if "[PRIMARY]" == g[0]: __UpperCamelCase :Any = g[1][0].strip() for i in range(len(SCREAMING_SNAKE_CASE ) ): if seq[i] not in residue_constants.restypes: __UpperCamelCase :List[Any] = '''X''' # FIXME: strings are immutable __UpperCamelCase :Dict = np.array( [residue_constants.restype_order.get(SCREAMING_SNAKE_CASE , residue_constants.restype_num ) for res_symbol in seq] ) elif "[TERTIARY]" == g[0]: __UpperCamelCase :List[List[float]] = [] for axis in range(3 ): tertiary.append(list(map(SCREAMING_SNAKE_CASE , g[1][axis].split() ) ) ) __UpperCamelCase :Optional[Any] = np.array(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = np.zeros((len(tertiary[0] ) // 3, residue_constants.atom_type_num, 3) ).astype(np.floataa ) for i, atom in enumerate(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Union[str, Any] = np.transpose(tertiary_np[:, i::3] ) atom_positions *= PICO_TO_ANGSTROM elif "[MASK]" == g[0]: __UpperCamelCase :List[str] = np.array(list(map({'''-''': 0, '''+''': 1}.get , g[1][0].strip() ) ) ) __UpperCamelCase :Union[str, Any] = np.zeros( ( len(SCREAMING_SNAKE_CASE ), residue_constants.atom_type_num, ) ).astype(np.floataa ) for i, atom in enumerate(SCREAMING_SNAKE_CASE ): __UpperCamelCase :List[Any] = 1 atom_mask *= mask[..., None] assert aatype is not None return Protein( atom_positions=SCREAMING_SNAKE_CASE , atom_mask=SCREAMING_SNAKE_CASE , aatype=SCREAMING_SNAKE_CASE , residue_index=np.arange(len(SCREAMING_SNAKE_CASE ) ) , b_factors=SCREAMING_SNAKE_CASE , ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 0 ): '''simple docstring''' __UpperCamelCase :List[str] = [] __UpperCamelCase :Optional[int] = prot.remark if remark is not None: pdb_headers.append(f"""REMARK {remark}""" ) __UpperCamelCase :Optional[int] = prot.parents __UpperCamelCase :Union[str, Any] = prot.parents_chain_index if parents is not None and parents_chain_index is not None: __UpperCamelCase :List[Any] = [p for i, p in zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if i == chain_id] if parents is None or len(SCREAMING_SNAKE_CASE ) == 0: __UpperCamelCase :Tuple = ['''N/A'''] pdb_headers.append(f"""PARENT {' '.join(SCREAMING_SNAKE_CASE )}""" ) return pdb_headers def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[str] = [] __UpperCamelCase :Dict = pdb_str.split('''\n''' ) __UpperCamelCase :int = prot.remark if remark is not None: out_pdb_lines.append(f"""REMARK {remark}""" ) __UpperCamelCase :List[List[str]] if prot.parents is not None and len(prot.parents ) > 0: __UpperCamelCase :Dict = [] if prot.parents_chain_index is not None: __UpperCamelCase :Dict[str, List[str]] = {} for p, i in zip(prot.parents , prot.parents_chain_index ): parent_dict.setdefault(str(SCREAMING_SNAKE_CASE ) , [] ) parent_dict[str(SCREAMING_SNAKE_CASE )].append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Union[str, Any] = max([int(SCREAMING_SNAKE_CASE ) for chain_idx in parent_dict] ) for i in range(max_idx + 1 ): __UpperCamelCase :Union[str, Any] = parent_dict.get(str(SCREAMING_SNAKE_CASE ) , ['''N/A'''] ) parents_per_chain.append(SCREAMING_SNAKE_CASE ) else: parents_per_chain.append(list(prot.parents ) ) else: __UpperCamelCase :List[Any] = [['''N/A''']] def make_parent_line(SCREAMING_SNAKE_CASE ) -> str: return f"""PARENT {' '.join(SCREAMING_SNAKE_CASE )}""" out_pdb_lines.append(make_parent_line(parents_per_chain[0] ) ) __UpperCamelCase :Optional[Any] = 0 for i, l in enumerate(SCREAMING_SNAKE_CASE ): if "PARENT" not in l and "REMARK" not in l: out_pdb_lines.append(SCREAMING_SNAKE_CASE ) if "TER" in l and "END" not in lines[i + 1]: chain_counter += 1 if not chain_counter >= len(SCREAMING_SNAKE_CASE ): __UpperCamelCase :List[str] = parents_per_chain[chain_counter] else: __UpperCamelCase :Optional[int] = ['''N/A'''] out_pdb_lines.append(make_parent_line(SCREAMING_SNAKE_CASE ) ) return "\n".join(SCREAMING_SNAKE_CASE ) def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = residue_constants.restypes + ['''X'''] def res_atoa(SCREAMING_SNAKE_CASE ) -> str: return residue_constants.restype_atoa.get(restypes[r] , '''UNK''' ) __UpperCamelCase :Optional[int] = residue_constants.atom_types __UpperCamelCase :List[str] = [] __UpperCamelCase :List[Any] = prot.atom_mask __UpperCamelCase :List[str] = prot.aatype __UpperCamelCase :Union[str, Any] = prot.atom_positions __UpperCamelCase :Any = prot.residue_index.astype(np.intaa ) __UpperCamelCase :List[str] = prot.b_factors __UpperCamelCase :List[str] = prot.chain_index if np.any(aatype > residue_constants.restype_num ): raise ValueError('''Invalid aatypes.''' ) __UpperCamelCase :int = get_pdb_headers(SCREAMING_SNAKE_CASE ) if len(SCREAMING_SNAKE_CASE ) > 0: pdb_lines.extend(SCREAMING_SNAKE_CASE ) __UpperCamelCase :str = aatype.shape[0] __UpperCamelCase :Optional[Any] = 1 __UpperCamelCase :List[Any] = 0 __UpperCamelCase :Any = string.ascii_uppercase __UpperCamelCase :Any = None # Add all atom sites. for i in range(SCREAMING_SNAKE_CASE ): __UpperCamelCase :List[Any] = res_atoa(aatype[i] ) for atom_name, pos, mask, b_factor in zip(SCREAMING_SNAKE_CASE , atom_positions[i] , atom_mask[i] , b_factors[i] ): if mask < 0.5: continue __UpperCamelCase :str = '''ATOM''' __UpperCamelCase :Tuple = atom_name if len(SCREAMING_SNAKE_CASE ) == 4 else f""" {atom_name}""" __UpperCamelCase :Optional[Any] = '''''' __UpperCamelCase :int = '''''' __UpperCamelCase :int = 1.00 __UpperCamelCase :int = atom_name[0] # Protein supports only C, N, O, S, this works. __UpperCamelCase :int = '''''' __UpperCamelCase :Dict = '''A''' if chain_index is not None: __UpperCamelCase :Union[str, Any] = chain_tags[chain_index[i]] # PDB is a columnar format, every space matters here! __UpperCamelCase :str = ( f"""{record_type:<6}{atom_index:>5} {name:<4}{alt_loc:>1}""" f"""{res_name_a:>3} {chain_tag:>1}""" f"""{residue_index[i]:>4}{insertion_code:>1} """ f"""{pos[0]:>8.3f}{pos[1]:>8.3f}{pos[2]:>8.3f}""" f"""{occupancy:>6.2f}{b_factor:>6.2f} """ f"""{element:>2}{charge:>2}""" ) pdb_lines.append(SCREAMING_SNAKE_CASE ) atom_index += 1 __UpperCamelCase :Dict = i == n - 1 if chain_index is not None: if i != n - 1 and chain_index[i + 1] != prev_chain_index: __UpperCamelCase :int = True __UpperCamelCase :List[str] = chain_index[i + 1] if should_terminate: # Close the chain. __UpperCamelCase :Optional[Any] = '''TER''' __UpperCamelCase :Optional[int] = ( f"""{chain_end:<6}{atom_index:>5} {res_atoa(aatype[i] ):>3} {chain_tag:>1}{residue_index[i]:>4}""" ) pdb_lines.append(SCREAMING_SNAKE_CASE ) atom_index += 1 if i != n - 1: # "prev" is a misnomer here. This happens at the beginning of # each new chain. pdb_lines.extend(get_pdb_headers(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) pdb_lines.append('''END''' ) pdb_lines.append('''''' ) return "\n".join(SCREAMING_SNAKE_CASE ) def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' return residue_constants.STANDARD_ATOM_MASK[prot.aatype] def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , ): '''simple docstring''' return Protein( aatype=features['''aatype'''] , atom_positions=result['''final_atom_positions'''] , atom_mask=result['''final_atom_mask'''] , residue_index=features['''residue_index'''] + 1 , b_factors=b_factors if b_factors is not None else np.zeros_like(result['''final_atom_mask'''] ) , chain_index=SCREAMING_SNAKE_CASE , remark=SCREAMING_SNAKE_CASE , parents=SCREAMING_SNAKE_CASE , parents_chain_index=SCREAMING_SNAKE_CASE , )
43
import math from typing import Any, Callable, List, Optional, Tuple, Union import numpy as np import torch from ...models import TaFilmDecoder from ...schedulers import DDPMScheduler from ...utils import is_onnx_available, logging, randn_tensor if is_onnx_available(): from ..onnx_utils import OnnxRuntimeModel from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline from .continous_encoder import SpectrogramContEncoder from .notes_encoder import SpectrogramNotesEncoder __lowercase = logging.get_logger(__name__) # pylint: disable=invalid-name __lowercase = 256 class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Tuple = ["""melgan"""] def __init__( self , __lowercase , __lowercase , __lowercase , __lowercase , __lowercase , ) -> None: super().__init__() # From MELGAN __UpperCamelCase :int = math.log(1E-5) # Matches MelGAN training. __UpperCamelCase :int = 4.0 # Largest value for most examples __UpperCamelCase :str = 128 self.register_modules( notes_encoder=__lowercase , continuous_encoder=__lowercase , decoder=__lowercase , scheduler=__lowercase , melgan=__lowercase , ) def UpperCamelCase__ ( self , __lowercase , __lowercase=(-1.0, 1.0) , __lowercase=False) -> Dict: __UpperCamelCase , __UpperCamelCase :str = output_range if clip: __UpperCamelCase :Union[str, Any] = torch.clip(__lowercase , self.min_value , self.max_value) # Scale to [0, 1]. __UpperCamelCase :Union[str, Any] = (features - self.min_value) / (self.max_value - self.min_value) # Scale to [min_out, max_out]. return zero_one * (max_out - min_out) + min_out def UpperCamelCase__ ( self , __lowercase , __lowercase=(-1.0, 1.0) , __lowercase=False) -> Optional[int]: __UpperCamelCase , __UpperCamelCase :int = input_range __UpperCamelCase :Optional[int] = torch.clip(__lowercase , __lowercase , __lowercase) if clip else outputs # Scale to [0, 1]. __UpperCamelCase :List[str] = (outputs - min_out) / (max_out - min_out) # Scale to [self.min_value, self.max_value]. return zero_one * (self.max_value - self.min_value) + self.min_value def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase) -> List[Any]: __UpperCamelCase :List[str] = input_tokens > 0 __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self.notes_encoder( encoder_input_tokens=__lowercase , encoder_inputs_mask=__lowercase) __UpperCamelCase , __UpperCamelCase :Union[str, Any] = self.continuous_encoder( encoder_inputs=__lowercase , encoder_inputs_mask=__lowercase) return [(tokens_encoded, tokens_mask), (continuous_encoded, continuous_mask)] def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase) -> str: __UpperCamelCase :Optional[int] = noise_time if not torch.is_tensor(__lowercase): __UpperCamelCase :str = torch.tensor([timesteps] , dtype=torch.long , device=input_tokens.device) elif torch.is_tensor(__lowercase) and len(timesteps.shape) == 0: __UpperCamelCase :Dict = timesteps[None].to(input_tokens.device) # broadcast to batch dimension in a way that's compatible with ONNX/Core ML __UpperCamelCase :List[str] = timesteps * torch.ones(input_tokens.shape[0] , dtype=timesteps.dtype , device=timesteps.device) __UpperCamelCase :Tuple = self.decoder( encodings_and_masks=__lowercase , decoder_input_tokens=__lowercase , decoder_noise_time=__lowercase) return logits @torch.no_grad() def __call__( self , __lowercase , __lowercase = None , __lowercase = 100 , __lowercase = True , __lowercase = "numpy" , __lowercase = None , __lowercase = 1 , ) -> Union[AudioPipelineOutput, Tuple]: if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__lowercase , __lowercase) or callback_steps <= 0) ): raise ValueError( f"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" f""" {type(__lowercase)}.""") __UpperCamelCase :Union[str, Any] = np.zeros([1, TARGET_FEATURE_LENGTH, self.n_dims] , dtype=np.floataa) __UpperCamelCase :Union[str, Any] = np.zeros([1, 0, self.n_dims] , np.floataa) __UpperCamelCase :Union[str, Any] = torch.ones((1, TARGET_FEATURE_LENGTH) , dtype=__lowercase , device=self.device) for i, encoder_input_tokens in enumerate(__lowercase): if i == 0: __UpperCamelCase :int = torch.from_numpy(pred_mel[:1].copy()).to( device=self.device , dtype=self.decoder.dtype) # The first chunk has no previous context. __UpperCamelCase :int = torch.zeros((1, TARGET_FEATURE_LENGTH) , dtype=__lowercase , device=self.device) else: # The full song pipeline does not feed in a context feature, so the mask # will be all 0s after the feature converter. Because we know we're # feeding in a full context chunk from the previous prediction, set it # to all 1s. __UpperCamelCase :Tuple = ones __UpperCamelCase :Optional[Any] = self.scale_features( __lowercase , output_range=[-1.0, 1.0] , clip=__lowercase) __UpperCamelCase :int = self.encode( input_tokens=torch.IntTensor([encoder_input_tokens]).to(device=self.device) , continuous_inputs=__lowercase , continuous_mask=__lowercase , ) # Sample encoder_continuous_inputs shaped gaussian noise to begin loop __UpperCamelCase :int = randn_tensor( shape=encoder_continuous_inputs.shape , generator=__lowercase , device=self.device , dtype=self.decoder.dtype , ) # set step values self.scheduler.set_timesteps(__lowercase) # Denoising diffusion loop for j, t in enumerate(self.progress_bar(self.scheduler.timesteps)): __UpperCamelCase :Optional[int] = self.decode( encodings_and_masks=__lowercase , input_tokens=__lowercase , noise_time=t / self.scheduler.config.num_train_timesteps , ) # Compute previous output: x_t -> x_t-1 __UpperCamelCase :int = self.scheduler.step(__lowercase , __lowercase , __lowercase , generator=__lowercase).prev_sample __UpperCamelCase :Tuple = self.scale_to_features(__lowercase , input_range=[-1.0, 1.0]) __UpperCamelCase :List[Any] = mel[:1] __UpperCamelCase :Optional[Any] = mel.cpu().float().numpy() __UpperCamelCase :Any = np.concatenate([full_pred_mel, pred_mel[:1]] , axis=1) # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__lowercase , __lowercase) logger.info('''Generated segment''' , __lowercase) if output_type == "numpy" and not is_onnx_available(): raise ValueError( '''Cannot return output in \'np\' format if ONNX is not available. Make sure to have ONNX installed or set \'output_type\' to \'mel\'.''') elif output_type == "numpy" and self.melgan is None: raise ValueError( '''Cannot return output in \'np\' format if melgan component is not defined. Make sure to define `self.melgan` or set \'output_type\' to \'mel\'.''') if output_type == "numpy": __UpperCamelCase :Optional[Any] = self.melgan(input_features=full_pred_mel.astype(np.floataa)) else: __UpperCamelCase :List[str] = full_pred_mel if not return_dict: return (output,) return AudioPipelineOutput(audios=__lowercase)
43
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = { '''alibaba-damo/mgp-str-base''': '''https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json''', } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[Any] = """mgp-str""" def __init__( self , __lowercase=[32, 128] , __lowercase=4 , __lowercase=3 , __lowercase=27 , __lowercase=38 , __lowercase=50_257 , __lowercase=30_522 , __lowercase=768 , __lowercase=12 , __lowercase=12 , __lowercase=4.0 , __lowercase=True , __lowercase=False , __lowercase=1E-5 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=False , __lowercase=0.02 , **__lowercase , ) -> List[Any]: super().__init__(**__lowercase) __UpperCamelCase :Any = image_size __UpperCamelCase :Optional[Any] = patch_size __UpperCamelCase :Optional[int] = num_channels __UpperCamelCase :int = max_token_length __UpperCamelCase :Any = num_character_labels __UpperCamelCase :Optional[Any] = num_bpe_labels __UpperCamelCase :int = num_wordpiece_labels __UpperCamelCase :Optional[int] = hidden_size __UpperCamelCase :int = num_hidden_layers __UpperCamelCase :Optional[Any] = num_attention_heads __UpperCamelCase :List[str] = mlp_ratio __UpperCamelCase :Dict = distilled __UpperCamelCase :Any = layer_norm_eps __UpperCamelCase :Dict = drop_rate __UpperCamelCase :str = qkv_bias __UpperCamelCase :Any = attn_drop_rate __UpperCamelCase :List[str] = drop_path_rate __UpperCamelCase :Optional[int] = output_aa_attentions __UpperCamelCase :Tuple = initializer_range
43
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( HubertConfig, HubertForCTC, HubertModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __lowercase = logging.get_logger(__name__) __lowercase = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' for attribute in key.split('''.''' ): __UpperCamelCase :str = getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if weight_type is not None: __UpperCamelCase :Any = getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).shape else: __UpperCamelCase :Union[str, Any] = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __UpperCamelCase :str = value elif weight_type == "weight_g": __UpperCamelCase :List[str] = value elif weight_type == "weight_v": __UpperCamelCase :str = value elif weight_type == "bias": __UpperCamelCase :Union[str, Any] = value else: __UpperCamelCase :str = value logger.info(f"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = [] __UpperCamelCase :int = fairseq_model.state_dict() __UpperCamelCase :List[Any] = hf_model.hubert.feature_extractor if is_finetuned else hf_model.feature_extractor for name, value in fairseq_dict.items(): __UpperCamelCase :List[Any] = False if "conv_layers" in name: load_conv_layer( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , hf_model.config.feat_extract_norm == '''group''' , ) __UpperCamelCase :List[str] = True else: for key, mapped_key in MAPPING.items(): __UpperCamelCase :Dict = '''hubert.''' + mapped_key if (is_finetuned and mapped_key != '''lm_head''') else mapped_key if key in name or (key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0] and not is_finetuned): __UpperCamelCase :Optional[Any] = True if "*" in mapped_key: __UpperCamelCase :List[str] = name.split(SCREAMING_SNAKE_CASE )[0].split('''.''' )[-2] __UpperCamelCase :Optional[int] = mapped_key.replace('''*''' , SCREAMING_SNAKE_CASE ) if "weight_g" in name: __UpperCamelCase :int = '''weight_g''' elif "weight_v" in name: __UpperCamelCase :List[Any] = '''weight_v''' elif "weight" in name: __UpperCamelCase :Dict = '''weight''' elif "bias" in name: __UpperCamelCase :Dict = '''bias''' else: __UpperCamelCase :Dict = None set_recursively(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(SCREAMING_SNAKE_CASE ) logger.warning(f"""Unused weights: {unused_weights}""" ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = full_name.split('''conv_layers.''' )[-1] __UpperCamelCase :Optional[int] = name.split('''.''' ) __UpperCamelCase :str = int(items[0] ) __UpperCamelCase :List[Any] = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) __UpperCamelCase :Dict = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) __UpperCamelCase :Any = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) __UpperCamelCase :int = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) __UpperCamelCase :Union[str, Any] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(SCREAMING_SNAKE_CASE ) @torch.no_grad() def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=True ): '''simple docstring''' if config_path is not None: __UpperCamelCase :Tuple = HubertConfig.from_pretrained(SCREAMING_SNAKE_CASE ) else: __UpperCamelCase :Optional[int] = HubertConfig() if is_finetuned: if dict_path: __UpperCamelCase :Optional[int] = Dictionary.load(SCREAMING_SNAKE_CASE ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq __UpperCamelCase :Optional[int] = target_dict.pad_index __UpperCamelCase :Dict = target_dict.bos_index __UpperCamelCase :str = target_dict.eos_index __UpperCamelCase :Dict = len(target_dict.symbols ) __UpperCamelCase :List[Any] = os.path.join(SCREAMING_SNAKE_CASE , '''vocab.json''' ) if not os.path.isdir(SCREAMING_SNAKE_CASE ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(SCREAMING_SNAKE_CASE ) ) return os.makedirs(SCREAMING_SNAKE_CASE , exist_ok=SCREAMING_SNAKE_CASE ) with open(SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(target_dict.indices , SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = WavaVecaCTCTokenizer( SCREAMING_SNAKE_CASE , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :Union[str, Any] = True if config.feat_extract_norm == '''layer''' else False __UpperCamelCase :Any = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0 , do_normalize=SCREAMING_SNAKE_CASE , return_attention_mask=SCREAMING_SNAKE_CASE , ) __UpperCamelCase :Any = WavaVecaProcessor(feature_extractor=SCREAMING_SNAKE_CASE , tokenizer=SCREAMING_SNAKE_CASE ) processor.save_pretrained(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[str] = HubertForCTC(SCREAMING_SNAKE_CASE ) else: __UpperCamelCase :str = HubertModel(SCREAMING_SNAKE_CASE ) if is_finetuned: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :int = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :Any = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) __UpperCamelCase :Dict = model[0].eval() recursively_load_weights(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) hf_wavavec.save_pretrained(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) __lowercase = parser.parse_args() convert_hubert_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
43
1
def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :list[list[float]] = [] for data in source_data: for i, el in enumerate(SCREAMING_SNAKE_CASE ): if len(SCREAMING_SNAKE_CASE ) < i + 1: data_lists.append([] ) data_lists[i].append(float(SCREAMING_SNAKE_CASE ) ) return data_lists def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :list[list[float]] = [] for dlist, weight in zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[int] = min(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Dict = max(SCREAMING_SNAKE_CASE ) __UpperCamelCase :list[float] = [] # for weight 0 score is 1 - actual score if weight == 0: for item in dlist: try: score.append(1 - ((item - mind) / (maxd - mind)) ) except ZeroDivisionError: score.append(1 ) elif weight == 1: for item in dlist: try: score.append((item - mind) / (maxd - mind) ) except ZeroDivisionError: score.append(0 ) # weight not 0 or 1 else: __UpperCamelCase :int = f"""Invalid weight of {weight:f} provided""" raise ValueError(SCREAMING_SNAKE_CASE ) score_lists.append(SCREAMING_SNAKE_CASE ) return score_lists def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :list[float] = [0 for i in range(len(score_lists[0] ) )] for slist in score_lists: for j, ele in enumerate(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = final_scores[j] + ele return final_scores def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Dict = get_data(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[Any] = calculate_each_score(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[str] = generate_final_scores(SCREAMING_SNAKE_CASE ) # append scores to source data for i, ele in enumerate(SCREAMING_SNAKE_CASE ): source_data[i].append(SCREAMING_SNAKE_CASE ) return source_data
43
import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters __lowercase = (720, 1280) # Height, Width __lowercase = (0.4, 0.6) # if height or width lower than this scale, drop it. __lowercase = 1 / 100 __lowercase = '''''' __lowercase = '''''' __lowercase = '''''' __lowercase = 250 def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase , __UpperCamelCase :List[Any] = get_dataset(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for index in range(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = random.sample(range(len(SCREAMING_SNAKE_CASE ) ) , 4 ) __UpperCamelCase , __UpperCamelCase , __UpperCamelCase :str = update_image_and_anno( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , filter_scale=SCREAMING_SNAKE_CASE , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __UpperCamelCase :List[Any] = random_chars(32 ) __UpperCamelCase :List[str] = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __UpperCamelCase :Tuple = f"""{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}""" cva.imwrite(f"""{file_root}.jpg""" , SCREAMING_SNAKE_CASE , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(f"""Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}""" ) __UpperCamelCase :Optional[Any] = [] for anno in new_annos: __UpperCamelCase :int = anno[3] - anno[1] __UpperCamelCase :Optional[int] = anno[4] - anno[2] __UpperCamelCase :int = anno[1] + width / 2 __UpperCamelCase :List[str] = anno[2] + height / 2 __UpperCamelCase :str = f"""{anno[0]} {x_center} {y_center} {width} {height}""" annos_list.append(SCREAMING_SNAKE_CASE ) with open(f"""{file_root}.txt""" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :str = [] __UpperCamelCase :str = [] for label_file in glob.glob(os.path.join(SCREAMING_SNAKE_CASE , '''*.txt''' ) ): __UpperCamelCase :Any = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(SCREAMING_SNAKE_CASE ) as in_file: __UpperCamelCase :str = in_file.readlines() __UpperCamelCase :Optional[int] = os.path.join(SCREAMING_SNAKE_CASE , f"""{label_name}.jpg""" ) __UpperCamelCase :int = [] for obj_list in obj_lists: __UpperCamelCase :Optional[int] = obj_list.rstrip('''\n''' ).split(''' ''' ) __UpperCamelCase :Any = float(obj[1] ) - float(obj[3] ) / 2 __UpperCamelCase :List[str] = float(obj[2] ) - float(obj[4] ) / 2 __UpperCamelCase :Dict = float(obj[1] ) + float(obj[3] ) / 2 __UpperCamelCase :List[str] = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(SCREAMING_SNAKE_CASE ) labels.append(SCREAMING_SNAKE_CASE ) return img_paths, labels def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = 0.0 , ): '''simple docstring''' __UpperCamelCase :List[str] = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __UpperCamelCase :List[Any] = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __UpperCamelCase :int = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __UpperCamelCase :Optional[int] = int(scale_x * output_size[1] ) __UpperCamelCase :Any = int(scale_y * output_size[0] ) __UpperCamelCase :List[str] = [] __UpperCamelCase :Dict = [] for i, index in enumerate(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Any = all_img_list[index] path_list.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = all_annos[index] __UpperCamelCase :Union[str, Any] = cva.imread(SCREAMING_SNAKE_CASE ) if i == 0: # top-left __UpperCamelCase :str = cva.resize(SCREAMING_SNAKE_CASE , (divid_point_x, divid_point_y) ) __UpperCamelCase :Union[str, Any] = img for bbox in img_annos: __UpperCamelCase :Union[str, Any] = bbox[1] * scale_x __UpperCamelCase :Optional[Any] = bbox[2] * scale_y __UpperCamelCase :int = bbox[3] * scale_x __UpperCamelCase :Union[str, Any] = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __UpperCamelCase :str = cva.resize(SCREAMING_SNAKE_CASE , (output_size[1] - divid_point_x, divid_point_y) ) __UpperCamelCase :List[str] = img for bbox in img_annos: __UpperCamelCase :str = scale_x + bbox[1] * (1 - scale_x) __UpperCamelCase :Dict = bbox[2] * scale_y __UpperCamelCase :Optional[Any] = scale_x + bbox[3] * (1 - scale_x) __UpperCamelCase :List[Any] = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __UpperCamelCase :str = cva.resize(SCREAMING_SNAKE_CASE , (divid_point_x, output_size[0] - divid_point_y) ) __UpperCamelCase :Optional[int] = img for bbox in img_annos: __UpperCamelCase :Tuple = bbox[1] * scale_x __UpperCamelCase :Optional[Any] = scale_y + bbox[2] * (1 - scale_y) __UpperCamelCase :Tuple = bbox[3] * scale_x __UpperCamelCase :Dict = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __UpperCamelCase :Optional[int] = cva.resize( SCREAMING_SNAKE_CASE , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __UpperCamelCase :Optional[int] = img for bbox in img_annos: __UpperCamelCase :Optional[Any] = scale_x + bbox[1] * (1 - scale_x) __UpperCamelCase :Optional[int] = scale_y + bbox[2] * (1 - scale_y) __UpperCamelCase :Optional[Any] = scale_x + bbox[3] * (1 - scale_x) __UpperCamelCase :int = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __UpperCamelCase :List[Any] = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' assert number_char > 1, "The number of character should greater than 1" __UpperCamelCase :Optional[Any] = ascii_lowercase + digits return "".join(random.choice(SCREAMING_SNAKE_CASE ) for _ in range(SCREAMING_SNAKE_CASE ) ) if __name__ == "__main__": main() print('''DONE ✅''')
43
1
import argparse import json from tqdm import tqdm def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--src_path''' , type=SCREAMING_SNAKE_CASE , default='''biencoder-nq-dev.json''' , help='''Path to raw DPR training data''' , ) parser.add_argument( '''--evaluation_set''' , type=SCREAMING_SNAKE_CASE , help='''where to store parsed evaluation_set file''' , ) parser.add_argument( '''--gold_data_path''' , type=SCREAMING_SNAKE_CASE , help='''where to store parsed gold_data_path file''' , ) __UpperCamelCase :str = parser.parse_args() with open(args.src_path , '''r''' ) as src_file, open(args.evaluation_set , '''w''' ) as eval_file, open( args.gold_data_path , '''w''' ) as gold_file: __UpperCamelCase :List[str] = json.load(SCREAMING_SNAKE_CASE ) for dpr_record in tqdm(SCREAMING_SNAKE_CASE ): __UpperCamelCase :List[str] = dpr_record['''question'''] __UpperCamelCase :Tuple = [context['''title'''] for context in dpr_record['''positive_ctxs''']] eval_file.write(question + '''\n''' ) gold_file.write('''\t'''.join(SCREAMING_SNAKE_CASE ) + '''\n''' ) if __name__ == "__main__": main()
43
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = { '''facebook/wav2vec2-base-960h''': '''https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/config.json''', # See all Wav2Vec2 models at https://huggingface.co/models?filter=wav2vec2 } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Union[str, Any] = """wav2vec2""" def __init__( self , __lowercase=32 , __lowercase=768 , __lowercase=12 , __lowercase=12 , __lowercase=3_072 , __lowercase="gelu" , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.1 , __lowercase=0.1 , __lowercase=0.02 , __lowercase=1E-5 , __lowercase="group" , __lowercase="gelu" , __lowercase=(512, 512, 512, 512, 512, 512, 512) , __lowercase=(5, 2, 2, 2, 2, 2, 2) , __lowercase=(10, 3, 3, 3, 3, 2, 2) , __lowercase=False , __lowercase=128 , __lowercase=16 , __lowercase=False , __lowercase=True , __lowercase=0.05 , __lowercase=10 , __lowercase=2 , __lowercase=0.0 , __lowercase=10 , __lowercase=0 , __lowercase=320 , __lowercase=2 , __lowercase=0.1 , __lowercase=100 , __lowercase=256 , __lowercase=256 , __lowercase=0.1 , __lowercase="sum" , __lowercase=False , __lowercase=False , __lowercase=256 , __lowercase=(512, 512, 512, 512, 1_500) , __lowercase=(5, 3, 3, 1, 1) , __lowercase=(1, 2, 3, 1, 1) , __lowercase=512 , __lowercase=0 , __lowercase=1 , __lowercase=2 , __lowercase=False , __lowercase=3 , __lowercase=2 , __lowercase=3 , __lowercase=None , __lowercase=None , **__lowercase , ) -> int: super().__init__(**__lowercase , pad_token_id=__lowercase , bos_token_id=__lowercase , eos_token_id=__lowercase) __UpperCamelCase :Any = hidden_size __UpperCamelCase :int = feat_extract_norm __UpperCamelCase :Tuple = feat_extract_activation __UpperCamelCase :Union[str, Any] = list(__lowercase) __UpperCamelCase :List[Any] = list(__lowercase) __UpperCamelCase :int = list(__lowercase) __UpperCamelCase :List[Any] = conv_bias __UpperCamelCase :Optional[int] = num_conv_pos_embeddings __UpperCamelCase :Dict = num_conv_pos_embedding_groups __UpperCamelCase :Any = len(self.conv_dim) __UpperCamelCase :List[str] = num_hidden_layers __UpperCamelCase :int = intermediate_size __UpperCamelCase :str = hidden_act __UpperCamelCase :Any = num_attention_heads __UpperCamelCase :int = hidden_dropout __UpperCamelCase :Tuple = attention_dropout __UpperCamelCase :List[str] = activation_dropout __UpperCamelCase :Optional[Any] = feat_proj_dropout __UpperCamelCase :Any = final_dropout __UpperCamelCase :Any = layerdrop __UpperCamelCase :str = layer_norm_eps __UpperCamelCase :Optional[Any] = initializer_range __UpperCamelCase :List[str] = vocab_size __UpperCamelCase :str = do_stable_layer_norm __UpperCamelCase :Union[str, Any] = use_weighted_layer_sum if ( (len(self.conv_stride) != self.num_feat_extract_layers) or (len(self.conv_kernel) != self.num_feat_extract_layers) or (len(self.conv_dim) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' f""" {len(self.conv_dim)}`, `len(config.conv_stride) = {len(self.conv_stride)}`,""" f""" `len(config.conv_kernel) = {len(self.conv_kernel)}`.""") # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __UpperCamelCase :List[Any] = apply_spec_augment __UpperCamelCase :Tuple = mask_time_prob __UpperCamelCase :int = mask_time_length __UpperCamelCase :Dict = mask_time_min_masks __UpperCamelCase :str = mask_feature_prob __UpperCamelCase :List[str] = mask_feature_length __UpperCamelCase :Union[str, Any] = mask_feature_min_masks # parameters for pretraining with codevector quantized representations __UpperCamelCase :Optional[Any] = num_codevectors_per_group __UpperCamelCase :List[Any] = num_codevector_groups __UpperCamelCase :Tuple = contrastive_logits_temperature __UpperCamelCase :Optional[int] = feat_quantizer_dropout __UpperCamelCase :Optional[int] = num_negatives __UpperCamelCase :List[Any] = codevector_dim __UpperCamelCase :str = proj_codevector_dim __UpperCamelCase :List[str] = diversity_loss_weight # ctc loss __UpperCamelCase :Tuple = ctc_loss_reduction __UpperCamelCase :Tuple = ctc_zero_infinity # adapter __UpperCamelCase :List[str] = add_adapter __UpperCamelCase :Tuple = adapter_kernel_size __UpperCamelCase :str = adapter_stride __UpperCamelCase :Tuple = num_adapter_layers __UpperCamelCase :Tuple = output_hidden_size or hidden_size __UpperCamelCase :Optional[Any] = adapter_attn_dim # SequenceClassification-specific parameter. Feel free to ignore for other classes. __UpperCamelCase :Optional[Any] = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. __UpperCamelCase :Optional[int] = list(__lowercase) __UpperCamelCase :List[Any] = list(__lowercase) __UpperCamelCase :List[Any] = list(__lowercase) __UpperCamelCase :str = xvector_output_dim @property def UpperCamelCase__ ( self) -> List[str]: return functools.reduce(operator.mul , self.conv_stride , 1)
43
1
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowercase = logging.get_logger(__name__) __lowercase = { '''nvidia/segformer-b0-finetuned-ade-512-512''': ( '''https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json''' ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : List[Any] = """segformer""" def __init__( self , __lowercase=3 , __lowercase=4 , __lowercase=[2, 2, 2, 2] , __lowercase=[8, 4, 2, 1] , __lowercase=[32, 64, 160, 256] , __lowercase=[7, 3, 3, 3] , __lowercase=[4, 2, 2, 2] , __lowercase=[1, 2, 5, 8] , __lowercase=[4, 4, 4, 4] , __lowercase="gelu" , __lowercase=0.0 , __lowercase=0.0 , __lowercase=0.1 , __lowercase=0.02 , __lowercase=0.1 , __lowercase=1E-6 , __lowercase=256 , __lowercase=255 , **__lowercase , ) -> List[Any]: super().__init__(**__lowercase) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( '''Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be''' ''' removed, as the behaviour will default to that of reshape_last_stage = True.''' , __lowercase , ) __UpperCamelCase :Tuple = num_channels __UpperCamelCase :int = num_encoder_blocks __UpperCamelCase :Tuple = depths __UpperCamelCase :Dict = sr_ratios __UpperCamelCase :List[Any] = hidden_sizes __UpperCamelCase :str = patch_sizes __UpperCamelCase :Tuple = strides __UpperCamelCase :str = mlp_ratios __UpperCamelCase :str = num_attention_heads __UpperCamelCase :Optional[Any] = hidden_act __UpperCamelCase :Optional[Any] = hidden_dropout_prob __UpperCamelCase :str = attention_probs_dropout_prob __UpperCamelCase :List[Any] = classifier_dropout_prob __UpperCamelCase :Optional[int] = initializer_range __UpperCamelCase :Dict = drop_path_rate __UpperCamelCase :List[Any] = layer_norm_eps __UpperCamelCase :Optional[int] = decoder_hidden_size __UpperCamelCase :Union[str, Any] = kwargs.get('''reshape_last_stage''' , __lowercase) __UpperCamelCase :Union[str, Any] = semantic_loss_ignore_index class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Any = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ]) @property def UpperCamelCase__ ( self) -> float: return 1E-4 @property def UpperCamelCase__ ( self) -> int: return 12
43
from typing import List, Optional, Union import numpy as np import PIL.Image from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import rescale, resize, to_channel_dimension_format from ...image_utils import ( ChannelDimension, PILImageResampling, get_image_size, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __lowercase = logging.get_logger(__name__) class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' a__ : Optional[Any] = ["""pixel_values"""] def __init__( self , __lowercase = True , __lowercase = 32 , __lowercase=PILImageResampling.BILINEAR , __lowercase = True , **__lowercase , ) -> None: __UpperCamelCase :Optional[int] = do_resize __UpperCamelCase :Any = do_rescale __UpperCamelCase :str = size_divisor __UpperCamelCase :Dict = resample super().__init__(**__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase , __lowercase = None , **__lowercase) -> np.ndarray: __UpperCamelCase , __UpperCamelCase :int = get_image_size(__lowercase) # Rounds the height and width down to the closest multiple of size_divisor __UpperCamelCase :List[Any] = height // size_divisor * size_divisor __UpperCamelCase :List[str] = width // size_divisor * size_divisor __UpperCamelCase :str = resize(__lowercase , (new_h, new_w) , resample=__lowercase , data_format=__lowercase , **__lowercase) return image def UpperCamelCase__ ( self , __lowercase , __lowercase , __lowercase = None , **__lowercase) -> np.ndarray: return rescale(image=__lowercase , scale=__lowercase , data_format=__lowercase , **__lowercase) def UpperCamelCase__ ( self , __lowercase , __lowercase = None , __lowercase = None , __lowercase=None , __lowercase = None , __lowercase = None , __lowercase = ChannelDimension.FIRST , **__lowercase , ) -> BatchFeature: __UpperCamelCase :Union[str, Any] = do_resize if do_resize is not None else self.do_resize __UpperCamelCase :Tuple = do_rescale if do_rescale is not None else self.do_rescale __UpperCamelCase :List[str] = size_divisor if size_divisor is not None else self.size_divisor __UpperCamelCase :List[Any] = resample if resample is not None else self.resample if do_resize and size_divisor is None: raise ValueError('''size_divisor is required for resizing''') __UpperCamelCase :List[Any] = make_list_of_images(__lowercase) if not valid_images(__lowercase): raise ValueError('''Invalid image(s)''') # All transformations expect numpy arrays. __UpperCamelCase :Optional[Any] = [to_numpy_array(__lowercase) for img in images] if do_resize: __UpperCamelCase :List[str] = [self.resize(__lowercase , size_divisor=__lowercase , resample=__lowercase) for image in images] if do_rescale: __UpperCamelCase :Dict = [self.rescale(__lowercase , scale=1 / 255) for image in images] __UpperCamelCase :str = [to_channel_dimension_format(__lowercase , __lowercase) for image in images] __UpperCamelCase :int = {'''pixel_values''': images} return BatchFeature(data=__lowercase , tensor_type=__lowercase)
43
1
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class lowerCamelCase_ ( UpperCAmelCase_ ): '''simple docstring''' def UpperCamelCase__ ( self , __lowercase) -> Optional[int]: with open(__lowercase , encoding='''utf-8''') as input_file: __UpperCamelCase :str = re.compile(r'''(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)''') __UpperCamelCase :str = input_file.read() __UpperCamelCase :Optional[Any] = regexp.search(__lowercase) return match def UpperCamelCase__ ( self , __lowercase) -> int: with open(__lowercase , encoding='''utf-8''') as input_file: __UpperCamelCase :Any = re.compile(r'''#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()''' , re.DOTALL) __UpperCamelCase :Optional[int] = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __UpperCamelCase :List[Any] = regexp.finditer(__lowercase) __UpperCamelCase :Optional[Any] = [match for match in matches if match is not None and match.group(1) is not None] return matches[0] if matches else None def UpperCamelCase__ ( self) -> int: __UpperCamelCase :Tuple = Path('''./datasets''') __UpperCamelCase :Tuple = list(dataset_paths.absolute().glob('''**/*.py''')) for dataset in dataset_files: if self._no_encoding_on_file_open(str(__lowercase)): raise AssertionError(f"""open(...) must use utf-8 encoding in {dataset}""") def UpperCamelCase__ ( self) -> Optional[int]: __UpperCamelCase :str = Path('''./datasets''') __UpperCamelCase :str = list(dataset_paths.absolute().glob('''**/*.py''')) for dataset in dataset_files: if self._no_print_statements(str(__lowercase)): raise AssertionError(f"""print statement found in {dataset}. Use datasets.logger/logging instead.""")
43
from __future__ import annotations from PIL import Image # Define glider example __lowercase = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example __lowercase = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :int = [] for i in range(len(SCREAMING_SNAKE_CASE ) ): __UpperCamelCase :Dict = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours __UpperCamelCase :List[str] = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(SCREAMING_SNAKE_CASE ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(SCREAMING_SNAKE_CASE ) - 1: neighbour_count += cells[i + 1][j] if i < len(SCREAMING_SNAKE_CASE ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. __UpperCamelCase :List[str] = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(SCREAMING_SNAKE_CASE ) return next_generation def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = [] for _ in range(SCREAMING_SNAKE_CASE ): # Create output image __UpperCamelCase :Dict = Image.new('''RGB''' , (len(cells[0] ), len(SCREAMING_SNAKE_CASE )) ) __UpperCamelCase :Any = img.load() # Save cells to image for x in range(len(SCREAMING_SNAKE_CASE ) ): for y in range(len(cells[0] ) ): __UpperCamelCase :Optional[Any] = 255 - cells[y][x] * 255 __UpperCamelCase :int = (colour, colour, colour) # Save image images.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Optional[int] = new_generation(SCREAMING_SNAKE_CASE ) return images if __name__ == "__main__": __lowercase = generate_images(GLIDER, 16) images[0].save('''out.gif''', save_all=True, append_images=images[1:])
43
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __lowercase = { '''configuration_efficientnet''': [ '''EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''EfficientNetConfig''', '''EfficientNetOnnxConfig''', ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = ['''EfficientNetImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''EfficientNetForImageClassification''', '''EfficientNetModel''', '''EfficientNetPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_efficientnet import ( EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientNetConfig, EfficientNetOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientnet import EfficientNetImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientnet import ( EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientNetForImageClassification, EfficientNetModel, EfficientNetPreTrainedModel, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
43
import re import jax.numpy as jnp from flax.traverse_util import flatten_dict, unflatten_dict from jax.random import PRNGKey from ..utils import logging __lowercase = logging.get_logger(__name__) def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = R'''\w+[.]\d+''' __UpperCamelCase :List[str] = re.findall(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for pat in pats: __UpperCamelCase :int = key.replace(SCREAMING_SNAKE_CASE , '''_'''.join(pat.split('''.''' ) ) ) return key def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Tuple = pt_tuple_key[:-1] + ('''scale''',) if ( any('''norm''' in str_ for str_ in pt_tuple_key ) and (pt_tuple_key[-1] == "bias") and (pt_tuple_key[:-1] + ("bias",) not in random_flax_state_dict) and (pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict) ): __UpperCamelCase :str = pt_tuple_key[:-1] + ('''scale''',) return renamed_pt_tuple_key, pt_tensor elif pt_tuple_key[-1] in ["weight", "gamma"] and pt_tuple_key[:-1] + ("scale",) in random_flax_state_dict: __UpperCamelCase :Any = pt_tuple_key[:-1] + ('''scale''',) return renamed_pt_tuple_key, pt_tensor # embedding if pt_tuple_key[-1] == "weight" and pt_tuple_key[:-1] + ("embedding",) in random_flax_state_dict: __UpperCamelCase :str = pt_tuple_key[:-1] + ('''embedding''',) return renamed_pt_tuple_key, pt_tensor # conv layer __UpperCamelCase :List[str] = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4: __UpperCamelCase :List[Any] = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer __UpperCamelCase :List[str] = pt_tuple_key[:-1] + ('''kernel''',) if pt_tuple_key[-1] == "weight": __UpperCamelCase :Any = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight __UpperCamelCase :int = pt_tuple_key[:-1] + ('''weight''',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias __UpperCamelCase :int = pt_tuple_key[:-1] + ('''bias''',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=42 ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = {k: v.numpy() for k, v in pt_state_dict.items()} # Step 2: Since the model is stateless, get random Flax params __UpperCamelCase :str = flax_model.init_weights(PRNGKey(SCREAMING_SNAKE_CASE ) ) __UpperCamelCase :int = flatten_dict(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = {} # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): __UpperCamelCase :List[Any] = rename_key(SCREAMING_SNAKE_CASE ) __UpperCamelCase :List[Any] = tuple(renamed_pt_key.split('''.''' ) ) # Correctly rename weight parameters __UpperCamelCase , __UpperCamelCase :Any = rename_key_and_reshape_tensor(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f"""PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape """ f"""{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.""" ) # also add unexpected weight so that warning is thrown __UpperCamelCase :str = jnp.asarray(SCREAMING_SNAKE_CASE ) return unflatten_dict(SCREAMING_SNAKE_CASE )
43
1
def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' try: __UpperCamelCase :int = float(SCREAMING_SNAKE_CASE ) except ValueError: raise ValueError('''Please enter a valid number''' ) __UpperCamelCase :int = decimal - int(SCREAMING_SNAKE_CASE ) if fractional_part == 0: return int(SCREAMING_SNAKE_CASE ), 1 else: __UpperCamelCase :Union[str, Any] = len(str(SCREAMING_SNAKE_CASE ).split('''.''' )[1] ) __UpperCamelCase :List[Any] = int(decimal * (10**number_of_frac_digits) ) __UpperCamelCase :Dict = 10**number_of_frac_digits __UpperCamelCase , __UpperCamelCase :Optional[int] = denominator, numerator while True: __UpperCamelCase :Optional[Any] = dividend % divisor if remainder == 0: break __UpperCamelCase , __UpperCamelCase :Optional[int] = divisor, remainder __UpperCamelCase , __UpperCamelCase :Union[str, Any] = numerator / divisor, denominator / divisor return int(SCREAMING_SNAKE_CASE ), int(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(F'{decimal_to_fraction(2) = }') print(F'{decimal_to_fraction(8_9.0) = }') print(F'{decimal_to_fraction("67") = }') print(F'{decimal_to_fraction("45.0") = }') print(F'{decimal_to_fraction(1.5) = }') print(F'{decimal_to_fraction("6.25") = }') print(F'{decimal_to_fraction("78td") = }')
43
import argparse import torch from ...utils import logging from . import AlbertConfig, AlbertForPreTraining, load_tf_weights_in_albert logging.set_verbosity_info() def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :List[Any] = AlbertConfig.from_json_file(SCREAMING_SNAKE_CASE ) print(f"""Building PyTorch model from configuration: {config}""" ) __UpperCamelCase :List[str] = AlbertForPreTraining(SCREAMING_SNAKE_CASE ) # Load weights from tf checkpoint load_tf_weights_in_albert(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--albert_config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained ALBERT model. \n''' '''This specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) __lowercase = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.albert_config_file, args.pytorch_dump_path)
43
1
def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' _validate_point(SCREAMING_SNAKE_CASE ) _validate_point(SCREAMING_SNAKE_CASE ) if len(SCREAMING_SNAKE_CASE ) != len(SCREAMING_SNAKE_CASE ): raise ValueError('''Both points must be in the same n-dimensional space''' ) return float(sum(abs(a - b ) for a, b in zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) ) def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' if point: if isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): for item in point: if not isinstance(SCREAMING_SNAKE_CASE , (int, float) ): __UpperCamelCase :Optional[int] = ( '''Expected a list of numbers as input, found ''' f"""{type(SCREAMING_SNAKE_CASE ).__name__}""" ) raise TypeError(SCREAMING_SNAKE_CASE ) else: __UpperCamelCase :List[str] = f"""Expected a list of numbers as input, found {type(SCREAMING_SNAKE_CASE ).__name__}""" raise TypeError(SCREAMING_SNAKE_CASE ) else: raise ValueError('''Missing an input''' ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' _validate_point(SCREAMING_SNAKE_CASE ) _validate_point(SCREAMING_SNAKE_CASE ) if len(SCREAMING_SNAKE_CASE ) != len(SCREAMING_SNAKE_CASE ): raise ValueError('''Both points must be in the same n-dimensional space''' ) return float(sum(abs(x - y ) for x, y in zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
43
import math import qiskit def lowerCamelCase ( SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1 ): '''simple docstring''' if ( isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) or isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) or isinstance(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) ): raise TypeError('''inputs must be integers.''' ) if (input_a < 0) or (input_a < 0) or (carry_in < 0): raise ValueError('''inputs must be positive.''' ) if ( (math.floor(SCREAMING_SNAKE_CASE ) != input_a) or (math.floor(SCREAMING_SNAKE_CASE ) != input_a) or (math.floor(SCREAMING_SNAKE_CASE ) != carry_in) ): raise ValueError('''inputs must be exact integers.''' ) if (input_a > 2) or (input_a > 2) or (carry_in > 2): raise ValueError('''inputs must be less or equal to 2.''' ) # build registers __UpperCamelCase :List[str] = qiskit.QuantumRegister(4 , '''qr''' ) __UpperCamelCase :str = qiskit.ClassicalRegister(2 , '''cr''' ) # list the entries __UpperCamelCase :Tuple = [input_a, input_a, carry_in] __UpperCamelCase :Optional[int] = qiskit.QuantumCircuit(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) for i in range(0 , 3 ): if entry[i] == 2: quantum_circuit.h(SCREAMING_SNAKE_CASE ) # for hadamard entries elif entry[i] == 1: quantum_circuit.x(SCREAMING_SNAKE_CASE ) # for 1 entries elif entry[i] == 0: quantum_circuit.i(SCREAMING_SNAKE_CASE ) # for 0 entries # build the circuit quantum_circuit.ccx(0 , 1 , 3 ) # ccx = toffoli gate quantum_circuit.cx(0 , 1 ) quantum_circuit.ccx(1 , 2 , 3 ) quantum_circuit.cx(1 , 2 ) quantum_circuit.cx(0 , 1 ) quantum_circuit.measure([2, 3] , SCREAMING_SNAKE_CASE ) # measure the last two qbits __UpperCamelCase :Optional[Any] = qiskit.Aer.get_backend('''aer_simulator''' ) __UpperCamelCase :Tuple = qiskit.execute(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , shots=1_000 ) return job.result().get_counts(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(F'Total sum count for state is: {quantum_full_adder(1, 1, 1)}')
43
1
def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' if not grid or not grid[0]: raise TypeError('''The grid does not contain the appropriate information''' ) for cell_n in range(1 , len(grid[0] ) ): grid[0][cell_n] += grid[0][cell_n - 1] __UpperCamelCase :List[Any] = grid[0] for row_n in range(1 , len(SCREAMING_SNAKE_CASE ) ): __UpperCamelCase :Dict = grid[row_n] __UpperCamelCase :List[str] = fill_row(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = grid[row_n] return grid[-1][-1] def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' current_row[0] += row_above[0] for cell_n in range(1 , len(SCREAMING_SNAKE_CASE ) ): current_row[cell_n] += min(current_row[cell_n - 1] , row_above[cell_n] ) return current_row if __name__ == "__main__": import doctest doctest.testmod()
43
import random def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Optional[Any] = a[left_index] __UpperCamelCase :Any = left_index + 1 for j in range(left_index + 1 , SCREAMING_SNAKE_CASE ): if a[j] < pivot: __UpperCamelCase , __UpperCamelCase :str = a[i], a[j] i += 1 __UpperCamelCase , __UpperCamelCase :Optional[int] = a[i - 1], a[left_index] return i - 1 def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' if left < right: __UpperCamelCase :int = random.randint(SCREAMING_SNAKE_CASE , right - 1 ) __UpperCamelCase , __UpperCamelCase :List[str] = ( a[left], a[pivot], ) # switches the pivot with the left most bound __UpperCamelCase :Dict = partition(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) quick_sort_random( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # recursive quicksort to the left of the pivot point quick_sort_random( SCREAMING_SNAKE_CASE , pivot_index + 1 , SCREAMING_SNAKE_CASE ) # recursive quicksort to the right of the pivot point def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = input('''Enter numbers separated by a comma:\n''' ).strip() __UpperCamelCase :Union[str, Any] = [int(SCREAMING_SNAKE_CASE ) for item in user_input.split(''',''' )] quick_sort_random(SCREAMING_SNAKE_CASE , 0 , len(SCREAMING_SNAKE_CASE ) ) print(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
43
1
from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time __lowercase = Lock() def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' global process_lock # we perform n swaps since after n swaps we know we are sorted # we *could* stop early if we are sorted already, but it takes as long to # find out we are sorted as it does to sort the list with this algorithm for i in range(0 , 10 ): if (i + position) % 2 == 0 and r_send is not None: # send your value to your right neighbor process_lock.acquire() r_send[1].send(SCREAMING_SNAKE_CASE ) process_lock.release() # receive your right neighbor's value process_lock.acquire() __UpperCamelCase :Any = rr_cv[0].recv() process_lock.release() # take the lower value since you are on the left __UpperCamelCase :Tuple = min(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) elif (i + position) % 2 != 0 and l_send is not None: # send your value to your left neighbor process_lock.acquire() l_send[1].send(SCREAMING_SNAKE_CASE ) process_lock.release() # receive your left neighbor's value process_lock.acquire() __UpperCamelCase :str = lr_cv[0].recv() process_lock.release() # take the higher value since you are on the right __UpperCamelCase :Tuple = max(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) # after all swaps are performed, send the values back to main result_pipe[1].send(SCREAMING_SNAKE_CASE ) def lowerCamelCase ( SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = [] __UpperCamelCase :Dict = [] # initialize the list of pipes where the values will be retrieved for _ in arr: result_pipe.append(Pipe() ) # creates the processes # the first and last process only have one neighbor so they are made outside # of the loop __UpperCamelCase :Optional[int] = Pipe() __UpperCamelCase :Optional[int] = Pipe() process_array_.append( Process( target=SCREAMING_SNAKE_CASE , args=(0, arr[0], None, temp_rs, None, temp_rr, result_pipe[0]) , ) ) __UpperCamelCase :Optional[int] = temp_rs __UpperCamelCase :Optional[Any] = temp_rr for i in range(1 , len(SCREAMING_SNAKE_CASE ) - 1 ): __UpperCamelCase :Any = Pipe() __UpperCamelCase :List[Any] = Pipe() process_array_.append( Process( target=SCREAMING_SNAKE_CASE , args=(i, arr[i], temp_ls, temp_rs, temp_lr, temp_rr, result_pipe[i]) , ) ) __UpperCamelCase :List[str] = temp_rs __UpperCamelCase :List[Any] = temp_rr process_array_.append( Process( target=SCREAMING_SNAKE_CASE , args=( len(SCREAMING_SNAKE_CASE ) - 1, arr[len(SCREAMING_SNAKE_CASE ) - 1], temp_ls, None, temp_lr, None, result_pipe[len(SCREAMING_SNAKE_CASE ) - 1], ) , ) ) # start the processes for p in process_array_: p.start() # wait for the processes to end and write their values to the list for p in range(0 , len(SCREAMING_SNAKE_CASE ) ): __UpperCamelCase :List[str] = result_pipe[p][0].recv() process_array_[p].join() return arr def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :int = list(range(10 , 0 , -1 ) ) print('''Initial List''' ) print(*SCREAMING_SNAKE_CASE ) __UpperCamelCase :Union[str, Any] = odd_even_transposition(SCREAMING_SNAKE_CASE ) print('''Sorted List\n''' ) print(*SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
43
def lowerCamelCase ( SCREAMING_SNAKE_CASE = 1 , SCREAMING_SNAKE_CASE = 1_000 ): '''simple docstring''' __UpperCamelCase :Union[str, Any] = 1 __UpperCamelCase :Any = 0 for divide_by_number in range(SCREAMING_SNAKE_CASE , digit + 1 ): __UpperCamelCase :list[int] = [] __UpperCamelCase :Optional[int] = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(SCREAMING_SNAKE_CASE ): __UpperCamelCase :Optional[Any] = len(SCREAMING_SNAKE_CASE ) __UpperCamelCase :int = divide_by_number else: has_been_divided.append(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Any = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
43
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowercase = { '''configuration_clap''': [ '''CLAP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ClapAudioConfig''', '''ClapConfig''', '''ClapTextConfig''', ], '''processing_clap''': ['''ClapProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowercase = [ '''CLAP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ClapModel''', '''ClapPreTrainedModel''', '''ClapTextModel''', '''ClapTextModelWithProjection''', '''ClapAudioModel''', '''ClapAudioModelWithProjection''', ] __lowercase = ['''ClapFeatureExtractor'''] if TYPE_CHECKING: from .configuration_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioConfig, ClapConfig, ClapTextConfig, ) from .processing_clap import ClapProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clap import ClapFeatureExtractor from .modeling_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioModel, ClapAudioModelWithProjection, ClapModel, ClapPreTrainedModel, ClapTextModel, ClapTextModelWithProjection, ) else: import sys __lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
43
import argparse import json from tqdm import tqdm def lowerCamelCase ( ): '''simple docstring''' __UpperCamelCase :Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--src_path''' , type=SCREAMING_SNAKE_CASE , default='''biencoder-nq-dev.json''' , help='''Path to raw DPR training data''' , ) parser.add_argument( '''--evaluation_set''' , type=SCREAMING_SNAKE_CASE , help='''where to store parsed evaluation_set file''' , ) parser.add_argument( '''--gold_data_path''' , type=SCREAMING_SNAKE_CASE , help='''where to store parsed gold_data_path file''' , ) __UpperCamelCase :str = parser.parse_args() with open(args.src_path , '''r''' ) as src_file, open(args.evaluation_set , '''w''' ) as eval_file, open( args.gold_data_path , '''w''' ) as gold_file: __UpperCamelCase :List[str] = json.load(SCREAMING_SNAKE_CASE ) for dpr_record in tqdm(SCREAMING_SNAKE_CASE ): __UpperCamelCase :List[str] = dpr_record['''question'''] __UpperCamelCase :Tuple = [context['''title'''] for context in dpr_record['''positive_ctxs''']] eval_file.write(question + '''\n''' ) gold_file.write('''\t'''.join(SCREAMING_SNAKE_CASE ) + '''\n''' ) if __name__ == "__main__": main()
43
1