code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. _lowercase = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. _lowercase = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. _lowercase = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : int = len([g for position, g in enumerate(snake_case__) if g == main_target[position]]) return (item, float(snake_case__)) def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : Any = random.randint(0 , len(snake_case__) - 1) lowerCAmelCase_ : Any = parent_a[:random_slice] + parent_a[random_slice:] lowerCAmelCase_ : Any = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : List[str] = list(snake_case__) if random.uniform(0 , 1) < MUTATION_PROBABILITY: lowerCAmelCase_ : Union[str, Any] = random.choice(snake_case__) return "".join(snake_case__) def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , ): lowerCAmelCase_ : Union[str, Any] = [] # Generate more children proportionally to the fitness score. lowerCAmelCase_ : List[Any] = int(parent_a[1] * 1_00) + 1 lowerCAmelCase_ : str = 10 if child_n >= 10 else child_n for _ in range(snake_case__): lowerCAmelCase_ : Union[str, Any] = population_score[random.randint(0 , snake_case__)][0] lowerCAmelCase_ , lowerCAmelCase_ : Tuple = crossover(parent_a[0] , snake_case__) # Append new string to the population list. pop.append(mutate(snake_case__ , snake_case__)) pop.append(mutate(snake_case__ , snake_case__)) return pop def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ = True): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: lowerCAmelCase_ : Dict = F'''{N_POPULATION} must be bigger than {N_SELECTED}''' raise ValueError(snake_case__) # Verify that the target contains no genes besides the ones inside genes variable. lowerCAmelCase_ : Tuple = sorted({c for c in target if c not in genes}) if not_in_genes_list: lowerCAmelCase_ : str = F'''{not_in_genes_list} is not in genes list, evolution cannot converge''' raise ValueError(snake_case__) # Generate random starting population. lowerCAmelCase_ : Tuple = [] for _ in range(snake_case__): population.append("".join([random.choice(snake_case__) for i in range(len(snake_case__))])) # Just some logs to know what the algorithms is doing. lowerCAmelCase_ , lowerCAmelCase_ : Dict = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(snake_case__) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. lowerCAmelCase_ : int = [evaluate(snake_case__ , snake_case__) for item in population] # Check if there is a matching evolution. lowerCAmelCase_ : List[Any] = sorted(snake_case__ , key=lambda snake_case__: x[1] , reverse=snake_case__) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( F'''\nGeneration: {generation}''' F'''\nTotal Population:{total_population}''' F'''\nBest score: {population_score[0][1]}''' F'''\nBest string: {population_score[0][0]}''') # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. lowerCAmelCase_ : str = population[: int(N_POPULATION / 3)] population.clear() population.extend(snake_case__) # Normalize population score to be between 0 and 1. lowerCAmelCase_ : Union[str, Any] = [ (item, score / len(snake_case__)) for item, score in population_score ] # This is selection for i in range(snake_case__): population.extend(select(population_score[int(snake_case__)] , snake_case__ , snake_case__)) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(snake_case__) > N_POPULATION: break if __name__ == "__main__": _lowercase = ( '''This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!''' ) _lowercase = list( ''' ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm''' '''nopqrstuvwxyz.,;!?+-*#@^\'èéòà€ù=)(&%$£/\\''' ) _lowercase , _lowercase , _lowercase = basic(target_str, genes_list) print( f"\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}" )
659
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''microsoft/swinv2-tiny-patch4-window8-256''': ( '''https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json''' ), } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = 'swinv2' UpperCamelCase_ = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : List[Any] ,lowerCAmelCase__ : Optional[int]=2_24 ,lowerCAmelCase__ : Dict=4 ,lowerCAmelCase__ : Dict=3 ,lowerCAmelCase__ : List[Any]=96 ,lowerCAmelCase__ : Optional[Any]=[2, 2, 6, 2] ,lowerCAmelCase__ : Optional[Any]=[3, 6, 12, 24] ,lowerCAmelCase__ : Optional[int]=7 ,lowerCAmelCase__ : Dict=4.0 ,lowerCAmelCase__ : Dict=True ,lowerCAmelCase__ : str=0.0 ,lowerCAmelCase__ : Tuple=0.0 ,lowerCAmelCase__ : str=0.1 ,lowerCAmelCase__ : List[str]="gelu" ,lowerCAmelCase__ : Union[str, Any]=False ,lowerCAmelCase__ : Dict=0.02 ,lowerCAmelCase__ : int=1e-5 ,lowerCAmelCase__ : List[str]=32 ,**lowerCAmelCase__ : Tuple ,) -> List[str]: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = image_size lowerCAmelCase_ : List[Any] = patch_size lowerCAmelCase_ : Dict = num_channels lowerCAmelCase_ : Optional[int] = embed_dim lowerCAmelCase_ : Optional[Any] = depths lowerCAmelCase_ : Any = len(lowerCAmelCase__ ) lowerCAmelCase_ : str = num_heads lowerCAmelCase_ : List[str] = window_size lowerCAmelCase_ : List[str] = mlp_ratio lowerCAmelCase_ : Dict = qkv_bias lowerCAmelCase_ : str = hidden_dropout_prob lowerCAmelCase_ : str = attention_probs_dropout_prob lowerCAmelCase_ : Union[str, Any] = drop_path_rate lowerCAmelCase_ : List[Any] = hidden_act lowerCAmelCase_ : Any = use_absolute_embeddings lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : int = initializer_range lowerCAmelCase_ : Union[str, Any] = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase_ : Tuple = int(embed_dim * 2 ** (len(lowerCAmelCase__ ) - 1) ) lowerCAmelCase_ : str = (0, 0, 0, 0)
659
1
from collections import defaultdict from pathlib import Path import pandas as pd from rouge_cli import calculate_rouge_path from utils import calculate_rouge _lowercase = [ '''Prosecutor: "No videos were used in the crash investigation" German papers say they saw a cell phone video of the''' ''' final seconds on board Flight 9525. The Germanwings co-pilot says he had a "previous episode of severe''' ''' depression\" German airline confirms it knew of Andreas Lubitz\'s depression years before he took control.''', '''The Palestinian Authority officially becomes the 123rd member of the International Criminal Court. The formal''' ''' accession was marked with a ceremony at The Hague, in the Netherlands. The Palestinians signed the ICC\'s''' ''' founding Rome Statute in January. Israel and the United States opposed the Palestinians\' efforts to join the''' ''' body.''', '''Amnesty International releases its annual report on the death penalty. The report catalogs the use of''' ''' state-sanctioned killing as a punitive measure across the globe. At least 607 people were executed around the''' ''' world in 2014, compared to 778 in 2013. The U.S. remains one of the worst offenders for imposing capital''' ''' punishment.''', ] _lowercase = [ '''Marseille prosecutor says "so far no videos were used in the crash investigation" despite media reports .''' ''' Journalists at Bild and Paris Match are "very confident" the video clip is real, an editor says . Andreas Lubitz''' ''' had informed his Lufthansa training school of an episode of severe depression, airline says .''', '''Membership gives the ICC jurisdiction over alleged crimes committed in Palestinian territories since last June .''' ''' Israel and the United States opposed the move, which could open the door to war crimes investigations against''' ''' Israelis .''', '''Amnesty\'s annual death penalty report catalogs encouraging signs, but setbacks in numbers of those sentenced to''' ''' death . Organization claims that governments around the world are using the threat of terrorism to advance''' ''' executions . The number of executions worldwide has gone down by almost 22% compared with 2013, but death''' ''' sentences up by 28% .''', ] def UpperCamelCase ( ): lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , bootstrap_aggregation=snake_case__ , rouge_keys=["rouge2", "rougeL"]) assert isinstance(snake_case__ , snake_case__) lowerCAmelCase_ : str = calculate_rouge(snake_case__ , snake_case__ , bootstrap_aggregation=snake_case__ , rouge_keys=["rouge2"]) assert ( pd.DataFrame(no_aggregation["rouge2"]).fmeasure.mean() == pd.DataFrame(no_aggregation_just_ra["rouge2"]).fmeasure.mean() ) def UpperCamelCase ( ): lowerCAmelCase_ : str = "rougeLsum" lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=[k])[k] lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=[k])[k] assert score > score_no_sep def UpperCamelCase ( ): lowerCAmelCase_ : int = ["rouge1", "rouge2", "rougeL"] lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=snake_case__) lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=snake_case__) assert score_sep == score_no_sep def UpperCamelCase ( ): lowerCAmelCase_ : List[str] = [ "Her older sister, Margot Frank, died in 1945, a month earlier than previously thought.", "Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports .", ] lowerCAmelCase_ : Dict = [ "Margot Frank, died in 1945, a month earlier than previously thought.", "Prosecutor: \"No videos were used in the crash investigation\" German papers say they saw a cell phone video of" " the final seconds on board Flight 9525.", ] assert calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__) == calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__) def UpperCamelCase ( ): lowerCAmelCase_ : Optional[int] = [ "\" \"a person who has such a video needs to immediately give it to the investigators,\" prosecutor says .<n> \"it is a very disturbing scene,\" editor-in-chief of bild online tells \"erin burnett: outfront\" " ] lowerCAmelCase_ : Any = [ " Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports . Journalists at Bild and Paris Match are \"very confident\" the video clip is real, an editor says . Andreas Lubitz had informed his Lufthansa training school of an episode of severe depression, airline says ." ] lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , rouge_keys=["rougeLsum"] , newline_sep=snake_case__)["rougeLsum"] lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , rouge_keys=["rougeLsum"])["rougeLsum"] assert new_score > prev_score def UpperCamelCase ( ): lowerCAmelCase_ : int = Path("examples/seq2seq/test_data/wmt_en_ro") lowerCAmelCase_ : Dict = calculate_rouge_path(data_dir.joinpath("test.source") , data_dir.joinpath("test.target")) assert isinstance(snake_case__ , snake_case__) lowerCAmelCase_ : Any = calculate_rouge_path( data_dir.joinpath("test.source") , data_dir.joinpath("test.target") , bootstrap_aggregation=snake_case__) assert isinstance(snake_case__ , snake_case__)
659
from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging _lowercase = logging.get_logger(__name__) class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = ['input_features', 'attention_mask'] def __init__( self : Optional[Any] ,lowerCAmelCase__ : Any=80 ,lowerCAmelCase__ : Optional[Any]=1_60_00 ,lowerCAmelCase__ : List[str]=0.0 ,lowerCAmelCase__ : Tuple=10 ,lowerCAmelCase__ : Optional[Any]=25 ,lowerCAmelCase__ : Any="hamming_window" ,lowerCAmelCase__ : List[str]=32_768.0 ,lowerCAmelCase__ : Union[str, Any]=0.97 ,lowerCAmelCase__ : Any=1.0 ,lowerCAmelCase__ : str=True ,lowerCAmelCase__ : int=True ,lowerCAmelCase__ : Tuple=False ,**lowerCAmelCase__ : Optional[int] ,) -> Optional[Any]: '''simple docstring''' super().__init__(feature_size=lowerCAmelCase__ ,sampling_rate=lowerCAmelCase__ ,padding_value=lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = feature_size lowerCAmelCase_ : List[Any] = sampling_rate lowerCAmelCase_ : Union[str, Any] = padding_value lowerCAmelCase_ : str = hop_length lowerCAmelCase_ : str = win_length lowerCAmelCase_ : str = frame_signal_scale lowerCAmelCase_ : Any = preemphasis_coeff lowerCAmelCase_ : Optional[Any] = mel_floor lowerCAmelCase_ : List[str] = normalize_means lowerCAmelCase_ : Optional[Any] = normalize_vars lowerCAmelCase_ : Dict = win_function lowerCAmelCase_ : List[Any] = return_attention_mask lowerCAmelCase_ : Tuple = win_length * sampling_rate // 10_00 lowerCAmelCase_ : str = hop_length * sampling_rate // 10_00 lowerCAmelCase_ : Dict = optimal_fft_length(self.sample_size ) lowerCAmelCase_ : Optional[int] = (self.n_fft // 2) + 1 def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : np.array ) -> np.ndarray: '''simple docstring''' if self.win_function == "hamming_window": lowerCAmelCase_ : int = window_function(window_length=self.sample_size ,name=self.win_function ,periodic=lowerCAmelCase__ ) else: lowerCAmelCase_ : Tuple = window_function(window_length=self.sample_size ,name=self.win_function ) lowerCAmelCase_ : List[str] = mel_filter_bank( num_frequency_bins=self.n_freqs ,num_mel_filters=self.feature_size ,min_frequency=0.0 ,max_frequency=self.sampling_rate / 2.0 ,sampling_rate=self.sampling_rate ,) lowerCAmelCase_ : Any = spectrogram( one_waveform * self.frame_signal_scale ,window=lowerCAmelCase__ ,frame_length=self.sample_size ,hop_length=self.sample_stride ,fft_length=self.n_fft ,center=lowerCAmelCase__ ,preemphasis=self.preemphasis_coeff ,mel_filters=lowerCAmelCase__ ,mel_floor=self.mel_floor ,log_mel="log" ,) return msfc_features.T def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : Tuple ) -> Optional[Any]: '''simple docstring''' if self.normalize_means: lowerCAmelCase_ : Optional[int] = x[:input_length].mean(axis=0 ) lowerCAmelCase_ : List[str] = np.subtract(lowerCAmelCase__ ,lowerCAmelCase__ ) if self.normalize_vars: lowerCAmelCase_ : Optional[Any] = x[:input_length].std(axis=0 ) lowerCAmelCase_ : Tuple = np.divide(lowerCAmelCase__ ,lowerCAmelCase__ ) if input_length < x.shape[0]: lowerCAmelCase_ : int = padding_value # make sure array is in float32 lowerCAmelCase_ : Any = x.astype(np.floataa ) return x def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[np.ndarray] ,lowerCAmelCase__ : Optional[np.ndarray] = None ) -> List[np.ndarray]: '''simple docstring''' lowerCAmelCase_ : List[Any] = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(lowerCAmelCase__ ,lowerCAmelCase__ ,self.padding_value ) for x, n in zip(lowerCAmelCase__ ,lowerCAmelCase__ )] def __call__( self : int ,lowerCAmelCase__ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,lowerCAmelCase__ : Union[bool, str, PaddingStrategy] = False ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : bool = False ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : Optional[bool] = None ,lowerCAmelCase__ : Optional[Union[str, TensorType]] = None ,lowerCAmelCase__ : Optional[int] = None ,**lowerCAmelCase__ : Union[str, Any] ,) -> BatchFeature: '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) lowerCAmelCase_ : List[Any] = isinstance(lowerCAmelCase__ ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase_ : str = is_batched_numpy or ( isinstance(lowerCAmelCase__ ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase_ : Tuple = [np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(lowerCAmelCase__ ,np.ndarray ): lowerCAmelCase_ : int = np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) elif isinstance(lowerCAmelCase__ ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase_ : Union[str, Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase_ : Optional[int] = [raw_speech] # extract fbank features lowerCAmelCase_ : Dict = [self._extract_mfsc_features(lowerCAmelCase__ ) for one_waveform in raw_speech] # convert into correct format for padding lowerCAmelCase_ : int = BatchFeature({"input_features": features} ) lowerCAmelCase_ : Union[str, Any] = self.pad( lowerCAmelCase__ ,padding=lowerCAmelCase__ ,max_length=lowerCAmelCase__ ,truncation=lowerCAmelCase__ ,pad_to_multiple_of=lowerCAmelCase__ ,return_attention_mask=lowerCAmelCase__ ,**lowerCAmelCase__ ,) # make sure list is in array format lowerCAmelCase_ : Optional[Any] = padded_inputs.get("input_features" ) if isinstance(input_features[0] ,lowerCAmelCase__ ): lowerCAmelCase_ : Optional[int] = [np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) for feature in input_features] lowerCAmelCase_ : List[Any] = padded_inputs.get("attention_mask" ) if attention_mask is not None: lowerCAmelCase_ : Dict = [np.asarray(lowerCAmelCase__ ,dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: lowerCAmelCase_ : Dict = ( np.array(lowerCAmelCase__ ,dtype=np.intaa ) if self._get_padding_strategies(lowerCAmelCase__ ,max_length=lowerCAmelCase__ ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) lowerCAmelCase_ : List[str] = self.normalize( padded_inputs["input_features"] ,attention_mask=lowerCAmelCase__ ) if return_tensors is not None: lowerCAmelCase_ : Dict = padded_inputs.convert_to_tensors(lowerCAmelCase__ ) return padded_inputs
659
1
def UpperCamelCase ( snake_case__): lowerCAmelCase_ : List[str] = 1 lowerCAmelCase_ : str = 2 while i * i <= n: lowerCAmelCase_ : str = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def UpperCamelCase ( ): lowerCAmelCase_ : int = 1 lowerCAmelCase_ : int = 1 while True: i += 1 t_num += i if count_divisors(snake_case__) > 5_00: break return t_num if __name__ == "__main__": print(solution())
659
from __future__ import annotations # This is the precision for this function which can be altered. # It is recommended for users to keep this number greater than or equal to 10. _lowercase = 10 def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): for i in range(snake_case__ , snake_case__): if array[i] == target: return i return -1 def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : List[str] = 0 lowerCAmelCase_ : Tuple = len(snake_case__) while left <= right: if right - left < precision: return lin_search(snake_case__ , snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : List[str] = (left + right) // 3 + 1 lowerCAmelCase_ : Tuple = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: lowerCAmelCase_ : str = one_third - 1 elif array[two_third] < target: lowerCAmelCase_ : Any = two_third + 1 else: lowerCAmelCase_ : List[str] = one_third + 1 lowerCAmelCase_ : Tuple = two_third - 1 else: return -1 def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): if left < right: if right - left < precision: return lin_search(snake_case__ , snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : Dict = (left + right) // 3 + 1 lowerCAmelCase_ : List[Any] = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: return rec_ternary_search(snake_case__ , one_third - 1 , snake_case__ , snake_case__) elif array[two_third] < target: return rec_ternary_search(two_third + 1 , snake_case__ , snake_case__ , snake_case__) else: return rec_ternary_search(one_third + 1 , two_third - 1 , snake_case__ , snake_case__) else: return -1 if __name__ == "__main__": import doctest doctest.testmod() _lowercase = input('''Enter numbers separated by comma:\n''').strip() _lowercase = [int(item.strip()) for item in user_input.split(''',''')] assert collection == sorted(collection), f"List must be ordered.\n{collection}." _lowercase = int(input('''Enter the number to be found in the list:\n''').strip()) _lowercase = ite_ternary_search(collection, target) _lowercase = rec_ternary_search(0, len(collection) - 1, collection, target) if resulta != -1: print(f"Iterative search: {target} found at positions: {resulta}") print(f"Recursive search: {target} found at positions: {resulta}") else: print('''Not found''')
659
1
_lowercase = [ 999, 800, 799, 600, 599, 500, 400, 399, 377, 355, 333, 311, 288, 266, 244, 222, 200, 199, 177, 155, 133, 111, 88, 66, 44, 22, 0, ] _lowercase = [ 999, 976, 952, 928, 905, 882, 858, 857, 810, 762, 715, 714, 572, 429, 428, 286, 285, 238, 190, 143, 142, 118, 95, 71, 47, 24, 0, ] _lowercase = [ 999, 988, 977, 966, 955, 944, 933, 922, 911, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 350, 300, 299, 266, 233, 200, 199, 179, 159, 140, 120, 100, 99, 88, 77, 66, 55, 44, 33, 22, 11, 0, ] _lowercase = [ 999, 995, 992, 989, 985, 981, 978, 975, 971, 967, 964, 961, 957, 956, 951, 947, 942, 937, 933, 928, 923, 919, 914, 913, 908, 903, 897, 892, 887, 881, 876, 871, 870, 864, 858, 852, 846, 840, 834, 828, 827, 820, 813, 806, 799, 792, 785, 784, 777, 770, 763, 756, 749, 742, 741, 733, 724, 716, 707, 699, 698, 688, 677, 666, 656, 655, 645, 634, 623, 613, 612, 598, 584, 570, 569, 555, 541, 527, 526, 505, 484, 483, 462, 440, 439, 396, 395, 352, 351, 308, 307, 264, 263, 220, 219, 176, 132, 88, 44, 0, ] _lowercase = [ 999, 997, 995, 992, 990, 988, 986, 984, 981, 979, 977, 975, 972, 970, 968, 966, 964, 961, 959, 957, 956, 954, 951, 949, 946, 944, 941, 939, 936, 934, 931, 929, 926, 924, 921, 919, 916, 914, 913, 910, 907, 905, 902, 899, 896, 893, 891, 888, 885, 882, 879, 877, 874, 871, 870, 867, 864, 861, 858, 855, 852, 849, 846, 843, 840, 837, 834, 831, 828, 827, 824, 821, 817, 814, 811, 808, 804, 801, 798, 795, 791, 788, 785, 784, 780, 777, 774, 770, 766, 763, 760, 756, 752, 749, 746, 742, 741, 737, 733, 730, 726, 722, 718, 714, 710, 707, 703, 699, 698, 694, 690, 685, 681, 677, 673, 669, 664, 660, 656, 655, 650, 646, 641, 636, 632, 627, 622, 618, 613, 612, 607, 602, 596, 591, 586, 580, 575, 570, 569, 563, 557, 551, 545, 539, 533, 527, 526, 519, 512, 505, 498, 491, 484, 483, 474, 466, 457, 449, 440, 439, 428, 418, 407, 396, 395, 381, 366, 352, 351, 330, 308, 307, 286, 264, 263, 242, 220, 219, 176, 175, 132, 131, 88, 44, 0, ] _lowercase = [ 999, 991, 982, 974, 966, 958, 950, 941, 933, 925, 916, 908, 900, 899, 874, 850, 825, 800, 799, 700, 600, 500, 400, 300, 200, 100, 0, ] _lowercase = [ 999, 992, 985, 978, 971, 964, 957, 949, 942, 935, 928, 921, 914, 907, 900, 899, 879, 859, 840, 820, 800, 799, 766, 733, 700, 699, 650, 600, 599, 500, 499, 400, 399, 300, 299, 200, 199, 100, 99, 0, ] _lowercase = [ 999, 996, 992, 989, 985, 982, 979, 975, 972, 968, 965, 961, 958, 955, 951, 948, 944, 941, 938, 934, 931, 927, 924, 920, 917, 914, 910, 907, 903, 900, 899, 891, 884, 876, 869, 861, 853, 846, 838, 830, 823, 815, 808, 800, 799, 788, 777, 766, 755, 744, 733, 722, 711, 700, 699, 688, 677, 666, 655, 644, 633, 622, 611, 600, 599, 585, 571, 557, 542, 528, 514, 500, 499, 485, 471, 457, 442, 428, 414, 400, 399, 379, 359, 340, 320, 300, 299, 279, 259, 240, 220, 200, 199, 166, 133, 100, 99, 66, 33, 0, ]
659
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer _lowercase = logging.get_logger(__name__) _lowercase = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } _lowercase = { '''vocab_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json''' }, '''merges_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt''' }, '''tokenizer_config_file''': { '''facebook/blenderbot_small-90M''': ( '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json''' ) }, } _lowercase = { '''facebook/blenderbot_small-90M''': 512, } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = BlenderbotSmallTokenizer def __init__( self : Optional[int] ,lowerCAmelCase__ : Optional[int]=None ,lowerCAmelCase__ : Union[str, Any]=None ,lowerCAmelCase__ : Any="<|endoftext|>" ,lowerCAmelCase__ : int="<|endoftext|>" ,lowerCAmelCase__ : Optional[Any]="<|endoftext|>" ,lowerCAmelCase__ : Union[str, Any]=False ,lowerCAmelCase__ : Optional[Any]=True ,**lowerCAmelCase__ : Union[str, Any] ,) -> str: '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=lowerCAmelCase__ ,merges=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ,) ,bos_token=lowerCAmelCase__ ,eos_token=lowerCAmelCase__ ,unk_token=lowerCAmelCase__ ,**lowerCAmelCase__ ,) lowerCAmelCase_ : Dict = add_prefix_space def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Tuple=None ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : str = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : Dict = [self.sep_token_id] lowerCAmelCase_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
659
1
import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor _lowercase = logging.get_logger(__name__) class __snake_case ( snake_case__ ): """simple docstring""" def __init__( self : Tuple ,*lowerCAmelCase__ : str ,**lowerCAmelCase__ : Tuple ) -> None: '''simple docstring''' warnings.warn( "The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use DPTImageProcessor instead." ,lowerCAmelCase__ ,) super().__init__(*lowerCAmelCase__ ,**lowerCAmelCase__ )
659
from collections.abc import Generator from math import sin def UpperCamelCase ( snake_case__): if len(snake_case__) != 32: raise ValueError("Input must be of length 32") lowerCAmelCase_ : Tuple = b"" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def UpperCamelCase ( snake_case__): if i < 0: raise ValueError("Input must be non-negative") lowerCAmelCase_ : List[str] = format(snake_case__ , "08x")[-8:] lowerCAmelCase_ : Any = b"" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("utf-8") return little_endian_hex def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Union[str, Any] = b"" for char in message: bit_string += format(snake_case__ , "08b").encode("utf-8") lowerCAmelCase_ : Optional[int] = format(len(snake_case__) , "064b").encode("utf-8") # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(snake_case__) % 5_12 != 4_48: bit_string += b"0" bit_string += to_little_endian(start_len[32:]) + to_little_endian(start_len[:32]) return bit_string def UpperCamelCase ( snake_case__): if len(snake_case__) % 5_12 != 0: raise ValueError("Input must have length that's a multiple of 512") for pos in range(0 , len(snake_case__) , 5_12): lowerCAmelCase_ : List[str] = bit_string[pos : pos + 5_12] lowerCAmelCase_ : Union[str, Any] = [] for i in range(0 , 5_12 , 32): block_words.append(int(to_little_endian(block[i : i + 32]) , 2)) yield block_words def UpperCamelCase ( snake_case__): if i < 0: raise ValueError("Input must be non-negative") lowerCAmelCase_ : Dict = format(snake_case__ , "032b") lowerCAmelCase_ : str = "" for c in i_str: new_str += "1" if c == "0" else "0" return int(snake_case__ , 2) def UpperCamelCase ( snake_case__ , snake_case__): return (a + b) % 2**32 def UpperCamelCase ( snake_case__ , snake_case__): if i < 0: raise ValueError("Input must be non-negative") if shift < 0: raise ValueError("Shift must be non-negative") return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Optional[Any] = preprocess(snake_case__) lowerCAmelCase_ : Optional[Any] = [int(2**32 * abs(sin(i + 1))) for i in range(64)] # Starting states lowerCAmelCase_ : List[str] = 0x67_45_23_01 lowerCAmelCase_ : Union[str, Any] = 0xef_cd_ab_89 lowerCAmelCase_ : List[Any] = 0x98_ba_dc_fe lowerCAmelCase_ : Tuple = 0x10_32_54_76 lowerCAmelCase_ : Any = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(snake_case__): lowerCAmelCase_ : Optional[int] = aa lowerCAmelCase_ : List[str] = ba lowerCAmelCase_ : Any = ca lowerCAmelCase_ : Union[str, Any] = da # Hash current chunk for i in range(64): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f lowerCAmelCase_ : Any = d ^ (b & (c ^ d)) lowerCAmelCase_ : Dict = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f lowerCAmelCase_ : Any = c ^ (d & (b ^ c)) lowerCAmelCase_ : List[str] = (5 * i + 1) % 16 elif i <= 47: lowerCAmelCase_ : int = b ^ c ^ d lowerCAmelCase_ : Optional[Any] = (3 * i + 5) % 16 else: lowerCAmelCase_ : List[Any] = c ^ (b | not_aa(snake_case__)) lowerCAmelCase_ : List[Any] = (7 * i) % 16 lowerCAmelCase_ : Optional[Any] = (f + a + added_consts[i] + block_words[g]) % 2**32 lowerCAmelCase_ : Optional[Any] = d lowerCAmelCase_ : Dict = c lowerCAmelCase_ : List[str] = b lowerCAmelCase_ : Any = sum_aa(snake_case__ , left_rotate_aa(snake_case__ , shift_amounts[i])) # Add hashed chunk to running total lowerCAmelCase_ : Dict = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : str = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : Optional[int] = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : int = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : Union[str, Any] = reformat_hex(snake_case__) + reformat_hex(snake_case__) + reformat_hex(snake_case__) + reformat_hex(snake_case__) return digest if __name__ == "__main__": import doctest doctest.testmod()
659
1
from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging _lowercase = logging.get_logger(__name__) class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = ['input_features', 'attention_mask'] def __init__( self : Optional[Any] ,lowerCAmelCase__ : Any=80 ,lowerCAmelCase__ : Optional[Any]=1_60_00 ,lowerCAmelCase__ : List[str]=0.0 ,lowerCAmelCase__ : Tuple=10 ,lowerCAmelCase__ : Optional[Any]=25 ,lowerCAmelCase__ : Any="hamming_window" ,lowerCAmelCase__ : List[str]=32_768.0 ,lowerCAmelCase__ : Union[str, Any]=0.97 ,lowerCAmelCase__ : Any=1.0 ,lowerCAmelCase__ : str=True ,lowerCAmelCase__ : int=True ,lowerCAmelCase__ : Tuple=False ,**lowerCAmelCase__ : Optional[int] ,) -> Optional[Any]: '''simple docstring''' super().__init__(feature_size=lowerCAmelCase__ ,sampling_rate=lowerCAmelCase__ ,padding_value=lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = feature_size lowerCAmelCase_ : List[Any] = sampling_rate lowerCAmelCase_ : Union[str, Any] = padding_value lowerCAmelCase_ : str = hop_length lowerCAmelCase_ : str = win_length lowerCAmelCase_ : str = frame_signal_scale lowerCAmelCase_ : Any = preemphasis_coeff lowerCAmelCase_ : Optional[Any] = mel_floor lowerCAmelCase_ : List[str] = normalize_means lowerCAmelCase_ : Optional[Any] = normalize_vars lowerCAmelCase_ : Dict = win_function lowerCAmelCase_ : List[Any] = return_attention_mask lowerCAmelCase_ : Tuple = win_length * sampling_rate // 10_00 lowerCAmelCase_ : str = hop_length * sampling_rate // 10_00 lowerCAmelCase_ : Dict = optimal_fft_length(self.sample_size ) lowerCAmelCase_ : Optional[int] = (self.n_fft // 2) + 1 def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : np.array ) -> np.ndarray: '''simple docstring''' if self.win_function == "hamming_window": lowerCAmelCase_ : int = window_function(window_length=self.sample_size ,name=self.win_function ,periodic=lowerCAmelCase__ ) else: lowerCAmelCase_ : Tuple = window_function(window_length=self.sample_size ,name=self.win_function ) lowerCAmelCase_ : List[str] = mel_filter_bank( num_frequency_bins=self.n_freqs ,num_mel_filters=self.feature_size ,min_frequency=0.0 ,max_frequency=self.sampling_rate / 2.0 ,sampling_rate=self.sampling_rate ,) lowerCAmelCase_ : Any = spectrogram( one_waveform * self.frame_signal_scale ,window=lowerCAmelCase__ ,frame_length=self.sample_size ,hop_length=self.sample_stride ,fft_length=self.n_fft ,center=lowerCAmelCase__ ,preemphasis=self.preemphasis_coeff ,mel_filters=lowerCAmelCase__ ,mel_floor=self.mel_floor ,log_mel="log" ,) return msfc_features.T def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : Tuple ) -> Optional[Any]: '''simple docstring''' if self.normalize_means: lowerCAmelCase_ : Optional[int] = x[:input_length].mean(axis=0 ) lowerCAmelCase_ : List[str] = np.subtract(lowerCAmelCase__ ,lowerCAmelCase__ ) if self.normalize_vars: lowerCAmelCase_ : Optional[Any] = x[:input_length].std(axis=0 ) lowerCAmelCase_ : Tuple = np.divide(lowerCAmelCase__ ,lowerCAmelCase__ ) if input_length < x.shape[0]: lowerCAmelCase_ : int = padding_value # make sure array is in float32 lowerCAmelCase_ : Any = x.astype(np.floataa ) return x def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[np.ndarray] ,lowerCAmelCase__ : Optional[np.ndarray] = None ) -> List[np.ndarray]: '''simple docstring''' lowerCAmelCase_ : List[Any] = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(lowerCAmelCase__ ,lowerCAmelCase__ ,self.padding_value ) for x, n in zip(lowerCAmelCase__ ,lowerCAmelCase__ )] def __call__( self : int ,lowerCAmelCase__ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,lowerCAmelCase__ : Union[bool, str, PaddingStrategy] = False ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : bool = False ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : Optional[bool] = None ,lowerCAmelCase__ : Optional[Union[str, TensorType]] = None ,lowerCAmelCase__ : Optional[int] = None ,**lowerCAmelCase__ : Union[str, Any] ,) -> BatchFeature: '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) lowerCAmelCase_ : List[Any] = isinstance(lowerCAmelCase__ ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase_ : str = is_batched_numpy or ( isinstance(lowerCAmelCase__ ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase_ : Tuple = [np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(lowerCAmelCase__ ,np.ndarray ): lowerCAmelCase_ : int = np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) elif isinstance(lowerCAmelCase__ ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase_ : Union[str, Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase_ : Optional[int] = [raw_speech] # extract fbank features lowerCAmelCase_ : Dict = [self._extract_mfsc_features(lowerCAmelCase__ ) for one_waveform in raw_speech] # convert into correct format for padding lowerCAmelCase_ : int = BatchFeature({"input_features": features} ) lowerCAmelCase_ : Union[str, Any] = self.pad( lowerCAmelCase__ ,padding=lowerCAmelCase__ ,max_length=lowerCAmelCase__ ,truncation=lowerCAmelCase__ ,pad_to_multiple_of=lowerCAmelCase__ ,return_attention_mask=lowerCAmelCase__ ,**lowerCAmelCase__ ,) # make sure list is in array format lowerCAmelCase_ : Optional[Any] = padded_inputs.get("input_features" ) if isinstance(input_features[0] ,lowerCAmelCase__ ): lowerCAmelCase_ : Optional[int] = [np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) for feature in input_features] lowerCAmelCase_ : List[Any] = padded_inputs.get("attention_mask" ) if attention_mask is not None: lowerCAmelCase_ : Dict = [np.asarray(lowerCAmelCase__ ,dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: lowerCAmelCase_ : Dict = ( np.array(lowerCAmelCase__ ,dtype=np.intaa ) if self._get_padding_strategies(lowerCAmelCase__ ,max_length=lowerCAmelCase__ ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) lowerCAmelCase_ : List[str] = self.normalize( padded_inputs["input_features"] ,attention_mask=lowerCAmelCase__ ) if return_tensors is not None: lowerCAmelCase_ : Dict = padded_inputs.convert_to_tensors(lowerCAmelCase__ ) return padded_inputs
659
import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): _lowercase = True from torch.cuda.amp import autocast _lowercase = logging.getLogger(__name__) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Whether to log verbose messages or not.'} , ) UpperCamelCase_ = field( default=2.0 , metadata={'help': 'Maximum temperature for gumbel softmax.'} ) UpperCamelCase_ = field( default=0.5 , metadata={'help': 'Minimum temperature for gumbel softmax.'} ) UpperCamelCase_ = field( default=0.99_99_95 , metadata={'help': 'Decay of gumbel temperature during training.'} ) def UpperCamelCase ( snake_case__ , snake_case__): logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout)] , ) lowerCAmelCase_ : str = logging.WARNING if model_args.verbose_logging: lowerCAmelCase_ : int = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank): lowerCAmelCase_ : Any = logging.INFO logger.setLevel(snake_case__) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The name of the dataset to use (via the datasets library).'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) UpperCamelCase_ = field( default='train' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) UpperCamelCase_ = field( default='validation' , metadata={ 'help': ( 'The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'' ) } , ) UpperCamelCase_ = field( default='file' , metadata={'help': 'Column in the dataset that contains speech file path. Defaults to \'file\''} , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Overwrite the cached preprocessed datasets or not.'} ) UpperCamelCase_ = field( default=1 , metadata={ 'help': 'The percentage of the train set used as validation set in case there\'s no validation split' } , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) UpperCamelCase_ = field( default=20.0 , metadata={'help': 'Filter audio files that are longer than `max_duration_in_seconds` seconds'} ) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = 42 UpperCamelCase_ = 42 UpperCamelCase_ = "longest" UpperCamelCase_ = None UpperCamelCase_ = None def __call__( self : str ,lowerCAmelCase__ : List[Dict[str, Union[List[int], torch.Tensor]]] ) -> Dict[str, torch.Tensor]: '''simple docstring''' lowerCAmelCase_ : Tuple = self.feature_extractor.pad( lowerCAmelCase__ ,max_length=self.max_length ,padding=self.padding ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors="pt" ,) lowerCAmelCase_ : Union[str, Any] = self.model._get_feat_extract_output_lengths(batch["input_values"].shape[-1] ) lowerCAmelCase_ : List[str] = batch["input_values"].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula lowerCAmelCase_ : Tuple = self.model._get_feat_extract_output_lengths(batch["attention_mask"].sum(-1 ) ).to( torch.long ) lowerCAmelCase_ : Optional[Any] = torch.zeros( (batch_size, mask_indices_seq_length) ,dtype=torch.long ,device=batch["input_values"].device ) # these two operations makes sure that all values # before the output lengths indices are attended to lowerCAmelCase_ : Tuple = 1 lowerCAmelCase_ : int = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices lowerCAmelCase_ : str = _compute_mask_indices( (batch_size, mask_indices_seq_length) ,self.model.config.mask_time_prob ,self.model.config.mask_time_length ,attention_mask=lowerCAmelCase__ ,min_masks=2 ,) return batch class __snake_case ( snake_case__ ): """simple docstring""" def __init__( self : List[str] ,*lowerCAmelCase__ : Optional[int] ,lowerCAmelCase__ : Tuple=1 ,lowerCAmelCase__ : Optional[int]=0 ,lowerCAmelCase__ : Optional[Any]=1.0 ,**lowerCAmelCase__ : Any ) -> str: '''simple docstring''' super().__init__(*lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = 0 lowerCAmelCase_ : int = max_gumbel_temp lowerCAmelCase_ : Union[str, Any] = min_gumbel_temp lowerCAmelCase_ : str = gumbel_temp_decay def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : nn.Module ,lowerCAmelCase__ : Dict[str, Union[torch.Tensor, Any]] ) -> torch.Tensor: '''simple docstring''' model.train() lowerCAmelCase_ : str = self._prepare_inputs(lowerCAmelCase__ ) if self.use_amp: with autocast(): lowerCAmelCase_ : List[Any] = self.compute_loss(lowerCAmelCase__ ,lowerCAmelCase__ ) else: lowerCAmelCase_ : List[Any] = self.compute_loss(lowerCAmelCase__ ,lowerCAmelCase__ ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": lowerCAmelCase_ : List[Any] = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": lowerCAmelCase_ : Optional[Any] = loss.sum() / (inputs["mask_time_indices"]).sum() else: raise ValueError(f'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: lowerCAmelCase_ : int = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(lowerCAmelCase__ ).backward() elif self.use_apex: with amp.scale_loss(lowerCAmelCase__ ,self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(lowerCAmelCase__ ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) return loss.detach() def UpperCamelCase ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase_ : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments)) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = parser.parse_args_into_dataclasses() configure_logger(snake_case__ , snake_case__) # Downloading and loading a dataset from the hub. lowerCAmelCase_ : List[str] = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" lowerCAmelCase_ : Any = DatasetDict() lowerCAmelCase_ : Union[str, Any] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : List[str] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" lowerCAmelCase_ : Union[str, Any] = DatasetDict() lowerCAmelCase_ : int = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split="validation" , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : Any = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported lowerCAmelCase_ : Dict = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=snake_case__) def prepare_dataset(snake_case__): # check that all files have the correct sampling rate lowerCAmelCase_ , lowerCAmelCase_ : str = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate) return batch # load audio files into numpy arrays lowerCAmelCase_ : int = datasets.map( snake_case__ , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets["train"].column_names) # filter audio files that are too long lowerCAmelCase_ : int = vectorized_datasets.filter( lambda snake_case__: len(data["speech"]) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate)) def normalize(snake_case__): return feature_extractor(batch["speech"] , sampling_rate=feature_extractor.sampling_rate) # normalize and transform to `BatchFeatures` lowerCAmelCase_ : str = vectorized_datasets.map( snake_case__ , batched=snake_case__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets["train"].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 lowerCAmelCase_ : Optional[Any] = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( "PreTraining is only supported for ``config.do_stable_layer_norm=True`` and" " ``config.feat_extract_norm='layer'") lowerCAmelCase_ : Dict = WavaVecaForPreTraining(snake_case__) lowerCAmelCase_ : int = DataCollatorForWavaVecaPretraining(model=snake_case__ , feature_extractor=snake_case__) lowerCAmelCase_ : List[Any] = WavaVecaPreTrainer( model=snake_case__ , data_collator=snake_case__ , args=snake_case__ , train_dataset=vectorized_datasets["train"] , eval_dataset=vectorized_datasets["validation"] , tokenizer=snake_case__ , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
659
1
from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class __snake_case ( unittest.TestCase ): """simple docstring""" @slow def UpperCAmelCase_ ( self : Dict ) -> int: '''simple docstring''' for model_name in ["bert-base-uncased"]: lowerCAmelCase_ : List[Any] = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Any = TFAutoModel.from_pretrained(lowerCAmelCase__ ,from_pt=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = AutoModel.from_pretrained(lowerCAmelCase__ ,from_tf=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) @slow def UpperCAmelCase_ ( self : Tuple ) -> int: '''simple docstring''' for model_name in ["bert-base-uncased"]: lowerCAmelCase_ : int = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = TFAutoModelForPreTraining.from_pretrained(lowerCAmelCase__ ,from_pt=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = AutoModelForPreTraining.from_pretrained(lowerCAmelCase__ ,from_tf=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) @slow def UpperCAmelCase_ ( self : Optional[int] ) -> Dict: '''simple docstring''' for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Tuple = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = TFAutoModelForCausalLM.from_pretrained(lowerCAmelCase__ ,from_pt=lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Dict = TFAutoModelForCausalLM.from_pretrained( lowerCAmelCase__ ,output_loading_info=lowerCAmelCase__ ,from_pt=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Any = AutoModelForCausalLM.from_pretrained(lowerCAmelCase__ ,from_tf=lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : str = AutoModelForCausalLM.from_pretrained( lowerCAmelCase__ ,output_loading_info=lowerCAmelCase__ ,from_tf=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) @slow def UpperCAmelCase_ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Optional[int] = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = TFAutoModelWithLMHead.from_pretrained(lowerCAmelCase__ ,from_pt=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : int = AutoModelWithLMHead.from_pretrained(lowerCAmelCase__ ,from_tf=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) @slow def UpperCAmelCase_ ( self : Any ) -> Optional[Any]: '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : List[str] = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = TFAutoModelForMaskedLM.from_pretrained(lowerCAmelCase__ ,from_pt=lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : str = TFAutoModelForMaskedLM.from_pretrained( lowerCAmelCase__ ,output_loading_info=lowerCAmelCase__ ,from_pt=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Any = AutoModelForMaskedLM.from_pretrained(lowerCAmelCase__ ,from_tf=lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Tuple = AutoModelForMaskedLM.from_pretrained( lowerCAmelCase__ ,output_loading_info=lowerCAmelCase__ ,from_tf=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) @slow def UpperCAmelCase_ ( self : List[str] ) -> Tuple: '''simple docstring''' for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Optional[Any] = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = TFAutoModelForSeqaSeqLM.from_pretrained(lowerCAmelCase__ ,from_pt=lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : str = TFAutoModelForSeqaSeqLM.from_pretrained( lowerCAmelCase__ ,output_loading_info=lowerCAmelCase__ ,from_pt=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = AutoModelForSeqaSeqLM.from_pretrained(lowerCAmelCase__ ,from_tf=lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Tuple = AutoModelForSeqaSeqLM.from_pretrained( lowerCAmelCase__ ,output_loading_info=lowerCAmelCase__ ,from_tf=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) @slow def UpperCAmelCase_ ( self : Dict ) -> Tuple: '''simple docstring''' for model_name in ["bert-base-uncased"]: lowerCAmelCase_ : Any = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = TFAutoModelForSequenceClassification.from_pretrained(lowerCAmelCase__ ,from_pt=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = AutoModelForSequenceClassification.from_pretrained(lowerCAmelCase__ ,from_tf=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) @slow def UpperCAmelCase_ ( self : Tuple ) -> Optional[int]: '''simple docstring''' for model_name in ["bert-base-uncased"]: lowerCAmelCase_ : Dict = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = TFAutoModelForQuestionAnswering.from_pretrained(lowerCAmelCase__ ,from_pt=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : int = AutoModelForQuestionAnswering.from_pretrained(lowerCAmelCase__ ,from_tf=lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : str ) -> Dict: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = TFAutoModelWithLMHead.from_pretrained(lowerCAmelCase__ ,from_pt=lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual(model.num_parameters() ,1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=lowerCAmelCase__ ) ,1_44_10 ) lowerCAmelCase_ : int = AutoModelWithLMHead.from_pretrained(lowerCAmelCase__ ,from_tf=lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual(model.num_parameters() ,1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=lowerCAmelCase__ ) ,1_44_10 ) def UpperCAmelCase_ ( self : Optional[int] ) -> int: '''simple docstring''' lowerCAmelCase_ : int = TFAutoModelWithLMHead.from_pretrained(lowerCAmelCase__ ,from_pt=lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual(model.num_parameters() ,1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=lowerCAmelCase__ ) ,1_44_10 ) lowerCAmelCase_ : Tuple = AutoModelWithLMHead.from_pretrained(lowerCAmelCase__ ,from_tf=lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual(model.num_parameters() ,1_44_10 ) self.assertEqual(model.num_parameters(only_trainable=lowerCAmelCase__ ) ,1_44_10 )
659
from __future__ import annotations from collections.abc import Callable def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1_00 , ): lowerCAmelCase_ : Any = x_start lowerCAmelCase_ : Optional[Any] = fnc(snake_case__) lowerCAmelCase_ : Union[str, Any] = 0.0 for _ in range(snake_case__): # Approximates small segments of curve as linear and solve # for trapezoidal area lowerCAmelCase_ : Any = (x_end - x_start) / steps + xa lowerCAmelCase_ : Dict = fnc(snake_case__) area += abs(fxa + fxa) * (xa - xa) / 2 # Increment step lowerCAmelCase_ : int = xa lowerCAmelCase_ : str = fxa return area if __name__ == "__main__": def UpperCamelCase ( snake_case__): return x**3 + x**2 print('''f(x) = x^3 + x^2''') print('''The area between the curve, x = -5, x = 5 and the x axis is:''') _lowercase = 10 while i <= 100000: print(f"with {i} steps: {trapezoidal_area(f, -5, 5, i)}") i *= 10
659
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_funnel import FunnelTokenizer _lowercase = logging.get_logger(__name__) _lowercase = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} _lowercase = [ '''small''', '''small-base''', '''medium''', '''medium-base''', '''intermediate''', '''intermediate-base''', '''large''', '''large-base''', '''xlarge''', '''xlarge-base''', ] _lowercase = { '''vocab_file''': { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/vocab.txt''', '''funnel-transformer/small-base''': '''https://huggingface.co/funnel-transformer/small-base/resolve/main/vocab.txt''', '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/vocab.txt''', '''funnel-transformer/medium-base''': ( '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/vocab.txt''' ), '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/vocab.txt''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/vocab.txt''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/vocab.txt''', '''funnel-transformer/large-base''': '''https://huggingface.co/funnel-transformer/large-base/resolve/main/vocab.txt''', '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/vocab.txt''', '''funnel-transformer/xlarge-base''': ( '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''funnel-transformer/small''': '''https://huggingface.co/funnel-transformer/small/resolve/main/tokenizer.json''', '''funnel-transformer/small-base''': ( '''https://huggingface.co/funnel-transformer/small-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/medium''': '''https://huggingface.co/funnel-transformer/medium/resolve/main/tokenizer.json''', '''funnel-transformer/medium-base''': ( '''https://huggingface.co/funnel-transformer/medium-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/intermediate''': ( '''https://huggingface.co/funnel-transformer/intermediate/resolve/main/tokenizer.json''' ), '''funnel-transformer/intermediate-base''': ( '''https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/large''': '''https://huggingface.co/funnel-transformer/large/resolve/main/tokenizer.json''', '''funnel-transformer/large-base''': ( '''https://huggingface.co/funnel-transformer/large-base/resolve/main/tokenizer.json''' ), '''funnel-transformer/xlarge''': '''https://huggingface.co/funnel-transformer/xlarge/resolve/main/tokenizer.json''', '''funnel-transformer/xlarge-base''': ( '''https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/tokenizer.json''' ), }, } _lowercase = {f"funnel-transformer/{name}": 512 for name in _model_names} _lowercase = {f"funnel-transformer/{name}": {'''do_lower_case''': True} for name in _model_names} class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_INIT_CONFIGURATION UpperCamelCase_ = FunnelTokenizer UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = 2 def __init__( self : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any]=None ,lowerCAmelCase__ : List[str]=None ,lowerCAmelCase__ : Optional[Any]=True ,lowerCAmelCase__ : int="<unk>" ,lowerCAmelCase__ : str="<sep>" ,lowerCAmelCase__ : str="<pad>" ,lowerCAmelCase__ : Optional[Any]="<cls>" ,lowerCAmelCase__ : Dict="<mask>" ,lowerCAmelCase__ : List[str]="<s>" ,lowerCAmelCase__ : Optional[Any]="</s>" ,lowerCAmelCase__ : Dict=True ,lowerCAmelCase__ : Tuple=True ,lowerCAmelCase__ : List[str]=None ,lowerCAmelCase__ : Optional[int]="##" ,**lowerCAmelCase__ : Optional[Any] ,) -> Tuple: '''simple docstring''' super().__init__( lowerCAmelCase__ ,tokenizer_file=lowerCAmelCase__ ,do_lower_case=lowerCAmelCase__ ,unk_token=lowerCAmelCase__ ,sep_token=lowerCAmelCase__ ,pad_token=lowerCAmelCase__ ,cls_token=lowerCAmelCase__ ,mask_token=lowerCAmelCase__ ,bos_token=lowerCAmelCase__ ,eos_token=lowerCAmelCase__ ,clean_text=lowerCAmelCase__ ,tokenize_chinese_chars=lowerCAmelCase__ ,strip_accents=lowerCAmelCase__ ,wordpieces_prefix=lowerCAmelCase__ ,**lowerCAmelCase__ ,) lowerCAmelCase_ : int = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" ,lowerCAmelCase__ ) != do_lower_case or normalizer_state.get("strip_accents" ,lowerCAmelCase__ ) != strip_accents or normalizer_state.get("handle_chinese_chars" ,lowerCAmelCase__ ) != tokenize_chinese_chars ): lowerCAmelCase_ : Any = getattr(lowerCAmelCase__ ,normalizer_state.pop("type" ) ) lowerCAmelCase_ : Optional[int] = do_lower_case lowerCAmelCase_ : Any = strip_accents lowerCAmelCase_ : Optional[Any] = tokenize_chinese_chars lowerCAmelCase_ : List[str] = normalizer_class(**lowerCAmelCase__ ) lowerCAmelCase_ : str = do_lower_case def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : List[str]=None ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = [self.sep_token_id] lowerCAmelCase_ : Dict = [self.cls_token_id] if token_ids_a is None: return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = self._tokenizer.model.save(lowerCAmelCase__ ,name=lowerCAmelCase__ ) return tuple(lowerCAmelCase__ )
659
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, PNDMScheduler, StableDiffusionLDMaDPipeline, UNetaDConditionModel, ) from diffusers.utils import nightly, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS enable_full_determinism() class __snake_case ( unittest.TestCase ): """simple docstring""" UpperCamelCase_ = StableDiffusionLDMaDPipeline UpperCamelCase_ = TEXT_TO_IMAGE_PARAMS UpperCamelCase_ = TEXT_TO_IMAGE_BATCH_PARAMS UpperCamelCase_ = TEXT_TO_IMAGE_IMAGE_PARAMS def UpperCAmelCase_ ( self : Tuple ) -> str: '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase_ : Optional[Any] = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") ,up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") ,cross_attention_dim=32 ,) lowerCAmelCase_ : Any = DDIMScheduler( beta_start=0.00_085 ,beta_end=0.012 ,beta_schedule="scaled_linear" ,clip_sample=lowerCAmelCase__ ,set_alpha_to_one=lowerCAmelCase__ ,) torch.manual_seed(0 ) lowerCAmelCase_ : str = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=6 ,out_channels=6 ,down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] ,up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] ,latent_channels=4 ,) torch.manual_seed(0 ) lowerCAmelCase_ : Optional[Any] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=10_00 ,) lowerCAmelCase_ : Optional[int] = CLIPTextModel(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) lowerCAmelCase_ : List[Any] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : List[str]=0 ) -> Dict: '''simple docstring''' if str(lowerCAmelCase__ ).startswith("mps" ): lowerCAmelCase_ : Optional[int] = torch.manual_seed(lowerCAmelCase__ ) else: lowerCAmelCase_ : Dict = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : str = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : Any ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Dict = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : List[str] = self.get_dummy_components() lowerCAmelCase_ : Union[str, Any] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Any = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Any = output.rgb, output.depth lowerCAmelCase_ : Dict = rgb[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) lowerCAmelCase_ : Optional[Any] = np.array( [0.37_338_176, 0.70_247, 0.74_203_193, 0.51_643_604, 0.58_256_793, 0.60_932_136, 0.4_181_095, 0.48_355_877, 0.46_535_262] ) lowerCAmelCase_ : Tuple = np.array([103.46_727, 85.812_004, 87.849_236] ) assert np.abs(image_slice_rgb.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(image_slice_depth.flatten() - expected_slice_depth ).max() < 1e-2 def UpperCAmelCase_ ( self : int ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Dict = self.get_dummy_components() lowerCAmelCase_ : List[str] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : str = 3 * [inputs["prompt"]] # forward lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = output.rgb, output.depth lowerCAmelCase_ : str = rgb_slice_a[0, -3:, -3:, -1] lowerCAmelCase_ : List[str] = depth_slice_a[0, -3:, -1] lowerCAmelCase_ : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = 3 * [inputs.pop("prompt" )] lowerCAmelCase_ : str = ldmad_pipe.tokenizer( lowerCAmelCase__ ,padding="max_length" ,max_length=ldmad_pipe.tokenizer.model_max_length ,truncation=lowerCAmelCase__ ,return_tensors="pt" ,) lowerCAmelCase_ : Union[str, Any] = text_inputs["input_ids"].to(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = ldmad_pipe.text_encoder(lowerCAmelCase__ )[0] lowerCAmelCase_ : Optional[int] = prompt_embeds # forward lowerCAmelCase_ : str = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : str = output.rgb, output.depth lowerCAmelCase_ : Optional[Any] = rgb_slice_a[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth_slice_a[0, -3:, -1] assert np.abs(rgb_slice_a.flatten() - rgb_slice_a.flatten() ).max() < 1e-4 assert np.abs(depth_slice_a.flatten() - depth_slice_a.flatten() ).max() < 1e-4 def UpperCAmelCase_ ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : Any = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : Optional[int] = self.get_dummy_components() lowerCAmelCase_ : Dict = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : Any = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = "french fries" lowerCAmelCase_ : Optional[int] = ldmad_pipe(**lowerCAmelCase__ ,negative_prompt=lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = output.rgb, output.depth lowerCAmelCase_ : Any = rgb[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) lowerCAmelCase_ : int = np.array( [0.37_044, 0.71_811_503, 0.7_223_251, 0.48_603_675, 0.5_638_391, 0.6_364_948, 0.42_833_704, 0.4_901_315, 0.47_926_217] ) lowerCAmelCase_ : Union[str, Any] = np.array([107.84_738, 84.62_802, 89.962_135] ) assert np.abs(rgb_slice.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(depth_slice.flatten() - expected_slice_depth ).max() < 1e-2 @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Dict="cpu" ,lowerCAmelCase__ : Union[str, Any]=torch.floataa ,lowerCAmelCase__ : List[str]=0 ) -> int: '''simple docstring''' lowerCAmelCase_ : Any = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = np.random.RandomState(lowerCAmelCase__ ).standard_normal((1, 4, 64, 64) ) lowerCAmelCase_ : Optional[Any] = torch.from_numpy(lowerCAmelCase__ ).to(device=lowerCAmelCase__ ,dtype=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : List[Any] ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ) lowerCAmelCase_ : List[str] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Dict = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Dict = output.rgb, output.depth lowerCAmelCase_ : List[str] = rgb[0, -3:, -3:, -1].flatten() lowerCAmelCase_ : Optional[int] = rgb[0, -3:, -1].flatten() assert rgb.shape == (1, 5_12, 5_12, 3) assert depth.shape == (1, 5_12, 5_12) lowerCAmelCase_ : int = np.array( [0.53_805_465, 0.56_707_305, 0.5_486_515, 0.57_012_236, 0.5_814_511, 0.56_253_487, 0.54_843_014, 0.55_092_263, 0.6_459_706] ) lowerCAmelCase_ : Optional[Any] = np.array( [0.9_263_781, 0.6_678_672, 0.5_486_515, 0.92_202_145, 0.67_831_135, 0.56_253_487, 0.9_241_694, 0.7_551_478, 0.6_459_706] ) assert np.abs(rgb_slice - expected_slice_rgb ).max() < 3e-3 assert np.abs(depth_slice - expected_slice_depth ).max() < 3e-3 @nightly @require_torch_gpu class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Dict="cpu" ,lowerCAmelCase__ : List[str]=torch.floataa ,lowerCAmelCase__ : Optional[int]=0 ) -> int: '''simple docstring''' lowerCAmelCase_ : Dict = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = np.random.RandomState(lowerCAmelCase__ ).standard_normal((1, 4, 64, 64) ) lowerCAmelCase_ : Any = torch.from_numpy(lowerCAmelCase__ ).to(device=lowerCAmelCase__ ,dtype=lowerCAmelCase__ ) lowerCAmelCase_ : int = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 50, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : Dict ) -> int: '''simple docstring''' lowerCAmelCase_ : List[Any] = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ).to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Any = output.rgb, output.depth lowerCAmelCase_ : Dict = 0.495_586 lowerCAmelCase_ : Optional[Any] = 0.33_795_515 lowerCAmelCase_ : Any = 112.48_518 lowerCAmelCase_ : List[Any] = 98.489_746 assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3 def UpperCAmelCase_ ( self : Tuple ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : int = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d-4c" ).to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : str = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Tuple = output.rgb, output.depth lowerCAmelCase_ : List[str] = 0.4_194_127 lowerCAmelCase_ : List[str] = 0.35_375_586 lowerCAmelCase_ : str = 0.5_638_502 lowerCAmelCase_ : Optional[Any] = 0.34_686_103 assert rgb.shape == (1, 5_12, 5_12, 3) assert depth.shape == (1, 5_12, 5_12, 1) assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3
659
1
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def UpperCamelCase ( snake_case__=None): if subparsers is not None: lowerCAmelCase_ : Optional[Any] = subparsers.add_parser("test") else: lowerCAmelCase_ : Optional[Any] = argparse.ArgumentParser("Accelerate test command") parser.add_argument( "--config_file" , default=snake_case__ , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=snake_case__) return parser def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Any = os.path.sep.join(__file__.split(os.path.sep)[:-2] + ["test_utils", "scripts", "test_script.py"]) if args.config_file is None: lowerCAmelCase_ : List[Any] = script_name else: lowerCAmelCase_ : Union[str, Any] = F'''--config_file={args.config_file} {script_name}''' lowerCAmelCase_ : Optional[int] = ["accelerate-launch"] + test_args.split() lowerCAmelCase_ : Optional[Any] = execute_subprocess_async(snake_case__ , env=os.environ.copy()) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!") def UpperCamelCase ( ): lowerCAmelCase_ : Any = test_command_parser() lowerCAmelCase_ : Optional[Any] = parser.parse_args() test_command(snake_case__) if __name__ == "__main__": main()
659
import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) _lowercase = { '''iou_prediction_head.layers.0''': '''iou_prediction_head.proj_in''', '''iou_prediction_head.layers.1''': '''iou_prediction_head.layers.0''', '''iou_prediction_head.layers.2''': '''iou_prediction_head.proj_out''', '''mask_decoder.output_upscaling.0''': '''mask_decoder.upscale_conv1''', '''mask_decoder.output_upscaling.1''': '''mask_decoder.upscale_layer_norm''', '''mask_decoder.output_upscaling.3''': '''mask_decoder.upscale_conv2''', '''mask_downscaling.0''': '''mask_embed.conv1''', '''mask_downscaling.1''': '''mask_embed.layer_norm1''', '''mask_downscaling.3''': '''mask_embed.conv2''', '''mask_downscaling.4''': '''mask_embed.layer_norm2''', '''mask_downscaling.6''': '''mask_embed.conv3''', '''point_embeddings''': '''point_embed''', '''pe_layer.positional_encoding_gaussian_matrix''': '''shared_embedding.positional_embedding''', '''image_encoder''': '''vision_encoder''', '''neck.0''': '''neck.conv1''', '''neck.1''': '''neck.layer_norm1''', '''neck.2''': '''neck.conv2''', '''neck.3''': '''neck.layer_norm2''', '''patch_embed.proj''': '''patch_embed.projection''', '''.norm''': '''.layer_norm''', '''blocks''': '''layers''', } def UpperCamelCase ( snake_case__): lowerCAmelCase_ : int = {} state_dict.pop("pixel_mean" , snake_case__) state_dict.pop("pixel_std" , snake_case__) lowerCAmelCase_ : List[Any] = R".*.output_hypernetworks_mlps.(\d+).layers.(\d+).*" for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: lowerCAmelCase_ : Dict = key.replace(snake_case__ , snake_case__) if re.match(snake_case__ , snake_case__): lowerCAmelCase_ : Any = int(re.match(snake_case__ , snake_case__).group(2)) if layer_nb == 0: lowerCAmelCase_ : List[Any] = key.replace("layers.0" , "proj_in") elif layer_nb == 1: lowerCAmelCase_ : List[Any] = key.replace("layers.1" , "layers.0") elif layer_nb == 2: lowerCAmelCase_ : int = key.replace("layers.2" , "proj_out") lowerCAmelCase_ : int = value lowerCAmelCase_ : Optional[int] = model_state_dict[ "prompt_encoder.shared_embedding.positional_embedding" ] return model_state_dict def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__="ybelkada/segment-anything"): lowerCAmelCase_ : Optional[int] = hf_hub_download(snake_case__ , F'''checkpoints/{model_name}.pth''') if "sam_vit_b" in model_name: lowerCAmelCase_ : Optional[Any] = SamConfig() elif "sam_vit_l" in model_name: lowerCAmelCase_ : Optional[int] = SamVisionConfig( hidden_size=10_24 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) lowerCAmelCase_ : Union[str, Any] = SamConfig( vision_config=snake_case__ , ) elif "sam_vit_h" in model_name: lowerCAmelCase_ : Optional[Any] = SamVisionConfig( hidden_size=12_80 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) lowerCAmelCase_ : Tuple = SamConfig( vision_config=snake_case__ , ) lowerCAmelCase_ : Optional[Any] = torch.load(snake_case__ , map_location="cpu") lowerCAmelCase_ : Union[str, Any] = replace_keys(snake_case__) lowerCAmelCase_ : List[Any] = SamImageProcessor() lowerCAmelCase_ : Any = SamProcessor(image_processor=snake_case__) lowerCAmelCase_ : Any = SamModel(snake_case__) hf_model.load_state_dict(snake_case__) lowerCAmelCase_ : Dict = hf_model.to("cuda") lowerCAmelCase_ : List[str] = "https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png" lowerCAmelCase_ : List[Any] = Image.open(requests.get(snake_case__ , stream=snake_case__).raw).convert("RGB") lowerCAmelCase_ : Optional[int] = [[[4_00, 6_50]]] lowerCAmelCase_ : int = [[1]] lowerCAmelCase_ : Optional[Any] = processor(images=np.array(snake_case__) , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Optional[Any] = hf_model(**snake_case__) lowerCAmelCase_ : Optional[int] = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579_890_251_159_668 lowerCAmelCase_ : Any = processor( images=np.array(snake_case__) , input_points=snake_case__ , input_labels=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Optional[Any] = hf_model(**snake_case__) lowerCAmelCase_ : Union[str, Any] = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_712_603_092_193_604 lowerCAmelCase_ : Tuple = ((75, 2_75, 17_25, 8_50),) lowerCAmelCase_ : Optional[Any] = processor(images=np.array(snake_case__) , input_boxes=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : List[Any] = hf_model(**snake_case__) lowerCAmelCase_ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.8_686_015_605_926_514 # Test with 2 points and 1 image. lowerCAmelCase_ : int = [[[4_00, 6_50], [8_00, 6_50]]] lowerCAmelCase_ : Optional[Any] = [[1, 1]] lowerCAmelCase_ : List[Any] = processor( images=np.array(snake_case__) , input_points=snake_case__ , input_labels=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Tuple = hf_model(**snake_case__) lowerCAmelCase_ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_936_047_792_434_692 if __name__ == "__main__": _lowercase = argparse.ArgumentParser() _lowercase = ['''sam_vit_b_01ec64''', '''sam_vit_h_4b8939''', '''sam_vit_l_0b3195'''] parser.add_argument( '''--model_name''', default='''sam_vit_h_4b8939''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model and processor to the hub after converting''', ) parser.add_argument( '''--model_hub_id''', default='''ybelkada/segment-anything''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) _lowercase = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
659
1
import random def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ = False): lowerCAmelCase_ : dict = {i: [] for i in range(snake_case__)} # if probability is greater or equal than 1, then generate a complete graph if probability >= 1: return complete_graph(snake_case__) # if probability is lower or equal than 0, then return a graph without edges if probability <= 0: return graph # for each couple of nodes, add an edge from u to v # if the number randomly generated is greater than probability probability for i in range(snake_case__): for j in range(i + 1 , snake_case__): if random.random() < probability: graph[i].append(snake_case__) if not directed: # if the graph is undirected, add an edge in from j to i, either graph[j].append(snake_case__) return graph def UpperCamelCase ( snake_case__): return { i: [j for j in range(snake_case__) if i != j] for i in range(snake_case__) } if __name__ == "__main__": import doctest doctest.testmod()
659
class __snake_case : """simple docstring""" def __init__( self : Union[str, Any] ,lowerCAmelCase__ : str = "" ,lowerCAmelCase__ : bool = False ) -> None: '''simple docstring''' lowerCAmelCase_ : dict[str, RadixNode] = {} # A node will be a leaf if the tree contains its word lowerCAmelCase_ : Optional[int] = is_leaf lowerCAmelCase_ : List[str] = prefix def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : str ) -> tuple[str, str, str]: '''simple docstring''' lowerCAmelCase_ : List[str] = 0 for q, w in zip(self.prefix ,lowerCAmelCase__ ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : list[str] ) -> None: '''simple docstring''' for word in words: self.insert(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : str ) -> None: '''simple docstring''' if self.prefix == word: lowerCAmelCase_ : Optional[Any] = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: lowerCAmelCase_ : Optional[int] = RadixNode(prefix=lowerCAmelCase__ ,is_leaf=lowerCAmelCase__ ) else: lowerCAmelCase_ : Optional[Any] = self.nodes[word[0]] lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = incoming_node.match( lowerCAmelCase__ ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: lowerCAmelCase_ : Dict = remaining_prefix lowerCAmelCase_ : str = self.nodes[matching_string[0]] lowerCAmelCase_ : Dict = RadixNode(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Any = aux_node if remaining_word == "": lowerCAmelCase_ : Optional[Any] = True else: self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : str ) -> bool: '''simple docstring''' lowerCAmelCase_ : List[str] = self.nodes.get(word[0] ,lowerCAmelCase__ ) if not incoming_node: return False else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : str ) -> bool: '''simple docstring''' lowerCAmelCase_ : int = self.nodes.get(word[0] ,lowerCAmelCase__ ) if not incoming_node: return False else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(lowerCAmelCase__ ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: lowerCAmelCase_ : int = list(self.nodes.values() )[0] lowerCAmelCase_ : List[Any] = merging_node.is_leaf self.prefix += merging_node.prefix lowerCAmelCase_ : int = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: lowerCAmelCase_ : List[str] = False # If there is 1 edge, we merge it with its child else: lowerCAmelCase_ : Union[str, Any] = list(incoming_node.nodes.values() )[0] lowerCAmelCase_ : Optional[int] = merging_node.is_leaf incoming_node.prefix += merging_node.prefix lowerCAmelCase_ : List[str] = merging_node.nodes return True def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : int = 0 ) -> None: '''simple docstring''' if self.prefix != "": print("-" * height ,self.prefix ," (leaf)" if self.is_leaf else "" ) for value in self.nodes.values(): value.print_tree(height + 1 ) def UpperCamelCase ( ): lowerCAmelCase_ : List[Any] = "banana bananas bandana band apple all beast".split() lowerCAmelCase_ : Optional[Any] = RadixNode() root.insert_many(snake_case__) assert all(root.find(snake_case__) for word in words) assert not root.find("bandanas") assert not root.find("apps") root.delete("all") assert not root.find("all") root.delete("banana") assert not root.find("banana") assert root.find("bananas") return True def UpperCamelCase ( ): assert test_trie() def UpperCamelCase ( ): lowerCAmelCase_ : str = RadixNode() lowerCAmelCase_ : str = "banana bananas bandanas bandana band apple all beast".split() root.insert_many(snake_case__) print("Words:" , snake_case__) print("Tree:") root.print_tree() if __name__ == "__main__": main()
659
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _lowercase = { '''configuration_mvp''': ['''MVP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MvpConfig''', '''MvpOnnxConfig'''], '''tokenization_mvp''': ['''MvpTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ['''MvpTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ '''MVP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MvpForCausalLM''', '''MvpForConditionalGeneration''', '''MvpForQuestionAnswering''', '''MvpForSequenceClassification''', '''MvpModel''', '''MvpPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mvp import MVP_PRETRAINED_CONFIG_ARCHIVE_MAP, MvpConfig, MvpOnnxConfig from .tokenization_mvp import MvpTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mvp_fast import MvpTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mvp import ( MVP_PRETRAINED_MODEL_ARCHIVE_LIST, MvpForCausalLM, MvpForConditionalGeneration, MvpForQuestionAnswering, MvpForSequenceClassification, MvpModel, MvpPreTrainedModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
659
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class __snake_case : """simple docstring""" def __init__( self : Tuple ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Optional[Any]=12 ,lowerCAmelCase__ : Union[str, Any]=7 ,lowerCAmelCase__ : Union[str, Any]=True ,lowerCAmelCase__ : List[str]=True ,lowerCAmelCase__ : Any=True ,lowerCAmelCase__ : Optional[Any]=99 ,lowerCAmelCase__ : List[str]=32 ,lowerCAmelCase__ : Dict=32 ,lowerCAmelCase__ : str=2 ,lowerCAmelCase__ : Optional[int]=4 ,lowerCAmelCase__ : str=37 ,lowerCAmelCase__ : Dict=0.1 ,lowerCAmelCase__ : List[str]=0.1 ,lowerCAmelCase__ : str=5_12 ,lowerCAmelCase__ : Union[str, Any]=0.02 ,lowerCAmelCase__ : Tuple=0 ,lowerCAmelCase__ : str=None ,) -> str: '''simple docstring''' lowerCAmelCase_ : int = parent lowerCAmelCase_ : str = batch_size lowerCAmelCase_ : int = seq_length lowerCAmelCase_ : Union[str, Any] = is_training lowerCAmelCase_ : int = use_input_mask lowerCAmelCase_ : List[Any] = use_labels lowerCAmelCase_ : Dict = vocab_size lowerCAmelCase_ : Union[str, Any] = hidden_size lowerCAmelCase_ : Union[str, Any] = projection_dim lowerCAmelCase_ : List[Any] = num_hidden_layers lowerCAmelCase_ : Any = num_attention_heads lowerCAmelCase_ : List[Any] = intermediate_size lowerCAmelCase_ : Any = dropout lowerCAmelCase_ : Optional[int] = attention_dropout lowerCAmelCase_ : int = max_position_embeddings lowerCAmelCase_ : Optional[int] = initializer_range lowerCAmelCase_ : Any = scope lowerCAmelCase_ : Tuple = bos_token_id def UpperCAmelCase_ ( self : str ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) lowerCAmelCase_ : Dict = None if self.use_input_mask: lowerCAmelCase_ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowerCAmelCase_ : List[Any] = input_mask.numpy() lowerCAmelCase_ , lowerCAmelCase_ : str = input_mask.shape lowerCAmelCase_ : Dict = np.random.randint(1 ,seq_length - 1 ,size=(batch_size,) ) for batch_idx, start_index in enumerate(lowerCAmelCase__ ): lowerCAmelCase_ : Union[str, Any] = 1 lowerCAmelCase_ : Optional[Any] = 0 lowerCAmelCase_ : List[Any] = self.get_config() return config, input_ids, tf.convert_to_tensor(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[str] ) -> str: '''simple docstring''' return BlipTextConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,projection_dim=self.projection_dim ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,dropout=self.dropout ,attention_dropout=self.attention_dropout ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,bos_token_id=self.bos_token_id ,) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Dict ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : List[Any] = TFBlipTextModel(config=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = model(lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,training=lowerCAmelCase__ ) lowerCAmelCase_ : str = model(lowerCAmelCase__ ,training=lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def UpperCAmelCase_ ( self : Optional[int] ) -> int: '''simple docstring''' lowerCAmelCase_ : List[str] = self.prepare_config_and_inputs() lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = config_and_inputs lowerCAmelCase_ : Tuple = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = (TFBlipTextModel,) if is_tf_available() else () UpperCamelCase_ = False UpperCamelCase_ = False UpperCamelCase_ = False def UpperCAmelCase_ ( self : Optional[Any] ) -> str: '''simple docstring''' lowerCAmelCase_ : List[str] = BlipTextModelTester(self ) lowerCAmelCase_ : Tuple = ConfigTester(self ,config_class=lowerCAmelCase__ ,hidden_size=37 ) def UpperCAmelCase_ ( self : str ) -> Any: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase_ ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' pass def UpperCAmelCase_ ( self : Union[str, Any] ) -> Any: '''simple docstring''' pass @unittest.skip(reason="Blip does not use inputs_embeds" ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCAmelCase_ ( self : int ) -> Optional[Any]: '''simple docstring''' pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCAmelCase_ ( self : Dict ) -> Union[str, Any]: '''simple docstring''' pass @slow def UpperCAmelCase_ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Tuple = TFBlipTextModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : str=True ) -> List[Any]: '''simple docstring''' super().test_pt_tf_model_equivalence(allow_missing_keys=lowerCAmelCase__ )
659
1
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''microsoft/swinv2-tiny-patch4-window8-256''': ( '''https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json''' ), } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = 'swinv2' UpperCamelCase_ = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : List[Any] ,lowerCAmelCase__ : Optional[int]=2_24 ,lowerCAmelCase__ : Dict=4 ,lowerCAmelCase__ : Dict=3 ,lowerCAmelCase__ : List[Any]=96 ,lowerCAmelCase__ : Optional[Any]=[2, 2, 6, 2] ,lowerCAmelCase__ : Optional[Any]=[3, 6, 12, 24] ,lowerCAmelCase__ : Optional[int]=7 ,lowerCAmelCase__ : Dict=4.0 ,lowerCAmelCase__ : Dict=True ,lowerCAmelCase__ : str=0.0 ,lowerCAmelCase__ : Tuple=0.0 ,lowerCAmelCase__ : str=0.1 ,lowerCAmelCase__ : List[str]="gelu" ,lowerCAmelCase__ : Union[str, Any]=False ,lowerCAmelCase__ : Dict=0.02 ,lowerCAmelCase__ : int=1e-5 ,lowerCAmelCase__ : List[str]=32 ,**lowerCAmelCase__ : Tuple ,) -> List[str]: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = image_size lowerCAmelCase_ : List[Any] = patch_size lowerCAmelCase_ : Dict = num_channels lowerCAmelCase_ : Optional[int] = embed_dim lowerCAmelCase_ : Optional[Any] = depths lowerCAmelCase_ : Any = len(lowerCAmelCase__ ) lowerCAmelCase_ : str = num_heads lowerCAmelCase_ : List[str] = window_size lowerCAmelCase_ : List[str] = mlp_ratio lowerCAmelCase_ : Dict = qkv_bias lowerCAmelCase_ : str = hidden_dropout_prob lowerCAmelCase_ : str = attention_probs_dropout_prob lowerCAmelCase_ : Union[str, Any] = drop_path_rate lowerCAmelCase_ : List[Any] = hidden_act lowerCAmelCase_ : Any = use_absolute_embeddings lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : int = initializer_range lowerCAmelCase_ : Union[str, Any] = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase_ : Tuple = int(embed_dim * 2 ** (len(lowerCAmelCase__ ) - 1) ) lowerCAmelCase_ : str = (0, 0, 0, 0)
659
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _lowercase = logging.get_logger(__name__) _lowercase = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} # See all LED models at https://huggingface.co/models?filter=LED _lowercase = { '''vocab_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json''', }, '''merges_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json''', }, } _lowercase = { '''allenai/led-base-16384''': 16384, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def UpperCamelCase ( ): lowerCAmelCase_ : Optional[int] = ( list(range(ord("!") , ord("~") + 1)) + list(range(ord("¡") , ord("¬") + 1)) + list(range(ord("®") , ord("ÿ") + 1)) ) lowerCAmelCase_ : List[Any] = bs[:] lowerCAmelCase_ : Optional[int] = 0 for b in range(2**8): if b not in bs: bs.append(snake_case__) cs.append(2**8 + n) n += 1 lowerCAmelCase_ : Tuple = [chr(snake_case__) for n in cs] return dict(zip(snake_case__ , snake_case__)) def UpperCamelCase ( snake_case__): lowerCAmelCase_ : str = set() lowerCAmelCase_ : List[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char)) lowerCAmelCase_ : Union[str, Any] = char return pairs class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = ['input_ids', 'attention_mask'] def __init__( self : int ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Tuple="replace" ,lowerCAmelCase__ : Optional[int]="<s>" ,lowerCAmelCase__ : Optional[int]="</s>" ,lowerCAmelCase__ : Tuple="</s>" ,lowerCAmelCase__ : int="<s>" ,lowerCAmelCase__ : Union[str, Any]="<unk>" ,lowerCAmelCase__ : str="<pad>" ,lowerCAmelCase__ : Tuple="<mask>" ,lowerCAmelCase__ : Optional[int]=False ,**lowerCAmelCase__ : Tuple ,) -> Any: '''simple docstring''' lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else bos_token lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else eos_token lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else sep_token lowerCAmelCase_ : Any = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else cls_token lowerCAmelCase_ : Tuple = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else unk_token lowerCAmelCase_ : Any = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase_ : Optional[int] = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else mask_token super().__init__( errors=lowerCAmelCase__ ,bos_token=lowerCAmelCase__ ,eos_token=lowerCAmelCase__ ,unk_token=lowerCAmelCase__ ,sep_token=lowerCAmelCase__ ,cls_token=lowerCAmelCase__ ,pad_token=lowerCAmelCase__ ,mask_token=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,**lowerCAmelCase__ ,) with open(lowerCAmelCase__ ,encoding="utf-8" ) as vocab_handle: lowerCAmelCase_ : List[str] = json.load(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = {v: k for k, v in self.encoder.items()} lowerCAmelCase_ : Optional[int] = errors # how to handle errors in decoding lowerCAmelCase_ : Optional[int] = bytes_to_unicode() lowerCAmelCase_ : str = {v: k for k, v in self.byte_encoder.items()} with open(lowerCAmelCase__ ,encoding="utf-8" ) as merges_handle: lowerCAmelCase_ : List[str] = merges_handle.read().split("\n" )[1:-1] lowerCAmelCase_ : List[Any] = [tuple(merge.split() ) for merge in bpe_merges] lowerCAmelCase_ : Union[str, Any] = dict(zip(lowerCAmelCase__ ,range(len(lowerCAmelCase__ ) ) ) ) lowerCAmelCase_ : Dict = {} lowerCAmelCase_ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowerCAmelCase_ : Any = re.compile(R"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def UpperCAmelCase_ ( self : Dict ) -> Dict: '''simple docstring''' return len(self.encoder ) def UpperCAmelCase_ ( self : Dict ) -> str: '''simple docstring''' return dict(self.encoder ,**self.added_tokens_encoder ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Dict ) -> Dict: '''simple docstring''' if token in self.cache: return self.cache[token] lowerCAmelCase_ : Union[str, Any] = tuple(lowerCAmelCase__ ) lowerCAmelCase_ : str = get_pairs(lowerCAmelCase__ ) if not pairs: return token while True: lowerCAmelCase_ : Optional[int] = min(lowerCAmelCase__ ,key=lambda lowerCAmelCase__ : self.bpe_ranks.get(lowerCAmelCase__ ,float("inf" ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = bigram lowerCAmelCase_ : Tuple = [] lowerCAmelCase_ : str = 0 while i < len(lowerCAmelCase__ ): try: lowerCAmelCase_ : Union[str, Any] = word.index(lowerCAmelCase__ ,lowerCAmelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase_ : List[str] = j if word[i] == first and i < len(lowerCAmelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase_ : Optional[int] = tuple(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = new_word if len(lowerCAmelCase__ ) == 1: break else: lowerCAmelCase_ : Dict = get_pairs(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = " ".join(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = word return word def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : Dict ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : Any = [] for token in re.findall(self.pat ,lowerCAmelCase__ ): lowerCAmelCase_ : Optional[int] = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCAmelCase__ ).split(" " ) ) return bpe_tokens def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any] ) -> Tuple: '''simple docstring''' return self.encoder.get(lowerCAmelCase__ ,self.encoder.get(self.unk_token ) ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Union[str, Any] ) -> Optional[int]: '''simple docstring''' return self.decoder.get(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[Any] ) -> Any: '''simple docstring''' lowerCAmelCase_ : int = "".join(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" ,errors=self.errors ) return text def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(lowerCAmelCase__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCAmelCase_ : Optional[int] = os.path.join( lowerCAmelCase__ ,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase_ : List[str] = os.path.join( lowerCAmelCase__ ,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(lowerCAmelCase__ ,"w" ,encoding="utf-8" ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=lowerCAmelCase__ ,ensure_ascii=lowerCAmelCase__ ) + "\n" ) lowerCAmelCase_ : Dict = 0 with open(lowerCAmelCase__ ,"w" ,encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda lowerCAmelCase__ : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' " Please check that the tokenizer is not corrupted!" ) lowerCAmelCase_ : List[Any] = token_index writer.write(" ".join(lowerCAmelCase__ ) + "\n" ) index += 1 return vocab_file, merge_file def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase_ : Union[str, Any] = [self.cls_token_id] lowerCAmelCase_ : str = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ,lowerCAmelCase__ : bool = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ ,token_ids_a=lowerCAmelCase__ ,already_has_special_tokens=lowerCAmelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase__ )) + [1] return [1] + ([0] * len(lowerCAmelCase__ )) + [1, 1] + ([0] * len(lowerCAmelCase__ )) + [1] def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = [self.sep_token_id] lowerCAmelCase_ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : Optional[int]=False ,**lowerCAmelCase__ : str ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = kwargs.pop("add_prefix_space" ,self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCAmelCase__ ) > 0 and not text[0].isspace()): lowerCAmelCase_ : List[str] = " " + text return (text, kwargs) def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : Union[Dict[str, EncodedInput], BatchEncoding] ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : Optional[bool] = None ,) -> dict: '''simple docstring''' lowerCAmelCase_ : int = super()._pad( encoded_inputs=lowerCAmelCase__ ,max_length=lowerCAmelCase__ ,padding_strategy=lowerCAmelCase__ ,pad_to_multiple_of=lowerCAmelCase__ ,return_attention_mask=lowerCAmelCase__ ,) # Load from model defaults if return_attention_mask is None: lowerCAmelCase_ : List[Any] = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowerCAmelCase_ : Dict = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowerCAmelCase_ : List[Any] = len(encoded_inputs["global_attention_mask"] ) != len(lowerCAmelCase__ ) if needs_to_be_padded: lowerCAmelCase_ : Union[str, Any] = len(lowerCAmelCase__ ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowerCAmelCase_ : Optional[int] = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": lowerCAmelCase_ : List[Any] = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
659
1
from importlib import import_module from .logging import get_logger _lowercase = get_logger(__name__) class __snake_case : """simple docstring""" def __init__( self : Dict ,lowerCAmelCase__ : int ,lowerCAmelCase__ : Tuple=None ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = attrs or [] if module is not None: for key in module.__dict__: if key in attrs or not key.startswith("__" ): setattr(self ,lowerCAmelCase__ ,getattr(lowerCAmelCase__ ,lowerCAmelCase__ ) ) lowerCAmelCase_ : str = module._original_module if isinstance(lowerCAmelCase__ ,_PatchedModuleObj ) else module class __snake_case : """simple docstring""" UpperCamelCase_ = [] def __init__( self : Any ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : str ,lowerCAmelCase__ : int ,lowerCAmelCase__ : List[Any]=None ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : List[str] = obj lowerCAmelCase_ : Optional[int] = target lowerCAmelCase_ : List[str] = new lowerCAmelCase_ : List[str] = target.split("." )[0] lowerCAmelCase_ : List[Any] = {} lowerCAmelCase_ : Union[str, Any] = attrs or [] def __enter__( self : str ) -> Dict: '''simple docstring''' *lowerCAmelCase_ , lowerCAmelCase_ : int = self.target.split("." ) # Patch modules: # it's used to patch attributes of submodules like "os.path.join"; # in this case we need to patch "os" and "os.path" for i in range(len(lowerCAmelCase__ ) ): try: lowerCAmelCase_ : List[Any] = import_module(".".join(submodules[: i + 1] ) ) except ModuleNotFoundError: continue # We iterate over all the globals in self.obj in case we find "os" or "os.path" for attr in self.obj.__dir__(): lowerCAmelCase_ : List[str] = getattr(self.obj ,lowerCAmelCase__ ) # We don't check for the name of the global, but rather if its value *is* "os" or "os.path". # This allows to patch renamed modules like "from os import path as ospath". if obj_attr is submodule or ( (isinstance(lowerCAmelCase__ ,_PatchedModuleObj ) and obj_attr._original_module is submodule) ): lowerCAmelCase_ : Dict = obj_attr # patch at top level setattr(self.obj ,lowerCAmelCase__ ,_PatchedModuleObj(lowerCAmelCase__ ,attrs=self.attrs ) ) lowerCAmelCase_ : str = getattr(self.obj ,lowerCAmelCase__ ) # construct lower levels patches for key in submodules[i + 1 :]: setattr(lowerCAmelCase__ ,lowerCAmelCase__ ,_PatchedModuleObj(getattr(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ) ,attrs=self.attrs ) ) lowerCAmelCase_ : Tuple = getattr(lowerCAmelCase__ ,lowerCAmelCase__ ) # finally set the target attribute setattr(lowerCAmelCase__ ,lowerCAmelCase__ ,self.new ) # Patch attribute itself: # it's used for builtins like "open", # and also to patch "os.path.join" we may also need to patch "join" # itself if it was imported as "from os.path import join". if submodules: # if it's an attribute of a submodule like "os.path.join" try: lowerCAmelCase_ : Optional[int] = getattr(import_module(".".join(lowerCAmelCase__ ) ) ,lowerCAmelCase__ ) except (AttributeError, ModuleNotFoundError): return # We iterate over all the globals in self.obj in case we find "os.path.join" for attr in self.obj.__dir__(): # We don't check for the name of the global, but rather if its value *is* "os.path.join". # This allows to patch renamed attributes like "from os.path import join as pjoin". if getattr(self.obj ,lowerCAmelCase__ ) is attr_value: lowerCAmelCase_ : Union[str, Any] = getattr(self.obj ,lowerCAmelCase__ ) setattr(self.obj ,lowerCAmelCase__ ,self.new ) elif target_attr in globals()["__builtins__"]: # if it'a s builtin like "open" lowerCAmelCase_ : List[Any] = globals()["__builtins__"][target_attr] setattr(self.obj ,lowerCAmelCase__ ,self.new ) else: raise RuntimeError(f'''Tried to patch attribute {target_attr} instead of a submodule.''' ) def __exit__( self : Optional[int] ,*lowerCAmelCase__ : Any ) -> int: '''simple docstring''' for attr in list(self.original ): setattr(self.obj ,lowerCAmelCase__ ,self.original.pop(lowerCAmelCase__ ) ) def UpperCAmelCase_ ( self : Tuple ) -> List[Any]: '''simple docstring''' self.__enter__() self._active_patches.append(self ) def UpperCAmelCase_ ( self : List[Any] ) -> str: '''simple docstring''' try: self._active_patches.remove(self ) except ValueError: # If the patch hasn't been started this will fail return None return self.__exit__()
659
import os _lowercase = {'''I''': 1, '''V''': 5, '''X''': 10, '''L''': 50, '''C''': 100, '''D''': 500, '''M''': 1000} def UpperCamelCase ( snake_case__): lowerCAmelCase_ : List[str] = 0 lowerCAmelCase_ : Any = 0 while index < len(snake_case__) - 1: lowerCAmelCase_ : Optional[Any] = SYMBOLS[numerals[index]] lowerCAmelCase_ : int = SYMBOLS[numerals[index + 1]] if current_value < next_value: total_value -= current_value else: total_value += current_value index += 1 total_value += SYMBOLS[numerals[index]] return total_value def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Optional[int] = "" lowerCAmelCase_ : Tuple = num // 10_00 numerals += m_count * "M" num %= 10_00 lowerCAmelCase_ : int = num // 1_00 if c_count == 9: numerals += "CM" c_count -= 9 elif c_count == 4: numerals += "CD" c_count -= 4 if c_count >= 5: numerals += "D" c_count -= 5 numerals += c_count * "C" num %= 1_00 lowerCAmelCase_ : int = num // 10 if x_count == 9: numerals += "XC" x_count -= 9 elif x_count == 4: numerals += "XL" x_count -= 4 if x_count >= 5: numerals += "L" x_count -= 5 numerals += x_count * "X" num %= 10 if num == 9: numerals += "IX" num -= 9 elif num == 4: numerals += "IV" num -= 4 if num >= 5: numerals += "V" num -= 5 numerals += num * "I" return numerals def UpperCamelCase ( snake_case__ = "/p089_roman.txt"): lowerCAmelCase_ : int = 0 with open(os.path.dirname(snake_case__) + roman_numerals_filename) as filea: lowerCAmelCase_ : List[Any] = filea.readlines() for line in lines: lowerCAmelCase_ : Any = line.strip() lowerCAmelCase_ : Tuple = parse_roman_numerals(snake_case__) lowerCAmelCase_ : List[Any] = generate_roman_numerals(snake_case__) savings += len(snake_case__) - len(snake_case__) return savings if __name__ == "__main__": print(f"{solution() = }")
659
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _lowercase = { '''configuration_perceiver''': ['''PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''PerceiverConfig''', '''PerceiverOnnxConfig'''], '''tokenization_perceiver''': ['''PerceiverTokenizer'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ['''PerceiverFeatureExtractor'''] _lowercase = ['''PerceiverImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ '''PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''PerceiverForImageClassificationConvProcessing''', '''PerceiverForImageClassificationFourier''', '''PerceiverForImageClassificationLearned''', '''PerceiverForMaskedLM''', '''PerceiverForMultimodalAutoencoding''', '''PerceiverForOpticalFlow''', '''PerceiverForSequenceClassification''', '''PerceiverLayer''', '''PerceiverModel''', '''PerceiverPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
659
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def UpperCamelCase ( ): lowerCAmelCase_ : Dict = HfArgumentParser(snake_case__) lowerCAmelCase_ : Dict = parser.parse_args_into_dataclasses()[0] lowerCAmelCase_ : List[Any] = TensorFlowBenchmark(args=snake_case__) try: lowerCAmelCase_ : str = parser.parse_args_into_dataclasses()[0] except ValueError as e: lowerCAmelCase_ : Optional[Any] = "Arg --no_{0} is no longer used, please use --no-{0} instead." lowerCAmelCase_ : Tuple = " ".join(str(snake_case__).split(" ")[:-1]) lowerCAmelCase_ : List[Any] = "" lowerCAmelCase_ : Optional[Any] = eval(str(snake_case__).split(" ")[-1]) lowerCAmelCase_ : List[Any] = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:]) else: wrong_args.append(snake_case__) if len(snake_case__) > 0: lowerCAmelCase_ : int = full_error_msg + begin_error_msg + str(snake_case__) raise ValueError(snake_case__) benchmark.run() if __name__ == "__main__": main()
659
1
from __future__ import annotations import numpy as np def UpperCamelCase ( snake_case__): return np.maximum(0 , snake_case__) if __name__ == "__main__": print(np.array(relu([-1, 0, 5]))) # --> [0, 0, 5]
659
from collections import defaultdict from pathlib import Path import pandas as pd from rouge_cli import calculate_rouge_path from utils import calculate_rouge _lowercase = [ '''Prosecutor: "No videos were used in the crash investigation" German papers say they saw a cell phone video of the''' ''' final seconds on board Flight 9525. The Germanwings co-pilot says he had a "previous episode of severe''' ''' depression\" German airline confirms it knew of Andreas Lubitz\'s depression years before he took control.''', '''The Palestinian Authority officially becomes the 123rd member of the International Criminal Court. The formal''' ''' accession was marked with a ceremony at The Hague, in the Netherlands. The Palestinians signed the ICC\'s''' ''' founding Rome Statute in January. Israel and the United States opposed the Palestinians\' efforts to join the''' ''' body.''', '''Amnesty International releases its annual report on the death penalty. The report catalogs the use of''' ''' state-sanctioned killing as a punitive measure across the globe. At least 607 people were executed around the''' ''' world in 2014, compared to 778 in 2013. The U.S. remains one of the worst offenders for imposing capital''' ''' punishment.''', ] _lowercase = [ '''Marseille prosecutor says "so far no videos were used in the crash investigation" despite media reports .''' ''' Journalists at Bild and Paris Match are "very confident" the video clip is real, an editor says . Andreas Lubitz''' ''' had informed his Lufthansa training school of an episode of severe depression, airline says .''', '''Membership gives the ICC jurisdiction over alleged crimes committed in Palestinian territories since last June .''' ''' Israel and the United States opposed the move, which could open the door to war crimes investigations against''' ''' Israelis .''', '''Amnesty\'s annual death penalty report catalogs encouraging signs, but setbacks in numbers of those sentenced to''' ''' death . Organization claims that governments around the world are using the threat of terrorism to advance''' ''' executions . The number of executions worldwide has gone down by almost 22% compared with 2013, but death''' ''' sentences up by 28% .''', ] def UpperCamelCase ( ): lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , bootstrap_aggregation=snake_case__ , rouge_keys=["rouge2", "rougeL"]) assert isinstance(snake_case__ , snake_case__) lowerCAmelCase_ : str = calculate_rouge(snake_case__ , snake_case__ , bootstrap_aggregation=snake_case__ , rouge_keys=["rouge2"]) assert ( pd.DataFrame(no_aggregation["rouge2"]).fmeasure.mean() == pd.DataFrame(no_aggregation_just_ra["rouge2"]).fmeasure.mean() ) def UpperCamelCase ( ): lowerCAmelCase_ : str = "rougeLsum" lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=[k])[k] lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=[k])[k] assert score > score_no_sep def UpperCamelCase ( ): lowerCAmelCase_ : int = ["rouge1", "rouge2", "rougeL"] lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=snake_case__) lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=snake_case__) assert score_sep == score_no_sep def UpperCamelCase ( ): lowerCAmelCase_ : List[str] = [ "Her older sister, Margot Frank, died in 1945, a month earlier than previously thought.", "Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports .", ] lowerCAmelCase_ : Dict = [ "Margot Frank, died in 1945, a month earlier than previously thought.", "Prosecutor: \"No videos were used in the crash investigation\" German papers say they saw a cell phone video of" " the final seconds on board Flight 9525.", ] assert calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__) == calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__) def UpperCamelCase ( ): lowerCAmelCase_ : Optional[int] = [ "\" \"a person who has such a video needs to immediately give it to the investigators,\" prosecutor says .<n> \"it is a very disturbing scene,\" editor-in-chief of bild online tells \"erin burnett: outfront\" " ] lowerCAmelCase_ : Any = [ " Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports . Journalists at Bild and Paris Match are \"very confident\" the video clip is real, an editor says . Andreas Lubitz had informed his Lufthansa training school of an episode of severe depression, airline says ." ] lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , rouge_keys=["rougeLsum"] , newline_sep=snake_case__)["rougeLsum"] lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , rouge_keys=["rougeLsum"])["rougeLsum"] assert new_score > prev_score def UpperCamelCase ( ): lowerCAmelCase_ : int = Path("examples/seq2seq/test_data/wmt_en_ro") lowerCAmelCase_ : Dict = calculate_rouge_path(data_dir.joinpath("test.source") , data_dir.joinpath("test.target")) assert isinstance(snake_case__ , snake_case__) lowerCAmelCase_ : Any = calculate_rouge_path( data_dir.joinpath("test.source") , data_dir.joinpath("test.target") , bootstrap_aggregation=snake_case__) assert isinstance(snake_case__ , snake_case__)
659
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _lowercase = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ['''BartphoTokenizer'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
659
import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = LEDTokenizer UpperCamelCase_ = LEDTokenizerFast UpperCamelCase_ = True def UpperCAmelCase_ ( self : List[Any] ) -> Optional[int]: '''simple docstring''' super().setUp() lowerCAmelCase_ : Union[str, Any] = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] lowerCAmelCase_ : Tuple = dict(zip(lowerCAmelCase__ ,range(len(lowerCAmelCase__ ) ) ) ) lowerCAmelCase_ : int = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] lowerCAmelCase_ : Union[str, Any] = {"unk_token": "<unk>"} lowerCAmelCase_ : List[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase_ : Any = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file ,"w" ,encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase__ ) + "\n" ) with open(self.merges_file ,"w" ,encoding="utf-8" ) as fp: fp.write("\n".join(lowerCAmelCase__ ) ) def UpperCAmelCase_ ( self : List[Any] ,**lowerCAmelCase__ : int ) -> Tuple: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ,**lowerCAmelCase__ : Optional[int] ) -> List[Any]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCAmelCase__ ) def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : int ) -> List[str]: '''simple docstring''' return "lower newer", "lower newer" @cached_property def UpperCAmelCase_ ( self : int ) -> Union[str, Any]: '''simple docstring''' return LEDTokenizer.from_pretrained("allenai/led-base-16384" ) @cached_property def UpperCAmelCase_ ( self : List[str] ) -> Dict: '''simple docstring''' return LEDTokenizerFast.from_pretrained("allenai/led-base-16384" ) @require_torch def UpperCAmelCase_ ( self : int ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = ["A long paragraph for summarization.", "Another paragraph for summarization."] lowerCAmelCase_ : int = [0, 2_50, 2_51, 1_78_18, 13, 3_91_86, 19_38, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Any = tokenizer(lowerCAmelCase__ ,max_length=len(lowerCAmelCase__ ) ,padding=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual((2, 9) ,batch.input_ids.shape ) self.assertEqual((2, 9) ,batch.attention_mask.shape ) lowerCAmelCase_ : int = batch.input_ids.tolist()[0] self.assertListEqual(lowerCAmelCase__ ,lowerCAmelCase__ ) @require_torch def UpperCAmelCase_ ( self : Dict ) -> Any: '''simple docstring''' lowerCAmelCase_ : int = ["A long paragraph for summarization.", "Another paragraph for summarization."] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[Any] = tokenizer(lowerCAmelCase__ ,padding=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIn("input_ids" ,lowerCAmelCase__ ) self.assertIn("attention_mask" ,lowerCAmelCase__ ) self.assertNotIn("labels" ,lowerCAmelCase__ ) self.assertNotIn("decoder_attention_mask" ,lowerCAmelCase__ ) @require_torch def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : int = [ "Summary of the text.", "Another summary.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[int] = tokenizer(text_target=lowerCAmelCase__ ,max_length=32 ,padding="max_length" ,return_tensors="pt" ) self.assertEqual(32 ,targets["input_ids"].shape[1] ) @require_torch def UpperCAmelCase_ ( self : Tuple ) -> List[str]: '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Tuple = tokenizer( ["I am a small frog" * 10_24, "I am a small frog"] ,padding=lowerCAmelCase__ ,truncation=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual(batch.input_ids.shape ,(2, 51_22) ) @require_torch def UpperCAmelCase_ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Tuple = ["A long paragraph for summarization."] lowerCAmelCase_ : Dict = [ "Summary of the text.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[Any] = tokenizer(lowerCAmelCase__ ,return_tensors="pt" ) lowerCAmelCase_ : Optional[Any] = tokenizer(text_target=lowerCAmelCase__ ,return_tensors="pt" ) lowerCAmelCase_ : List[str] = inputs["input_ids"] lowerCAmelCase_ : Any = targets["input_ids"] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) @require_torch def UpperCAmelCase_ ( self : str ) -> Tuple: '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : str = ["Summary of the text.", "Another summary."] lowerCAmelCase_ : str = [[0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, -1, -1]] lowerCAmelCase_ : List[Any] = tokenizer(lowerCAmelCase__ ,padding=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = [[0] * len(lowerCAmelCase__ ) for x in encoded_output["input_ids"]] lowerCAmelCase_ : Optional[int] = tokenizer.pad(lowerCAmelCase__ ) self.assertSequenceEqual(outputs["global_attention_mask"] ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' pass def UpperCAmelCase_ ( self : str ) -> Union[str, Any]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCAmelCase_ : Dict = self.rust_tokenizer_class.from_pretrained(lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = self.tokenizer_class.from_pretrained(lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Dict = "A, <mask> AllenNLP sentence." lowerCAmelCase_ : Tuple = tokenizer_r.encode_plus(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,return_token_type_ids=lowerCAmelCase__ ) lowerCAmelCase_ : int = tokenizer_p.encode_plus(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,return_token_type_ids=lowerCAmelCase__ ) self.assertEqual(sum(tokens_r["token_type_ids"] ) ,sum(tokens_p["token_type_ids"] ) ) self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) ,sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) ,) lowerCAmelCase_ : Any = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) lowerCAmelCase_ : Union[str, Any] = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) self.assertSequenceEqual(tokens_p["input_ids"] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( lowerCAmelCase__ ,["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( lowerCAmelCase__ ,["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] )
659
1
from __future__ import annotations from collections.abc import Callable def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1_00 , ): lowerCAmelCase_ : Any = x_start lowerCAmelCase_ : Optional[Any] = fnc(snake_case__) lowerCAmelCase_ : Union[str, Any] = 0.0 for _ in range(snake_case__): # Approximates small segments of curve as linear and solve # for trapezoidal area lowerCAmelCase_ : Any = (x_end - x_start) / steps + xa lowerCAmelCase_ : Dict = fnc(snake_case__) area += abs(fxa + fxa) * (xa - xa) / 2 # Increment step lowerCAmelCase_ : int = xa lowerCAmelCase_ : str = fxa return area if __name__ == "__main__": def UpperCamelCase ( snake_case__): return x**3 + x**2 print('''f(x) = x^3 + x^2''') print('''The area between the curve, x = -5, x = 5 and the x axis is:''') _lowercase = 10 while i <= 100000: print(f"with {i} steps: {trapezoidal_area(f, -5, 5, i)}") i *= 10
659
from ....configuration_utils import PretrainedConfig from ....utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''Visual-Attention-Network/van-base''': ( '''https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json''' ), } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = 'van' def __init__( self : List[str] ,lowerCAmelCase__ : int=2_24 ,lowerCAmelCase__ : Optional[int]=3 ,lowerCAmelCase__ : Dict=[7, 3, 3, 3] ,lowerCAmelCase__ : List[str]=[4, 2, 2, 2] ,lowerCAmelCase__ : Union[str, Any]=[64, 1_28, 3_20, 5_12] ,lowerCAmelCase__ : Union[str, Any]=[3, 3, 12, 3] ,lowerCAmelCase__ : Any=[8, 8, 4, 4] ,lowerCAmelCase__ : Optional[int]="gelu" ,lowerCAmelCase__ : List[str]=0.02 ,lowerCAmelCase__ : Optional[Any]=1e-6 ,lowerCAmelCase__ : Dict=1e-2 ,lowerCAmelCase__ : Union[str, Any]=0.0 ,lowerCAmelCase__ : Optional[Any]=0.0 ,**lowerCAmelCase__ : List[str] ,) -> Tuple: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = image_size lowerCAmelCase_ : List[str] = num_channels lowerCAmelCase_ : str = patch_sizes lowerCAmelCase_ : Optional[Any] = strides lowerCAmelCase_ : List[Any] = hidden_sizes lowerCAmelCase_ : int = depths lowerCAmelCase_ : int = mlp_ratios lowerCAmelCase_ : str = hidden_act lowerCAmelCase_ : List[str] = initializer_range lowerCAmelCase_ : Dict = layer_norm_eps lowerCAmelCase_ : str = layer_scale_init_value lowerCAmelCase_ : Tuple = drop_path_rate lowerCAmelCase_ : Dict = dropout_rate
659
1
import argparse import struct import unittest class __snake_case : """simple docstring""" def __init__( self : Union[str, Any] ,lowerCAmelCase__ : bytes ) -> None: '''simple docstring''' lowerCAmelCase_ : Tuple = data # Initialize hash values lowerCAmelCase_ : int = [ 0X6a09e667, 0Xbb67ae85, 0X3c6ef372, 0Xa54ff53a, 0X510e527f, 0X9b05688c, 0X1f83d9ab, 0X5be0cd19, ] # Initialize round constants lowerCAmelCase_ : Tuple = [ 0X428a2f98, 0X71374491, 0Xb5c0fbcf, 0Xe9b5dba5, 0X3956c25b, 0X59f111f1, 0X923f82a4, 0Xab1c5ed5, 0Xd807aa98, 0X12835b01, 0X243185be, 0X550c7dc3, 0X72be5d74, 0X80deb1fe, 0X9bdc06a7, 0Xc19bf174, 0Xe49b69c1, 0Xefbe4786, 0X0fc19dc6, 0X240ca1cc, 0X2de92c6f, 0X4a7484aa, 0X5cb0a9dc, 0X76f988da, 0X983e5152, 0Xa831c66d, 0Xb00327c8, 0Xbf597fc7, 0Xc6e00bf3, 0Xd5a79147, 0X06ca6351, 0X14292967, 0X27b70a85, 0X2e1b2138, 0X4d2c6dfc, 0X53380d13, 0X650a7354, 0X766a0abb, 0X81c2c92e, 0X92722c85, 0Xa2bfe8a1, 0Xa81a664b, 0Xc24b8b70, 0Xc76c51a3, 0Xd192e819, 0Xd6990624, 0Xf40e3585, 0X106aa070, 0X19a4c116, 0X1e376c08, 0X2748774c, 0X34b0bcb5, 0X391c0cb3, 0X4ed8aa4a, 0X5b9cca4f, 0X682e6ff3, 0X748f82ee, 0X78a5636f, 0X84c87814, 0X8cc70208, 0X90befffa, 0Xa4506ceb, 0Xbef9a3f7, 0Xc67178f2, ] lowerCAmelCase_ : List[Any] = self.preprocessing(self.data ) self.final_hash() @staticmethod def UpperCAmelCase_ ( lowerCAmelCase__ : bytes ) -> bytes: '''simple docstring''' lowerCAmelCase_ : List[Any] = B"\x80" + (B"\x00" * (63 - (len(lowerCAmelCase__ ) + 8) % 64)) lowerCAmelCase_ : List[Any] = struct.pack(">Q" ,(len(lowerCAmelCase__ ) * 8) ) return data + padding + big_endian_integer def UpperCAmelCase_ ( self : int ) -> None: '''simple docstring''' lowerCAmelCase_ : Dict = [ self.preprocessed_data[x : x + 64] for x in range(0 ,len(self.preprocessed_data ) ,64 ) ] for block in self.blocks: # Convert the given block into a list of 4 byte integers lowerCAmelCase_ : Optional[int] = list(struct.unpack(">16L" ,lowerCAmelCase__ ) ) # add 48 0-ed integers words += [0] * 48 lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = self.hashes for index in range(0 ,64 ): if index > 15: # modify the zero-ed indexes at the end of the array lowerCAmelCase_ : Any = ( self.ror(words[index - 15] ,7 ) ^ self.ror(words[index - 15] ,18 ) ^ (words[index - 15] >> 3) ) lowerCAmelCase_ : Optional[int] = ( self.ror(words[index - 2] ,17 ) ^ self.ror(words[index - 2] ,19 ) ^ (words[index - 2] >> 10) ) lowerCAmelCase_ : Dict = ( words[index - 16] + sa + words[index - 7] + sa ) % 0X100000000 # Compression lowerCAmelCase_ : Union[str, Any] = self.ror(lowerCAmelCase__ ,6 ) ^ self.ror(lowerCAmelCase__ ,11 ) ^ self.ror(lowerCAmelCase__ ,25 ) lowerCAmelCase_ : List[Any] = (e & f) ^ ((~e & 0Xffffffff) & g) lowerCAmelCase_ : Optional[Any] = ( h + sa + ch + self.round_constants[index] + words[index] ) % 0X100000000 lowerCAmelCase_ : Tuple = self.ror(lowerCAmelCase__ ,2 ) ^ self.ror(lowerCAmelCase__ ,13 ) ^ self.ror(lowerCAmelCase__ ,22 ) lowerCAmelCase_ : Dict = (a & b) ^ (a & c) ^ (b & c) lowerCAmelCase_ : Optional[int] = (sa + maj) % 0X100000000 lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = ( g, f, e, ((d + tempa) % 0X100000000), c, b, a, ((tempa + tempa) % 0X100000000), ) lowerCAmelCase_ : List[str] = [a, b, c, d, e, f, g, h] # Modify final values lowerCAmelCase_ : Union[str, Any] = [ ((element + mutated_hash_values[index]) % 0X100000000) for index, element in enumerate(self.hashes ) ] lowerCAmelCase_ : List[Any] = "".join([hex(lowerCAmelCase__ )[2:].zfill(8 ) for value in self.hashes] ) def UpperCAmelCase_ ( self : Optional[int] ,lowerCAmelCase__ : int ,lowerCAmelCase__ : int ) -> int: '''simple docstring''' return 0Xffffffff & (value << (32 - rotations)) | (value >> rotations) class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : int ) -> None: '''simple docstring''' import hashlib lowerCAmelCase_ : Optional[int] = bytes("Test String" ,"utf-8" ) self.assertEqual(SHAaaa(lowerCAmelCase__ ).hash ,hashlib.shaaaa(lowerCAmelCase__ ).hexdigest() ) def UpperCamelCase ( ): import doctest doctest.testmod() lowerCAmelCase_ : str = argparse.ArgumentParser() parser.add_argument( "-s" , "--string" , dest="input_string" , default="Hello World!! Welcome to Cryptography" , help="Hash the string" , ) parser.add_argument( "-f" , "--file" , dest="input_file" , help="Hash contents of a file") lowerCAmelCase_ : Dict = parser.parse_args() lowerCAmelCase_ : Any = args.input_string # hash input should be a bytestring if args.input_file: with open(args.input_file , "rb") as f: lowerCAmelCase_ : Dict = f.read() else: lowerCAmelCase_ : Optional[int] = bytes(snake_case__ , "utf-8") print(SHAaaa(snake_case__).hash) if __name__ == "__main__": main()
659
from math import factorial def UpperCamelCase ( snake_case__ , snake_case__): # If either of the conditions are true, the function is being asked # to calculate a factorial of a negative number, which is not possible if n < k or k < 0: raise ValueError("Please enter positive integers for n and k where n >= k") return factorial(snake_case__) // (factorial(snake_case__) * factorial(n - k)) if __name__ == "__main__": print( '''The number of five-card hands possible from a standard''', f"fifty-two card deck is: {combinations(52, 5)}\n", ) print( '''If a class of 40 students must be arranged into groups of''', f"4 for group projects, there are {combinations(40, 4)} ways", '''to arrange them.\n''', ) print( '''If 10 teams are competing in a Formula One race, there''', f"are {combinations(10, 3)} ways that first, second and", '''third place can be awarded.''', )
659
1
import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): _lowercase = True from torch.cuda.amp import autocast _lowercase = logging.getLogger(__name__) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Whether to log verbose messages or not.'} , ) UpperCamelCase_ = field( default=2.0 , metadata={'help': 'Maximum temperature for gumbel softmax.'} ) UpperCamelCase_ = field( default=0.5 , metadata={'help': 'Minimum temperature for gumbel softmax.'} ) UpperCamelCase_ = field( default=0.99_99_95 , metadata={'help': 'Decay of gumbel temperature during training.'} ) def UpperCamelCase ( snake_case__ , snake_case__): logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout)] , ) lowerCAmelCase_ : str = logging.WARNING if model_args.verbose_logging: lowerCAmelCase_ : int = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank): lowerCAmelCase_ : Any = logging.INFO logger.setLevel(snake_case__) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The name of the dataset to use (via the datasets library).'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) UpperCamelCase_ = field( default='train' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) UpperCamelCase_ = field( default='validation' , metadata={ 'help': ( 'The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'' ) } , ) UpperCamelCase_ = field( default='file' , metadata={'help': 'Column in the dataset that contains speech file path. Defaults to \'file\''} , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Overwrite the cached preprocessed datasets or not.'} ) UpperCamelCase_ = field( default=1 , metadata={ 'help': 'The percentage of the train set used as validation set in case there\'s no validation split' } , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) UpperCamelCase_ = field( default=20.0 , metadata={'help': 'Filter audio files that are longer than `max_duration_in_seconds` seconds'} ) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = 42 UpperCamelCase_ = 42 UpperCamelCase_ = "longest" UpperCamelCase_ = None UpperCamelCase_ = None def __call__( self : str ,lowerCAmelCase__ : List[Dict[str, Union[List[int], torch.Tensor]]] ) -> Dict[str, torch.Tensor]: '''simple docstring''' lowerCAmelCase_ : Tuple = self.feature_extractor.pad( lowerCAmelCase__ ,max_length=self.max_length ,padding=self.padding ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors="pt" ,) lowerCAmelCase_ : Union[str, Any] = self.model._get_feat_extract_output_lengths(batch["input_values"].shape[-1] ) lowerCAmelCase_ : List[str] = batch["input_values"].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula lowerCAmelCase_ : Tuple = self.model._get_feat_extract_output_lengths(batch["attention_mask"].sum(-1 ) ).to( torch.long ) lowerCAmelCase_ : Optional[Any] = torch.zeros( (batch_size, mask_indices_seq_length) ,dtype=torch.long ,device=batch["input_values"].device ) # these two operations makes sure that all values # before the output lengths indices are attended to lowerCAmelCase_ : Tuple = 1 lowerCAmelCase_ : int = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices lowerCAmelCase_ : str = _compute_mask_indices( (batch_size, mask_indices_seq_length) ,self.model.config.mask_time_prob ,self.model.config.mask_time_length ,attention_mask=lowerCAmelCase__ ,min_masks=2 ,) return batch class __snake_case ( snake_case__ ): """simple docstring""" def __init__( self : List[str] ,*lowerCAmelCase__ : Optional[int] ,lowerCAmelCase__ : Tuple=1 ,lowerCAmelCase__ : Optional[int]=0 ,lowerCAmelCase__ : Optional[Any]=1.0 ,**lowerCAmelCase__ : Any ) -> str: '''simple docstring''' super().__init__(*lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = 0 lowerCAmelCase_ : int = max_gumbel_temp lowerCAmelCase_ : Union[str, Any] = min_gumbel_temp lowerCAmelCase_ : str = gumbel_temp_decay def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : nn.Module ,lowerCAmelCase__ : Dict[str, Union[torch.Tensor, Any]] ) -> torch.Tensor: '''simple docstring''' model.train() lowerCAmelCase_ : str = self._prepare_inputs(lowerCAmelCase__ ) if self.use_amp: with autocast(): lowerCAmelCase_ : List[Any] = self.compute_loss(lowerCAmelCase__ ,lowerCAmelCase__ ) else: lowerCAmelCase_ : List[Any] = self.compute_loss(lowerCAmelCase__ ,lowerCAmelCase__ ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": lowerCAmelCase_ : List[Any] = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": lowerCAmelCase_ : Optional[Any] = loss.sum() / (inputs["mask_time_indices"]).sum() else: raise ValueError(f'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: lowerCAmelCase_ : int = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(lowerCAmelCase__ ).backward() elif self.use_apex: with amp.scale_loss(lowerCAmelCase__ ,self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(lowerCAmelCase__ ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) return loss.detach() def UpperCamelCase ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase_ : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments)) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = parser.parse_args_into_dataclasses() configure_logger(snake_case__ , snake_case__) # Downloading and loading a dataset from the hub. lowerCAmelCase_ : List[str] = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" lowerCAmelCase_ : Any = DatasetDict() lowerCAmelCase_ : Union[str, Any] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : List[str] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" lowerCAmelCase_ : Union[str, Any] = DatasetDict() lowerCAmelCase_ : int = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split="validation" , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : Any = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported lowerCAmelCase_ : Dict = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=snake_case__) def prepare_dataset(snake_case__): # check that all files have the correct sampling rate lowerCAmelCase_ , lowerCAmelCase_ : str = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate) return batch # load audio files into numpy arrays lowerCAmelCase_ : int = datasets.map( snake_case__ , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets["train"].column_names) # filter audio files that are too long lowerCAmelCase_ : int = vectorized_datasets.filter( lambda snake_case__: len(data["speech"]) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate)) def normalize(snake_case__): return feature_extractor(batch["speech"] , sampling_rate=feature_extractor.sampling_rate) # normalize and transform to `BatchFeatures` lowerCAmelCase_ : str = vectorized_datasets.map( snake_case__ , batched=snake_case__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets["train"].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 lowerCAmelCase_ : Optional[Any] = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( "PreTraining is only supported for ``config.do_stable_layer_norm=True`` and" " ``config.feat_extract_norm='layer'") lowerCAmelCase_ : Dict = WavaVecaForPreTraining(snake_case__) lowerCAmelCase_ : int = DataCollatorForWavaVecaPretraining(model=snake_case__ , feature_extractor=snake_case__) lowerCAmelCase_ : List[Any] = WavaVecaPreTrainer( model=snake_case__ , data_collator=snake_case__ , args=snake_case__ , train_dataset=vectorized_datasets["train"] , eval_dataset=vectorized_datasets["validation"] , tokenizer=snake_case__ , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
659
import argparse import json from tqdm import tqdm def UpperCamelCase ( ): lowerCAmelCase_ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--src_path" , type=snake_case__ , default="biencoder-nq-dev.json" , help="Path to raw DPR training data" , ) parser.add_argument( "--evaluation_set" , type=snake_case__ , help="where to store parsed evaluation_set file" , ) parser.add_argument( "--gold_data_path" , type=snake_case__ , help="where to store parsed gold_data_path file" , ) lowerCAmelCase_ : Dict = parser.parse_args() with open(args.src_path , "r") as src_file, open(args.evaluation_set , "w") as eval_file, open( args.gold_data_path , "w") as gold_file: lowerCAmelCase_ : Optional[int] = json.load(snake_case__) for dpr_record in tqdm(snake_case__): lowerCAmelCase_ : str = dpr_record["question"] lowerCAmelCase_ : Dict = [context["title"] for context in dpr_record["positive_ctxs"]] eval_file.write(question + "\n") gold_file.write("\t".join(snake_case__) + "\n") if __name__ == "__main__": main()
659
1
from PIL import Image def UpperCamelCase ( snake_case__ , snake_case__): def brightness(snake_case__) -> float: return 1_28 + level + (c - 1_28) if not -255.0 <= level <= 255.0: raise ValueError("level must be between -255.0 (black) and 255.0 (white)") return img.point(snake_case__) if __name__ == "__main__": # Load image with Image.open('''image_data/lena.jpg''') as img: # Change brightness to 100 _lowercase = change_brightness(img, 100) brigt_img.save('''image_data/lena_brightness.png''', format='''png''')
659
from collections.abc import Sequence def UpperCamelCase ( snake_case__ = None): if nums is None or not nums: raise ValueError("Input sequence should not be empty") lowerCAmelCase_ : Dict = nums[0] for i in range(1 , len(snake_case__)): lowerCAmelCase_ : Optional[int] = nums[i] lowerCAmelCase_ : Optional[int] = max(snake_case__ , ans + num , snake_case__) return ans if __name__ == "__main__": import doctest doctest.testmod() # Try on a sample input from the user _lowercase = int(input('''Enter number of elements : ''').strip()) _lowercase = list(map(int, input('''\nEnter the numbers : ''').strip().split()))[:n] print(max_subsequence_sum(array))
659
1
def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): def update_area_of_max_square(snake_case__ , snake_case__) -> int: # BASE CASE if row >= rows or col >= cols: return 0 lowerCAmelCase_ : Tuple = update_area_of_max_square(snake_case__ , col + 1) lowerCAmelCase_ : Optional[int] = update_area_of_max_square(row + 1 , col + 1) lowerCAmelCase_ : Optional[int] = update_area_of_max_square(row + 1 , snake_case__) if mat[row][col]: lowerCAmelCase_ : List[str] = 1 + min([right, diagonal, down]) lowerCAmelCase_ : Tuple = max(largest_square_area[0] , snake_case__) return sub_problem_sol else: return 0 lowerCAmelCase_ : str = [0] update_area_of_max_square(0 , 0) return largest_square_area[0] def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): def update_area_of_max_square_using_dp_array( snake_case__ , snake_case__ , snake_case__) -> int: if row >= rows or col >= cols: return 0 if dp_array[row][col] != -1: return dp_array[row][col] lowerCAmelCase_ : str = update_area_of_max_square_using_dp_array(snake_case__ , col + 1 , snake_case__) lowerCAmelCase_ : Dict = update_area_of_max_square_using_dp_array(row + 1 , col + 1 , snake_case__) lowerCAmelCase_ : Dict = update_area_of_max_square_using_dp_array(row + 1 , snake_case__ , snake_case__) if mat[row][col]: lowerCAmelCase_ : Tuple = 1 + min([right, diagonal, down]) lowerCAmelCase_ : List[Any] = max(largest_square_area[0] , snake_case__) lowerCAmelCase_ : int = sub_problem_sol return sub_problem_sol else: return 0 lowerCAmelCase_ : Tuple = [0] lowerCAmelCase_ : Union[str, Any] = [[-1] * cols for _ in range(snake_case__)] update_area_of_max_square_using_dp_array(0 , 0 , snake_case__) return largest_square_area[0] def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : Union[str, Any] = [[0] * (cols + 1) for _ in range(rows + 1)] lowerCAmelCase_ : str = 0 for row in range(rows - 1 , -1 , -1): for col in range(cols - 1 , -1 , -1): lowerCAmelCase_ : Union[str, Any] = dp_array[row][col + 1] lowerCAmelCase_ : int = dp_array[row + 1][col + 1] lowerCAmelCase_ : Union[str, Any] = dp_array[row + 1][col] if mat[row][col] == 1: lowerCAmelCase_ : Union[str, Any] = 1 + min(snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : Union[str, Any] = max(dp_array[row][col] , snake_case__) else: lowerCAmelCase_ : Optional[int] = 0 return largest_square_area def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : int = [0] * (cols + 1) lowerCAmelCase_ : Tuple = [0] * (cols + 1) lowerCAmelCase_ : str = 0 for row in range(rows - 1 , -1 , -1): for col in range(cols - 1 , -1 , -1): lowerCAmelCase_ : Optional[Any] = current_row[col + 1] lowerCAmelCase_ : List[Any] = next_row[col + 1] lowerCAmelCase_ : Union[str, Any] = next_row[col] if mat[row][col] == 1: lowerCAmelCase_ : Optional[int] = 1 + min(snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : List[str] = max(current_row[col] , snake_case__) else: lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : Optional[Any] = current_row return largest_square_area if __name__ == "__main__": import doctest doctest.testmod() print(largest_square_area_in_matrix_bottom_up(2, 2, [[1, 1], [1, 1]]))
659
from typing import TYPE_CHECKING from ....utils import _LazyModule _lowercase = {'''tokenization_tapex''': ['''TapexTokenizer''']} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
659
1
import qiskit def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : List[Any] = qiskit.Aer.get_backend("aer_simulator") lowerCAmelCase_ : List[str] = qiskit.QuantumCircuit(4 , 2) # encode inputs in qubits 0 and 1 if bita == 1: qc_ha.x(0) if bita == 1: qc_ha.x(1) qc_ha.barrier() # use cnots to write XOR of the inputs on qubit2 qc_ha.cx(0 , 2) qc_ha.cx(1 , 2) # use ccx / toffoli gate to write AND of the inputs on qubit3 qc_ha.ccx(0 , 1 , 3) qc_ha.barrier() # extract outputs qc_ha.measure(2 , 0) # extract XOR value qc_ha.measure(3 , 1) # extract AND value # Execute the circuit on the qasm simulator lowerCAmelCase_ : int = qiskit.execute(snake_case__ , snake_case__ , shots=10_00) # Return the histogram data of the results of the experiment return job.result().get_counts(snake_case__) if __name__ == "__main__": _lowercase = half_adder(1, 1) print(f"Half Adder Output Qubit Counts: {counts}")
659
import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _lowercase = '''src/diffusers''' _lowercase = '''.''' # This is to make sure the diffusers module imported is the one in the repo. _lowercase = importlib.util.spec_from_file_location( '''diffusers''', os.path.join(DIFFUSERS_PATH, '''__init__.py'''), submodule_search_locations=[DIFFUSERS_PATH], ) _lowercase = spec.loader.load_module() def UpperCamelCase ( snake_case__ , snake_case__): return line.startswith(snake_case__) or len(snake_case__) <= 1 or re.search(R"^\s*\)(\s*->.*:|:)\s*$" , snake_case__) is not None def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Tuple = object_name.split(".") lowerCAmelCase_ : Union[str, Any] = 0 # First let's find the module where our object lives. lowerCAmelCase_ : Union[str, Any] = parts[i] while i < len(snake_case__) and not os.path.isfile(os.path.join(snake_case__ , F'''{module}.py''')): i += 1 if i < len(snake_case__): lowerCAmelCase_ : Dict = os.path.join(snake_case__ , parts[i]) if i >= len(snake_case__): raise ValueError(F'''`object_name` should begin with the name of a module of diffusers but got {object_name}.''') with open(os.path.join(snake_case__ , F'''{module}.py''') , "r" , encoding="utf-8" , newline="\n") as f: lowerCAmelCase_ : Optional[Any] = f.readlines() # Now let's find the class / func in the code! lowerCAmelCase_ : Union[str, Any] = "" lowerCAmelCase_ : int = 0 for name in parts[i + 1 :]: while ( line_index < len(snake_case__) and re.search(RF'''^{indent}(class|def)\s+{name}(\(|\:)''' , lines[line_index]) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(snake_case__): raise ValueError(F''' {object_name} does not match any function or class in {module}.''') # We found the beginning of the class / func, now let's find the end (when the indent diminishes). lowerCAmelCase_ : Union[str, Any] = line_index while line_index < len(snake_case__) and _should_continue(lines[line_index] , snake_case__): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1]) <= 1: line_index -= 1 lowerCAmelCase_ : List[str] = lines[start_index:line_index] return "".join(snake_case__) _lowercase = re.compile(r'''^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)''') _lowercase = re.compile(r'''^\s*(\S+)->(\S+)(\s+.*|$)''') _lowercase = re.compile(r'''<FILL\s+[^>]*>''') def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Any = code.split("\n") lowerCAmelCase_ : Any = 0 while idx < len(snake_case__) and len(lines[idx]) == 0: idx += 1 if idx < len(snake_case__): return re.search(R"^(\s*)\S" , lines[idx]).groups()[0] return "" def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Dict = len(get_indent(snake_case__)) > 0 if has_indent: lowerCAmelCase_ : Dict = F'''class Bla:\n{code}''' lowerCAmelCase_ : Optional[int] = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=1_19 , preview=snake_case__) lowerCAmelCase_ : Optional[Any] = black.format_str(snake_case__ , mode=snake_case__) lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = style_docstrings_in_code(snake_case__) return result[len("class Bla:\n") :] if has_indent else result def UpperCamelCase ( snake_case__ , snake_case__=False): with open(snake_case__ , "r" , encoding="utf-8" , newline="\n") as f: lowerCAmelCase_ : Tuple = f.readlines() lowerCAmelCase_ : Tuple = [] lowerCAmelCase_ : Union[str, Any] = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(snake_case__): lowerCAmelCase_ : Optional[int] = _re_copy_warning.search(lines[line_index]) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = search.groups() lowerCAmelCase_ : int = find_code_in_diffusers(snake_case__) lowerCAmelCase_ : Dict = get_indent(snake_case__) lowerCAmelCase_ : Union[str, Any] = line_index + 1 if indent == theoretical_indent else line_index + 2 lowerCAmelCase_ : str = theoretical_indent lowerCAmelCase_ : Union[str, Any] = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. lowerCAmelCase_ : Optional[int] = True while line_index < len(snake_case__) and should_continue: line_index += 1 if line_index >= len(snake_case__): break lowerCAmelCase_ : Dict = lines[line_index] lowerCAmelCase_ : List[str] = _should_continue(snake_case__ , snake_case__) and re.search(F'''^{indent}# End copy''' , snake_case__) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1]) <= 1: line_index -= 1 lowerCAmelCase_ : Dict = lines[start_index:line_index] lowerCAmelCase_ : Optional[int] = "".join(snake_case__) # Remove any nested `Copied from` comments to avoid circular copies lowerCAmelCase_ : List[Any] = [line for line in theoretical_code.split("\n") if _re_copy_warning.search(snake_case__) is None] lowerCAmelCase_ : Optional[Any] = "\n".join(snake_case__) # Before comparing, use the `replace_pattern` on the original code. if len(snake_case__) > 0: lowerCAmelCase_ : List[str] = replace_pattern.replace("with" , "").split(",") lowerCAmelCase_ : Tuple = [_re_replace_pattern.search(snake_case__) for p in patterns] for pattern in patterns: if pattern is None: continue lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[str] = pattern.groups() lowerCAmelCase_ : int = re.sub(snake_case__ , snake_case__ , snake_case__) if option.strip() == "all-casing": lowerCAmelCase_ : List[str] = re.sub(obja.lower() , obja.lower() , snake_case__) lowerCAmelCase_ : int = re.sub(obja.upper() , obja.upper() , snake_case__) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line lowerCAmelCase_ : List[Any] = blackify(lines[start_index - 1] + theoretical_code) lowerCAmelCase_ : Union[str, Any] = theoretical_code[len(lines[start_index - 1]) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index]) if overwrite: lowerCAmelCase_ : List[Any] = lines[:start_index] + [theoretical_code] + lines[line_index:] lowerCAmelCase_ : Union[str, Any] = start_index + 1 if overwrite and len(snake_case__) > 0: # Warn the user a file has been modified. print(F'''Detected changes, rewriting {filename}.''') with open(snake_case__ , "w" , encoding="utf-8" , newline="\n") as f: f.writelines(snake_case__) return diffs def UpperCamelCase ( snake_case__ = False): lowerCAmelCase_ : Tuple = glob.glob(os.path.join(snake_case__ , "**/*.py") , recursive=snake_case__) lowerCAmelCase_ : int = [] for filename in all_files: lowerCAmelCase_ : Union[str, Any] = is_copy_consistent(snake_case__ , snake_case__) diffs += [F'''- {filename}: copy does not match {d[0]} at line {d[1]}''' for d in new_diffs] if not overwrite and len(snake_case__) > 0: lowerCAmelCase_ : Optional[Any] = "\n".join(snake_case__) raise Exception( "Found the following copy inconsistencies:\n" + diff + "\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.") if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') _lowercase = parser.parse_args() check_copies(args.fix_and_overwrite)
659
1
import os from typing import Dict, List, Tuple, TypeVar, Union _lowercase = TypeVar('''T''') _lowercase = Union[List[T], Tuple[T, ...]] _lowercase = Union[T, List[T], Dict[str, T]] _lowercase = Union[str, bytes, os.PathLike]
659
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''microsoft/swinv2-tiny-patch4-window8-256''': ( '''https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json''' ), } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = 'swinv2' UpperCamelCase_ = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : List[Any] ,lowerCAmelCase__ : Optional[int]=2_24 ,lowerCAmelCase__ : Dict=4 ,lowerCAmelCase__ : Dict=3 ,lowerCAmelCase__ : List[Any]=96 ,lowerCAmelCase__ : Optional[Any]=[2, 2, 6, 2] ,lowerCAmelCase__ : Optional[Any]=[3, 6, 12, 24] ,lowerCAmelCase__ : Optional[int]=7 ,lowerCAmelCase__ : Dict=4.0 ,lowerCAmelCase__ : Dict=True ,lowerCAmelCase__ : str=0.0 ,lowerCAmelCase__ : Tuple=0.0 ,lowerCAmelCase__ : str=0.1 ,lowerCAmelCase__ : List[str]="gelu" ,lowerCAmelCase__ : Union[str, Any]=False ,lowerCAmelCase__ : Dict=0.02 ,lowerCAmelCase__ : int=1e-5 ,lowerCAmelCase__ : List[str]=32 ,**lowerCAmelCase__ : Tuple ,) -> List[str]: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = image_size lowerCAmelCase_ : List[Any] = patch_size lowerCAmelCase_ : Dict = num_channels lowerCAmelCase_ : Optional[int] = embed_dim lowerCAmelCase_ : Optional[Any] = depths lowerCAmelCase_ : Any = len(lowerCAmelCase__ ) lowerCAmelCase_ : str = num_heads lowerCAmelCase_ : List[str] = window_size lowerCAmelCase_ : List[str] = mlp_ratio lowerCAmelCase_ : Dict = qkv_bias lowerCAmelCase_ : str = hidden_dropout_prob lowerCAmelCase_ : str = attention_probs_dropout_prob lowerCAmelCase_ : Union[str, Any] = drop_path_rate lowerCAmelCase_ : List[Any] = hidden_act lowerCAmelCase_ : Any = use_absolute_embeddings lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : int = initializer_range lowerCAmelCase_ : Union[str, Any] = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase_ : Tuple = int(embed_dim * 2 ** (len(lowerCAmelCase__ ) - 1) ) lowerCAmelCase_ : str = (0, 0, 0, 0)
659
1
import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class __snake_case ( snake_case__ ): """simple docstring""" def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : str ) -> Union[str, Any]: '''simple docstring''' with open(lowerCAmelCase__ ,encoding="utf-8" ) as input_file: lowerCAmelCase_ : List[str] = re.compile(R"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) lowerCAmelCase_ : Tuple = input_file.read() lowerCAmelCase_ : Dict = regexp.search(lowerCAmelCase__ ) return match def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : str ) -> Optional[int]: '''simple docstring''' with open(lowerCAmelCase__ ,encoding="utf-8" ) as input_file: lowerCAmelCase_ : Union[str, Any] = re.compile(R"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" ,re.DOTALL ) lowerCAmelCase_ : str = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` lowerCAmelCase_ : Optional[Any] = regexp.finditer(lowerCAmelCase__ ) lowerCAmelCase_ : int = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def UpperCAmelCase_ ( self : Optional[int] ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : List[str] = Path("./datasets" ) lowerCAmelCase_ : Optional[int] = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(lowerCAmelCase__ ) ): raise AssertionError(f'''open(...) must use utf-8 encoding in {dataset}''' ) def UpperCAmelCase_ ( self : Any ) -> Dict: '''simple docstring''' lowerCAmelCase_ : int = Path("./datasets" ) lowerCAmelCase_ : Any = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(lowerCAmelCase__ ) ): raise AssertionError(f'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
659
from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging _lowercase = logging.get_logger(__name__) class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = ['input_features', 'attention_mask'] def __init__( self : Optional[Any] ,lowerCAmelCase__ : Any=80 ,lowerCAmelCase__ : Optional[Any]=1_60_00 ,lowerCAmelCase__ : List[str]=0.0 ,lowerCAmelCase__ : Tuple=10 ,lowerCAmelCase__ : Optional[Any]=25 ,lowerCAmelCase__ : Any="hamming_window" ,lowerCAmelCase__ : List[str]=32_768.0 ,lowerCAmelCase__ : Union[str, Any]=0.97 ,lowerCAmelCase__ : Any=1.0 ,lowerCAmelCase__ : str=True ,lowerCAmelCase__ : int=True ,lowerCAmelCase__ : Tuple=False ,**lowerCAmelCase__ : Optional[int] ,) -> Optional[Any]: '''simple docstring''' super().__init__(feature_size=lowerCAmelCase__ ,sampling_rate=lowerCAmelCase__ ,padding_value=lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = feature_size lowerCAmelCase_ : List[Any] = sampling_rate lowerCAmelCase_ : Union[str, Any] = padding_value lowerCAmelCase_ : str = hop_length lowerCAmelCase_ : str = win_length lowerCAmelCase_ : str = frame_signal_scale lowerCAmelCase_ : Any = preemphasis_coeff lowerCAmelCase_ : Optional[Any] = mel_floor lowerCAmelCase_ : List[str] = normalize_means lowerCAmelCase_ : Optional[Any] = normalize_vars lowerCAmelCase_ : Dict = win_function lowerCAmelCase_ : List[Any] = return_attention_mask lowerCAmelCase_ : Tuple = win_length * sampling_rate // 10_00 lowerCAmelCase_ : str = hop_length * sampling_rate // 10_00 lowerCAmelCase_ : Dict = optimal_fft_length(self.sample_size ) lowerCAmelCase_ : Optional[int] = (self.n_fft // 2) + 1 def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : np.array ) -> np.ndarray: '''simple docstring''' if self.win_function == "hamming_window": lowerCAmelCase_ : int = window_function(window_length=self.sample_size ,name=self.win_function ,periodic=lowerCAmelCase__ ) else: lowerCAmelCase_ : Tuple = window_function(window_length=self.sample_size ,name=self.win_function ) lowerCAmelCase_ : List[str] = mel_filter_bank( num_frequency_bins=self.n_freqs ,num_mel_filters=self.feature_size ,min_frequency=0.0 ,max_frequency=self.sampling_rate / 2.0 ,sampling_rate=self.sampling_rate ,) lowerCAmelCase_ : Any = spectrogram( one_waveform * self.frame_signal_scale ,window=lowerCAmelCase__ ,frame_length=self.sample_size ,hop_length=self.sample_stride ,fft_length=self.n_fft ,center=lowerCAmelCase__ ,preemphasis=self.preemphasis_coeff ,mel_filters=lowerCAmelCase__ ,mel_floor=self.mel_floor ,log_mel="log" ,) return msfc_features.T def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : Tuple ) -> Optional[Any]: '''simple docstring''' if self.normalize_means: lowerCAmelCase_ : Optional[int] = x[:input_length].mean(axis=0 ) lowerCAmelCase_ : List[str] = np.subtract(lowerCAmelCase__ ,lowerCAmelCase__ ) if self.normalize_vars: lowerCAmelCase_ : Optional[Any] = x[:input_length].std(axis=0 ) lowerCAmelCase_ : Tuple = np.divide(lowerCAmelCase__ ,lowerCAmelCase__ ) if input_length < x.shape[0]: lowerCAmelCase_ : int = padding_value # make sure array is in float32 lowerCAmelCase_ : Any = x.astype(np.floataa ) return x def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[np.ndarray] ,lowerCAmelCase__ : Optional[np.ndarray] = None ) -> List[np.ndarray]: '''simple docstring''' lowerCAmelCase_ : List[Any] = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(lowerCAmelCase__ ,lowerCAmelCase__ ,self.padding_value ) for x, n in zip(lowerCAmelCase__ ,lowerCAmelCase__ )] def __call__( self : int ,lowerCAmelCase__ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,lowerCAmelCase__ : Union[bool, str, PaddingStrategy] = False ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : bool = False ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : Optional[bool] = None ,lowerCAmelCase__ : Optional[Union[str, TensorType]] = None ,lowerCAmelCase__ : Optional[int] = None ,**lowerCAmelCase__ : Union[str, Any] ,) -> BatchFeature: '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) lowerCAmelCase_ : List[Any] = isinstance(lowerCAmelCase__ ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase_ : str = is_batched_numpy or ( isinstance(lowerCAmelCase__ ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase_ : Tuple = [np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(lowerCAmelCase__ ,np.ndarray ): lowerCAmelCase_ : int = np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) elif isinstance(lowerCAmelCase__ ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase_ : Union[str, Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase_ : Optional[int] = [raw_speech] # extract fbank features lowerCAmelCase_ : Dict = [self._extract_mfsc_features(lowerCAmelCase__ ) for one_waveform in raw_speech] # convert into correct format for padding lowerCAmelCase_ : int = BatchFeature({"input_features": features} ) lowerCAmelCase_ : Union[str, Any] = self.pad( lowerCAmelCase__ ,padding=lowerCAmelCase__ ,max_length=lowerCAmelCase__ ,truncation=lowerCAmelCase__ ,pad_to_multiple_of=lowerCAmelCase__ ,return_attention_mask=lowerCAmelCase__ ,**lowerCAmelCase__ ,) # make sure list is in array format lowerCAmelCase_ : Optional[Any] = padded_inputs.get("input_features" ) if isinstance(input_features[0] ,lowerCAmelCase__ ): lowerCAmelCase_ : Optional[int] = [np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) for feature in input_features] lowerCAmelCase_ : List[Any] = padded_inputs.get("attention_mask" ) if attention_mask is not None: lowerCAmelCase_ : Dict = [np.asarray(lowerCAmelCase__ ,dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: lowerCAmelCase_ : Dict = ( np.array(lowerCAmelCase__ ,dtype=np.intaa ) if self._get_padding_strategies(lowerCAmelCase__ ,max_length=lowerCAmelCase__ ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) lowerCAmelCase_ : List[str] = self.normalize( padded_inputs["input_features"] ,attention_mask=lowerCAmelCase__ ) if return_tensors is not None: lowerCAmelCase_ : Dict = padded_inputs.convert_to_tensors(lowerCAmelCase__ ) return padded_inputs
659
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowercase = { '''configuration_lilt''': ['''LILT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LiltConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ '''LILT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LiltForQuestionAnswering''', '''LiltForSequenceClassification''', '''LiltForTokenClassification''', '''LiltModel''', '''LiltPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_lilt import LILT_PRETRAINED_CONFIG_ARCHIVE_MAP, LiltConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lilt import ( LILT_PRETRAINED_MODEL_ARCHIVE_LIST, LiltForQuestionAnswering, LiltForSequenceClassification, LiltForTokenClassification, LiltModel, LiltPreTrainedModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
659
from __future__ import annotations # This is the precision for this function which can be altered. # It is recommended for users to keep this number greater than or equal to 10. _lowercase = 10 def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): for i in range(snake_case__ , snake_case__): if array[i] == target: return i return -1 def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : List[str] = 0 lowerCAmelCase_ : Tuple = len(snake_case__) while left <= right: if right - left < precision: return lin_search(snake_case__ , snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : List[str] = (left + right) // 3 + 1 lowerCAmelCase_ : Tuple = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: lowerCAmelCase_ : str = one_third - 1 elif array[two_third] < target: lowerCAmelCase_ : Any = two_third + 1 else: lowerCAmelCase_ : List[str] = one_third + 1 lowerCAmelCase_ : Tuple = two_third - 1 else: return -1 def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): if left < right: if right - left < precision: return lin_search(snake_case__ , snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : Dict = (left + right) // 3 + 1 lowerCAmelCase_ : List[Any] = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: return rec_ternary_search(snake_case__ , one_third - 1 , snake_case__ , snake_case__) elif array[two_third] < target: return rec_ternary_search(two_third + 1 , snake_case__ , snake_case__ , snake_case__) else: return rec_ternary_search(one_third + 1 , two_third - 1 , snake_case__ , snake_case__) else: return -1 if __name__ == "__main__": import doctest doctest.testmod() _lowercase = input('''Enter numbers separated by comma:\n''').strip() _lowercase = [int(item.strip()) for item in user_input.split(''',''')] assert collection == sorted(collection), f"List must be ordered.\n{collection}." _lowercase = int(input('''Enter the number to be found in the list:\n''').strip()) _lowercase = ite_ternary_search(collection, target) _lowercase = rec_ternary_search(0, len(collection) - 1, collection, target) if resulta != -1: print(f"Iterative search: {target} found at positions: {resulta}") print(f"Recursive search: {target} found at positions: {resulta}") else: print('''Not found''')
659
1
import unittest import numpy as np import torch from diffusers import PNDMPipeline, PNDMScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class __snake_case ( unittest.TestCase ): """simple docstring""" @property def UpperCAmelCase_ ( self : Dict ) -> Optional[Any]: '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase_ : List[Any] = UNetaDModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=3 ,out_channels=3 ,down_block_types=("DownBlock2D", "AttnDownBlock2D") ,up_block_types=("AttnUpBlock2D", "UpBlock2D") ,) return model def UpperCAmelCase_ ( self : Optional[Any] ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : int = self.dummy_uncond_unet lowerCAmelCase_ : Optional[Any] = PNDMScheduler() lowerCAmelCase_ : Tuple = PNDMPipeline(unet=lowerCAmelCase__ ,scheduler=lowerCAmelCase__ ) pndm.to(lowerCAmelCase__ ) pndm.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = torch.manual_seed(0 ) lowerCAmelCase_ : List[str] = pndm(generator=lowerCAmelCase__ ,num_inference_steps=20 ,output_type="numpy" ).images lowerCAmelCase_ : int = torch.manual_seed(0 ) lowerCAmelCase_ : Tuple = pndm(generator=lowerCAmelCase__ ,num_inference_steps=20 ,output_type="numpy" ,return_dict=lowerCAmelCase__ )[0] lowerCAmelCase_ : Any = image[0, -3:, -3:, -1] lowerCAmelCase_ : Optional[int] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCAmelCase_ : Optional[Any] = np.array([1.0, 1.0, 0.0, 1.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Tuple = "google/ddpm-cifar10-32" lowerCAmelCase_ : Optional[Any] = UNetaDModel.from_pretrained(lowerCAmelCase__ ) lowerCAmelCase_ : Any = PNDMScheduler() lowerCAmelCase_ : str = PNDMPipeline(unet=lowerCAmelCase__ ,scheduler=lowerCAmelCase__ ) pndm.to(lowerCAmelCase__ ) pndm.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Any = torch.manual_seed(0 ) lowerCAmelCase_ : List[str] = pndm(generator=lowerCAmelCase__ ,output_type="numpy" ).images lowerCAmelCase_ : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCAmelCase_ : List[Any] = np.array([0.1_564, 0.14_645, 0.1_406, 0.14_715, 0.12_425, 0.14_045, 0.13_115, 0.12_175, 0.125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
659
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer _lowercase = logging.get_logger(__name__) _lowercase = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } _lowercase = { '''vocab_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json''' }, '''merges_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt''' }, '''tokenizer_config_file''': { '''facebook/blenderbot_small-90M''': ( '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json''' ) }, } _lowercase = { '''facebook/blenderbot_small-90M''': 512, } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = BlenderbotSmallTokenizer def __init__( self : Optional[int] ,lowerCAmelCase__ : Optional[int]=None ,lowerCAmelCase__ : Union[str, Any]=None ,lowerCAmelCase__ : Any="<|endoftext|>" ,lowerCAmelCase__ : int="<|endoftext|>" ,lowerCAmelCase__ : Optional[Any]="<|endoftext|>" ,lowerCAmelCase__ : Union[str, Any]=False ,lowerCAmelCase__ : Optional[Any]=True ,**lowerCAmelCase__ : Union[str, Any] ,) -> str: '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=lowerCAmelCase__ ,merges=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ,) ,bos_token=lowerCAmelCase__ ,eos_token=lowerCAmelCase__ ,unk_token=lowerCAmelCase__ ,**lowerCAmelCase__ ,) lowerCAmelCase_ : Dict = add_prefix_space def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Tuple=None ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : str = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : Dict = [self.sep_token_id] lowerCAmelCase_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
659
1
from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar _lowercase = TypeVar('''KEY''') _lowercase = TypeVar('''VAL''') @dataclass(frozen=snake_case__ , slots=snake_case__ ) class __snake_case ( Generic[KEY, VAL] ): """simple docstring""" UpperCamelCase_ = 42 UpperCamelCase_ = 42 class __snake_case ( _Item ): """simple docstring""" def __init__( self : int ) -> None: '''simple docstring''' super().__init__(lowerCAmelCase__ ,lowerCAmelCase__ ) def __bool__( self : Tuple ) -> bool: '''simple docstring''' return False _lowercase = _DeletedItem() class __snake_case ( MutableMapping[KEY, VAL] ): """simple docstring""" def __init__( self : Tuple ,lowerCAmelCase__ : int = 8 ,lowerCAmelCase__ : float = 0.75 ) -> None: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = initial_block_size lowerCAmelCase_ : list[_Item | None] = [None] * initial_block_size assert 0.0 < capacity_factor < 1.0 lowerCAmelCase_ : str = capacity_factor lowerCAmelCase_ : List[Any] = 0 def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : KEY ) -> int: '''simple docstring''' return hash(lowerCAmelCase__ ) % len(self._buckets ) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : int ) -> int: '''simple docstring''' return (ind + 1) % len(self._buckets ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : int ,lowerCAmelCase__ : KEY ,lowerCAmelCase__ : VAL ) -> bool: '''simple docstring''' lowerCAmelCase_ : List[Any] = self._buckets[ind] if not stored: lowerCAmelCase_ : List[str] = _Item(lowerCAmelCase__ ,lowerCAmelCase__ ) self._len += 1 return True elif stored.key == key: lowerCAmelCase_ : Any = _Item(lowerCAmelCase__ ,lowerCAmelCase__ ) return True else: return False def UpperCAmelCase_ ( self : Dict ) -> bool: '''simple docstring''' lowerCAmelCase_ : Any = len(self._buckets ) * self._capacity_factor return len(self ) >= int(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : int ) -> bool: '''simple docstring''' if len(self._buckets ) <= self._initial_block_size: return False lowerCAmelCase_ : List[str] = len(self._buckets ) * self._capacity_factor / 2 return len(self ) < limit def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : int ) -> None: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = self._buckets lowerCAmelCase_ : Tuple = [None] * new_size lowerCAmelCase_ : List[str] = 0 for item in old_buckets: if item: self._add_item(item.key ,item.val ) def UpperCAmelCase_ ( self : Optional[Any] ) -> None: '''simple docstring''' self._resize(len(self._buckets ) * 2 ) def UpperCAmelCase_ ( self : Dict ) -> None: '''simple docstring''' self._resize(len(self._buckets ) // 2 ) def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : KEY ) -> Iterator[int]: '''simple docstring''' lowerCAmelCase_ : int = self._get_bucket_index(lowerCAmelCase__ ) for _ in range(len(self._buckets ) ): yield ind lowerCAmelCase_ : str = self._get_next_ind(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : KEY ,lowerCAmelCase__ : VAL ) -> None: '''simple docstring''' for ind in self._iterate_buckets(lowerCAmelCase__ ): if self._try_set(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ): break def __setitem__( self : Dict ,lowerCAmelCase__ : KEY ,lowerCAmelCase__ : VAL ) -> None: '''simple docstring''' if self._is_full(): self._size_up() self._add_item(lowerCAmelCase__ ,lowerCAmelCase__ ) def __delitem__( self : List[Any] ,lowerCAmelCase__ : KEY ) -> None: '''simple docstring''' for ind in self._iterate_buckets(lowerCAmelCase__ ): lowerCAmelCase_ : List[str] = self._buckets[ind] if item is None: raise KeyError(lowerCAmelCase__ ) if item is _deleted: continue if item.key == key: lowerCAmelCase_ : str = _deleted self._len -= 1 break if self._is_sparse(): self._size_down() def __getitem__( self : Dict ,lowerCAmelCase__ : KEY ) -> VAL: '''simple docstring''' for ind in self._iterate_buckets(lowerCAmelCase__ ): lowerCAmelCase_ : List[str] = self._buckets[ind] if item is None: break if item is _deleted: continue if item.key == key: return item.val raise KeyError(lowerCAmelCase__ ) def __len__( self : Optional[int] ) -> int: '''simple docstring''' return self._len def __iter__( self : str ) -> Iterator[KEY]: '''simple docstring''' yield from (item.key for item in self._buckets if item) def __repr__( self : List[Any] ) -> str: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = " ,".join( f'''{item.key}: {item.val}''' for item in self._buckets if item ) return f'''HashMap({val_string})'''
659
from collections.abc import Generator from math import sin def UpperCamelCase ( snake_case__): if len(snake_case__) != 32: raise ValueError("Input must be of length 32") lowerCAmelCase_ : Tuple = b"" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def UpperCamelCase ( snake_case__): if i < 0: raise ValueError("Input must be non-negative") lowerCAmelCase_ : List[str] = format(snake_case__ , "08x")[-8:] lowerCAmelCase_ : Any = b"" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("utf-8") return little_endian_hex def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Union[str, Any] = b"" for char in message: bit_string += format(snake_case__ , "08b").encode("utf-8") lowerCAmelCase_ : Optional[int] = format(len(snake_case__) , "064b").encode("utf-8") # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(snake_case__) % 5_12 != 4_48: bit_string += b"0" bit_string += to_little_endian(start_len[32:]) + to_little_endian(start_len[:32]) return bit_string def UpperCamelCase ( snake_case__): if len(snake_case__) % 5_12 != 0: raise ValueError("Input must have length that's a multiple of 512") for pos in range(0 , len(snake_case__) , 5_12): lowerCAmelCase_ : List[str] = bit_string[pos : pos + 5_12] lowerCAmelCase_ : Union[str, Any] = [] for i in range(0 , 5_12 , 32): block_words.append(int(to_little_endian(block[i : i + 32]) , 2)) yield block_words def UpperCamelCase ( snake_case__): if i < 0: raise ValueError("Input must be non-negative") lowerCAmelCase_ : Dict = format(snake_case__ , "032b") lowerCAmelCase_ : str = "" for c in i_str: new_str += "1" if c == "0" else "0" return int(snake_case__ , 2) def UpperCamelCase ( snake_case__ , snake_case__): return (a + b) % 2**32 def UpperCamelCase ( snake_case__ , snake_case__): if i < 0: raise ValueError("Input must be non-negative") if shift < 0: raise ValueError("Shift must be non-negative") return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Optional[Any] = preprocess(snake_case__) lowerCAmelCase_ : Optional[Any] = [int(2**32 * abs(sin(i + 1))) for i in range(64)] # Starting states lowerCAmelCase_ : List[str] = 0x67_45_23_01 lowerCAmelCase_ : Union[str, Any] = 0xef_cd_ab_89 lowerCAmelCase_ : List[Any] = 0x98_ba_dc_fe lowerCAmelCase_ : Tuple = 0x10_32_54_76 lowerCAmelCase_ : Any = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(snake_case__): lowerCAmelCase_ : Optional[int] = aa lowerCAmelCase_ : List[str] = ba lowerCAmelCase_ : Any = ca lowerCAmelCase_ : Union[str, Any] = da # Hash current chunk for i in range(64): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f lowerCAmelCase_ : Any = d ^ (b & (c ^ d)) lowerCAmelCase_ : Dict = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f lowerCAmelCase_ : Any = c ^ (d & (b ^ c)) lowerCAmelCase_ : List[str] = (5 * i + 1) % 16 elif i <= 47: lowerCAmelCase_ : int = b ^ c ^ d lowerCAmelCase_ : Optional[Any] = (3 * i + 5) % 16 else: lowerCAmelCase_ : List[Any] = c ^ (b | not_aa(snake_case__)) lowerCAmelCase_ : List[Any] = (7 * i) % 16 lowerCAmelCase_ : Optional[Any] = (f + a + added_consts[i] + block_words[g]) % 2**32 lowerCAmelCase_ : Optional[Any] = d lowerCAmelCase_ : Dict = c lowerCAmelCase_ : List[str] = b lowerCAmelCase_ : Any = sum_aa(snake_case__ , left_rotate_aa(snake_case__ , shift_amounts[i])) # Add hashed chunk to running total lowerCAmelCase_ : Dict = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : str = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : Optional[int] = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : int = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : Union[str, Any] = reformat_hex(snake_case__) + reformat_hex(snake_case__) + reformat_hex(snake_case__) + reformat_hex(snake_case__) return digest if __name__ == "__main__": import doctest doctest.testmod()
659
1
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DiffusionPipeline, EulerDiscreteScheduler, StableDiffusionXLImgaImgPipeline, UNetaDConditionModel, ) from diffusers.utils import floats_tensor, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __snake_case ( snake_case__ , snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = StableDiffusionXLImgaImgPipeline UpperCamelCase_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} UpperCamelCase_ = PipelineTesterMixin.required_optional_params - {'latents'} UpperCamelCase_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS UpperCamelCase_ = IMAGE_TO_IMAGE_IMAGE_PARAMS UpperCamelCase_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def UpperCAmelCase_ ( self : Dict ) -> int: '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase_ : Union[str, Any] = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") ,up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") ,attention_head_dim=(2, 4) ,use_linear_projection=lowerCAmelCase__ ,addition_embed_type="text_time" ,addition_time_embed_dim=8 ,transformer_layers_per_block=(1, 2) ,projection_class_embeddings_input_dim=80 ,cross_attention_dim=64 ,) lowerCAmelCase_ : Optional[int] = EulerDiscreteScheduler( beta_start=0.00_085 ,beta_end=0.012 ,steps_offset=1 ,beta_schedule="scaled_linear" ,timestep_spacing="leading" ,) torch.manual_seed(0 ) lowerCAmelCase_ : Any = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] ,up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] ,latent_channels=4 ,sample_size=1_28 ,) torch.manual_seed(0 ) lowerCAmelCase_ : str = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=10_00 ,hidden_act="gelu" ,projection_dim=32 ,) lowerCAmelCase_ : Dict = CLIPTextModel(lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ,local_files_only=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = CLIPTextModelWithProjection(lowerCAmelCase__ ) lowerCAmelCase_ : int = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ,local_files_only=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_encoder_2": text_encoder_a, "tokenizer_2": tokenizer_a, # "safety_checker": None, # "feature_extractor": None, } return components def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Optional[Any]=0 ) -> Any: '''simple docstring''' lowerCAmelCase_ : str = floats_tensor((1, 3, 32, 32) ,rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = image / 2 + 0.5 if str(lowerCAmelCase__ ).startswith("mps" ): lowerCAmelCase_ : str = torch.manual_seed(lowerCAmelCase__ ) else: lowerCAmelCase_ : Optional[int] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = { "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 2, "guidance_scale": 5.0, "output_type": "numpy", "strength": 0.75, } return inputs def UpperCAmelCase_ ( self : int ) -> str: '''simple docstring''' lowerCAmelCase_ : List[str] = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : Dict = self.get_dummy_components() lowerCAmelCase_ : Tuple = StableDiffusionXLImgaImgPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = sd_pipe.to(lowerCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = sd_pipe(**lowerCAmelCase__ ).images lowerCAmelCase_ : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCAmelCase_ : Optional[int] = np.array([0.4_656, 0.4_840, 0.4_439, 0.6_698, 0.5_574, 0.4_524, 0.5_799, 0.5_943, 0.5_165] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCAmelCase_ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' super().test_attention_slicing_forward_pass(expected_max_diff=3e-3 ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Optional[Any]: '''simple docstring''' super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def UpperCAmelCase_ ( self : Any ) -> str: '''simple docstring''' pass def UpperCAmelCase_ ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = self.get_dummy_components() lowerCAmelCase_ : int = StableDiffusionXLImgaImgPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = sd_pipe.to(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = sd_pipe.to(lowerCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) # forward without prompt embeds lowerCAmelCase_ : Dict = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Any = 3 * ["this is a negative prompt"] lowerCAmelCase_ : Optional[Any] = negative_prompt lowerCAmelCase_ : str = 3 * [inputs["prompt"]] lowerCAmelCase_ : Optional[Any] = sd_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ : str = output.images[0, -3:, -3:, -1] # forward with prompt embeds lowerCAmelCase_ : str = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : str = 3 * ["this is a negative prompt"] lowerCAmelCase_ : int = 3 * [inputs.pop("prompt" )] ( ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ) : List[Any] = sd_pipe.encode_prompt(lowerCAmelCase__ ,negative_prompt=lowerCAmelCase__ ) lowerCAmelCase_ : Dict = sd_pipe( **lowerCAmelCase__ ,prompt_embeds=lowerCAmelCase__ ,negative_prompt_embeds=lowerCAmelCase__ ,pooled_prompt_embeds=lowerCAmelCase__ ,negative_pooled_prompt_embeds=lowerCAmelCase__ ,) lowerCAmelCase_ : int = output.images[0, -3:, -3:, -1] # make sure that it's equal assert np.abs(image_slice_a.flatten() - image_slice_a.flatten() ).max() < 1e-4 @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : Any ) -> Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : str="cpu" ,lowerCAmelCase__ : Any=torch.floataa ,lowerCAmelCase__ : int=0 ) -> Dict: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : int = np.random.RandomState(lowerCAmelCase__ ).standard_normal((1, 4, 64, 64) ) lowerCAmelCase_ : List[Any] = torch.from_numpy(lowerCAmelCase__ ).to(device=lowerCAmelCase__ ,dtype=lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : Dict ) -> int: '''simple docstring''' lowerCAmelCase_ : Any = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-2-base" ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : str = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = pipe(**lowerCAmelCase__ ).images lowerCAmelCase_ : int = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 5_12, 3) lowerCAmelCase_ : Optional[Any] = np.array([0.49_493, 0.47_896, 0.40_798, 0.54_214, 0.53_212, 0.48_202, 0.47_656, 0.46_329, 0.48_506] ) assert np.abs(image_slice - expected_slice ).max() < 7e-3
659
import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): _lowercase = True from torch.cuda.amp import autocast _lowercase = logging.getLogger(__name__) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Whether to log verbose messages or not.'} , ) UpperCamelCase_ = field( default=2.0 , metadata={'help': 'Maximum temperature for gumbel softmax.'} ) UpperCamelCase_ = field( default=0.5 , metadata={'help': 'Minimum temperature for gumbel softmax.'} ) UpperCamelCase_ = field( default=0.99_99_95 , metadata={'help': 'Decay of gumbel temperature during training.'} ) def UpperCamelCase ( snake_case__ , snake_case__): logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout)] , ) lowerCAmelCase_ : str = logging.WARNING if model_args.verbose_logging: lowerCAmelCase_ : int = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank): lowerCAmelCase_ : Any = logging.INFO logger.setLevel(snake_case__) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The name of the dataset to use (via the datasets library).'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) UpperCamelCase_ = field( default='train' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) UpperCamelCase_ = field( default='validation' , metadata={ 'help': ( 'The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'' ) } , ) UpperCamelCase_ = field( default='file' , metadata={'help': 'Column in the dataset that contains speech file path. Defaults to \'file\''} , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Overwrite the cached preprocessed datasets or not.'} ) UpperCamelCase_ = field( default=1 , metadata={ 'help': 'The percentage of the train set used as validation set in case there\'s no validation split' } , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) UpperCamelCase_ = field( default=20.0 , metadata={'help': 'Filter audio files that are longer than `max_duration_in_seconds` seconds'} ) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = 42 UpperCamelCase_ = 42 UpperCamelCase_ = "longest" UpperCamelCase_ = None UpperCamelCase_ = None def __call__( self : str ,lowerCAmelCase__ : List[Dict[str, Union[List[int], torch.Tensor]]] ) -> Dict[str, torch.Tensor]: '''simple docstring''' lowerCAmelCase_ : Tuple = self.feature_extractor.pad( lowerCAmelCase__ ,max_length=self.max_length ,padding=self.padding ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors="pt" ,) lowerCAmelCase_ : Union[str, Any] = self.model._get_feat_extract_output_lengths(batch["input_values"].shape[-1] ) lowerCAmelCase_ : List[str] = batch["input_values"].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula lowerCAmelCase_ : Tuple = self.model._get_feat_extract_output_lengths(batch["attention_mask"].sum(-1 ) ).to( torch.long ) lowerCAmelCase_ : Optional[Any] = torch.zeros( (batch_size, mask_indices_seq_length) ,dtype=torch.long ,device=batch["input_values"].device ) # these two operations makes sure that all values # before the output lengths indices are attended to lowerCAmelCase_ : Tuple = 1 lowerCAmelCase_ : int = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices lowerCAmelCase_ : str = _compute_mask_indices( (batch_size, mask_indices_seq_length) ,self.model.config.mask_time_prob ,self.model.config.mask_time_length ,attention_mask=lowerCAmelCase__ ,min_masks=2 ,) return batch class __snake_case ( snake_case__ ): """simple docstring""" def __init__( self : List[str] ,*lowerCAmelCase__ : Optional[int] ,lowerCAmelCase__ : Tuple=1 ,lowerCAmelCase__ : Optional[int]=0 ,lowerCAmelCase__ : Optional[Any]=1.0 ,**lowerCAmelCase__ : Any ) -> str: '''simple docstring''' super().__init__(*lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = 0 lowerCAmelCase_ : int = max_gumbel_temp lowerCAmelCase_ : Union[str, Any] = min_gumbel_temp lowerCAmelCase_ : str = gumbel_temp_decay def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : nn.Module ,lowerCAmelCase__ : Dict[str, Union[torch.Tensor, Any]] ) -> torch.Tensor: '''simple docstring''' model.train() lowerCAmelCase_ : str = self._prepare_inputs(lowerCAmelCase__ ) if self.use_amp: with autocast(): lowerCAmelCase_ : List[Any] = self.compute_loss(lowerCAmelCase__ ,lowerCAmelCase__ ) else: lowerCAmelCase_ : List[Any] = self.compute_loss(lowerCAmelCase__ ,lowerCAmelCase__ ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": lowerCAmelCase_ : List[Any] = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": lowerCAmelCase_ : Optional[Any] = loss.sum() / (inputs["mask_time_indices"]).sum() else: raise ValueError(f'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: lowerCAmelCase_ : int = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(lowerCAmelCase__ ).backward() elif self.use_apex: with amp.scale_loss(lowerCAmelCase__ ,self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(lowerCAmelCase__ ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) return loss.detach() def UpperCamelCase ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase_ : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments)) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = parser.parse_args_into_dataclasses() configure_logger(snake_case__ , snake_case__) # Downloading and loading a dataset from the hub. lowerCAmelCase_ : List[str] = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" lowerCAmelCase_ : Any = DatasetDict() lowerCAmelCase_ : Union[str, Any] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : List[str] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" lowerCAmelCase_ : Union[str, Any] = DatasetDict() lowerCAmelCase_ : int = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split="validation" , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : Any = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported lowerCAmelCase_ : Dict = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=snake_case__) def prepare_dataset(snake_case__): # check that all files have the correct sampling rate lowerCAmelCase_ , lowerCAmelCase_ : str = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate) return batch # load audio files into numpy arrays lowerCAmelCase_ : int = datasets.map( snake_case__ , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets["train"].column_names) # filter audio files that are too long lowerCAmelCase_ : int = vectorized_datasets.filter( lambda snake_case__: len(data["speech"]) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate)) def normalize(snake_case__): return feature_extractor(batch["speech"] , sampling_rate=feature_extractor.sampling_rate) # normalize and transform to `BatchFeatures` lowerCAmelCase_ : str = vectorized_datasets.map( snake_case__ , batched=snake_case__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets["train"].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 lowerCAmelCase_ : Optional[Any] = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( "PreTraining is only supported for ``config.do_stable_layer_norm=True`` and" " ``config.feat_extract_norm='layer'") lowerCAmelCase_ : Dict = WavaVecaForPreTraining(snake_case__) lowerCAmelCase_ : int = DataCollatorForWavaVecaPretraining(model=snake_case__ , feature_extractor=snake_case__) lowerCAmelCase_ : List[Any] = WavaVecaPreTrainer( model=snake_case__ , data_collator=snake_case__ , args=snake_case__ , train_dataset=vectorized_datasets["train"] , eval_dataset=vectorized_datasets["validation"] , tokenizer=snake_case__ , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
659
1
class __snake_case : """simple docstring""" def __init__( self : Optional[int] ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Tuple ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : Tuple = name lowerCAmelCase_ : Tuple = value lowerCAmelCase_ : str = weight def __repr__( self : Any ) -> int: '''simple docstring''' return f'''{self.__class__.__name__}({self.name}, {self.value}, {self.weight})''' def UpperCAmelCase_ ( self : Any ) -> str: '''simple docstring''' return self.value def UpperCAmelCase_ ( self : Any ) -> List[Any]: '''simple docstring''' return self.name def UpperCAmelCase_ ( self : Any ) -> Optional[int]: '''simple docstring''' return self.weight def UpperCAmelCase_ ( self : List[Any] ) -> List[str]: '''simple docstring''' return self.value / self.weight def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : int = [] for i in range(len(snake_case__)): menu.append(Things(name[i] , value[i] , weight[i])) return menu def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : Tuple = sorted(snake_case__ , key=snake_case__ , reverse=snake_case__) lowerCAmelCase_ : int = [] lowerCAmelCase_ , lowerCAmelCase_ : Any = 0.0, 0.0 for i in range(len(snake_case__)): if (total_cost + items_copy[i].get_weight()) <= max_cost: result.append(items_copy[i]) total_cost += items_copy[i].get_weight() total_value += items_copy[i].get_value() return (result, total_value) def UpperCamelCase ( ): pass if __name__ == "__main__": import doctest doctest.testmod()
659
from __future__ import annotations from collections.abc import Callable def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1_00 , ): lowerCAmelCase_ : Any = x_start lowerCAmelCase_ : Optional[Any] = fnc(snake_case__) lowerCAmelCase_ : Union[str, Any] = 0.0 for _ in range(snake_case__): # Approximates small segments of curve as linear and solve # for trapezoidal area lowerCAmelCase_ : Any = (x_end - x_start) / steps + xa lowerCAmelCase_ : Dict = fnc(snake_case__) area += abs(fxa + fxa) * (xa - xa) / 2 # Increment step lowerCAmelCase_ : int = xa lowerCAmelCase_ : str = fxa return area if __name__ == "__main__": def UpperCamelCase ( snake_case__): return x**3 + x**2 print('''f(x) = x^3 + x^2''') print('''The area between the curve, x = -5, x = 5 and the x axis is:''') _lowercase = 10 while i <= 100000: print(f"with {i} steps: {trapezoidal_area(f, -5, 5, i)}") i *= 10
659
1
import argparse import json from tqdm import tqdm def UpperCamelCase ( ): lowerCAmelCase_ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--src_path" , type=snake_case__ , default="biencoder-nq-dev.json" , help="Path to raw DPR training data" , ) parser.add_argument( "--evaluation_set" , type=snake_case__ , help="where to store parsed evaluation_set file" , ) parser.add_argument( "--gold_data_path" , type=snake_case__ , help="where to store parsed gold_data_path file" , ) lowerCAmelCase_ : Dict = parser.parse_args() with open(args.src_path , "r") as src_file, open(args.evaluation_set , "w") as eval_file, open( args.gold_data_path , "w") as gold_file: lowerCAmelCase_ : Optional[int] = json.load(snake_case__) for dpr_record in tqdm(snake_case__): lowerCAmelCase_ : str = dpr_record["question"] lowerCAmelCase_ : Dict = [context["title"] for context in dpr_record["positive_ctxs"]] eval_file.write(question + "\n") gold_file.write("\t".join(snake_case__) + "\n") if __name__ == "__main__": main()
659
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, PNDMScheduler, StableDiffusionLDMaDPipeline, UNetaDConditionModel, ) from diffusers.utils import nightly, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS enable_full_determinism() class __snake_case ( unittest.TestCase ): """simple docstring""" UpperCamelCase_ = StableDiffusionLDMaDPipeline UpperCamelCase_ = TEXT_TO_IMAGE_PARAMS UpperCamelCase_ = TEXT_TO_IMAGE_BATCH_PARAMS UpperCamelCase_ = TEXT_TO_IMAGE_IMAGE_PARAMS def UpperCAmelCase_ ( self : Tuple ) -> str: '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase_ : Optional[Any] = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") ,up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") ,cross_attention_dim=32 ,) lowerCAmelCase_ : Any = DDIMScheduler( beta_start=0.00_085 ,beta_end=0.012 ,beta_schedule="scaled_linear" ,clip_sample=lowerCAmelCase__ ,set_alpha_to_one=lowerCAmelCase__ ,) torch.manual_seed(0 ) lowerCAmelCase_ : str = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=6 ,out_channels=6 ,down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] ,up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] ,latent_channels=4 ,) torch.manual_seed(0 ) lowerCAmelCase_ : Optional[Any] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=10_00 ,) lowerCAmelCase_ : Optional[int] = CLIPTextModel(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) lowerCAmelCase_ : List[Any] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : List[str]=0 ) -> Dict: '''simple docstring''' if str(lowerCAmelCase__ ).startswith("mps" ): lowerCAmelCase_ : Optional[int] = torch.manual_seed(lowerCAmelCase__ ) else: lowerCAmelCase_ : Dict = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : str = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : Any ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Dict = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : List[str] = self.get_dummy_components() lowerCAmelCase_ : Union[str, Any] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Any = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Any = output.rgb, output.depth lowerCAmelCase_ : Dict = rgb[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) lowerCAmelCase_ : Optional[Any] = np.array( [0.37_338_176, 0.70_247, 0.74_203_193, 0.51_643_604, 0.58_256_793, 0.60_932_136, 0.4_181_095, 0.48_355_877, 0.46_535_262] ) lowerCAmelCase_ : Tuple = np.array([103.46_727, 85.812_004, 87.849_236] ) assert np.abs(image_slice_rgb.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(image_slice_depth.flatten() - expected_slice_depth ).max() < 1e-2 def UpperCAmelCase_ ( self : int ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Dict = self.get_dummy_components() lowerCAmelCase_ : List[str] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : str = 3 * [inputs["prompt"]] # forward lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = output.rgb, output.depth lowerCAmelCase_ : str = rgb_slice_a[0, -3:, -3:, -1] lowerCAmelCase_ : List[str] = depth_slice_a[0, -3:, -1] lowerCAmelCase_ : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = 3 * [inputs.pop("prompt" )] lowerCAmelCase_ : str = ldmad_pipe.tokenizer( lowerCAmelCase__ ,padding="max_length" ,max_length=ldmad_pipe.tokenizer.model_max_length ,truncation=lowerCAmelCase__ ,return_tensors="pt" ,) lowerCAmelCase_ : Union[str, Any] = text_inputs["input_ids"].to(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = ldmad_pipe.text_encoder(lowerCAmelCase__ )[0] lowerCAmelCase_ : Optional[int] = prompt_embeds # forward lowerCAmelCase_ : str = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : str = output.rgb, output.depth lowerCAmelCase_ : Optional[Any] = rgb_slice_a[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth_slice_a[0, -3:, -1] assert np.abs(rgb_slice_a.flatten() - rgb_slice_a.flatten() ).max() < 1e-4 assert np.abs(depth_slice_a.flatten() - depth_slice_a.flatten() ).max() < 1e-4 def UpperCAmelCase_ ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : Any = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : Optional[int] = self.get_dummy_components() lowerCAmelCase_ : Dict = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : Any = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = "french fries" lowerCAmelCase_ : Optional[int] = ldmad_pipe(**lowerCAmelCase__ ,negative_prompt=lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = output.rgb, output.depth lowerCAmelCase_ : Any = rgb[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) lowerCAmelCase_ : int = np.array( [0.37_044, 0.71_811_503, 0.7_223_251, 0.48_603_675, 0.5_638_391, 0.6_364_948, 0.42_833_704, 0.4_901_315, 0.47_926_217] ) lowerCAmelCase_ : Union[str, Any] = np.array([107.84_738, 84.62_802, 89.962_135] ) assert np.abs(rgb_slice.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(depth_slice.flatten() - expected_slice_depth ).max() < 1e-2 @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Dict="cpu" ,lowerCAmelCase__ : Union[str, Any]=torch.floataa ,lowerCAmelCase__ : List[str]=0 ) -> int: '''simple docstring''' lowerCAmelCase_ : Any = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = np.random.RandomState(lowerCAmelCase__ ).standard_normal((1, 4, 64, 64) ) lowerCAmelCase_ : Optional[Any] = torch.from_numpy(lowerCAmelCase__ ).to(device=lowerCAmelCase__ ,dtype=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : List[Any] ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ) lowerCAmelCase_ : List[str] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Dict = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Dict = output.rgb, output.depth lowerCAmelCase_ : List[str] = rgb[0, -3:, -3:, -1].flatten() lowerCAmelCase_ : Optional[int] = rgb[0, -3:, -1].flatten() assert rgb.shape == (1, 5_12, 5_12, 3) assert depth.shape == (1, 5_12, 5_12) lowerCAmelCase_ : int = np.array( [0.53_805_465, 0.56_707_305, 0.5_486_515, 0.57_012_236, 0.5_814_511, 0.56_253_487, 0.54_843_014, 0.55_092_263, 0.6_459_706] ) lowerCAmelCase_ : Optional[Any] = np.array( [0.9_263_781, 0.6_678_672, 0.5_486_515, 0.92_202_145, 0.67_831_135, 0.56_253_487, 0.9_241_694, 0.7_551_478, 0.6_459_706] ) assert np.abs(rgb_slice - expected_slice_rgb ).max() < 3e-3 assert np.abs(depth_slice - expected_slice_depth ).max() < 3e-3 @nightly @require_torch_gpu class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Dict="cpu" ,lowerCAmelCase__ : List[str]=torch.floataa ,lowerCAmelCase__ : Optional[int]=0 ) -> int: '''simple docstring''' lowerCAmelCase_ : Dict = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = np.random.RandomState(lowerCAmelCase__ ).standard_normal((1, 4, 64, 64) ) lowerCAmelCase_ : Any = torch.from_numpy(lowerCAmelCase__ ).to(device=lowerCAmelCase__ ,dtype=lowerCAmelCase__ ) lowerCAmelCase_ : int = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 50, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : Dict ) -> int: '''simple docstring''' lowerCAmelCase_ : List[Any] = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ).to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Any = output.rgb, output.depth lowerCAmelCase_ : Dict = 0.495_586 lowerCAmelCase_ : Optional[Any] = 0.33_795_515 lowerCAmelCase_ : Any = 112.48_518 lowerCAmelCase_ : List[Any] = 98.489_746 assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3 def UpperCAmelCase_ ( self : Tuple ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : int = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d-4c" ).to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : str = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Tuple = output.rgb, output.depth lowerCAmelCase_ : List[str] = 0.4_194_127 lowerCAmelCase_ : List[str] = 0.35_375_586 lowerCAmelCase_ : str = 0.5_638_502 lowerCAmelCase_ : Optional[Any] = 0.34_686_103 assert rgb.shape == (1, 5_12, 5_12, 3) assert depth.shape == (1, 5_12, 5_12, 1) assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3
659
1
from math import factorial def UpperCamelCase ( snake_case__ = 20): lowerCAmelCase_ : Optional[int] = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... lowerCAmelCase_ : Dict = n // 2 return int(factorial(snake_case__) / (factorial(snake_case__) * factorial(n - k))) if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution(20)) else: try: _lowercase = int(sys.argv[1]) print(solution(n)) except ValueError: print('''Invalid entry - please enter a number.''')
659
import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) _lowercase = { '''iou_prediction_head.layers.0''': '''iou_prediction_head.proj_in''', '''iou_prediction_head.layers.1''': '''iou_prediction_head.layers.0''', '''iou_prediction_head.layers.2''': '''iou_prediction_head.proj_out''', '''mask_decoder.output_upscaling.0''': '''mask_decoder.upscale_conv1''', '''mask_decoder.output_upscaling.1''': '''mask_decoder.upscale_layer_norm''', '''mask_decoder.output_upscaling.3''': '''mask_decoder.upscale_conv2''', '''mask_downscaling.0''': '''mask_embed.conv1''', '''mask_downscaling.1''': '''mask_embed.layer_norm1''', '''mask_downscaling.3''': '''mask_embed.conv2''', '''mask_downscaling.4''': '''mask_embed.layer_norm2''', '''mask_downscaling.6''': '''mask_embed.conv3''', '''point_embeddings''': '''point_embed''', '''pe_layer.positional_encoding_gaussian_matrix''': '''shared_embedding.positional_embedding''', '''image_encoder''': '''vision_encoder''', '''neck.0''': '''neck.conv1''', '''neck.1''': '''neck.layer_norm1''', '''neck.2''': '''neck.conv2''', '''neck.3''': '''neck.layer_norm2''', '''patch_embed.proj''': '''patch_embed.projection''', '''.norm''': '''.layer_norm''', '''blocks''': '''layers''', } def UpperCamelCase ( snake_case__): lowerCAmelCase_ : int = {} state_dict.pop("pixel_mean" , snake_case__) state_dict.pop("pixel_std" , snake_case__) lowerCAmelCase_ : List[Any] = R".*.output_hypernetworks_mlps.(\d+).layers.(\d+).*" for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: lowerCAmelCase_ : Dict = key.replace(snake_case__ , snake_case__) if re.match(snake_case__ , snake_case__): lowerCAmelCase_ : Any = int(re.match(snake_case__ , snake_case__).group(2)) if layer_nb == 0: lowerCAmelCase_ : List[Any] = key.replace("layers.0" , "proj_in") elif layer_nb == 1: lowerCAmelCase_ : List[Any] = key.replace("layers.1" , "layers.0") elif layer_nb == 2: lowerCAmelCase_ : int = key.replace("layers.2" , "proj_out") lowerCAmelCase_ : int = value lowerCAmelCase_ : Optional[int] = model_state_dict[ "prompt_encoder.shared_embedding.positional_embedding" ] return model_state_dict def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__="ybelkada/segment-anything"): lowerCAmelCase_ : Optional[int] = hf_hub_download(snake_case__ , F'''checkpoints/{model_name}.pth''') if "sam_vit_b" in model_name: lowerCAmelCase_ : Optional[Any] = SamConfig() elif "sam_vit_l" in model_name: lowerCAmelCase_ : Optional[int] = SamVisionConfig( hidden_size=10_24 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) lowerCAmelCase_ : Union[str, Any] = SamConfig( vision_config=snake_case__ , ) elif "sam_vit_h" in model_name: lowerCAmelCase_ : Optional[Any] = SamVisionConfig( hidden_size=12_80 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) lowerCAmelCase_ : Tuple = SamConfig( vision_config=snake_case__ , ) lowerCAmelCase_ : Optional[Any] = torch.load(snake_case__ , map_location="cpu") lowerCAmelCase_ : Union[str, Any] = replace_keys(snake_case__) lowerCAmelCase_ : List[Any] = SamImageProcessor() lowerCAmelCase_ : Any = SamProcessor(image_processor=snake_case__) lowerCAmelCase_ : Any = SamModel(snake_case__) hf_model.load_state_dict(snake_case__) lowerCAmelCase_ : Dict = hf_model.to("cuda") lowerCAmelCase_ : List[str] = "https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png" lowerCAmelCase_ : List[Any] = Image.open(requests.get(snake_case__ , stream=snake_case__).raw).convert("RGB") lowerCAmelCase_ : Optional[int] = [[[4_00, 6_50]]] lowerCAmelCase_ : int = [[1]] lowerCAmelCase_ : Optional[Any] = processor(images=np.array(snake_case__) , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Optional[Any] = hf_model(**snake_case__) lowerCAmelCase_ : Optional[int] = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579_890_251_159_668 lowerCAmelCase_ : Any = processor( images=np.array(snake_case__) , input_points=snake_case__ , input_labels=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Optional[Any] = hf_model(**snake_case__) lowerCAmelCase_ : Union[str, Any] = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_712_603_092_193_604 lowerCAmelCase_ : Tuple = ((75, 2_75, 17_25, 8_50),) lowerCAmelCase_ : Optional[Any] = processor(images=np.array(snake_case__) , input_boxes=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : List[Any] = hf_model(**snake_case__) lowerCAmelCase_ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.8_686_015_605_926_514 # Test with 2 points and 1 image. lowerCAmelCase_ : int = [[[4_00, 6_50], [8_00, 6_50]]] lowerCAmelCase_ : Optional[Any] = [[1, 1]] lowerCAmelCase_ : List[Any] = processor( images=np.array(snake_case__) , input_points=snake_case__ , input_labels=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Tuple = hf_model(**snake_case__) lowerCAmelCase_ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_936_047_792_434_692 if __name__ == "__main__": _lowercase = argparse.ArgumentParser() _lowercase = ['''sam_vit_b_01ec64''', '''sam_vit_h_4b8939''', '''sam_vit_l_0b3195'''] parser.add_argument( '''--model_name''', default='''sam_vit_h_4b8939''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model and processor to the hub after converting''', ) parser.add_argument( '''--model_hub_id''', default='''ybelkada/segment-anything''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) _lowercase = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
659
1
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Optional[int] = {} lowerCAmelCase_ : Dict = tokenizer(example["content"] , truncation=snake_case__)["input_ids"] lowerCAmelCase_ : str = len(example["content"]) / len(output["input_ids"]) return output _lowercase = HfArgumentParser(PretokenizationArguments) _lowercase = parser.parse_args() if args.num_workers is None: _lowercase = multiprocessing.cpu_count() _lowercase = AutoTokenizer.from_pretrained(args.tokenizer_dir) _lowercase = time.time() _lowercase = load_dataset(args.dataset_name, split='''train''') print(f"Dataset loaded in {time.time()-t_start:.2f}s") _lowercase = time.time() _lowercase = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ '''repo_name''', '''path''', '''copies''', '''size''', '''content''', '''license''', '''hash''', '''line_mean''', '''line_max''', '''alpha_frac''', '''autogenerated''', ], ) print(f"Dataset tokenized in {time.time()-t_start:.2f}s") _lowercase = time.time() ds.push_to_hub(args.tokenized_data_repo) print(f"Data pushed to the hub in {time.time()-t_start:.2f}s")
659
class __snake_case : """simple docstring""" def __init__( self : Union[str, Any] ,lowerCAmelCase__ : str = "" ,lowerCAmelCase__ : bool = False ) -> None: '''simple docstring''' lowerCAmelCase_ : dict[str, RadixNode] = {} # A node will be a leaf if the tree contains its word lowerCAmelCase_ : Optional[int] = is_leaf lowerCAmelCase_ : List[str] = prefix def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : str ) -> tuple[str, str, str]: '''simple docstring''' lowerCAmelCase_ : List[str] = 0 for q, w in zip(self.prefix ,lowerCAmelCase__ ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : list[str] ) -> None: '''simple docstring''' for word in words: self.insert(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : str ) -> None: '''simple docstring''' if self.prefix == word: lowerCAmelCase_ : Optional[Any] = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: lowerCAmelCase_ : Optional[int] = RadixNode(prefix=lowerCAmelCase__ ,is_leaf=lowerCAmelCase__ ) else: lowerCAmelCase_ : Optional[Any] = self.nodes[word[0]] lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = incoming_node.match( lowerCAmelCase__ ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: lowerCAmelCase_ : Dict = remaining_prefix lowerCAmelCase_ : str = self.nodes[matching_string[0]] lowerCAmelCase_ : Dict = RadixNode(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Any = aux_node if remaining_word == "": lowerCAmelCase_ : Optional[Any] = True else: self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : str ) -> bool: '''simple docstring''' lowerCAmelCase_ : List[str] = self.nodes.get(word[0] ,lowerCAmelCase__ ) if not incoming_node: return False else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : str ) -> bool: '''simple docstring''' lowerCAmelCase_ : int = self.nodes.get(word[0] ,lowerCAmelCase__ ) if not incoming_node: return False else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(lowerCAmelCase__ ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: lowerCAmelCase_ : int = list(self.nodes.values() )[0] lowerCAmelCase_ : List[Any] = merging_node.is_leaf self.prefix += merging_node.prefix lowerCAmelCase_ : int = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: lowerCAmelCase_ : List[str] = False # If there is 1 edge, we merge it with its child else: lowerCAmelCase_ : Union[str, Any] = list(incoming_node.nodes.values() )[0] lowerCAmelCase_ : Optional[int] = merging_node.is_leaf incoming_node.prefix += merging_node.prefix lowerCAmelCase_ : List[str] = merging_node.nodes return True def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : int = 0 ) -> None: '''simple docstring''' if self.prefix != "": print("-" * height ,self.prefix ," (leaf)" if self.is_leaf else "" ) for value in self.nodes.values(): value.print_tree(height + 1 ) def UpperCamelCase ( ): lowerCAmelCase_ : List[Any] = "banana bananas bandana band apple all beast".split() lowerCAmelCase_ : Optional[Any] = RadixNode() root.insert_many(snake_case__) assert all(root.find(snake_case__) for word in words) assert not root.find("bandanas") assert not root.find("apps") root.delete("all") assert not root.find("all") root.delete("banana") assert not root.find("banana") assert root.find("bananas") return True def UpperCamelCase ( ): assert test_trie() def UpperCamelCase ( ): lowerCAmelCase_ : str = RadixNode() lowerCAmelCase_ : str = "banana bananas bandanas bandana band apple all beast".split() root.insert_many(snake_case__) print("Words:" , snake_case__) print("Tree:") root.print_tree() if __name__ == "__main__": main()
659
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowercase = {'''configuration_yolos''': ['''YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''YolosConfig''', '''YolosOnnxConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ['''YolosFeatureExtractor'''] _lowercase = ['''YolosImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ '''YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''YolosForObjectDetection''', '''YolosModel''', '''YolosPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
659
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class __snake_case : """simple docstring""" def __init__( self : Tuple ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Optional[Any]=12 ,lowerCAmelCase__ : Union[str, Any]=7 ,lowerCAmelCase__ : Union[str, Any]=True ,lowerCAmelCase__ : List[str]=True ,lowerCAmelCase__ : Any=True ,lowerCAmelCase__ : Optional[Any]=99 ,lowerCAmelCase__ : List[str]=32 ,lowerCAmelCase__ : Dict=32 ,lowerCAmelCase__ : str=2 ,lowerCAmelCase__ : Optional[int]=4 ,lowerCAmelCase__ : str=37 ,lowerCAmelCase__ : Dict=0.1 ,lowerCAmelCase__ : List[str]=0.1 ,lowerCAmelCase__ : str=5_12 ,lowerCAmelCase__ : Union[str, Any]=0.02 ,lowerCAmelCase__ : Tuple=0 ,lowerCAmelCase__ : str=None ,) -> str: '''simple docstring''' lowerCAmelCase_ : int = parent lowerCAmelCase_ : str = batch_size lowerCAmelCase_ : int = seq_length lowerCAmelCase_ : Union[str, Any] = is_training lowerCAmelCase_ : int = use_input_mask lowerCAmelCase_ : List[Any] = use_labels lowerCAmelCase_ : Dict = vocab_size lowerCAmelCase_ : Union[str, Any] = hidden_size lowerCAmelCase_ : Union[str, Any] = projection_dim lowerCAmelCase_ : List[Any] = num_hidden_layers lowerCAmelCase_ : Any = num_attention_heads lowerCAmelCase_ : List[Any] = intermediate_size lowerCAmelCase_ : Any = dropout lowerCAmelCase_ : Optional[int] = attention_dropout lowerCAmelCase_ : int = max_position_embeddings lowerCAmelCase_ : Optional[int] = initializer_range lowerCAmelCase_ : Any = scope lowerCAmelCase_ : Tuple = bos_token_id def UpperCAmelCase_ ( self : str ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) lowerCAmelCase_ : Dict = None if self.use_input_mask: lowerCAmelCase_ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowerCAmelCase_ : List[Any] = input_mask.numpy() lowerCAmelCase_ , lowerCAmelCase_ : str = input_mask.shape lowerCAmelCase_ : Dict = np.random.randint(1 ,seq_length - 1 ,size=(batch_size,) ) for batch_idx, start_index in enumerate(lowerCAmelCase__ ): lowerCAmelCase_ : Union[str, Any] = 1 lowerCAmelCase_ : Optional[Any] = 0 lowerCAmelCase_ : List[Any] = self.get_config() return config, input_ids, tf.convert_to_tensor(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[str] ) -> str: '''simple docstring''' return BlipTextConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,projection_dim=self.projection_dim ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,dropout=self.dropout ,attention_dropout=self.attention_dropout ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,bos_token_id=self.bos_token_id ,) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Dict ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : List[Any] = TFBlipTextModel(config=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = model(lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,training=lowerCAmelCase__ ) lowerCAmelCase_ : str = model(lowerCAmelCase__ ,training=lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def UpperCAmelCase_ ( self : Optional[int] ) -> int: '''simple docstring''' lowerCAmelCase_ : List[str] = self.prepare_config_and_inputs() lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = config_and_inputs lowerCAmelCase_ : Tuple = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = (TFBlipTextModel,) if is_tf_available() else () UpperCamelCase_ = False UpperCamelCase_ = False UpperCamelCase_ = False def UpperCAmelCase_ ( self : Optional[Any] ) -> str: '''simple docstring''' lowerCAmelCase_ : List[str] = BlipTextModelTester(self ) lowerCAmelCase_ : Tuple = ConfigTester(self ,config_class=lowerCAmelCase__ ,hidden_size=37 ) def UpperCAmelCase_ ( self : str ) -> Any: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase_ ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' pass def UpperCAmelCase_ ( self : Union[str, Any] ) -> Any: '''simple docstring''' pass @unittest.skip(reason="Blip does not use inputs_embeds" ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCAmelCase_ ( self : int ) -> Optional[Any]: '''simple docstring''' pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCAmelCase_ ( self : Dict ) -> Union[str, Any]: '''simple docstring''' pass @slow def UpperCAmelCase_ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Tuple = TFBlipTextModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : str=True ) -> List[Any]: '''simple docstring''' super().test_pt_tf_model_equivalence(allow_missing_keys=lowerCAmelCase__ )
659
1
from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class __snake_case : """simple docstring""" def __init__( self : List[Any] ,lowerCAmelCase__ : Optional[int] ,) -> str: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = parent lowerCAmelCase_ : int = 13 lowerCAmelCase_ : Union[str, Any] = 7 lowerCAmelCase_ : List[str] = 30 lowerCAmelCase_ : int = self.seq_length + self.mem_len lowerCAmelCase_ : List[str] = 15 lowerCAmelCase_ : Any = True lowerCAmelCase_ : Any = True lowerCAmelCase_ : List[Any] = 99 lowerCAmelCase_ : Tuple = [10, 50, 80] lowerCAmelCase_ : int = 32 lowerCAmelCase_ : Dict = 32 lowerCAmelCase_ : Optional[Any] = 4 lowerCAmelCase_ : Optional[Any] = 8 lowerCAmelCase_ : int = 1_28 lowerCAmelCase_ : Optional[Any] = 2 lowerCAmelCase_ : List[str] = 2 lowerCAmelCase_ : List[Any] = None lowerCAmelCase_ : Tuple = 1 lowerCAmelCase_ : Any = 0 lowerCAmelCase_ : List[Any] = 3 lowerCAmelCase_ : Optional[int] = self.vocab_size - 1 lowerCAmelCase_ : List[Any] = 0.01 def UpperCAmelCase_ ( self : Optional[int] ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : int = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) lowerCAmelCase_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) lowerCAmelCase_ : List[Any] = None if self.use_labels: lowerCAmelCase_ : int = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) lowerCAmelCase_ : Optional[Any] = TransfoXLConfig( vocab_size=self.vocab_size ,mem_len=self.mem_len ,clamp_len=self.clamp_len ,cutoffs=self.cutoffs ,d_model=self.hidden_size ,d_embed=self.d_embed ,n_head=self.num_attention_heads ,d_head=self.d_head ,d_inner=self.d_inner ,div_val=self.div_val ,n_layer=self.num_hidden_layers ,eos_token_id=self.eos_token_id ,pad_token_id=self.vocab_size - 1 ,init_range=self.init_range ,num_labels=self.num_labels ,) return (config, input_ids_a, input_ids_a, lm_labels) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Any: '''simple docstring''' random.seed(self.seed ) tf.random.set_seed(self.seed ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : int ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : str ,lowerCAmelCase__ : List[str] ) -> str: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = TFTransfoXLModel(lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : int = model(lowerCAmelCase__ ).to_tuple() lowerCAmelCase_ : str = {"input_ids": input_ids_a, "mems": mems_a} lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = model(lowerCAmelCase__ ).to_tuple() self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Optional[Any] ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : List[str] = TFTransfoXLLMHeadModel(lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : List[str] = model(lowerCAmelCase__ ).to_tuple() lowerCAmelCase_ : int = {"input_ids": input_ids_a, "labels": lm_labels} lowerCAmelCase_ , lowerCAmelCase_ : List[str] = model(lowerCAmelCase__ ).to_tuple() lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = model([input_ids_a, mems_a] ).to_tuple() lowerCAmelCase_ : Optional[Any] = {"input_ids": input_ids_a, "mems": mems_a, "labels": lm_labels} lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = model(lowerCAmelCase__ ).to_tuple() self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) self.parent.assertEqual(lm_logits_a.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] ,[(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers ,) def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : str ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : Union[str, Any] ) -> Dict: '''simple docstring''' lowerCAmelCase_ : List[str] = TFTransfoXLForSequenceClassification(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = model(lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self : Optional[int] ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : Dict = self.prepare_config_and_inputs() ((lowerCAmelCase_) , (lowerCAmelCase_) , (lowerCAmelCase_) , (lowerCAmelCase_)) : Optional[int] = config_and_inputs lowerCAmelCase_ : Optional[Any] = {"input_ids": input_ids_a} return config, inputs_dict @require_tf class __snake_case ( snake_case__ , snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) UpperCamelCase_ = () if is_tf_available() else () UpperCamelCase_ = ( { 'feature-extraction': TFTransfoXLModel, 'text-classification': TFTransfoXLForSequenceClassification, 'text-generation': TFTransfoXLLMHeadModel, 'zero-shot': TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented UpperCamelCase_ = False UpperCamelCase_ = False UpperCamelCase_ = False UpperCamelCase_ = False def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Tuple ) -> List[Any]: '''simple docstring''' if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def UpperCAmelCase_ ( self : List[str] ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : int = TFTransfoXLModelTester(self ) lowerCAmelCase_ : List[Any] = ConfigTester(self ,config_class=lowerCAmelCase__ ,d_embed=37 ) def UpperCAmelCase_ ( self : int ) -> Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase_ ( self : Tuple ) -> List[Any]: '''simple docstring''' self.model_tester.set_seed() lowerCAmelCase_ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[str]: '''simple docstring''' self.model_tester.set_seed() lowerCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> Any: '''simple docstring''' lowerCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Any: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase_ : List[str] = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: lowerCAmelCase_ : Optional[int] = model_class(lowerCAmelCase__ ) assert isinstance(model.get_input_embeddings() ,tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: lowerCAmelCase_ : str = model.get_output_embeddings() assert isinstance(lowerCAmelCase__ ,tf.keras.layers.Layer ) lowerCAmelCase_ : Optional[Any] = model.get_bias() assert name is None else: lowerCAmelCase_ : Dict = model.get_output_embeddings() assert x is None lowerCAmelCase_ : int = model.get_bias() assert name is None def UpperCAmelCase_ ( self : Dict ) -> List[str]: '''simple docstring''' pass @slow def UpperCAmelCase_ ( self : Tuple ) -> Any: '''simple docstring''' for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Tuple = TFTransfoXLModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) @unittest.skip(reason="This model doesn't play well with fit() due to not returning a single loss." ) def UpperCAmelCase_ ( self : Any ) -> Tuple: '''simple docstring''' pass @require_tf class __snake_case ( unittest.TestCase ): """simple docstring""" @unittest.skip("Skip test until #12651 is resolved." ) @slow def UpperCAmelCase_ ( self : List[Any] ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : Tuple = TFTransfoXLLMHeadModel.from_pretrained("transfo-xl-wt103" ) # fmt: off lowerCAmelCase_ : Optional[Any] = tf.convert_to_tensor([[33,12_97,2,1,10_09,4,11_09,1_17_39,47_62,3_58,5,25,2_45,22,17_06,17,2_00_98,5,32_15,21,37,11_10,3,13,10_41,4,24,6_03,4_90,2,7_14_77,2_00_98,10_44_47,2,2_09_61,1,26_04,4,1,3_29,3,62_24,8_31,1_60_02,2,8,6_03,7_89_67,2_95_46,23,8_03,20,25,4_16,5,8,2_32,4,2_77,6,18_55,46_01,3,2_95_46,54,8,36_09,5,5_72_11,49,4,1,2_77,18,8,17_55,1_56_91,3,3_41,25,4_16,6_93,4_25_73,71,17,4_01,94,31,1_79_19,2,2_95_46,78_73,18,1,4_35,23,1_10_11,7_55,5,51_67,3,79_83,98,84,2,2_95_46,32_67,8,36_09,4,1,48_65,10_75,2,60_87,71,6,3_46,8,58_54,3,2_95_46,8_24,14_00,18_68,2,19,1_60,2,3_11,8,54_96,2,2_09_20,17,25,1_50_97,3,24,24,0]] ,dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off lowerCAmelCase_ : Optional[int] = [33,12_97,2,1,10_09,4,11_09,1_17_39,47_62,3_58,5,25,2_45,22,17_06,17,2_00_98,5,32_15,21,37,11_10,3,13,10_41,4,24,6_03,4_90,2,7_14_77,2_00_98,10_44_47,2,2_09_61,1,26_04,4,1,3_29,3,62_24,8_31,1_60_02,2,8,6_03,7_89_67,2_95_46,23,8_03,20,25,4_16,5,8,2_32,4,2_77,6,18_55,46_01,3,2_95_46,54,8,36_09,5,5_72_11,49,4,1,2_77,18,8,17_55,1_56_91,3,3_41,25,4_16,6_93,4_25_73,71,17,4_01,94,31,1_79_19,2,2_95_46,78_73,18,1,4_35,23,1_10_11,7_55,5,51_67,3,79_83,98,84,2,2_95_46,32_67,8,36_09,4,1,48_65,10_75,2,60_87,71,6,3_46,8,58_54,3,2_95_46,8_24,14_00,18_68,2,19,1_60,2,3_11,8,54_96,2,2_09_20,17,25,1_50_97,3,24,24,0,33,1,18_57,2,1,10_09,4,11_09,1_17_39,47_62,3_58,5,25,2_45,28,11_10,3,13,10_41,4,24,6_03,4_90,2,7_14_77,2_00_98,10_44_47,2,2_09_61,1,26_04,4,1,3_29,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> lowerCAmelCase_ : List[str] = model.generate(lowerCAmelCase__ ,max_length=2_00 ,do_sample=lowerCAmelCase__ ) self.assertListEqual(output_ids[0].numpy().tolist() ,lowerCAmelCase__ )
659
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _lowercase = logging.get_logger(__name__) _lowercase = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} # See all LED models at https://huggingface.co/models?filter=LED _lowercase = { '''vocab_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json''', }, '''merges_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json''', }, } _lowercase = { '''allenai/led-base-16384''': 16384, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def UpperCamelCase ( ): lowerCAmelCase_ : Optional[int] = ( list(range(ord("!") , ord("~") + 1)) + list(range(ord("¡") , ord("¬") + 1)) + list(range(ord("®") , ord("ÿ") + 1)) ) lowerCAmelCase_ : List[Any] = bs[:] lowerCAmelCase_ : Optional[int] = 0 for b in range(2**8): if b not in bs: bs.append(snake_case__) cs.append(2**8 + n) n += 1 lowerCAmelCase_ : Tuple = [chr(snake_case__) for n in cs] return dict(zip(snake_case__ , snake_case__)) def UpperCamelCase ( snake_case__): lowerCAmelCase_ : str = set() lowerCAmelCase_ : List[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char)) lowerCAmelCase_ : Union[str, Any] = char return pairs class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = ['input_ids', 'attention_mask'] def __init__( self : int ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Tuple="replace" ,lowerCAmelCase__ : Optional[int]="<s>" ,lowerCAmelCase__ : Optional[int]="</s>" ,lowerCAmelCase__ : Tuple="</s>" ,lowerCAmelCase__ : int="<s>" ,lowerCAmelCase__ : Union[str, Any]="<unk>" ,lowerCAmelCase__ : str="<pad>" ,lowerCAmelCase__ : Tuple="<mask>" ,lowerCAmelCase__ : Optional[int]=False ,**lowerCAmelCase__ : Tuple ,) -> Any: '''simple docstring''' lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else bos_token lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else eos_token lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else sep_token lowerCAmelCase_ : Any = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else cls_token lowerCAmelCase_ : Tuple = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else unk_token lowerCAmelCase_ : Any = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase_ : Optional[int] = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else mask_token super().__init__( errors=lowerCAmelCase__ ,bos_token=lowerCAmelCase__ ,eos_token=lowerCAmelCase__ ,unk_token=lowerCAmelCase__ ,sep_token=lowerCAmelCase__ ,cls_token=lowerCAmelCase__ ,pad_token=lowerCAmelCase__ ,mask_token=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,**lowerCAmelCase__ ,) with open(lowerCAmelCase__ ,encoding="utf-8" ) as vocab_handle: lowerCAmelCase_ : List[str] = json.load(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = {v: k for k, v in self.encoder.items()} lowerCAmelCase_ : Optional[int] = errors # how to handle errors in decoding lowerCAmelCase_ : Optional[int] = bytes_to_unicode() lowerCAmelCase_ : str = {v: k for k, v in self.byte_encoder.items()} with open(lowerCAmelCase__ ,encoding="utf-8" ) as merges_handle: lowerCAmelCase_ : List[str] = merges_handle.read().split("\n" )[1:-1] lowerCAmelCase_ : List[Any] = [tuple(merge.split() ) for merge in bpe_merges] lowerCAmelCase_ : Union[str, Any] = dict(zip(lowerCAmelCase__ ,range(len(lowerCAmelCase__ ) ) ) ) lowerCAmelCase_ : Dict = {} lowerCAmelCase_ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowerCAmelCase_ : Any = re.compile(R"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def UpperCAmelCase_ ( self : Dict ) -> Dict: '''simple docstring''' return len(self.encoder ) def UpperCAmelCase_ ( self : Dict ) -> str: '''simple docstring''' return dict(self.encoder ,**self.added_tokens_encoder ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Dict ) -> Dict: '''simple docstring''' if token in self.cache: return self.cache[token] lowerCAmelCase_ : Union[str, Any] = tuple(lowerCAmelCase__ ) lowerCAmelCase_ : str = get_pairs(lowerCAmelCase__ ) if not pairs: return token while True: lowerCAmelCase_ : Optional[int] = min(lowerCAmelCase__ ,key=lambda lowerCAmelCase__ : self.bpe_ranks.get(lowerCAmelCase__ ,float("inf" ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = bigram lowerCAmelCase_ : Tuple = [] lowerCAmelCase_ : str = 0 while i < len(lowerCAmelCase__ ): try: lowerCAmelCase_ : Union[str, Any] = word.index(lowerCAmelCase__ ,lowerCAmelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase_ : List[str] = j if word[i] == first and i < len(lowerCAmelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase_ : Optional[int] = tuple(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = new_word if len(lowerCAmelCase__ ) == 1: break else: lowerCAmelCase_ : Dict = get_pairs(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = " ".join(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = word return word def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : Dict ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : Any = [] for token in re.findall(self.pat ,lowerCAmelCase__ ): lowerCAmelCase_ : Optional[int] = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCAmelCase__ ).split(" " ) ) return bpe_tokens def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any] ) -> Tuple: '''simple docstring''' return self.encoder.get(lowerCAmelCase__ ,self.encoder.get(self.unk_token ) ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Union[str, Any] ) -> Optional[int]: '''simple docstring''' return self.decoder.get(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[Any] ) -> Any: '''simple docstring''' lowerCAmelCase_ : int = "".join(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" ,errors=self.errors ) return text def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(lowerCAmelCase__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCAmelCase_ : Optional[int] = os.path.join( lowerCAmelCase__ ,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase_ : List[str] = os.path.join( lowerCAmelCase__ ,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(lowerCAmelCase__ ,"w" ,encoding="utf-8" ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=lowerCAmelCase__ ,ensure_ascii=lowerCAmelCase__ ) + "\n" ) lowerCAmelCase_ : Dict = 0 with open(lowerCAmelCase__ ,"w" ,encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda lowerCAmelCase__ : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' " Please check that the tokenizer is not corrupted!" ) lowerCAmelCase_ : List[Any] = token_index writer.write(" ".join(lowerCAmelCase__ ) + "\n" ) index += 1 return vocab_file, merge_file def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase_ : Union[str, Any] = [self.cls_token_id] lowerCAmelCase_ : str = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ,lowerCAmelCase__ : bool = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ ,token_ids_a=lowerCAmelCase__ ,already_has_special_tokens=lowerCAmelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase__ )) + [1] return [1] + ([0] * len(lowerCAmelCase__ )) + [1, 1] + ([0] * len(lowerCAmelCase__ )) + [1] def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = [self.sep_token_id] lowerCAmelCase_ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : Optional[int]=False ,**lowerCAmelCase__ : str ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = kwargs.pop("add_prefix_space" ,self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCAmelCase__ ) > 0 and not text[0].isspace()): lowerCAmelCase_ : List[str] = " " + text return (text, kwargs) def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : Union[Dict[str, EncodedInput], BatchEncoding] ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : Optional[bool] = None ,) -> dict: '''simple docstring''' lowerCAmelCase_ : int = super()._pad( encoded_inputs=lowerCAmelCase__ ,max_length=lowerCAmelCase__ ,padding_strategy=lowerCAmelCase__ ,pad_to_multiple_of=lowerCAmelCase__ ,return_attention_mask=lowerCAmelCase__ ,) # Load from model defaults if return_attention_mask is None: lowerCAmelCase_ : List[Any] = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowerCAmelCase_ : Dict = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowerCAmelCase_ : List[Any] = len(encoded_inputs["global_attention_mask"] ) != len(lowerCAmelCase__ ) if needs_to_be_padded: lowerCAmelCase_ : Union[str, Any] = len(lowerCAmelCase__ ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowerCAmelCase_ : Optional[int] = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": lowerCAmelCase_ : List[Any] = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
659
1
_lowercase = {str(digit): digit**5 for digit in range(10)} def UpperCamelCase ( snake_case__): return sum(DIGITS_FIFTH_POWER[digit] for digit in str(snake_case__)) def UpperCamelCase ( ): return sum( number for number in range(10_00 , 1_00_00_00) if number == digits_fifth_powers_sum(snake_case__)) if __name__ == "__main__": print(solution())
659
import os _lowercase = {'''I''': 1, '''V''': 5, '''X''': 10, '''L''': 50, '''C''': 100, '''D''': 500, '''M''': 1000} def UpperCamelCase ( snake_case__): lowerCAmelCase_ : List[str] = 0 lowerCAmelCase_ : Any = 0 while index < len(snake_case__) - 1: lowerCAmelCase_ : Optional[Any] = SYMBOLS[numerals[index]] lowerCAmelCase_ : int = SYMBOLS[numerals[index + 1]] if current_value < next_value: total_value -= current_value else: total_value += current_value index += 1 total_value += SYMBOLS[numerals[index]] return total_value def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Optional[int] = "" lowerCAmelCase_ : Tuple = num // 10_00 numerals += m_count * "M" num %= 10_00 lowerCAmelCase_ : int = num // 1_00 if c_count == 9: numerals += "CM" c_count -= 9 elif c_count == 4: numerals += "CD" c_count -= 4 if c_count >= 5: numerals += "D" c_count -= 5 numerals += c_count * "C" num %= 1_00 lowerCAmelCase_ : int = num // 10 if x_count == 9: numerals += "XC" x_count -= 9 elif x_count == 4: numerals += "XL" x_count -= 4 if x_count >= 5: numerals += "L" x_count -= 5 numerals += x_count * "X" num %= 10 if num == 9: numerals += "IX" num -= 9 elif num == 4: numerals += "IV" num -= 4 if num >= 5: numerals += "V" num -= 5 numerals += num * "I" return numerals def UpperCamelCase ( snake_case__ = "/p089_roman.txt"): lowerCAmelCase_ : int = 0 with open(os.path.dirname(snake_case__) + roman_numerals_filename) as filea: lowerCAmelCase_ : List[Any] = filea.readlines() for line in lines: lowerCAmelCase_ : Any = line.strip() lowerCAmelCase_ : Tuple = parse_roman_numerals(snake_case__) lowerCAmelCase_ : List[Any] = generate_roman_numerals(snake_case__) savings += len(snake_case__) - len(snake_case__) return savings if __name__ == "__main__": print(f"{solution() = }")
659
1
from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''EleutherAI/gpt-j-6B''': '''https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json''', # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = 'gptj' UpperCamelCase_ = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : List[str] ,lowerCAmelCase__ : Union[str, Any]=5_04_00 ,lowerCAmelCase__ : Union[str, Any]=20_48 ,lowerCAmelCase__ : Union[str, Any]=40_96 ,lowerCAmelCase__ : Any=28 ,lowerCAmelCase__ : int=16 ,lowerCAmelCase__ : List[str]=64 ,lowerCAmelCase__ : int=None ,lowerCAmelCase__ : Dict="gelu_new" ,lowerCAmelCase__ : Dict=0.0 ,lowerCAmelCase__ : List[Any]=0.0 ,lowerCAmelCase__ : Union[str, Any]=0.0 ,lowerCAmelCase__ : Tuple=1e-5 ,lowerCAmelCase__ : Optional[int]=0.02 ,lowerCAmelCase__ : str=True ,lowerCAmelCase__ : Union[str, Any]=5_02_56 ,lowerCAmelCase__ : str=5_02_56 ,lowerCAmelCase__ : str=False ,**lowerCAmelCase__ : Optional[int] ,) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = vocab_size lowerCAmelCase_ : str = n_positions lowerCAmelCase_ : List[str] = n_embd lowerCAmelCase_ : Tuple = n_layer lowerCAmelCase_ : Dict = n_head lowerCAmelCase_ : List[Any] = n_inner lowerCAmelCase_ : Tuple = rotary_dim lowerCAmelCase_ : Optional[Any] = activation_function lowerCAmelCase_ : Any = resid_pdrop lowerCAmelCase_ : Union[str, Any] = embd_pdrop lowerCAmelCase_ : int = attn_pdrop lowerCAmelCase_ : Optional[int] = layer_norm_epsilon lowerCAmelCase_ : Union[str, Any] = initializer_range lowerCAmelCase_ : str = use_cache lowerCAmelCase_ : List[Any] = bos_token_id lowerCAmelCase_ : str = eos_token_id super().__init__( bos_token_id=lowerCAmelCase__ ,eos_token_id=lowerCAmelCase__ ,tie_word_embeddings=lowerCAmelCase__ ,**lowerCAmelCase__ ) class __snake_case ( snake_case__ ): """simple docstring""" def __init__( self : str ,lowerCAmelCase__ : PretrainedConfig ,lowerCAmelCase__ : str = "default" ,lowerCAmelCase__ : List[PatchingSpec] = None ,lowerCAmelCase__ : bool = False ,) -> Optional[int]: '''simple docstring''' super().__init__(lowerCAmelCase__ ,task=lowerCAmelCase__ ,patching_specs=lowerCAmelCase__ ,use_past=lowerCAmelCase__ ) if not getattr(self._config ,"pad_token_id" ,lowerCAmelCase__ ): # TODO: how to do that better? lowerCAmelCase_ : Union[str, Any] = 0 @property def UpperCAmelCase_ ( self : int ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = OrderedDict({"input_ids": {0: "batch", 1: "sequence"}} ) if self.use_past: self.fill_with_past_key_values_(lowerCAmelCase__ ,direction="inputs" ) lowerCAmelCase_ : int = {0: "batch", 1: "past_sequence + sequence"} else: lowerCAmelCase_ : int = {0: "batch", 1: "sequence"} return common_inputs @property def UpperCAmelCase_ ( self : Any ) -> int: '''simple docstring''' return self._config.n_layer @property def UpperCAmelCase_ ( self : List[str] ) -> int: '''simple docstring''' return self._config.n_head def UpperCAmelCase_ ( self : Optional[int] ,lowerCAmelCase__ : PreTrainedTokenizer ,lowerCAmelCase__ : int = -1 ,lowerCAmelCase__ : int = -1 ,lowerCAmelCase__ : bool = False ,lowerCAmelCase__ : Optional[TensorType] = None ,) -> Mapping[str, Any]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = super(lowerCAmelCase__ ,self ).generate_dummy_inputs( lowerCAmelCase__ ,batch_size=lowerCAmelCase__ ,seq_length=lowerCAmelCase__ ,is_pair=lowerCAmelCase__ ,framework=lowerCAmelCase__ ) # We need to order the input in the way they appears in the forward() lowerCAmelCase_ : Any = OrderedDict({"input_ids": common_inputs["input_ids"]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = common_inputs["input_ids"].shape # Not using the same length for past_key_values lowerCAmelCase_ : Tuple = seqlen + 2 lowerCAmelCase_ : Optional[int] = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) lowerCAmelCase_ : Dict = [ (torch.zeros(lowerCAmelCase__ ), torch.zeros(lowerCAmelCase__ )) for _ in range(self.num_layers ) ] lowerCAmelCase_ : List[str] = common_inputs["attention_mask"] if self.use_past: lowerCAmelCase_ : Tuple = ordered_inputs["attention_mask"].dtype lowerCAmelCase_ : str = torch.cat( [ordered_inputs["attention_mask"], torch.ones(lowerCAmelCase__ ,lowerCAmelCase__ ,dtype=lowerCAmelCase__ )] ,dim=1 ) return ordered_inputs @property def UpperCAmelCase_ ( self : str ) -> int: '''simple docstring''' return 13
659
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def UpperCamelCase ( ): lowerCAmelCase_ : Dict = HfArgumentParser(snake_case__) lowerCAmelCase_ : Dict = parser.parse_args_into_dataclasses()[0] lowerCAmelCase_ : List[Any] = TensorFlowBenchmark(args=snake_case__) try: lowerCAmelCase_ : str = parser.parse_args_into_dataclasses()[0] except ValueError as e: lowerCAmelCase_ : Optional[Any] = "Arg --no_{0} is no longer used, please use --no-{0} instead." lowerCAmelCase_ : Tuple = " ".join(str(snake_case__).split(" ")[:-1]) lowerCAmelCase_ : List[Any] = "" lowerCAmelCase_ : Optional[Any] = eval(str(snake_case__).split(" ")[-1]) lowerCAmelCase_ : List[Any] = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:]) else: wrong_args.append(snake_case__) if len(snake_case__) > 0: lowerCAmelCase_ : int = full_error_msg + begin_error_msg + str(snake_case__) raise ValueError(snake_case__) benchmark.run() if __name__ == "__main__": main()
659
1
import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) _lowercase = { '''iou_prediction_head.layers.0''': '''iou_prediction_head.proj_in''', '''iou_prediction_head.layers.1''': '''iou_prediction_head.layers.0''', '''iou_prediction_head.layers.2''': '''iou_prediction_head.proj_out''', '''mask_decoder.output_upscaling.0''': '''mask_decoder.upscale_conv1''', '''mask_decoder.output_upscaling.1''': '''mask_decoder.upscale_layer_norm''', '''mask_decoder.output_upscaling.3''': '''mask_decoder.upscale_conv2''', '''mask_downscaling.0''': '''mask_embed.conv1''', '''mask_downscaling.1''': '''mask_embed.layer_norm1''', '''mask_downscaling.3''': '''mask_embed.conv2''', '''mask_downscaling.4''': '''mask_embed.layer_norm2''', '''mask_downscaling.6''': '''mask_embed.conv3''', '''point_embeddings''': '''point_embed''', '''pe_layer.positional_encoding_gaussian_matrix''': '''shared_embedding.positional_embedding''', '''image_encoder''': '''vision_encoder''', '''neck.0''': '''neck.conv1''', '''neck.1''': '''neck.layer_norm1''', '''neck.2''': '''neck.conv2''', '''neck.3''': '''neck.layer_norm2''', '''patch_embed.proj''': '''patch_embed.projection''', '''.norm''': '''.layer_norm''', '''blocks''': '''layers''', } def UpperCamelCase ( snake_case__): lowerCAmelCase_ : int = {} state_dict.pop("pixel_mean" , snake_case__) state_dict.pop("pixel_std" , snake_case__) lowerCAmelCase_ : List[Any] = R".*.output_hypernetworks_mlps.(\d+).layers.(\d+).*" for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: lowerCAmelCase_ : Dict = key.replace(snake_case__ , snake_case__) if re.match(snake_case__ , snake_case__): lowerCAmelCase_ : Any = int(re.match(snake_case__ , snake_case__).group(2)) if layer_nb == 0: lowerCAmelCase_ : List[Any] = key.replace("layers.0" , "proj_in") elif layer_nb == 1: lowerCAmelCase_ : List[Any] = key.replace("layers.1" , "layers.0") elif layer_nb == 2: lowerCAmelCase_ : int = key.replace("layers.2" , "proj_out") lowerCAmelCase_ : int = value lowerCAmelCase_ : Optional[int] = model_state_dict[ "prompt_encoder.shared_embedding.positional_embedding" ] return model_state_dict def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__="ybelkada/segment-anything"): lowerCAmelCase_ : Optional[int] = hf_hub_download(snake_case__ , F'''checkpoints/{model_name}.pth''') if "sam_vit_b" in model_name: lowerCAmelCase_ : Optional[Any] = SamConfig() elif "sam_vit_l" in model_name: lowerCAmelCase_ : Optional[int] = SamVisionConfig( hidden_size=10_24 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) lowerCAmelCase_ : Union[str, Any] = SamConfig( vision_config=snake_case__ , ) elif "sam_vit_h" in model_name: lowerCAmelCase_ : Optional[Any] = SamVisionConfig( hidden_size=12_80 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) lowerCAmelCase_ : Tuple = SamConfig( vision_config=snake_case__ , ) lowerCAmelCase_ : Optional[Any] = torch.load(snake_case__ , map_location="cpu") lowerCAmelCase_ : Union[str, Any] = replace_keys(snake_case__) lowerCAmelCase_ : List[Any] = SamImageProcessor() lowerCAmelCase_ : Any = SamProcessor(image_processor=snake_case__) lowerCAmelCase_ : Any = SamModel(snake_case__) hf_model.load_state_dict(snake_case__) lowerCAmelCase_ : Dict = hf_model.to("cuda") lowerCAmelCase_ : List[str] = "https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png" lowerCAmelCase_ : List[Any] = Image.open(requests.get(snake_case__ , stream=snake_case__).raw).convert("RGB") lowerCAmelCase_ : Optional[int] = [[[4_00, 6_50]]] lowerCAmelCase_ : int = [[1]] lowerCAmelCase_ : Optional[Any] = processor(images=np.array(snake_case__) , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Optional[Any] = hf_model(**snake_case__) lowerCAmelCase_ : Optional[int] = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579_890_251_159_668 lowerCAmelCase_ : Any = processor( images=np.array(snake_case__) , input_points=snake_case__ , input_labels=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Optional[Any] = hf_model(**snake_case__) lowerCAmelCase_ : Union[str, Any] = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_712_603_092_193_604 lowerCAmelCase_ : Tuple = ((75, 2_75, 17_25, 8_50),) lowerCAmelCase_ : Optional[Any] = processor(images=np.array(snake_case__) , input_boxes=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : List[Any] = hf_model(**snake_case__) lowerCAmelCase_ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.8_686_015_605_926_514 # Test with 2 points and 1 image. lowerCAmelCase_ : int = [[[4_00, 6_50], [8_00, 6_50]]] lowerCAmelCase_ : Optional[Any] = [[1, 1]] lowerCAmelCase_ : List[Any] = processor( images=np.array(snake_case__) , input_points=snake_case__ , input_labels=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Tuple = hf_model(**snake_case__) lowerCAmelCase_ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_936_047_792_434_692 if __name__ == "__main__": _lowercase = argparse.ArgumentParser() _lowercase = ['''sam_vit_b_01ec64''', '''sam_vit_h_4b8939''', '''sam_vit_l_0b3195'''] parser.add_argument( '''--model_name''', default='''sam_vit_h_4b8939''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model and processor to the hub after converting''', ) parser.add_argument( '''--model_hub_id''', default='''ybelkada/segment-anything''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) _lowercase = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
659
from collections import defaultdict from pathlib import Path import pandas as pd from rouge_cli import calculate_rouge_path from utils import calculate_rouge _lowercase = [ '''Prosecutor: "No videos were used in the crash investigation" German papers say they saw a cell phone video of the''' ''' final seconds on board Flight 9525. The Germanwings co-pilot says he had a "previous episode of severe''' ''' depression\" German airline confirms it knew of Andreas Lubitz\'s depression years before he took control.''', '''The Palestinian Authority officially becomes the 123rd member of the International Criminal Court. The formal''' ''' accession was marked with a ceremony at The Hague, in the Netherlands. The Palestinians signed the ICC\'s''' ''' founding Rome Statute in January. Israel and the United States opposed the Palestinians\' efforts to join the''' ''' body.''', '''Amnesty International releases its annual report on the death penalty. The report catalogs the use of''' ''' state-sanctioned killing as a punitive measure across the globe. At least 607 people were executed around the''' ''' world in 2014, compared to 778 in 2013. The U.S. remains one of the worst offenders for imposing capital''' ''' punishment.''', ] _lowercase = [ '''Marseille prosecutor says "so far no videos were used in the crash investigation" despite media reports .''' ''' Journalists at Bild and Paris Match are "very confident" the video clip is real, an editor says . Andreas Lubitz''' ''' had informed his Lufthansa training school of an episode of severe depression, airline says .''', '''Membership gives the ICC jurisdiction over alleged crimes committed in Palestinian territories since last June .''' ''' Israel and the United States opposed the move, which could open the door to war crimes investigations against''' ''' Israelis .''', '''Amnesty\'s annual death penalty report catalogs encouraging signs, but setbacks in numbers of those sentenced to''' ''' death . Organization claims that governments around the world are using the threat of terrorism to advance''' ''' executions . The number of executions worldwide has gone down by almost 22% compared with 2013, but death''' ''' sentences up by 28% .''', ] def UpperCamelCase ( ): lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , bootstrap_aggregation=snake_case__ , rouge_keys=["rouge2", "rougeL"]) assert isinstance(snake_case__ , snake_case__) lowerCAmelCase_ : str = calculate_rouge(snake_case__ , snake_case__ , bootstrap_aggregation=snake_case__ , rouge_keys=["rouge2"]) assert ( pd.DataFrame(no_aggregation["rouge2"]).fmeasure.mean() == pd.DataFrame(no_aggregation_just_ra["rouge2"]).fmeasure.mean() ) def UpperCamelCase ( ): lowerCAmelCase_ : str = "rougeLsum" lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=[k])[k] lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=[k])[k] assert score > score_no_sep def UpperCamelCase ( ): lowerCAmelCase_ : int = ["rouge1", "rouge2", "rougeL"] lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=snake_case__) lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=snake_case__) assert score_sep == score_no_sep def UpperCamelCase ( ): lowerCAmelCase_ : List[str] = [ "Her older sister, Margot Frank, died in 1945, a month earlier than previously thought.", "Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports .", ] lowerCAmelCase_ : Dict = [ "Margot Frank, died in 1945, a month earlier than previously thought.", "Prosecutor: \"No videos were used in the crash investigation\" German papers say they saw a cell phone video of" " the final seconds on board Flight 9525.", ] assert calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__) == calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__) def UpperCamelCase ( ): lowerCAmelCase_ : Optional[int] = [ "\" \"a person who has such a video needs to immediately give it to the investigators,\" prosecutor says .<n> \"it is a very disturbing scene,\" editor-in-chief of bild online tells \"erin burnett: outfront\" " ] lowerCAmelCase_ : Any = [ " Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports . Journalists at Bild and Paris Match are \"very confident\" the video clip is real, an editor says . Andreas Lubitz had informed his Lufthansa training school of an episode of severe depression, airline says ." ] lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , rouge_keys=["rougeLsum"] , newline_sep=snake_case__)["rougeLsum"] lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , rouge_keys=["rougeLsum"])["rougeLsum"] assert new_score > prev_score def UpperCamelCase ( ): lowerCAmelCase_ : int = Path("examples/seq2seq/test_data/wmt_en_ro") lowerCAmelCase_ : Dict = calculate_rouge_path(data_dir.joinpath("test.source") , data_dir.joinpath("test.target")) assert isinstance(snake_case__ , snake_case__) lowerCAmelCase_ : Any = calculate_rouge_path( data_dir.joinpath("test.source") , data_dir.joinpath("test.target") , bootstrap_aggregation=snake_case__) assert isinstance(snake_case__ , snake_case__)
659
1
# This model implementation is heavily inspired by https://github.com/haofanwang/ControlNet-for-Diffusers/ import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlNetImgaImgPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel from diffusers.utils import floats_tensor, load_image, load_numpy, randn_tensor, slow, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, ) enable_full_determinism() class __snake_case ( snake_case__ , snake_case__ , snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = StableDiffusionControlNetImgaImgPipeline UpperCamelCase_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} UpperCamelCase_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS UpperCamelCase_ = IMAGE_TO_IMAGE_IMAGE_PARAMS.union({'control_image'} ) UpperCamelCase_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def UpperCAmelCase_ ( self : Optional[int] ) -> Union[str, Any]: '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase_ : List[str] = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") ,up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") ,cross_attention_dim=32 ,) torch.manual_seed(0 ) lowerCAmelCase_ : str = ControlNetModel( block_out_channels=(32, 64) ,layers_per_block=2 ,in_channels=4 ,down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") ,cross_attention_dim=32 ,conditioning_embedding_out_channels=(16, 32) ,) torch.manual_seed(0 ) lowerCAmelCase_ : int = DDIMScheduler( beta_start=0.00_085 ,beta_end=0.012 ,beta_schedule="scaled_linear" ,clip_sample=lowerCAmelCase__ ,set_alpha_to_one=lowerCAmelCase__ ,) torch.manual_seed(0 ) lowerCAmelCase_ : Tuple = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] ,up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] ,latent_channels=4 ,) torch.manual_seed(0 ) lowerCAmelCase_ : Union[str, Any] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=10_00 ,) lowerCAmelCase_ : str = CLIPTextModel(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) lowerCAmelCase_ : Tuple = { "unet": unet, "controlnet": controlnet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Optional[Any]=0 ) -> Dict: '''simple docstring''' if str(lowerCAmelCase__ ).startswith("mps" ): lowerCAmelCase_ : Optional[Any] = torch.manual_seed(lowerCAmelCase__ ) else: lowerCAmelCase_ : Optional[Any] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : str = 2 lowerCAmelCase_ : Optional[Any] = randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) ,generator=lowerCAmelCase__ ,device=torch.device(lowerCAmelCase__ ) ,) lowerCAmelCase_ : Tuple = floats_tensor(control_image.shape ,rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = image.cpu().permute(0 ,2 ,3 ,1 )[0] lowerCAmelCase_ : Optional[Any] = Image.fromarray(np.uinta(lowerCAmelCase__ ) ).convert("RGB" ).resize((64, 64) ) lowerCAmelCase_ : List[Any] = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", "image": image, "control_image": control_image, } return inputs def UpperCAmelCase_ ( self : List[str] ) -> Tuple: '''simple docstring''' return self._test_attention_slicing_forward_pass(expected_max_diff=2e-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() ,reason="XFormers attention is only available with CUDA and `xformers` installed" ,) def UpperCAmelCase_ ( self : Any ) -> Dict: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2e-3 ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> str: '''simple docstring''' self._test_inference_batch_single_identical(expected_max_diff=2e-3 ) class __snake_case ( snake_case__ , snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = StableDiffusionControlNetImgaImgPipeline UpperCamelCase_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'height', 'width'} UpperCamelCase_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS UpperCamelCase_ = frozenset([] ) # TO_DO: add image_params once refactored VaeImageProcessor.preprocess def UpperCAmelCase_ ( self : Optional[int] ) -> List[str]: '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase_ : Optional[int] = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") ,up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") ,cross_attention_dim=32 ,) torch.manual_seed(0 ) def init_weights(lowerCAmelCase__ : int ): if isinstance(lowerCAmelCase__ ,torch.nn.Convad ): torch.nn.init.normal(m.weight ) m.bias.data.fill_(1.0 ) lowerCAmelCase_ : List[Any] = ControlNetModel( block_out_channels=(32, 64) ,layers_per_block=2 ,in_channels=4 ,down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") ,cross_attention_dim=32 ,conditioning_embedding_out_channels=(16, 32) ,) controlneta.controlnet_down_blocks.apply(lowerCAmelCase__ ) torch.manual_seed(0 ) lowerCAmelCase_ : List[str] = ControlNetModel( block_out_channels=(32, 64) ,layers_per_block=2 ,in_channels=4 ,down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") ,cross_attention_dim=32 ,conditioning_embedding_out_channels=(16, 32) ,) controlneta.controlnet_down_blocks.apply(lowerCAmelCase__ ) torch.manual_seed(0 ) lowerCAmelCase_ : Dict = DDIMScheduler( beta_start=0.00_085 ,beta_end=0.012 ,beta_schedule="scaled_linear" ,clip_sample=lowerCAmelCase__ ,set_alpha_to_one=lowerCAmelCase__ ,) torch.manual_seed(0 ) lowerCAmelCase_ : Dict = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] ,up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] ,latent_channels=4 ,) torch.manual_seed(0 ) lowerCAmelCase_ : Optional[int] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=10_00 ,) lowerCAmelCase_ : str = CLIPTextModel(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) lowerCAmelCase_ : List[str] = MultiControlNetModel([controlneta, controlneta] ) lowerCAmelCase_ : Tuple = { "unet": unet, "controlnet": controlnet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : str ,lowerCAmelCase__ : str=0 ) -> Optional[Any]: '''simple docstring''' if str(lowerCAmelCase__ ).startswith("mps" ): lowerCAmelCase_ : List[Any] = torch.manual_seed(lowerCAmelCase__ ) else: lowerCAmelCase_ : List[Any] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = 2 lowerCAmelCase_ : Union[str, Any] = [ randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) ,generator=lowerCAmelCase__ ,device=torch.device(lowerCAmelCase__ ) ,), randn_tensor( (1, 3, 32 * controlnet_embedder_scale_factor, 32 * controlnet_embedder_scale_factor) ,generator=lowerCAmelCase__ ,device=torch.device(lowerCAmelCase__ ) ,), ] lowerCAmelCase_ : Dict = floats_tensor(control_image[0].shape ,rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) lowerCAmelCase_ : str = image.cpu().permute(0 ,2 ,3 ,1 )[0] lowerCAmelCase_ : Union[str, Any] = Image.fromarray(np.uinta(lowerCAmelCase__ ) ).convert("RGB" ).resize((64, 64) ) lowerCAmelCase_ : Any = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", "image": image, "control_image": control_image, } return inputs def UpperCAmelCase_ ( self : Any ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : int = self.get_dummy_components() lowerCAmelCase_ : Optional[int] = self.pipeline_class(**lowerCAmelCase__ ) pipe.to(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = 10.0 lowerCAmelCase_ : Tuple = 4 lowerCAmelCase_ : Any = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Any = steps lowerCAmelCase_ : Tuple = scale lowerCAmelCase_ : List[Any] = pipe(**lowerCAmelCase__ )[0] lowerCAmelCase_ : Optional[Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : str = steps lowerCAmelCase_ : Tuple = scale lowerCAmelCase_ : Optional[int] = pipe(**lowerCAmelCase__ ,control_guidance_start=0.1 ,control_guidance_end=0.2 )[0] lowerCAmelCase_ : Optional[Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = steps lowerCAmelCase_ : Optional[Any] = scale lowerCAmelCase_ : int = pipe(**lowerCAmelCase__ ,control_guidance_start=[0.1, 0.3] ,control_guidance_end=[0.2, 0.7] )[0] lowerCAmelCase_ : int = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = steps lowerCAmelCase_ : Optional[Any] = scale lowerCAmelCase_ : Union[str, Any] = pipe(**lowerCAmelCase__ ,control_guidance_start=0.4 ,control_guidance_end=[0.5, 0.8] )[0] # make sure that all outputs are different assert np.sum(np.abs(output_a - output_a ) ) > 1e-3 assert np.sum(np.abs(output_a - output_a ) ) > 1e-3 assert np.sum(np.abs(output_a - output_a ) ) > 1e-3 def UpperCAmelCase_ ( self : str ) -> Tuple: '''simple docstring''' return self._test_attention_slicing_forward_pass(expected_max_diff=2e-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() ,reason="XFormers attention is only available with CUDA and `xformers` installed" ,) def UpperCAmelCase_ ( self : List[str] ) -> Any: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2e-3 ) def UpperCAmelCase_ ( self : Dict ) -> int: '''simple docstring''' self._test_inference_batch_single_identical(expected_max_diff=2e-3 ) def UpperCAmelCase_ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = self.get_dummy_components() lowerCAmelCase_ : str = self.pipeline_class(**lowerCAmelCase__ ) pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) with tempfile.TemporaryDirectory() as tmpdir: try: # save_pretrained is not implemented for Multi-ControlNet pipe.save_pretrained(lowerCAmelCase__ ) except NotImplementedError: pass @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : str ) -> Optional[Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : Optional[int] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = ControlNetModel.from_pretrained("lllyasviel/sd-controlnet-canny" ) lowerCAmelCase_ : Union[str, Any] = StableDiffusionControlNetImgaImgPipeline.from_pretrained( "runwayml/stable-diffusion-v1-5" ,safety_checker=lowerCAmelCase__ ,controlnet=lowerCAmelCase__ ) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = torch.Generator(device="cpu" ).manual_seed(0 ) lowerCAmelCase_ : Any = "evil space-punk bird" lowerCAmelCase_ : str = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png" ).resize((5_12, 5_12) ) lowerCAmelCase_ : Any = load_image( "https://huggingface.co/lllyasviel/sd-controlnet-canny/resolve/main/images/bird.png" ).resize((5_12, 5_12) ) lowerCAmelCase_ : Union[str, Any] = pipe( lowerCAmelCase__ ,lowerCAmelCase__ ,control_image=lowerCAmelCase__ ,generator=lowerCAmelCase__ ,output_type="np" ,num_inference_steps=50 ,strength=0.6 ,) lowerCAmelCase_ : Optional[int] = output.images[0] assert image.shape == (5_12, 5_12, 3) lowerCAmelCase_ : Optional[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/img2img.npy" ) assert np.abs(expected_image - image ).max() < 9e-2
659
import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = LEDTokenizer UpperCamelCase_ = LEDTokenizerFast UpperCamelCase_ = True def UpperCAmelCase_ ( self : List[Any] ) -> Optional[int]: '''simple docstring''' super().setUp() lowerCAmelCase_ : Union[str, Any] = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] lowerCAmelCase_ : Tuple = dict(zip(lowerCAmelCase__ ,range(len(lowerCAmelCase__ ) ) ) ) lowerCAmelCase_ : int = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] lowerCAmelCase_ : Union[str, Any] = {"unk_token": "<unk>"} lowerCAmelCase_ : List[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase_ : Any = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file ,"w" ,encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase__ ) + "\n" ) with open(self.merges_file ,"w" ,encoding="utf-8" ) as fp: fp.write("\n".join(lowerCAmelCase__ ) ) def UpperCAmelCase_ ( self : List[Any] ,**lowerCAmelCase__ : int ) -> Tuple: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ,**lowerCAmelCase__ : Optional[int] ) -> List[Any]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCAmelCase__ ) def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : int ) -> List[str]: '''simple docstring''' return "lower newer", "lower newer" @cached_property def UpperCAmelCase_ ( self : int ) -> Union[str, Any]: '''simple docstring''' return LEDTokenizer.from_pretrained("allenai/led-base-16384" ) @cached_property def UpperCAmelCase_ ( self : List[str] ) -> Dict: '''simple docstring''' return LEDTokenizerFast.from_pretrained("allenai/led-base-16384" ) @require_torch def UpperCAmelCase_ ( self : int ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = ["A long paragraph for summarization.", "Another paragraph for summarization."] lowerCAmelCase_ : int = [0, 2_50, 2_51, 1_78_18, 13, 3_91_86, 19_38, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Any = tokenizer(lowerCAmelCase__ ,max_length=len(lowerCAmelCase__ ) ,padding=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual((2, 9) ,batch.input_ids.shape ) self.assertEqual((2, 9) ,batch.attention_mask.shape ) lowerCAmelCase_ : int = batch.input_ids.tolist()[0] self.assertListEqual(lowerCAmelCase__ ,lowerCAmelCase__ ) @require_torch def UpperCAmelCase_ ( self : Dict ) -> Any: '''simple docstring''' lowerCAmelCase_ : int = ["A long paragraph for summarization.", "Another paragraph for summarization."] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[Any] = tokenizer(lowerCAmelCase__ ,padding=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIn("input_ids" ,lowerCAmelCase__ ) self.assertIn("attention_mask" ,lowerCAmelCase__ ) self.assertNotIn("labels" ,lowerCAmelCase__ ) self.assertNotIn("decoder_attention_mask" ,lowerCAmelCase__ ) @require_torch def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : int = [ "Summary of the text.", "Another summary.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[int] = tokenizer(text_target=lowerCAmelCase__ ,max_length=32 ,padding="max_length" ,return_tensors="pt" ) self.assertEqual(32 ,targets["input_ids"].shape[1] ) @require_torch def UpperCAmelCase_ ( self : Tuple ) -> List[str]: '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Tuple = tokenizer( ["I am a small frog" * 10_24, "I am a small frog"] ,padding=lowerCAmelCase__ ,truncation=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual(batch.input_ids.shape ,(2, 51_22) ) @require_torch def UpperCAmelCase_ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Tuple = ["A long paragraph for summarization."] lowerCAmelCase_ : Dict = [ "Summary of the text.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[Any] = tokenizer(lowerCAmelCase__ ,return_tensors="pt" ) lowerCAmelCase_ : Optional[Any] = tokenizer(text_target=lowerCAmelCase__ ,return_tensors="pt" ) lowerCAmelCase_ : List[str] = inputs["input_ids"] lowerCAmelCase_ : Any = targets["input_ids"] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) @require_torch def UpperCAmelCase_ ( self : str ) -> Tuple: '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : str = ["Summary of the text.", "Another summary."] lowerCAmelCase_ : str = [[0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, -1, -1]] lowerCAmelCase_ : List[Any] = tokenizer(lowerCAmelCase__ ,padding=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = [[0] * len(lowerCAmelCase__ ) for x in encoded_output["input_ids"]] lowerCAmelCase_ : Optional[int] = tokenizer.pad(lowerCAmelCase__ ) self.assertSequenceEqual(outputs["global_attention_mask"] ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' pass def UpperCAmelCase_ ( self : str ) -> Union[str, Any]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCAmelCase_ : Dict = self.rust_tokenizer_class.from_pretrained(lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = self.tokenizer_class.from_pretrained(lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Dict = "A, <mask> AllenNLP sentence." lowerCAmelCase_ : Tuple = tokenizer_r.encode_plus(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,return_token_type_ids=lowerCAmelCase__ ) lowerCAmelCase_ : int = tokenizer_p.encode_plus(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,return_token_type_ids=lowerCAmelCase__ ) self.assertEqual(sum(tokens_r["token_type_ids"] ) ,sum(tokens_p["token_type_ids"] ) ) self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) ,sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) ,) lowerCAmelCase_ : Any = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) lowerCAmelCase_ : Union[str, Any] = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) self.assertSequenceEqual(tokens_p["input_ids"] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( lowerCAmelCase__ ,["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( lowerCAmelCase__ ,["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] )
659
1
class __snake_case : """simple docstring""" def __init__( self : Union[str, Any] ,lowerCAmelCase__ : str = "" ,lowerCAmelCase__ : bool = False ) -> None: '''simple docstring''' lowerCAmelCase_ : dict[str, RadixNode] = {} # A node will be a leaf if the tree contains its word lowerCAmelCase_ : Optional[int] = is_leaf lowerCAmelCase_ : List[str] = prefix def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : str ) -> tuple[str, str, str]: '''simple docstring''' lowerCAmelCase_ : List[str] = 0 for q, w in zip(self.prefix ,lowerCAmelCase__ ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : list[str] ) -> None: '''simple docstring''' for word in words: self.insert(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : str ) -> None: '''simple docstring''' if self.prefix == word: lowerCAmelCase_ : Optional[Any] = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: lowerCAmelCase_ : Optional[int] = RadixNode(prefix=lowerCAmelCase__ ,is_leaf=lowerCAmelCase__ ) else: lowerCAmelCase_ : Optional[Any] = self.nodes[word[0]] lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = incoming_node.match( lowerCAmelCase__ ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: lowerCAmelCase_ : Dict = remaining_prefix lowerCAmelCase_ : str = self.nodes[matching_string[0]] lowerCAmelCase_ : Dict = RadixNode(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Any = aux_node if remaining_word == "": lowerCAmelCase_ : Optional[Any] = True else: self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : str ) -> bool: '''simple docstring''' lowerCAmelCase_ : List[str] = self.nodes.get(word[0] ,lowerCAmelCase__ ) if not incoming_node: return False else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : str ) -> bool: '''simple docstring''' lowerCAmelCase_ : int = self.nodes.get(word[0] ,lowerCAmelCase__ ) if not incoming_node: return False else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(lowerCAmelCase__ ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: lowerCAmelCase_ : int = list(self.nodes.values() )[0] lowerCAmelCase_ : List[Any] = merging_node.is_leaf self.prefix += merging_node.prefix lowerCAmelCase_ : int = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: lowerCAmelCase_ : List[str] = False # If there is 1 edge, we merge it with its child else: lowerCAmelCase_ : Union[str, Any] = list(incoming_node.nodes.values() )[0] lowerCAmelCase_ : Optional[int] = merging_node.is_leaf incoming_node.prefix += merging_node.prefix lowerCAmelCase_ : List[str] = merging_node.nodes return True def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : int = 0 ) -> None: '''simple docstring''' if self.prefix != "": print("-" * height ,self.prefix ," (leaf)" if self.is_leaf else "" ) for value in self.nodes.values(): value.print_tree(height + 1 ) def UpperCamelCase ( ): lowerCAmelCase_ : List[Any] = "banana bananas bandana band apple all beast".split() lowerCAmelCase_ : Optional[Any] = RadixNode() root.insert_many(snake_case__) assert all(root.find(snake_case__) for word in words) assert not root.find("bandanas") assert not root.find("apps") root.delete("all") assert not root.find("all") root.delete("banana") assert not root.find("banana") assert root.find("bananas") return True def UpperCamelCase ( ): assert test_trie() def UpperCamelCase ( ): lowerCAmelCase_ : str = RadixNode() lowerCAmelCase_ : str = "banana bananas bandanas bandana band apple all beast".split() root.insert_many(snake_case__) print("Words:" , snake_case__) print("Tree:") root.print_tree() if __name__ == "__main__": main()
659
from ....configuration_utils import PretrainedConfig from ....utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''Visual-Attention-Network/van-base''': ( '''https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json''' ), } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = 'van' def __init__( self : List[str] ,lowerCAmelCase__ : int=2_24 ,lowerCAmelCase__ : Optional[int]=3 ,lowerCAmelCase__ : Dict=[7, 3, 3, 3] ,lowerCAmelCase__ : List[str]=[4, 2, 2, 2] ,lowerCAmelCase__ : Union[str, Any]=[64, 1_28, 3_20, 5_12] ,lowerCAmelCase__ : Union[str, Any]=[3, 3, 12, 3] ,lowerCAmelCase__ : Any=[8, 8, 4, 4] ,lowerCAmelCase__ : Optional[int]="gelu" ,lowerCAmelCase__ : List[str]=0.02 ,lowerCAmelCase__ : Optional[Any]=1e-6 ,lowerCAmelCase__ : Dict=1e-2 ,lowerCAmelCase__ : Union[str, Any]=0.0 ,lowerCAmelCase__ : Optional[Any]=0.0 ,**lowerCAmelCase__ : List[str] ,) -> Tuple: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = image_size lowerCAmelCase_ : List[str] = num_channels lowerCAmelCase_ : str = patch_sizes lowerCAmelCase_ : Optional[Any] = strides lowerCAmelCase_ : List[Any] = hidden_sizes lowerCAmelCase_ : int = depths lowerCAmelCase_ : int = mlp_ratios lowerCAmelCase_ : str = hidden_act lowerCAmelCase_ : List[str] = initializer_range lowerCAmelCase_ : Dict = layer_norm_eps lowerCAmelCase_ : str = layer_scale_init_value lowerCAmelCase_ : Tuple = drop_path_rate lowerCAmelCase_ : Dict = dropout_rate
659
1
import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch _lowercase = logging.get_logger(__name__) class __snake_case : """simple docstring""" def __init__( self : Dict ,lowerCAmelCase__ : str = None ,lowerCAmelCase__ : uuid.UUID = None ,lowerCAmelCase__ : Optional[Any]=None ,lowerCAmelCase__ : List[str]=None ) -> List[str]: '''simple docstring''' if not conversation_id: lowerCAmelCase_ : List[Any] = uuid.uuida() if past_user_inputs is None: lowerCAmelCase_ : List[Any] = [] if generated_responses is None: lowerCAmelCase_ : str = [] lowerCAmelCase_ : uuid.UUID = conversation_id lowerCAmelCase_ : List[str] = past_user_inputs lowerCAmelCase_ : List[str] = generated_responses lowerCAmelCase_ : Optional[str] = text def __eq__( self : Tuple ,lowerCAmelCase__ : List[str] ) -> Optional[Any]: '''simple docstring''' if not isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : str ,lowerCAmelCase__ : bool = False ) -> int: '''simple docstring''' if self.new_user_input: if overwrite: logger.warning( f'''User input added while unprocessed input was existing: "{self.new_user_input}" was overwritten ''' f'''with: "{text}".''' ) lowerCAmelCase_ : List[Any] = text else: logger.warning( f'''User input added while unprocessed input was existing: "{self.new_user_input}" new input ''' f'''ignored: "{text}". Set `overwrite` to True to overwrite unprocessed user input''' ) else: lowerCAmelCase_ : List[Any] = text def UpperCAmelCase_ ( self : List[Any] ) -> List[str]: '''simple docstring''' if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) lowerCAmelCase_ : str = None def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : str ) -> List[str]: '''simple docstring''' self.generated_responses.append(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Dict ) -> List[str]: '''simple docstring''' for user_input, generated_response in zip(self.past_user_inputs ,self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : List[str] = f'''Conversation id: {self.uuid} \n''' for is_user, text in self.iter_texts(): lowerCAmelCase_ : Optional[Any] = "user" if is_user else "bot" output += f'''{name} >> {text} \n''' return output @add_end_docstrings( snake_case__ , r'\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n ' , ) class __snake_case ( snake_case__ ): """simple docstring""" def __init__( self : Optional[int] ,*lowerCAmelCase__ : str ,**lowerCAmelCase__ : Optional[int] ) -> Union[str, Any]: '''simple docstring''' super().__init__(*lowerCAmelCase__ ,**lowerCAmelCase__ ) if self.tokenizer.pad_token_id is None: lowerCAmelCase_ : List[Any] = self.tokenizer.eos_token def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : Any=None ,lowerCAmelCase__ : List[str]=None ,lowerCAmelCase__ : Optional[Any]=None ,**lowerCAmelCase__ : Tuple ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : List[Any] = {} lowerCAmelCase_ : List[Any] = {} lowerCAmelCase_ : str = {} if min_length_for_response is not None: lowerCAmelCase_ : Dict = min_length_for_response if minimum_tokens is not None: lowerCAmelCase_ : List[str] = minimum_tokens if "max_length" in generate_kwargs: lowerCAmelCase_ : List[Any] = generate_kwargs["max_length"] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: lowerCAmelCase_ : List[str] = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(lowerCAmelCase__ ) return preprocess_params, forward_params, postprocess_params def __call__( self : Optional[int] ,lowerCAmelCase__ : Union[Conversation, List[Conversation]] ,lowerCAmelCase__ : Union[str, Any]=0 ,**lowerCAmelCase__ : Optional[Any] ) -> Dict: '''simple docstring''' lowerCAmelCase_ : List[str] = super().__call__(lowerCAmelCase__ ,num_workers=lowerCAmelCase__ ,**lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) and len(lowerCAmelCase__ ) == 1: return outputs[0] return outputs def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : Conversation ,lowerCAmelCase__ : Dict=32 ) -> Dict[str, Any]: '''simple docstring''' if not isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ): raise ValueError("ConversationalPipeline, expects Conversation as inputs" ) if conversation.new_user_input is None: raise ValueError( f'''Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. ''' "Add user inputs with the conversation's `add_user_input` method" ) if hasattr(self.tokenizer ,"_build_conversation_input_ids" ): lowerCAmelCase_ : Optional[Any] = self.tokenizer._build_conversation_input_ids(lowerCAmelCase__ ) else: # If the tokenizer cannot handle conversations, we default to only the old version lowerCAmelCase_ : List[Any] = self._legacy_parse_and_tokenize(lowerCAmelCase__ ) if self.framework == "pt": lowerCAmelCase_ : int = torch.LongTensor([input_ids] ) elif self.framework == "tf": lowerCAmelCase_ : str = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : str ,lowerCAmelCase__ : int=10 ,**lowerCAmelCase__ : Optional[int] ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = generate_kwargs.get("max_length" ,self.model.config.max_length ) lowerCAmelCase_ : str = model_inputs["input_ids"].shape[1] if max_length - minimum_tokens < n: logger.warning(f'''Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})''' ) lowerCAmelCase_ : str = max_length - minimum_tokens lowerCAmelCase_ : int = model_inputs["input_ids"][:, -trim:] if "attention_mask" in model_inputs: lowerCAmelCase_ : Any = model_inputs["attention_mask"][:, -trim:] lowerCAmelCase_ : List[str] = model_inputs.pop("conversation" ) lowerCAmelCase_ : Any = max_length lowerCAmelCase_ : Tuple = self.model.generate(**lowerCAmelCase__ ,**lowerCAmelCase__ ) if self.model.config.is_encoder_decoder: lowerCAmelCase_ : Any = 1 else: lowerCAmelCase_ : Union[str, Any] = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def UpperCAmelCase_ ( self : Dict ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : Optional[Any]=True ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : str = model_outputs["output_ids"] lowerCAmelCase_ : Union[str, Any] = self.tokenizer.decode( output_ids[0] ,skip_special_tokens=lowerCAmelCase__ ,clean_up_tokenization_spaces=lowerCAmelCase__ ,) lowerCAmelCase_ : Optional[Any] = model_outputs["conversation"] conversation.mark_processed() conversation.append_response(lowerCAmelCase__ ) return conversation def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : Conversation ) -> Dict: '''simple docstring''' lowerCAmelCase_ : Dict = self.tokenizer.eos_token_id lowerCAmelCase_ : Optional[int] = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ) ) if len(lowerCAmelCase__ ) > self.tokenizer.model_max_length: lowerCAmelCase_ : int = input_ids[-self.tokenizer.model_max_length :] return input_ids
659
from math import factorial def UpperCamelCase ( snake_case__ , snake_case__): # If either of the conditions are true, the function is being asked # to calculate a factorial of a negative number, which is not possible if n < k or k < 0: raise ValueError("Please enter positive integers for n and k where n >= k") return factorial(snake_case__) // (factorial(snake_case__) * factorial(n - k)) if __name__ == "__main__": print( '''The number of five-card hands possible from a standard''', f"fifty-two card deck is: {combinations(52, 5)}\n", ) print( '''If a class of 40 students must be arranged into groups of''', f"4 for group projects, there are {combinations(40, 4)} ways", '''to arrange them.\n''', ) print( '''If 10 teams are competing in a Formula One race, there''', f"are {combinations(10, 3)} ways that first, second and", '''third place can be awarded.''', )
659
1
import argparse import torch from transformers import ( SpeechTaConfig, SpeechTaFeatureExtractor, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaProcessor, SpeechTaTokenizer, logging, ) from transformers.tokenization_utils import AddedToken logging.set_verbosity_info() _lowercase = logging.get_logger('''transformers.models.speecht5''') _lowercase = { '''speech_encoder_prenet.layer_norm''': '''speecht5.encoder.prenet.feature_projection.layer_norm''', '''speech_encoder_prenet.post_extract_proj''': '''speecht5.encoder.prenet.feature_projection.projection''', '''speech_encoder_prenet.pos_conv.0''': '''speecht5.encoder.prenet.pos_conv_embed.conv''', '''speech_encoder_prenet.mask_emb''': '''speecht5.encoder.prenet.masked_spec_embed''', } _lowercase = { '''text_encoder_prenet.encoder_prenet.0''': '''speecht5.encoder.prenet.embed_tokens''', '''text_encoder_prenet.encoder_prenet.1.alpha''': '''speecht5.encoder.prenet.encode_positions.alpha''', } _lowercase = { '''speech_decoder_prenet.decoder_prenet.0.0.prenet.0.0''': '''speecht5.decoder.prenet.layers.0''', '''speech_decoder_prenet.decoder_prenet.0.0.prenet.1.0''': '''speecht5.decoder.prenet.layers.1''', '''speech_decoder_prenet.decoder_prenet.0.1''': '''speecht5.decoder.prenet.final_layer''', '''speech_decoder_prenet.decoder_prenet.1.alpha''': '''speecht5.decoder.prenet.encode_positions.alpha''', '''speech_decoder_prenet.spkembs_layer.0''': '''speecht5.decoder.prenet.speaker_embeds_layer''', } _lowercase = { '''speech_decoder_postnet.feat_out''': '''speech_decoder_postnet.feat_out''', '''speech_decoder_postnet.prob_out''': '''speech_decoder_postnet.prob_out''', '''speech_decoder_postnet.postnet.postnet.0.0''': '''speech_decoder_postnet.layers.0.conv''', '''speech_decoder_postnet.postnet.postnet.0.1''': '''speech_decoder_postnet.layers.0.batch_norm''', '''speech_decoder_postnet.postnet.postnet.1.0''': '''speech_decoder_postnet.layers.1.conv''', '''speech_decoder_postnet.postnet.postnet.1.1''': '''speech_decoder_postnet.layers.1.batch_norm''', '''speech_decoder_postnet.postnet.postnet.2.0''': '''speech_decoder_postnet.layers.2.conv''', '''speech_decoder_postnet.postnet.postnet.2.1''': '''speech_decoder_postnet.layers.2.batch_norm''', '''speech_decoder_postnet.postnet.postnet.3.0''': '''speech_decoder_postnet.layers.3.conv''', '''speech_decoder_postnet.postnet.postnet.3.1''': '''speech_decoder_postnet.layers.3.batch_norm''', '''speech_decoder_postnet.postnet.postnet.4.0''': '''speech_decoder_postnet.layers.4.conv''', '''speech_decoder_postnet.postnet.postnet.4.1''': '''speech_decoder_postnet.layers.4.batch_norm''', } _lowercase = { '''text_decoder_prenet.embed_tokens''': '''speecht5.decoder.prenet.embed_tokens''', } _lowercase = { '''text_decoder_postnet.output_projection''': '''text_decoder_postnet.lm_head''', } _lowercase = { '''encoder.layers.*.self_attn.k_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.k_proj''', '''encoder.layers.*.self_attn.v_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.v_proj''', '''encoder.layers.*.self_attn.q_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.q_proj''', '''encoder.layers.*.self_attn.out_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.out_proj''', '''encoder.layers.*.self_attn_layer_norm''': '''speecht5.encoder.wrapped_encoder.layers.*.layer_norm''', '''encoder.layers.*.fc1''': '''speecht5.encoder.wrapped_encoder.layers.*.feed_forward.intermediate_dense''', '''encoder.layers.*.fc2''': '''speecht5.encoder.wrapped_encoder.layers.*.feed_forward.output_dense''', '''encoder.layers.*.final_layer_norm''': '''speecht5.encoder.wrapped_encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''speecht5.encoder.wrapped_encoder.layer_norm''', '''encoder.pos_emb.pe_k''': '''speecht5.encoder.wrapped_encoder.embed_positions.pe_k''', } _lowercase = { '''decoder.layers.*.self_attn.k_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.k_proj''', '''decoder.layers.*.self_attn.v_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.v_proj''', '''decoder.layers.*.self_attn.q_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.q_proj''', '''decoder.layers.*.self_attn.out_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.out_proj''', '''decoder.layers.*.self_attn_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn_layer_norm''', '''decoder.layers.*.encoder_attn.k_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.k_proj''', '''decoder.layers.*.encoder_attn.v_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.v_proj''', '''decoder.layers.*.encoder_attn.q_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.q_proj''', '''decoder.layers.*.encoder_attn.out_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.out_proj''', '''decoder.layers.*.encoder_attn_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn_layer_norm''', '''decoder.layers.*.fc1''': '''speecht5.decoder.wrapped_decoder.layers.*.feed_forward.intermediate_dense''', '''decoder.layers.*.fc2''': '''speecht5.decoder.wrapped_decoder.layers.*.feed_forward.output_dense''', '''decoder.layers.*.final_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.final_layer_norm''', } _lowercase = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_TEXT_DECODER_PRENET, **MAPPING_TEXT_DECODER_POSTNET, } _lowercase = { **MAPPING_TEXT_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } _lowercase = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } _lowercase = [] _lowercase = [ '''encoder.version''', '''encoder.layers.*.norm_k.weight''', '''encoder.layers.*.norm_k.bias''', '''decoder.version''', '''decoder.layers.*.norm_k.weight''', '''decoder.layers.*.norm_k.bias''', '''decoder.pos_emb.pe_k''', '''speech_encoder_prenet.embed_positions._float_tensor''', '''text_decoder_prenet.embed_positions._float_tensor''', ] _lowercase = IGNORE_KEYS + [ '''encoder.proj''', '''text_encoder_prenet.*''', '''speech_decoder_prenet.*''', '''speech_decoder_postnet.*''', ] _lowercase = IGNORE_KEYS + [ '''encoder.proj''', '''speech_encoder_prenet.*''', '''text_decoder_prenet.*''', '''text_decoder_postnet.*''', ] _lowercase = IGNORE_KEYS + [ '''encoder.proj''', '''text_encoder_prenet.*''', '''text_decoder_prenet.*''', '''text_decoder_postnet.*''', ] def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__): for attribute in key.split("."): lowerCAmelCase_ : Tuple = getattr(snake_case__ , snake_case__) if weight_type is not None: lowerCAmelCase_ : Any = getattr(snake_case__ , snake_case__).shape else: lowerCAmelCase_ : int = hf_pointer.shape if hf_shape != value.shape: raise ValueError( F'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' F''' {value.shape} for {full_name}''') if weight_type == "weight": lowerCAmelCase_ : int = value elif weight_type == "weight_g": lowerCAmelCase_ : List[Any] = value elif weight_type == "weight_v": lowerCAmelCase_ : Dict = value elif weight_type == "bias": lowerCAmelCase_ : Optional[int] = value elif weight_type == "running_mean": lowerCAmelCase_ : Optional[int] = value elif weight_type == "running_var": lowerCAmelCase_ : Tuple = value elif weight_type == "num_batches_tracked": lowerCAmelCase_ : Union[str, Any] = value else: lowerCAmelCase_ : List[Any] = value logger.info(F'''{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.''') def UpperCamelCase ( snake_case__ , snake_case__): for key in ignore_keys: if key.endswith(".*"): if name.startswith(key[:-1]): return True elif ".*." in key: lowerCAmelCase_ , lowerCAmelCase_ : str = key.split(".*.") if prefix in name and suffix in name: return True elif key in name: return True return False def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : Optional[int] = [] if task == "s2t": lowerCAmelCase_ : Optional[int] = hf_model.speechta.encoder.prenet.feature_encoder lowerCAmelCase_ : int = MAPPING_S2T lowerCAmelCase_ : Union[str, Any] = IGNORE_KEYS_S2T elif task == "t2s": lowerCAmelCase_ : Any = None lowerCAmelCase_ : Any = MAPPING_T2S lowerCAmelCase_ : List[Any] = IGNORE_KEYS_T2S elif task == "s2s": lowerCAmelCase_ : List[str] = hf_model.speechta.encoder.prenet.feature_encoder lowerCAmelCase_ : Any = MAPPING_S2S lowerCAmelCase_ : Any = IGNORE_KEYS_S2S else: raise ValueError(F'''Unsupported task: {task}''') for name, value in fairseq_dict.items(): if should_ignore(snake_case__ , snake_case__): logger.info(F'''{name} was ignored''') continue lowerCAmelCase_ : Optional[int] = False if "conv_layers" in name: load_conv_layer( snake_case__ , snake_case__ , snake_case__ , snake_case__ , hf_model.config.feat_extract_norm == "group" , ) lowerCAmelCase_ : int = True else: for key, mapped_key in MAPPING.items(): # mapped_key = "speecht5." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if "*" in key: lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = key.split(".*.") if prefix in name and suffix in name: lowerCAmelCase_ : int = suffix # if key in name or key.split("w2v_model.")[-1] == name.split(".")[0]: if key in name: lowerCAmelCase_ : str = True if "*" in mapped_key: lowerCAmelCase_ : Optional[Any] = name.split(snake_case__)[0].split(".")[-2] lowerCAmelCase_ : Tuple = mapped_key.replace("*" , snake_case__) if "weight_g" in name: lowerCAmelCase_ : str = "weight_g" elif "weight_v" in name: lowerCAmelCase_ : Dict = "weight_v" elif "bias" in name: lowerCAmelCase_ : int = "bias" elif "weight" in name: lowerCAmelCase_ : str = "weight" elif "running_mean" in name: lowerCAmelCase_ : str = "running_mean" elif "running_var" in name: lowerCAmelCase_ : Any = "running_var" elif "num_batches_tracked" in name: lowerCAmelCase_ : List[str] = "num_batches_tracked" else: lowerCAmelCase_ : int = None set_recursively(snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__) continue if not is_used: unused_weights.append(snake_case__) logger.warning(F'''Unused weights: {unused_weights}''') def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : str = full_name.split("conv_layers.")[-1] lowerCAmelCase_ : Optional[Any] = name.split(".") lowerCAmelCase_ : List[str] = int(items[0]) lowerCAmelCase_ : List[str] = int(items[1]) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''') lowerCAmelCase_ : int = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''') elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''') lowerCAmelCase_ : Tuple = value logger.info(F'''Feat extract conv layer {layer_id} was initialized from {full_name}.''') elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.''') lowerCAmelCase_ : str = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''') elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( F'''{full_name} has size {value.shape}, but''' F''' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.''') lowerCAmelCase_ : Optional[int] = value logger.info(F'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''') else: unused_weights.append(snake_case__) @torch.no_grad() def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__=None , snake_case__=None , snake_case__=None , ): if config_path is not None: lowerCAmelCase_ : Any = SpeechTaConfig.from_pretrained(snake_case__) else: lowerCAmelCase_ : Optional[int] = SpeechTaConfig() if task == "s2t": lowerCAmelCase_ : Optional[int] = config.max_text_positions lowerCAmelCase_ : Optional[Any] = SpeechTaForSpeechToText(snake_case__) elif task == "t2s": lowerCAmelCase_ : Optional[int] = 18_76 lowerCAmelCase_ : Tuple = 6_00 lowerCAmelCase_ : Dict = config.max_speech_positions lowerCAmelCase_ : int = SpeechTaForTextToSpeech(snake_case__) elif task == "s2s": lowerCAmelCase_ : Union[str, Any] = 18_76 lowerCAmelCase_ : Optional[Any] = config.max_speech_positions lowerCAmelCase_ : str = SpeechTaForSpeechToSpeech(snake_case__) else: raise ValueError(F'''Unknown task name: {task}''') if vocab_path: lowerCAmelCase_ : Dict = SpeechTaTokenizer(snake_case__ , model_max_length=config.max_text_positions) # Mask token behaves like a normal word, i.e. include the space before it lowerCAmelCase_ : List[str] = AddedToken("<mask>" , lstrip=snake_case__ , rstrip=snake_case__) lowerCAmelCase_ : Dict = mask_token tokenizer.add_special_tokens({"mask_token": mask_token}) tokenizer.add_tokens(["<ctc_blank>"]) lowerCAmelCase_ : List[str] = SpeechTaFeatureExtractor() lowerCAmelCase_ : Optional[Any] = SpeechTaProcessor(tokenizer=snake_case__ , feature_extractor=snake_case__) processor.save_pretrained(snake_case__) lowerCAmelCase_ : List[Any] = torch.load(snake_case__) recursively_load_weights(fairseq_checkpoint["model"] , snake_case__ , snake_case__) model.save_pretrained(snake_case__) if repo_id: print("Pushing to the hub...") processor.push_to_hub(snake_case__) model.push_to_hub(snake_case__) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument( '''--task''', default='''s2t''', type=str, help='''Type of the SpeechT5 model you\'d like to convert. Should be one of \'s2t\', \'t2s\', \'s2s\'.''', ) parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--vocab_path''', default=None, type=str, help='''Path to SentencePiece model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) _lowercase = parser.parse_args() convert_speechta_checkpoint( args.task, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.vocab_path, args.push_to_hub, )
659
import argparse import json from tqdm import tqdm def UpperCamelCase ( ): lowerCAmelCase_ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--src_path" , type=snake_case__ , default="biencoder-nq-dev.json" , help="Path to raw DPR training data" , ) parser.add_argument( "--evaluation_set" , type=snake_case__ , help="where to store parsed evaluation_set file" , ) parser.add_argument( "--gold_data_path" , type=snake_case__ , help="where to store parsed gold_data_path file" , ) lowerCAmelCase_ : Dict = parser.parse_args() with open(args.src_path , "r") as src_file, open(args.evaluation_set , "w") as eval_file, open( args.gold_data_path , "w") as gold_file: lowerCAmelCase_ : Optional[int] = json.load(snake_case__) for dpr_record in tqdm(snake_case__): lowerCAmelCase_ : str = dpr_record["question"] lowerCAmelCase_ : Dict = [context["title"] for context in dpr_record["positive_ctxs"]] eval_file.write(question + "\n") gold_file.write("\t".join(snake_case__) + "\n") if __name__ == "__main__": main()
659
1
from __future__ import annotations def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : list[list[int]] = [] create_all_state(1 , snake_case__ , snake_case__ , [] , snake_case__) return result def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ): if level == 0: total_list.append(current_list[:]) return for i in range(snake_case__ , total_number - level + 2): current_list.append(snake_case__) create_all_state(i + 1 , snake_case__ , level - 1 , snake_case__ , snake_case__) current_list.pop() def UpperCamelCase ( snake_case__): for i in total_list: print(*snake_case__) if __name__ == "__main__": _lowercase = 4 _lowercase = 2 _lowercase = generate_all_combinations(n, k) print_all_state(total_list)
659
from collections.abc import Sequence def UpperCamelCase ( snake_case__ = None): if nums is None or not nums: raise ValueError("Input sequence should not be empty") lowerCAmelCase_ : Dict = nums[0] for i in range(1 , len(snake_case__)): lowerCAmelCase_ : Optional[int] = nums[i] lowerCAmelCase_ : Optional[int] = max(snake_case__ , ans + num , snake_case__) return ans if __name__ == "__main__": import doctest doctest.testmod() # Try on a sample input from the user _lowercase = int(input('''Enter number of elements : ''').strip()) _lowercase = list(map(int, input('''\nEnter the numbers : ''').strip().split()))[:n] print(max_subsequence_sum(array))
659
1
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaPriorEmbaEmbPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = KandinskyVaaControlnetImgaImgPipeline UpperCamelCase_ = ['image_embeds', 'negative_image_embeds', 'image', 'hint'] UpperCamelCase_ = ['image_embeds', 'negative_image_embeds', 'image', 'hint'] UpperCamelCase_ = [ 'generator', 'height', 'width', 'strength', 'guidance_scale', 'num_inference_steps', 'return_dict', 'guidance_scale', 'num_images_per_prompt', 'output_type', 'return_dict', ] UpperCamelCase_ = False @property def UpperCAmelCase_ ( self : List[Any] ) -> List[Any]: '''simple docstring''' return 32 @property def UpperCAmelCase_ ( self : Any ) -> List[str]: '''simple docstring''' return 32 @property def UpperCAmelCase_ ( self : Tuple ) -> Tuple: '''simple docstring''' return self.time_input_dim @property def UpperCAmelCase_ ( self : Tuple ) -> Any: '''simple docstring''' return self.time_input_dim * 4 @property def UpperCAmelCase_ ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' return 1_00 @property def UpperCAmelCase_ ( self : Optional[int] ) -> int: '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase_ : Tuple = { "in_channels": 8, # Out channels is double in channels because predicts mean and variance "out_channels": 8, "addition_embed_type": "image_hint", "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "encoder_hid_dim": self.text_embedder_hidden_size, "encoder_hid_dim_type": "image_proj", "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": None, } lowerCAmelCase_ : Any = UNetaDConditionModel(**lowerCAmelCase__ ) return model @property def UpperCAmelCase_ ( self : Any ) -> str: '''simple docstring''' return { "block_out_channels": [32, 32, 64, 64], "down_block_types": [ "DownEncoderBlock2D", "DownEncoderBlock2D", "DownEncoderBlock2D", "AttnDownEncoderBlock2D", ], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": ["AttnUpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D", "UpDecoderBlock2D"], "vq_embed_dim": 4, } @property def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[Any]: '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase_ : Any = VQModel(**self.dummy_movq_kwargs ) return model def UpperCAmelCase_ ( self : Dict ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : List[Any] = self.dummy_unet lowerCAmelCase_ : Tuple = self.dummy_movq lowerCAmelCase_ : Tuple = { "num_train_timesteps": 10_00, "beta_schedule": "linear", "beta_start": 0.00_085, "beta_end": 0.012, "clip_sample": False, "set_alpha_to_one": False, "steps_offset": 0, "prediction_type": "epsilon", "thresholding": False, } lowerCAmelCase_ : Dict = DDIMScheduler(**lowerCAmelCase__ ) lowerCAmelCase_ : Any = { "unet": unet, "scheduler": scheduler, "movq": movq, } return components def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Optional[int]=0 ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Any = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = floats_tensor((1, self.text_embedder_hidden_size) ,rng=random.Random(seed + 1 ) ).to( lowerCAmelCase__ ) # create init_image lowerCAmelCase_ : Optional[Any] = floats_tensor((1, 3, 64, 64) ,rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) lowerCAmelCase_ : int = image.cpu().permute(0 ,2 ,3 ,1 )[0] lowerCAmelCase_ : Dict = Image.fromarray(np.uinta(lowerCAmelCase__ ) ).convert("RGB" ).resize((2_56, 2_56) ) # create hint lowerCAmelCase_ : Union[str, Any] = floats_tensor((1, 3, 64, 64) ,rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) if str(lowerCAmelCase__ ).startswith("mps" ): lowerCAmelCase_ : Optional[Any] = torch.manual_seed(lowerCAmelCase__ ) else: lowerCAmelCase_ : List[str] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : Any = { "image": init_image, "image_embeds": image_embeds, "negative_image_embeds": negative_image_embeds, "hint": hint, "generator": generator, "height": 64, "width": 64, "num_inference_steps": 10, "guidance_scale": 7.0, "strength": 0.2, "output_type": "np", } return inputs def UpperCAmelCase_ ( self : Optional[Any] ) -> str: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = "cpu" lowerCAmelCase_ : Any = self.get_dummy_components() lowerCAmelCase_ : List[str] = self.pipeline_class(**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = pipe(**self.get_dummy_inputs(lowerCAmelCase__ ) ) lowerCAmelCase_ : Optional[Any] = output.images lowerCAmelCase_ : Optional[Any] = pipe( **self.get_dummy_inputs(lowerCAmelCase__ ) ,return_dict=lowerCAmelCase__ ,)[0] lowerCAmelCase_ : Optional[int] = image[0, -3:, -3:, -1] lowerCAmelCase_ : Optional[Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) lowerCAmelCase_ : str = np.array( [0.54_985_034, 0.55_509_365, 0.52_561_504, 0.5_570_494, 0.5_593_818, 0.5_263_979, 0.50_285_643, 0.5_069_846, 0.51_196_736] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 ), f''' expected_slice {expected_slice}, but got {image_slice.flatten()}''' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 ), f''' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}''' @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : int ) -> Optional[int]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : str ) -> Any: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/kandinskyv22_controlnet_img2img_robotcat_fp16.npy" ) lowerCAmelCase_ : int = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinsky/cat.png" ) lowerCAmelCase_ : str = init_image.resize((5_12, 5_12) ) lowerCAmelCase_ : Any = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/kandinskyv22/hint_image_cat.png" ) lowerCAmelCase_ : List[str] = torch.from_numpy(np.array(lowerCAmelCase__ ) ).float() / 255.0 lowerCAmelCase_ : Optional[Any] = hint.permute(2 ,0 ,1 ).unsqueeze(0 ) lowerCAmelCase_ : Any = "A robot, 4k photo" lowerCAmelCase_ : Union[str, Any] = KandinskyVaaPriorEmbaEmbPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-prior" ,torch_dtype=torch.floataa ) pipe_prior.to(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = KandinskyVaaControlnetImgaImgPipeline.from_pretrained( "kandinsky-community/kandinsky-2-2-controlnet-depth" ,torch_dtype=torch.floataa ) lowerCAmelCase_ : int = pipeline.to(lowerCAmelCase__ ) pipeline.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = torch.Generator(device="cpu" ).manual_seed(0 ) lowerCAmelCase_ , lowerCAmelCase_ : str = pipe_prior( lowerCAmelCase__ ,image=lowerCAmelCase__ ,strength=0.85 ,generator=lowerCAmelCase__ ,negative_prompt="" ,).to_tuple() lowerCAmelCase_ : str = pipeline( image=lowerCAmelCase__ ,image_embeds=lowerCAmelCase__ ,negative_image_embeds=lowerCAmelCase__ ,hint=lowerCAmelCase__ ,generator=lowerCAmelCase__ ,num_inference_steps=1_00 ,height=5_12 ,width=5_12 ,strength=0.5 ,output_type="np" ,) lowerCAmelCase_ : Any = output.images[0] assert image.shape == (5_12, 5_12, 3) assert_mean_pixel_difference(lowerCAmelCase__ ,lowerCAmelCase__ )
659
from typing import TYPE_CHECKING from ....utils import _LazyModule _lowercase = {'''tokenization_tapex''': ['''TapexTokenizer''']} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
659
1
import importlib import json import os import sys import tempfile import unittest from pathlib import Path import transformers import transformers.models.auto from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.bert.configuration_bert import BertConfig from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 _lowercase = get_tests_dir('''fixtures/dummy-config.json''') class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : Dict ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = 0 def UpperCAmelCase_ ( self : str ) -> Tuple: '''simple docstring''' self.assertIsNotNone(transformers.models.auto.__spec__ ) self.assertIsNotNone(importlib.util.find_spec("transformers.models.auto" ) ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : str = AutoConfig.from_pretrained("bert-base-uncased" ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Any ) -> str: '''simple docstring''' lowerCAmelCase_ : List[str] = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Tuple = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Tuple ) -> Any: '''simple docstring''' lowerCAmelCase_ : List[Any] = AutoConfig.for_model("roberta" ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Dict ) -> List[str]: '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: # This model name contains bert and roberta, but roberta ends up being picked. lowerCAmelCase_ : Dict = os.path.join(lowerCAmelCase__ ,"fake-roberta" ) os.makedirs(lowerCAmelCase__ ,exist_ok=lowerCAmelCase__ ) with open(os.path.join(lowerCAmelCase__ ,"config.json" ) ,"w" ) as f: f.write(json.dumps({} ) ) lowerCAmelCase_ : Any = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertEqual(type(lowerCAmelCase__ ) ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Tuple ) -> Optional[int]: '''simple docstring''' try: AutoConfig.register("custom" ,lowerCAmelCase__ ) # Wrong model type will raise an error with self.assertRaises(lowerCAmelCase__ ): AutoConfig.register("model" ,lowerCAmelCase__ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(lowerCAmelCase__ ): AutoConfig.register("bert" ,lowerCAmelCase__ ) # Now that the config is registered, it can be used as any other config with the auto-API lowerCAmelCase_ : List[Any] = CustomConfig() with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCAmelCase__ ) lowerCAmelCase_ : str = AutoConfig.from_pretrained(lowerCAmelCase__ ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] def UpperCAmelCase_ ( self : List[Any] ) -> List[Any]: '''simple docstring''' with self.assertRaisesRegex( lowerCAmelCase__ ,"bert-base is not a local folder and is not a valid model identifier" ): lowerCAmelCase_ : Optional[int] = AutoConfig.from_pretrained("bert-base" ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' with self.assertRaisesRegex( lowerCAmelCase__ ,R"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): lowerCAmelCase_ : Tuple = AutoConfig.from_pretrained(lowerCAmelCase__ ,revision="aaaaaa" ) def UpperCAmelCase_ ( self : List[str] ) -> Optional[Any]: '''simple docstring''' with self.assertRaisesRegex( lowerCAmelCase__ ,"hf-internal-testing/no-config-test-repo does not appear to have a file named config.json." ,): lowerCAmelCase_ : Tuple = AutoConfig.from_pretrained("hf-internal-testing/no-config-test-repo" ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Tuple: '''simple docstring''' with self.assertRaises(lowerCAmelCase__ ): lowerCAmelCase_ : Tuple = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" ) # If remote code is disabled, we can't load this config. with self.assertRaises(lowerCAmelCase__ ): lowerCAmelCase_ : List[str] = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" ,trust_remote_code=lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" ,trust_remote_code=lowerCAmelCase__ ) self.assertEqual(config.__class__.__name__ ,"NewModelConfig" ) # Test config can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = AutoConfig.from_pretrained(lowerCAmelCase__ ,trust_remote_code=lowerCAmelCase__ ) self.assertEqual(reloaded_config.__class__.__name__ ,"NewModelConfig" ) def UpperCAmelCase_ ( self : Dict ) -> Any: '''simple docstring''' class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = 'new-model' try: AutoConfig.register("new-model" ,lowerCAmelCase__ ) # If remote code is not set, the default is to use local lowerCAmelCase_ : str = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" ) self.assertEqual(config.__class__.__name__ ,"NewModelConfigLocal" ) # If remote code is disabled, we load the local one. lowerCAmelCase_ : List[Any] = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" ,trust_remote_code=lowerCAmelCase__ ) self.assertEqual(config.__class__.__name__ ,"NewModelConfigLocal" ) # If remote is enabled, we load from the Hub lowerCAmelCase_ : Optional[Any] = AutoConfig.from_pretrained("hf-internal-testing/test_dynamic_model" ,trust_remote_code=lowerCAmelCase__ ) self.assertEqual(config.__class__.__name__ ,"NewModelConfig" ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"]
659
import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _lowercase = '''src/diffusers''' _lowercase = '''.''' # This is to make sure the diffusers module imported is the one in the repo. _lowercase = importlib.util.spec_from_file_location( '''diffusers''', os.path.join(DIFFUSERS_PATH, '''__init__.py'''), submodule_search_locations=[DIFFUSERS_PATH], ) _lowercase = spec.loader.load_module() def UpperCamelCase ( snake_case__ , snake_case__): return line.startswith(snake_case__) or len(snake_case__) <= 1 or re.search(R"^\s*\)(\s*->.*:|:)\s*$" , snake_case__) is not None def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Tuple = object_name.split(".") lowerCAmelCase_ : Union[str, Any] = 0 # First let's find the module where our object lives. lowerCAmelCase_ : Union[str, Any] = parts[i] while i < len(snake_case__) and not os.path.isfile(os.path.join(snake_case__ , F'''{module}.py''')): i += 1 if i < len(snake_case__): lowerCAmelCase_ : Dict = os.path.join(snake_case__ , parts[i]) if i >= len(snake_case__): raise ValueError(F'''`object_name` should begin with the name of a module of diffusers but got {object_name}.''') with open(os.path.join(snake_case__ , F'''{module}.py''') , "r" , encoding="utf-8" , newline="\n") as f: lowerCAmelCase_ : Optional[Any] = f.readlines() # Now let's find the class / func in the code! lowerCAmelCase_ : Union[str, Any] = "" lowerCAmelCase_ : int = 0 for name in parts[i + 1 :]: while ( line_index < len(snake_case__) and re.search(RF'''^{indent}(class|def)\s+{name}(\(|\:)''' , lines[line_index]) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(snake_case__): raise ValueError(F''' {object_name} does not match any function or class in {module}.''') # We found the beginning of the class / func, now let's find the end (when the indent diminishes). lowerCAmelCase_ : Union[str, Any] = line_index while line_index < len(snake_case__) and _should_continue(lines[line_index] , snake_case__): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1]) <= 1: line_index -= 1 lowerCAmelCase_ : List[str] = lines[start_index:line_index] return "".join(snake_case__) _lowercase = re.compile(r'''^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)''') _lowercase = re.compile(r'''^\s*(\S+)->(\S+)(\s+.*|$)''') _lowercase = re.compile(r'''<FILL\s+[^>]*>''') def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Any = code.split("\n") lowerCAmelCase_ : Any = 0 while idx < len(snake_case__) and len(lines[idx]) == 0: idx += 1 if idx < len(snake_case__): return re.search(R"^(\s*)\S" , lines[idx]).groups()[0] return "" def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Dict = len(get_indent(snake_case__)) > 0 if has_indent: lowerCAmelCase_ : Dict = F'''class Bla:\n{code}''' lowerCAmelCase_ : Optional[int] = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=1_19 , preview=snake_case__) lowerCAmelCase_ : Optional[Any] = black.format_str(snake_case__ , mode=snake_case__) lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = style_docstrings_in_code(snake_case__) return result[len("class Bla:\n") :] if has_indent else result def UpperCamelCase ( snake_case__ , snake_case__=False): with open(snake_case__ , "r" , encoding="utf-8" , newline="\n") as f: lowerCAmelCase_ : Tuple = f.readlines() lowerCAmelCase_ : Tuple = [] lowerCAmelCase_ : Union[str, Any] = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(snake_case__): lowerCAmelCase_ : Optional[int] = _re_copy_warning.search(lines[line_index]) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = search.groups() lowerCAmelCase_ : int = find_code_in_diffusers(snake_case__) lowerCAmelCase_ : Dict = get_indent(snake_case__) lowerCAmelCase_ : Union[str, Any] = line_index + 1 if indent == theoretical_indent else line_index + 2 lowerCAmelCase_ : str = theoretical_indent lowerCAmelCase_ : Union[str, Any] = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. lowerCAmelCase_ : Optional[int] = True while line_index < len(snake_case__) and should_continue: line_index += 1 if line_index >= len(snake_case__): break lowerCAmelCase_ : Dict = lines[line_index] lowerCAmelCase_ : List[str] = _should_continue(snake_case__ , snake_case__) and re.search(F'''^{indent}# End copy''' , snake_case__) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1]) <= 1: line_index -= 1 lowerCAmelCase_ : Dict = lines[start_index:line_index] lowerCAmelCase_ : Optional[int] = "".join(snake_case__) # Remove any nested `Copied from` comments to avoid circular copies lowerCAmelCase_ : List[Any] = [line for line in theoretical_code.split("\n") if _re_copy_warning.search(snake_case__) is None] lowerCAmelCase_ : Optional[Any] = "\n".join(snake_case__) # Before comparing, use the `replace_pattern` on the original code. if len(snake_case__) > 0: lowerCAmelCase_ : List[str] = replace_pattern.replace("with" , "").split(",") lowerCAmelCase_ : Tuple = [_re_replace_pattern.search(snake_case__) for p in patterns] for pattern in patterns: if pattern is None: continue lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[str] = pattern.groups() lowerCAmelCase_ : int = re.sub(snake_case__ , snake_case__ , snake_case__) if option.strip() == "all-casing": lowerCAmelCase_ : List[str] = re.sub(obja.lower() , obja.lower() , snake_case__) lowerCAmelCase_ : int = re.sub(obja.upper() , obja.upper() , snake_case__) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line lowerCAmelCase_ : List[Any] = blackify(lines[start_index - 1] + theoretical_code) lowerCAmelCase_ : Union[str, Any] = theoretical_code[len(lines[start_index - 1]) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index]) if overwrite: lowerCAmelCase_ : List[Any] = lines[:start_index] + [theoretical_code] + lines[line_index:] lowerCAmelCase_ : Union[str, Any] = start_index + 1 if overwrite and len(snake_case__) > 0: # Warn the user a file has been modified. print(F'''Detected changes, rewriting {filename}.''') with open(snake_case__ , "w" , encoding="utf-8" , newline="\n") as f: f.writelines(snake_case__) return diffs def UpperCamelCase ( snake_case__ = False): lowerCAmelCase_ : Tuple = glob.glob(os.path.join(snake_case__ , "**/*.py") , recursive=snake_case__) lowerCAmelCase_ : int = [] for filename in all_files: lowerCAmelCase_ : Union[str, Any] = is_copy_consistent(snake_case__ , snake_case__) diffs += [F'''- {filename}: copy does not match {d[0]} at line {d[1]}''' for d in new_diffs] if not overwrite and len(snake_case__) > 0: lowerCAmelCase_ : Optional[Any] = "\n".join(snake_case__) raise Exception( "Found the following copy inconsistencies:\n" + diff + "\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.") if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') _lowercase = parser.parse_args() check_copies(args.fix_and_overwrite)
659
1
from typing import TYPE_CHECKING from ....utils import _LazyModule _lowercase = {'''tokenization_tapex''': ['''TapexTokenizer''']} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
659
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''microsoft/swinv2-tiny-patch4-window8-256''': ( '''https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json''' ), } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = 'swinv2' UpperCamelCase_ = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : List[Any] ,lowerCAmelCase__ : Optional[int]=2_24 ,lowerCAmelCase__ : Dict=4 ,lowerCAmelCase__ : Dict=3 ,lowerCAmelCase__ : List[Any]=96 ,lowerCAmelCase__ : Optional[Any]=[2, 2, 6, 2] ,lowerCAmelCase__ : Optional[Any]=[3, 6, 12, 24] ,lowerCAmelCase__ : Optional[int]=7 ,lowerCAmelCase__ : Dict=4.0 ,lowerCAmelCase__ : Dict=True ,lowerCAmelCase__ : str=0.0 ,lowerCAmelCase__ : Tuple=0.0 ,lowerCAmelCase__ : str=0.1 ,lowerCAmelCase__ : List[str]="gelu" ,lowerCAmelCase__ : Union[str, Any]=False ,lowerCAmelCase__ : Dict=0.02 ,lowerCAmelCase__ : int=1e-5 ,lowerCAmelCase__ : List[str]=32 ,**lowerCAmelCase__ : Tuple ,) -> List[str]: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = image_size lowerCAmelCase_ : List[Any] = patch_size lowerCAmelCase_ : Dict = num_channels lowerCAmelCase_ : Optional[int] = embed_dim lowerCAmelCase_ : Optional[Any] = depths lowerCAmelCase_ : Any = len(lowerCAmelCase__ ) lowerCAmelCase_ : str = num_heads lowerCAmelCase_ : List[str] = window_size lowerCAmelCase_ : List[str] = mlp_ratio lowerCAmelCase_ : Dict = qkv_bias lowerCAmelCase_ : str = hidden_dropout_prob lowerCAmelCase_ : str = attention_probs_dropout_prob lowerCAmelCase_ : Union[str, Any] = drop_path_rate lowerCAmelCase_ : List[Any] = hidden_act lowerCAmelCase_ : Any = use_absolute_embeddings lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : int = initializer_range lowerCAmelCase_ : Union[str, Any] = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase_ : Tuple = int(embed_dim * 2 ** (len(lowerCAmelCase__ ) - 1) ) lowerCAmelCase_ : str = (0, 0, 0, 0)
659
1
import operator def UpperCamelCase ( snake_case__ , snake_case__ = False , snake_case__ = None): lowerCAmelCase_ : List[Any] = operator.lt if reverse else operator.gt lowerCAmelCase_ : Tuple = solution or [] if not arr: return solution lowerCAmelCase_ : Union[str, Any] = [arr.pop(0)] for i, item in enumerate(snake_case__): if _operator(snake_case__ , sublist[-1]): sublist.append(snake_case__) arr.pop(snake_case__) # merging sublist into solution list if not solution: solution.extend(snake_case__) else: while sublist: lowerCAmelCase_ : Union[str, Any] = sublist.pop(0) for i, xx in enumerate(snake_case__): if not _operator(snake_case__ , snake_case__): solution.insert(snake_case__ , snake_case__) break else: solution.append(snake_case__) strand_sort(snake_case__ , snake_case__ , snake_case__) return solution if __name__ == "__main__": assert strand_sort([4, 3, 5, 1, 2]) == [1, 2, 3, 4, 5] assert strand_sort([4, 3, 5, 1, 2], reverse=True) == [5, 4, 3, 2, 1]
659
from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging _lowercase = logging.get_logger(__name__) class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = ['input_features', 'attention_mask'] def __init__( self : Optional[Any] ,lowerCAmelCase__ : Any=80 ,lowerCAmelCase__ : Optional[Any]=1_60_00 ,lowerCAmelCase__ : List[str]=0.0 ,lowerCAmelCase__ : Tuple=10 ,lowerCAmelCase__ : Optional[Any]=25 ,lowerCAmelCase__ : Any="hamming_window" ,lowerCAmelCase__ : List[str]=32_768.0 ,lowerCAmelCase__ : Union[str, Any]=0.97 ,lowerCAmelCase__ : Any=1.0 ,lowerCAmelCase__ : str=True ,lowerCAmelCase__ : int=True ,lowerCAmelCase__ : Tuple=False ,**lowerCAmelCase__ : Optional[int] ,) -> Optional[Any]: '''simple docstring''' super().__init__(feature_size=lowerCAmelCase__ ,sampling_rate=lowerCAmelCase__ ,padding_value=lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = feature_size lowerCAmelCase_ : List[Any] = sampling_rate lowerCAmelCase_ : Union[str, Any] = padding_value lowerCAmelCase_ : str = hop_length lowerCAmelCase_ : str = win_length lowerCAmelCase_ : str = frame_signal_scale lowerCAmelCase_ : Any = preemphasis_coeff lowerCAmelCase_ : Optional[Any] = mel_floor lowerCAmelCase_ : List[str] = normalize_means lowerCAmelCase_ : Optional[Any] = normalize_vars lowerCAmelCase_ : Dict = win_function lowerCAmelCase_ : List[Any] = return_attention_mask lowerCAmelCase_ : Tuple = win_length * sampling_rate // 10_00 lowerCAmelCase_ : str = hop_length * sampling_rate // 10_00 lowerCAmelCase_ : Dict = optimal_fft_length(self.sample_size ) lowerCAmelCase_ : Optional[int] = (self.n_fft // 2) + 1 def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : np.array ) -> np.ndarray: '''simple docstring''' if self.win_function == "hamming_window": lowerCAmelCase_ : int = window_function(window_length=self.sample_size ,name=self.win_function ,periodic=lowerCAmelCase__ ) else: lowerCAmelCase_ : Tuple = window_function(window_length=self.sample_size ,name=self.win_function ) lowerCAmelCase_ : List[str] = mel_filter_bank( num_frequency_bins=self.n_freqs ,num_mel_filters=self.feature_size ,min_frequency=0.0 ,max_frequency=self.sampling_rate / 2.0 ,sampling_rate=self.sampling_rate ,) lowerCAmelCase_ : Any = spectrogram( one_waveform * self.frame_signal_scale ,window=lowerCAmelCase__ ,frame_length=self.sample_size ,hop_length=self.sample_stride ,fft_length=self.n_fft ,center=lowerCAmelCase__ ,preemphasis=self.preemphasis_coeff ,mel_filters=lowerCAmelCase__ ,mel_floor=self.mel_floor ,log_mel="log" ,) return msfc_features.T def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : Tuple ) -> Optional[Any]: '''simple docstring''' if self.normalize_means: lowerCAmelCase_ : Optional[int] = x[:input_length].mean(axis=0 ) lowerCAmelCase_ : List[str] = np.subtract(lowerCAmelCase__ ,lowerCAmelCase__ ) if self.normalize_vars: lowerCAmelCase_ : Optional[Any] = x[:input_length].std(axis=0 ) lowerCAmelCase_ : Tuple = np.divide(lowerCAmelCase__ ,lowerCAmelCase__ ) if input_length < x.shape[0]: lowerCAmelCase_ : int = padding_value # make sure array is in float32 lowerCAmelCase_ : Any = x.astype(np.floataa ) return x def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[np.ndarray] ,lowerCAmelCase__ : Optional[np.ndarray] = None ) -> List[np.ndarray]: '''simple docstring''' lowerCAmelCase_ : List[Any] = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(lowerCAmelCase__ ,lowerCAmelCase__ ,self.padding_value ) for x, n in zip(lowerCAmelCase__ ,lowerCAmelCase__ )] def __call__( self : int ,lowerCAmelCase__ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,lowerCAmelCase__ : Union[bool, str, PaddingStrategy] = False ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : bool = False ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : Optional[bool] = None ,lowerCAmelCase__ : Optional[Union[str, TensorType]] = None ,lowerCAmelCase__ : Optional[int] = None ,**lowerCAmelCase__ : Union[str, Any] ,) -> BatchFeature: '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) lowerCAmelCase_ : List[Any] = isinstance(lowerCAmelCase__ ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase_ : str = is_batched_numpy or ( isinstance(lowerCAmelCase__ ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase_ : Tuple = [np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(lowerCAmelCase__ ,np.ndarray ): lowerCAmelCase_ : int = np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) elif isinstance(lowerCAmelCase__ ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase_ : Union[str, Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase_ : Optional[int] = [raw_speech] # extract fbank features lowerCAmelCase_ : Dict = [self._extract_mfsc_features(lowerCAmelCase__ ) for one_waveform in raw_speech] # convert into correct format for padding lowerCAmelCase_ : int = BatchFeature({"input_features": features} ) lowerCAmelCase_ : Union[str, Any] = self.pad( lowerCAmelCase__ ,padding=lowerCAmelCase__ ,max_length=lowerCAmelCase__ ,truncation=lowerCAmelCase__ ,pad_to_multiple_of=lowerCAmelCase__ ,return_attention_mask=lowerCAmelCase__ ,**lowerCAmelCase__ ,) # make sure list is in array format lowerCAmelCase_ : Optional[Any] = padded_inputs.get("input_features" ) if isinstance(input_features[0] ,lowerCAmelCase__ ): lowerCAmelCase_ : Optional[int] = [np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) for feature in input_features] lowerCAmelCase_ : List[Any] = padded_inputs.get("attention_mask" ) if attention_mask is not None: lowerCAmelCase_ : Dict = [np.asarray(lowerCAmelCase__ ,dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: lowerCAmelCase_ : Dict = ( np.array(lowerCAmelCase__ ,dtype=np.intaa ) if self._get_padding_strategies(lowerCAmelCase__ ,max_length=lowerCAmelCase__ ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) lowerCAmelCase_ : List[str] = self.normalize( padded_inputs["input_features"] ,attention_mask=lowerCAmelCase__ ) if return_tensors is not None: lowerCAmelCase_ : Dict = padded_inputs.convert_to_tensors(lowerCAmelCase__ ) return padded_inputs
659
1
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Tuple = 3_84 if "tiny" in model_name: lowerCAmelCase_ : Union[str, Any] = [3, 3, 9, 3] lowerCAmelCase_ : str = [96, 1_92, 3_84, 7_68] if "small" in model_name: lowerCAmelCase_ : Any = [3, 3, 27, 3] lowerCAmelCase_ : Optional[int] = [96, 1_92, 3_84, 7_68] if "base" in model_name: lowerCAmelCase_ : int = [3, 3, 27, 3] lowerCAmelCase_ : Tuple = [1_28, 2_56, 5_12, 10_24] lowerCAmelCase_ : Optional[Any] = 5_12 if "large" in model_name: lowerCAmelCase_ : List[str] = [3, 3, 27, 3] lowerCAmelCase_ : Optional[int] = [1_92, 3_84, 7_68, 15_36] lowerCAmelCase_ : Dict = 7_68 if "xlarge" in model_name: lowerCAmelCase_ : Optional[int] = [3, 3, 27, 3] lowerCAmelCase_ : Union[str, Any] = [2_56, 5_12, 10_24, 20_48] lowerCAmelCase_ : Dict = 10_24 # set label information lowerCAmelCase_ : List[Any] = 1_50 lowerCAmelCase_ : List[Any] = "huggingface/label-files" lowerCAmelCase_ : Any = "ade20k-id2label.json" lowerCAmelCase_ : Optional[int] = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type="dataset") , "r")) lowerCAmelCase_ : str = {int(snake_case__): v for k, v in idalabel.items()} lowerCAmelCase_ : str = {v: k for k, v in idalabel.items()} lowerCAmelCase_ : Optional[Any] = ConvNextConfig( depths=snake_case__ , hidden_sizes=snake_case__ , out_features=["stage1", "stage2", "stage3", "stage4"]) lowerCAmelCase_ : List[Any] = UperNetConfig( backbone_config=snake_case__ , auxiliary_in_channels=snake_case__ , num_labels=snake_case__ , idalabel=snake_case__ , labelaid=snake_case__ , ) return config def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Tuple = [] # fmt: off # stem rename_keys.append(("backbone.downsample_layers.0.0.weight", "backbone.embeddings.patch_embeddings.weight")) rename_keys.append(("backbone.downsample_layers.0.0.bias", "backbone.embeddings.patch_embeddings.bias")) rename_keys.append(("backbone.downsample_layers.0.1.weight", "backbone.embeddings.layernorm.weight")) rename_keys.append(("backbone.downsample_layers.0.1.bias", "backbone.embeddings.layernorm.bias")) # stages for i in range(len(config.backbone_config.depths)): for j in range(config.backbone_config.depths[i]): rename_keys.append((F'''backbone.stages.{i}.{j}.gamma''', F'''backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter''')) rename_keys.append((F'''backbone.stages.{i}.{j}.depthwise_conv.weight''', F'''backbone.encoder.stages.{i}.layers.{j}.dwconv.weight''')) rename_keys.append((F'''backbone.stages.{i}.{j}.depthwise_conv.bias''', F'''backbone.encoder.stages.{i}.layers.{j}.dwconv.bias''')) rename_keys.append((F'''backbone.stages.{i}.{j}.norm.weight''', F'''backbone.encoder.stages.{i}.layers.{j}.layernorm.weight''')) rename_keys.append((F'''backbone.stages.{i}.{j}.norm.bias''', F'''backbone.encoder.stages.{i}.layers.{j}.layernorm.bias''')) rename_keys.append((F'''backbone.stages.{i}.{j}.pointwise_conv1.weight''', F'''backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight''')) rename_keys.append((F'''backbone.stages.{i}.{j}.pointwise_conv1.bias''', F'''backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias''')) rename_keys.append((F'''backbone.stages.{i}.{j}.pointwise_conv2.weight''', F'''backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight''')) rename_keys.append((F'''backbone.stages.{i}.{j}.pointwise_conv2.bias''', F'''backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias''')) if i > 0: rename_keys.append((F'''backbone.downsample_layers.{i}.0.weight''', F'''backbone.encoder.stages.{i}.downsampling_layer.0.weight''')) rename_keys.append((F'''backbone.downsample_layers.{i}.0.bias''', F'''backbone.encoder.stages.{i}.downsampling_layer.0.bias''')) rename_keys.append((F'''backbone.downsample_layers.{i}.1.weight''', F'''backbone.encoder.stages.{i}.downsampling_layer.1.weight''')) rename_keys.append((F'''backbone.downsample_layers.{i}.1.bias''', F'''backbone.encoder.stages.{i}.downsampling_layer.1.bias''')) rename_keys.append((F'''backbone.norm{i}.weight''', F'''backbone.hidden_states_norms.stage{i+1}.weight''')) rename_keys.append((F'''backbone.norm{i}.bias''', F'''backbone.hidden_states_norms.stage{i+1}.bias''')) # decode head rename_keys.extend( [ ("decode_head.conv_seg.weight", "decode_head.classifier.weight"), ("decode_head.conv_seg.bias", "decode_head.classifier.bias"), ("auxiliary_head.conv_seg.weight", "auxiliary_head.classifier.weight"), ("auxiliary_head.conv_seg.bias", "auxiliary_head.classifier.bias"), ]) # fmt: on return rename_keys def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : str = dct.pop(snake_case__) lowerCAmelCase_ : Optional[Any] = val def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : List[str] = { "upernet-convnext-tiny": "https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth", "upernet-convnext-small": "https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth", "upernet-convnext-base": "https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth", "upernet-convnext-large": "https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth", "upernet-convnext-xlarge": "https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth", } lowerCAmelCase_ : str = model_name_to_url[model_name] lowerCAmelCase_ : int = torch.hub.load_state_dict_from_url(snake_case__ , map_location="cpu")["state_dict"] lowerCAmelCase_ : Dict = get_upernet_config(snake_case__) lowerCAmelCase_ : Dict = UperNetForSemanticSegmentation(snake_case__) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): lowerCAmelCase_ : Any = state_dict.pop(snake_case__) if "bn" in key: lowerCAmelCase_ : Optional[Any] = key.replace("bn" , "batch_norm") lowerCAmelCase_ : List[str] = val # rename keys lowerCAmelCase_ : Union[str, Any] = create_rename_keys(snake_case__) for src, dest in rename_keys: rename_key(snake_case__ , snake_case__ , snake_case__) model.load_state_dict(snake_case__) # verify on image lowerCAmelCase_ : List[str] = "https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg" lowerCAmelCase_ : Dict = Image.open(requests.get(snake_case__ , stream=snake_case__).raw).convert("RGB") lowerCAmelCase_ : Optional[int] = SegformerImageProcessor() lowerCAmelCase_ : List[Any] = processor(snake_case__ , return_tensors="pt").pixel_values with torch.no_grad(): lowerCAmelCase_ : Union[str, Any] = model(snake_case__) if model_name == "upernet-convnext-tiny": lowerCAmelCase_ : Union[str, Any] = torch.tensor( [[-8.8_110, -8.8_110, -8.6_521], [-8.8_110, -8.8_110, -8.6_521], [-8.7_746, -8.7_746, -8.6_130]]) elif model_name == "upernet-convnext-small": lowerCAmelCase_ : Dict = torch.tensor( [[-8.8_236, -8.8_236, -8.6_771], [-8.8_236, -8.8_236, -8.6_771], [-8.7_638, -8.7_638, -8.6_240]]) elif model_name == "upernet-convnext-base": lowerCAmelCase_ : Union[str, Any] = torch.tensor( [[-8.8_558, -8.8_558, -8.6_905], [-8.8_558, -8.8_558, -8.6_905], [-8.7_669, -8.7_669, -8.6_021]]) elif model_name == "upernet-convnext-large": lowerCAmelCase_ : Optional[Any] = torch.tensor( [[-8.6_660, -8.6_660, -8.6_210], [-8.6_660, -8.6_660, -8.6_210], [-8.6_310, -8.6_310, -8.5_964]]) elif model_name == "upernet-convnext-xlarge": lowerCAmelCase_ : Union[str, Any] = torch.tensor( [[-8.4_980, -8.4_980, -8.3_977], [-8.4_980, -8.4_980, -8.3_977], [-8.4_379, -8.4_379, -8.3_412]]) print("Logits:" , outputs.logits[0, 0, :3, :3]) assert torch.allclose(outputs.logits[0, 0, :3, :3] , snake_case__ , atol=1e-4) print("Looks ok!") if pytorch_dump_folder_path is not None: print(F'''Saving model {model_name} to {pytorch_dump_folder_path}''') model.save_pretrained(snake_case__) print(F'''Saving processor to {pytorch_dump_folder_path}''') processor.save_pretrained(snake_case__) if push_to_hub: print(F'''Pushing model and processor for {model_name} to hub''') model.push_to_hub(F'''openmmlab/{model_name}''') processor.push_to_hub(F'''openmmlab/{model_name}''') if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''upernet-convnext-tiny''', type=str, choices=[f"upernet-convnext-{size}" for size in ['''tiny''', '''small''', '''base''', '''large''', '''xlarge''']], help='''Name of the ConvNext UperNet model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) _lowercase = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
659
from __future__ import annotations # This is the precision for this function which can be altered. # It is recommended for users to keep this number greater than or equal to 10. _lowercase = 10 def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): for i in range(snake_case__ , snake_case__): if array[i] == target: return i return -1 def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : List[str] = 0 lowerCAmelCase_ : Tuple = len(snake_case__) while left <= right: if right - left < precision: return lin_search(snake_case__ , snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : List[str] = (left + right) // 3 + 1 lowerCAmelCase_ : Tuple = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: lowerCAmelCase_ : str = one_third - 1 elif array[two_third] < target: lowerCAmelCase_ : Any = two_third + 1 else: lowerCAmelCase_ : List[str] = one_third + 1 lowerCAmelCase_ : Tuple = two_third - 1 else: return -1 def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): if left < right: if right - left < precision: return lin_search(snake_case__ , snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : Dict = (left + right) // 3 + 1 lowerCAmelCase_ : List[Any] = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: return rec_ternary_search(snake_case__ , one_third - 1 , snake_case__ , snake_case__) elif array[two_third] < target: return rec_ternary_search(two_third + 1 , snake_case__ , snake_case__ , snake_case__) else: return rec_ternary_search(one_third + 1 , two_third - 1 , snake_case__ , snake_case__) else: return -1 if __name__ == "__main__": import doctest doctest.testmod() _lowercase = input('''Enter numbers separated by comma:\n''').strip() _lowercase = [int(item.strip()) for item in user_input.split(''',''')] assert collection == sorted(collection), f"List must be ordered.\n{collection}." _lowercase = int(input('''Enter the number to be found in the list:\n''').strip()) _lowercase = ite_ternary_search(collection, target) _lowercase = rec_ternary_search(0, len(collection) - 1, collection, target) if resulta != -1: print(f"Iterative search: {target} found at positions: {resulta}") print(f"Recursive search: {target} found at positions: {resulta}") else: print('''Not found''')
659
1
import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : str = s.rsplit(snake_case__ , snake_case__) return new.join(snake_case__) def UpperCamelCase ( snake_case__): # encoder.embeddings are double copied in original FLAVA return sum(param.float().sum() if "encoder.embeddings" not in key else 0 for key, param in state_dict.items()) def UpperCamelCase ( snake_case__): lowerCAmelCase_ : int = {} lowerCAmelCase_ : Optional[int] = ["group_1", "group_2", "group_3", "group_4"] for key, value in state_dict.items(): for group_key in group_keys: if group_key in key: lowerCAmelCase_ : str = key.replace(F'''{group_key}.''' , F'''{group_key}.group.''') if "res_path" in key: lowerCAmelCase_ : Dict = key.replace("res_path." , "res_path.path.") if key.endswith(".w"): lowerCAmelCase_ : List[Any] = rreplace(snake_case__ , ".w" , ".weight" , 1) if key.endswith(".b"): lowerCAmelCase_ : List[str] = rreplace(snake_case__ , ".b" , ".bias" , 1) lowerCAmelCase_ : Dict = value.float() return upgrade @torch.no_grad() def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__=None , snake_case__=True): from dall_e import Encoder lowerCAmelCase_ : Dict = Encoder() if os.path.exists(snake_case__): lowerCAmelCase_ : List[str] = torch.load(snake_case__) else: lowerCAmelCase_ : Dict = torch.hub.load_state_dict_from_url(snake_case__) if isinstance(snake_case__ , snake_case__): lowerCAmelCase_ : Optional[Any] = ckpt.state_dict() encoder.load_state_dict(snake_case__) if config_path is not None: lowerCAmelCase_ : Any = FlavaImageCodebookConfig.from_pretrained(snake_case__) else: lowerCAmelCase_ : Union[str, Any] = FlavaImageCodebookConfig() lowerCAmelCase_ : str = FlavaImageCodebook(snake_case__).eval() lowerCAmelCase_ : Union[str, Any] = encoder.state_dict() lowerCAmelCase_ : Union[str, Any] = upgrade_state_dict(snake_case__) hf_model.load_state_dict(snake_case__) lowerCAmelCase_ : Union[str, Any] = hf_model.state_dict() lowerCAmelCase_ : Any = count_parameters(snake_case__) lowerCAmelCase_ : Dict = count_parameters(snake_case__) assert torch.allclose(snake_case__ , snake_case__ , atol=1e-3) if save_checkpoint: hf_model.save_pretrained(snake_case__) else: return hf_state_dict if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to flava checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') _lowercase = parser.parse_args() convert_dalle_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
659
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer _lowercase = logging.get_logger(__name__) _lowercase = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } _lowercase = { '''vocab_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json''' }, '''merges_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt''' }, '''tokenizer_config_file''': { '''facebook/blenderbot_small-90M''': ( '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json''' ) }, } _lowercase = { '''facebook/blenderbot_small-90M''': 512, } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = BlenderbotSmallTokenizer def __init__( self : Optional[int] ,lowerCAmelCase__ : Optional[int]=None ,lowerCAmelCase__ : Union[str, Any]=None ,lowerCAmelCase__ : Any="<|endoftext|>" ,lowerCAmelCase__ : int="<|endoftext|>" ,lowerCAmelCase__ : Optional[Any]="<|endoftext|>" ,lowerCAmelCase__ : Union[str, Any]=False ,lowerCAmelCase__ : Optional[Any]=True ,**lowerCAmelCase__ : Union[str, Any] ,) -> str: '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=lowerCAmelCase__ ,merges=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ,) ,bos_token=lowerCAmelCase__ ,eos_token=lowerCAmelCase__ ,unk_token=lowerCAmelCase__ ,**lowerCAmelCase__ ,) lowerCAmelCase_ : Dict = add_prefix_space def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Tuple=None ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : str = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : Dict = [self.sep_token_id] lowerCAmelCase_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
659
1
import os from datetime import datetime as dt from github import Github _lowercase = [ '''good first issue''', '''good second issue''', '''good difficult issue''', '''enhancement''', '''new pipeline/model''', '''new scheduler''', '''wip''', ] def UpperCamelCase ( ): lowerCAmelCase_ : List[str] = Github(os.environ["GITHUB_TOKEN"]) lowerCAmelCase_ : Optional[int] = g.get_repo("huggingface/diffusers") lowerCAmelCase_ : Optional[Any] = repo.get_issues(state="open") for issue in open_issues: lowerCAmelCase_ : str = sorted(issue.get_comments() , key=lambda snake_case__: i.created_at , reverse=snake_case__) lowerCAmelCase_ : Optional[Any] = comments[0] if len(snake_case__) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state="closed") elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state="open") issue.remove_from_labels("stale") elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) " "are likely to be ignored.") issue.add_to_labels("stale") if __name__ == "__main__": main()
659
from collections.abc import Generator from math import sin def UpperCamelCase ( snake_case__): if len(snake_case__) != 32: raise ValueError("Input must be of length 32") lowerCAmelCase_ : Tuple = b"" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def UpperCamelCase ( snake_case__): if i < 0: raise ValueError("Input must be non-negative") lowerCAmelCase_ : List[str] = format(snake_case__ , "08x")[-8:] lowerCAmelCase_ : Any = b"" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("utf-8") return little_endian_hex def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Union[str, Any] = b"" for char in message: bit_string += format(snake_case__ , "08b").encode("utf-8") lowerCAmelCase_ : Optional[int] = format(len(snake_case__) , "064b").encode("utf-8") # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(snake_case__) % 5_12 != 4_48: bit_string += b"0" bit_string += to_little_endian(start_len[32:]) + to_little_endian(start_len[:32]) return bit_string def UpperCamelCase ( snake_case__): if len(snake_case__) % 5_12 != 0: raise ValueError("Input must have length that's a multiple of 512") for pos in range(0 , len(snake_case__) , 5_12): lowerCAmelCase_ : List[str] = bit_string[pos : pos + 5_12] lowerCAmelCase_ : Union[str, Any] = [] for i in range(0 , 5_12 , 32): block_words.append(int(to_little_endian(block[i : i + 32]) , 2)) yield block_words def UpperCamelCase ( snake_case__): if i < 0: raise ValueError("Input must be non-negative") lowerCAmelCase_ : Dict = format(snake_case__ , "032b") lowerCAmelCase_ : str = "" for c in i_str: new_str += "1" if c == "0" else "0" return int(snake_case__ , 2) def UpperCamelCase ( snake_case__ , snake_case__): return (a + b) % 2**32 def UpperCamelCase ( snake_case__ , snake_case__): if i < 0: raise ValueError("Input must be non-negative") if shift < 0: raise ValueError("Shift must be non-negative") return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Optional[Any] = preprocess(snake_case__) lowerCAmelCase_ : Optional[Any] = [int(2**32 * abs(sin(i + 1))) for i in range(64)] # Starting states lowerCAmelCase_ : List[str] = 0x67_45_23_01 lowerCAmelCase_ : Union[str, Any] = 0xef_cd_ab_89 lowerCAmelCase_ : List[Any] = 0x98_ba_dc_fe lowerCAmelCase_ : Tuple = 0x10_32_54_76 lowerCAmelCase_ : Any = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(snake_case__): lowerCAmelCase_ : Optional[int] = aa lowerCAmelCase_ : List[str] = ba lowerCAmelCase_ : Any = ca lowerCAmelCase_ : Union[str, Any] = da # Hash current chunk for i in range(64): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f lowerCAmelCase_ : Any = d ^ (b & (c ^ d)) lowerCAmelCase_ : Dict = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f lowerCAmelCase_ : Any = c ^ (d & (b ^ c)) lowerCAmelCase_ : List[str] = (5 * i + 1) % 16 elif i <= 47: lowerCAmelCase_ : int = b ^ c ^ d lowerCAmelCase_ : Optional[Any] = (3 * i + 5) % 16 else: lowerCAmelCase_ : List[Any] = c ^ (b | not_aa(snake_case__)) lowerCAmelCase_ : List[Any] = (7 * i) % 16 lowerCAmelCase_ : Optional[Any] = (f + a + added_consts[i] + block_words[g]) % 2**32 lowerCAmelCase_ : Optional[Any] = d lowerCAmelCase_ : Dict = c lowerCAmelCase_ : List[str] = b lowerCAmelCase_ : Any = sum_aa(snake_case__ , left_rotate_aa(snake_case__ , shift_amounts[i])) # Add hashed chunk to running total lowerCAmelCase_ : Dict = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : str = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : Optional[int] = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : int = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : Union[str, Any] = reformat_hex(snake_case__) + reformat_hex(snake_case__) + reformat_hex(snake_case__) + reformat_hex(snake_case__) return digest if __name__ == "__main__": import doctest doctest.testmod()
659
1
import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder _lowercase = '''base_with_context''' def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : int = nn.Parameter(torch.FloatTensor(weights["token_embedder"]["embedding"])) lowerCAmelCase_ : Union[str, Any] = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"]) , requires_grad=snake_case__) for lyr_num, lyr in enumerate(model.encoders): lowerCAmelCase_ : Dict = weights[F'''layers_{lyr_num}'''] lowerCAmelCase_ : Optional[int] = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"])) lowerCAmelCase_ : List[str] = ly_weight["attention"] lowerCAmelCase_ : Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T)) lowerCAmelCase_ : Optional[Any] = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T)) lowerCAmelCase_ : Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T)) lowerCAmelCase_ : int = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T)) lowerCAmelCase_ : Dict = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"])) lowerCAmelCase_ : Dict = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T)) lowerCAmelCase_ : Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T)) lowerCAmelCase_ : Optional[int] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T)) lowerCAmelCase_ : Any = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"])) return model def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : str = nn.Parameter(torch.FloatTensor(weights["input_proj"]["kernel"].T)) lowerCAmelCase_ : Tuple = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"]) , requires_grad=snake_case__) for lyr_num, lyr in enumerate(model.encoders): lowerCAmelCase_ : Optional[int] = weights[F'''layers_{lyr_num}'''] lowerCAmelCase_ : Union[str, Any] = ly_weight["attention"] lowerCAmelCase_ : int = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T)) lowerCAmelCase_ : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T)) lowerCAmelCase_ : int = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T)) lowerCAmelCase_ : Any = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T)) lowerCAmelCase_ : str = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"])) lowerCAmelCase_ : List[Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T)) lowerCAmelCase_ : List[Any] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T)) lowerCAmelCase_ : int = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T)) lowerCAmelCase_ : Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"])) lowerCAmelCase_ : List[Any] = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"])) return model def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : str = nn.Parameter(torch.FloatTensor(weights["time_emb_dense0"]["kernel"].T)) lowerCAmelCase_ : Optional[int] = nn.Parameter(torch.FloatTensor(weights["time_emb_dense1"]["kernel"].T)) lowerCAmelCase_ : Union[str, Any] = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"]) , requires_grad=snake_case__) lowerCAmelCase_ : Tuple = nn.Parameter( torch.FloatTensor(weights["continuous_inputs_projection"]["kernel"].T)) for lyr_num, lyr in enumerate(model.decoders): lowerCAmelCase_ : Dict = weights[F'''layers_{lyr_num}'''] lowerCAmelCase_ : str = nn.Parameter( torch.FloatTensor(ly_weight["pre_self_attention_layer_norm"]["scale"])) lowerCAmelCase_ : List[Any] = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_0"]["DenseGeneral_0"]["kernel"].T)) lowerCAmelCase_ : List[Any] = ly_weight["self_attention"] lowerCAmelCase_ : List[Any] = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T)) lowerCAmelCase_ : str = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T)) lowerCAmelCase_ : str = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T)) lowerCAmelCase_ : Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T)) lowerCAmelCase_ : List[Any] = ly_weight["MultiHeadDotProductAttention_0"] lowerCAmelCase_ : str = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T)) lowerCAmelCase_ : Any = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T)) lowerCAmelCase_ : Tuple = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T)) lowerCAmelCase_ : Any = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T)) lowerCAmelCase_ : str = nn.Parameter( torch.FloatTensor(ly_weight["pre_cross_attention_layer_norm"]["scale"])) lowerCAmelCase_ : Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"])) lowerCAmelCase_ : Dict = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_1"]["DenseGeneral_0"]["kernel"].T)) lowerCAmelCase_ : Tuple = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T)) lowerCAmelCase_ : List[str] = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T)) lowerCAmelCase_ : str = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T)) lowerCAmelCase_ : str = nn.Parameter(torch.FloatTensor(weights["decoder_norm"]["scale"])) lowerCAmelCase_ : Optional[int] = nn.Parameter(torch.FloatTensor(weights["spec_out_dense"]["kernel"].T)) return model def UpperCamelCase ( snake_case__): lowerCAmelCase_ : int = checkpoints.load_tax_checkpoint(args.checkpoint_path) lowerCAmelCase_ : str = jnp.tree_util.tree_map(onp.array , snake_case__) lowerCAmelCase_ : Optional[int] = [ "from __gin__ import dynamic_registration", "from music_spectrogram_diffusion.models.diffusion import diffusion_utils", "diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0", "diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()", ] lowerCAmelCase_ : str = os.path.join(args.checkpoint_path , ".." , "config.gin") lowerCAmelCase_ : List[Any] = inference.parse_training_gin_file(snake_case__ , snake_case__) lowerCAmelCase_ : Any = inference.InferenceModel(args.checkpoint_path , snake_case__) lowerCAmelCase_ : int = DDPMScheduler(beta_schedule="squaredcos_cap_v2" , variance_type="fixed_large") lowerCAmelCase_ : Tuple = SpectrogramNotesEncoder( max_length=synth_model.sequence_length["inputs"] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj="gated-gelu" , ) lowerCAmelCase_ : int = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length["targets_context"] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj="gated-gelu" , ) lowerCAmelCase_ : int = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length["targets_context"] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) lowerCAmelCase_ : Optional[int] = load_notes_encoder(ta_checkpoint["target"]["token_encoder"] , snake_case__) lowerCAmelCase_ : Optional[Any] = load_continuous_encoder(ta_checkpoint["target"]["continuous_encoder"] , snake_case__) lowerCAmelCase_ : Dict = load_decoder(ta_checkpoint["target"]["decoder"] , snake_case__) lowerCAmelCase_ : Tuple = OnnxRuntimeModel.from_pretrained("kashif/soundstream_mel_decoder") lowerCAmelCase_ : Optional[Any] = SpectrogramDiffusionPipeline( notes_encoder=snake_case__ , continuous_encoder=snake_case__ , decoder=snake_case__ , scheduler=snake_case__ , melgan=snake_case__ , ) if args.save: pipe.save_pretrained(args.output_path) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument('''--output_path''', default=None, type=str, required=True, help='''Path to the converted model.''') parser.add_argument( '''--save''', default=True, type=bool, required=False, help='''Whether to save the converted model or not.''' ) parser.add_argument( '''--checkpoint_path''', default=f"{MODEL}/checkpoint_500000", type=str, required=False, help='''Path to the original jax model checkpoint.''', ) _lowercase = parser.parse_args() main(args)
659
import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): _lowercase = True from torch.cuda.amp import autocast _lowercase = logging.getLogger(__name__) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Whether to log verbose messages or not.'} , ) UpperCamelCase_ = field( default=2.0 , metadata={'help': 'Maximum temperature for gumbel softmax.'} ) UpperCamelCase_ = field( default=0.5 , metadata={'help': 'Minimum temperature for gumbel softmax.'} ) UpperCamelCase_ = field( default=0.99_99_95 , metadata={'help': 'Decay of gumbel temperature during training.'} ) def UpperCamelCase ( snake_case__ , snake_case__): logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout)] , ) lowerCAmelCase_ : str = logging.WARNING if model_args.verbose_logging: lowerCAmelCase_ : int = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank): lowerCAmelCase_ : Any = logging.INFO logger.setLevel(snake_case__) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The name of the dataset to use (via the datasets library).'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) UpperCamelCase_ = field( default='train' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) UpperCamelCase_ = field( default='validation' , metadata={ 'help': ( 'The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'' ) } , ) UpperCamelCase_ = field( default='file' , metadata={'help': 'Column in the dataset that contains speech file path. Defaults to \'file\''} , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Overwrite the cached preprocessed datasets or not.'} ) UpperCamelCase_ = field( default=1 , metadata={ 'help': 'The percentage of the train set used as validation set in case there\'s no validation split' } , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) UpperCamelCase_ = field( default=20.0 , metadata={'help': 'Filter audio files that are longer than `max_duration_in_seconds` seconds'} ) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = 42 UpperCamelCase_ = 42 UpperCamelCase_ = "longest" UpperCamelCase_ = None UpperCamelCase_ = None def __call__( self : str ,lowerCAmelCase__ : List[Dict[str, Union[List[int], torch.Tensor]]] ) -> Dict[str, torch.Tensor]: '''simple docstring''' lowerCAmelCase_ : Tuple = self.feature_extractor.pad( lowerCAmelCase__ ,max_length=self.max_length ,padding=self.padding ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors="pt" ,) lowerCAmelCase_ : Union[str, Any] = self.model._get_feat_extract_output_lengths(batch["input_values"].shape[-1] ) lowerCAmelCase_ : List[str] = batch["input_values"].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula lowerCAmelCase_ : Tuple = self.model._get_feat_extract_output_lengths(batch["attention_mask"].sum(-1 ) ).to( torch.long ) lowerCAmelCase_ : Optional[Any] = torch.zeros( (batch_size, mask_indices_seq_length) ,dtype=torch.long ,device=batch["input_values"].device ) # these two operations makes sure that all values # before the output lengths indices are attended to lowerCAmelCase_ : Tuple = 1 lowerCAmelCase_ : int = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices lowerCAmelCase_ : str = _compute_mask_indices( (batch_size, mask_indices_seq_length) ,self.model.config.mask_time_prob ,self.model.config.mask_time_length ,attention_mask=lowerCAmelCase__ ,min_masks=2 ,) return batch class __snake_case ( snake_case__ ): """simple docstring""" def __init__( self : List[str] ,*lowerCAmelCase__ : Optional[int] ,lowerCAmelCase__ : Tuple=1 ,lowerCAmelCase__ : Optional[int]=0 ,lowerCAmelCase__ : Optional[Any]=1.0 ,**lowerCAmelCase__ : Any ) -> str: '''simple docstring''' super().__init__(*lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = 0 lowerCAmelCase_ : int = max_gumbel_temp lowerCAmelCase_ : Union[str, Any] = min_gumbel_temp lowerCAmelCase_ : str = gumbel_temp_decay def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : nn.Module ,lowerCAmelCase__ : Dict[str, Union[torch.Tensor, Any]] ) -> torch.Tensor: '''simple docstring''' model.train() lowerCAmelCase_ : str = self._prepare_inputs(lowerCAmelCase__ ) if self.use_amp: with autocast(): lowerCAmelCase_ : List[Any] = self.compute_loss(lowerCAmelCase__ ,lowerCAmelCase__ ) else: lowerCAmelCase_ : List[Any] = self.compute_loss(lowerCAmelCase__ ,lowerCAmelCase__ ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": lowerCAmelCase_ : List[Any] = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": lowerCAmelCase_ : Optional[Any] = loss.sum() / (inputs["mask_time_indices"]).sum() else: raise ValueError(f'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: lowerCAmelCase_ : int = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(lowerCAmelCase__ ).backward() elif self.use_apex: with amp.scale_loss(lowerCAmelCase__ ,self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(lowerCAmelCase__ ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) return loss.detach() def UpperCamelCase ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase_ : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments)) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = parser.parse_args_into_dataclasses() configure_logger(snake_case__ , snake_case__) # Downloading and loading a dataset from the hub. lowerCAmelCase_ : List[str] = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" lowerCAmelCase_ : Any = DatasetDict() lowerCAmelCase_ : Union[str, Any] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : List[str] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" lowerCAmelCase_ : Union[str, Any] = DatasetDict() lowerCAmelCase_ : int = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split="validation" , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : Any = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported lowerCAmelCase_ : Dict = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=snake_case__) def prepare_dataset(snake_case__): # check that all files have the correct sampling rate lowerCAmelCase_ , lowerCAmelCase_ : str = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate) return batch # load audio files into numpy arrays lowerCAmelCase_ : int = datasets.map( snake_case__ , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets["train"].column_names) # filter audio files that are too long lowerCAmelCase_ : int = vectorized_datasets.filter( lambda snake_case__: len(data["speech"]) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate)) def normalize(snake_case__): return feature_extractor(batch["speech"] , sampling_rate=feature_extractor.sampling_rate) # normalize and transform to `BatchFeatures` lowerCAmelCase_ : str = vectorized_datasets.map( snake_case__ , batched=snake_case__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets["train"].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 lowerCAmelCase_ : Optional[Any] = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( "PreTraining is only supported for ``config.do_stable_layer_norm=True`` and" " ``config.feat_extract_norm='layer'") lowerCAmelCase_ : Dict = WavaVecaForPreTraining(snake_case__) lowerCAmelCase_ : int = DataCollatorForWavaVecaPretraining(model=snake_case__ , feature_extractor=snake_case__) lowerCAmelCase_ : List[Any] = WavaVecaPreTrainer( model=snake_case__ , data_collator=snake_case__ , args=snake_case__ , train_dataset=vectorized_datasets["train"] , eval_dataset=vectorized_datasets["validation"] , tokenizer=snake_case__ , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
659
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowercase = { '''configuration_jukebox''': [ '''JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''JukeboxConfig''', '''JukeboxPriorConfig''', '''JukeboxVQVAEConfig''', ], '''tokenization_jukebox''': ['''JukeboxTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ '''JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST''', '''JukeboxModel''', '''JukeboxPreTrainedModel''', '''JukeboxVQVAE''', '''JukeboxPrior''', ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
659
from __future__ import annotations from collections.abc import Callable def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1_00 , ): lowerCAmelCase_ : Any = x_start lowerCAmelCase_ : Optional[Any] = fnc(snake_case__) lowerCAmelCase_ : Union[str, Any] = 0.0 for _ in range(snake_case__): # Approximates small segments of curve as linear and solve # for trapezoidal area lowerCAmelCase_ : Any = (x_end - x_start) / steps + xa lowerCAmelCase_ : Dict = fnc(snake_case__) area += abs(fxa + fxa) * (xa - xa) / 2 # Increment step lowerCAmelCase_ : int = xa lowerCAmelCase_ : str = fxa return area if __name__ == "__main__": def UpperCamelCase ( snake_case__): return x**3 + x**2 print('''f(x) = x^3 + x^2''') print('''The area between the curve, x = -5, x = 5 and the x axis is:''') _lowercase = 10 while i <= 100000: print(f"with {i} steps: {trapezoidal_area(f, -5, 5, i)}") i *= 10
659
1
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model _lowercase = '''0.12''' # assumed parallelism: 8 if is_torch_available(): import torch def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__=None): if rng is None: lowerCAmelCase_ : List[str] = random.Random() lowerCAmelCase_ : Dict = 1 for dim in shape: total_dims *= dim lowerCAmelCase_ : Optional[int] = [] for _ in range(snake_case__): values.append(rng.randint(0 , vocab_size - 1)) lowerCAmelCase_ : Any = np.array(snake_case__ , dtype=jnp.intaa).reshape(snake_case__) return output def UpperCamelCase ( snake_case__ , snake_case__=None): lowerCAmelCase_ : Optional[int] = ids_tensor(snake_case__ , vocab_size=2 , rng=snake_case__) # make sure that at least one token is attended to for each batch lowerCAmelCase_ : Tuple = 1 return attn_mask @require_flax class __snake_case : """simple docstring""" UpperCamelCase_ = None UpperCamelCase_ = () def UpperCAmelCase_ ( self : Optional[Any] ) -> Any: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 lowerCAmelCase_ : Optional[int] = 2 lowerCAmelCase_ : Union[str, Any] = inputs["input_ids"].shape[-1] // 2 lowerCAmelCase_ : Tuple = inputs["input_ids"][:max_batch_size, :sequence_length] lowerCAmelCase_ : Tuple = jnp.ones_like(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens lowerCAmelCase_ : Optional[Any] = input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` lowerCAmelCase_ : Optional[int] = config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def UpperCAmelCase_ ( self : str ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = self._get_input_ids_and_config() lowerCAmelCase_ : str = False lowerCAmelCase_ : List[Any] = max_length lowerCAmelCase_ : Dict = 0 for model_class in self.all_generative_model_classes: lowerCAmelCase_ : Tuple = model_class(lowerCAmelCase__ ) lowerCAmelCase_ : int = model_class.__name__[4:] # Skip the "Flax" at the beginning lowerCAmelCase_ : List[Any] = getattr(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Any = pt_model_class(lowerCAmelCase__ ).eval() lowerCAmelCase_ : List[Any] = load_flax_weights_in_pytorch_model(lowerCAmelCase__ ,flax_model.params ) lowerCAmelCase_ : List[str] = flax_model.generate(lowerCAmelCase__ ).sequences lowerCAmelCase_ : Union[str, Any] = pt_model.generate(torch.tensor(lowerCAmelCase__ ,dtype=torch.long ) ) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: lowerCAmelCase_ : List[Any] = flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() ,flax_generation_outputs.tolist() ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[str]: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = self._get_input_ids_and_config() lowerCAmelCase_ : List[Any] = False lowerCAmelCase_ : Optional[Any] = max_length for model_class in self.all_generative_model_classes: lowerCAmelCase_ : Tuple = model_class(lowerCAmelCase__ ) lowerCAmelCase_ : Any = model.generate(lowerCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] ,lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = jit(model.generate ) lowerCAmelCase_ : Tuple = jit_generate(lowerCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist() ) def UpperCAmelCase_ ( self : Tuple ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = self._get_input_ids_and_config() lowerCAmelCase_ : List[Any] = True lowerCAmelCase_ : Union[str, Any] = max_length for model_class in self.all_generative_model_classes: lowerCAmelCase_ : Any = model_class(lowerCAmelCase__ ) lowerCAmelCase_ : int = model.generate(lowerCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] ,lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = jit(model.generate ) lowerCAmelCase_ : Optional[int] = jit_generate(lowerCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist() ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[str]: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = self._get_input_ids_and_config() lowerCAmelCase_ : int = False lowerCAmelCase_ : int = max_length lowerCAmelCase_ : int = 2 for model_class in self.all_generative_model_classes: lowerCAmelCase_ : Optional[int] = model_class(lowerCAmelCase__ ) lowerCAmelCase_ : str = model.generate(lowerCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] ,lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = jit(model.generate ) lowerCAmelCase_ : Union[str, Any] = jit_generate(lowerCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist() ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Dict: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Tuple = self._get_input_ids_and_config() lowerCAmelCase_ : Optional[int] = False lowerCAmelCase_ : Union[str, Any] = max_length lowerCAmelCase_ : int = 2 lowerCAmelCase_ : List[Any] = 2 for model_class in self.all_generative_model_classes: lowerCAmelCase_ : str = model_class(lowerCAmelCase__ ) lowerCAmelCase_ : str = model.generate(lowerCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[0] ,input_ids.shape[0] * config.num_return_sequences ) def UpperCAmelCase_ ( self : Dict ) -> int: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = self._get_input_ids_and_config() lowerCAmelCase_ : Dict = True lowerCAmelCase_ : int = max_length lowerCAmelCase_ : str = 0.8 lowerCAmelCase_ : Dict = 10 lowerCAmelCase_ : List[Any] = 0.3 lowerCAmelCase_ : str = 1 lowerCAmelCase_ : Tuple = 8 lowerCAmelCase_ : List[Any] = 9 for model_class in self.all_generative_model_classes: lowerCAmelCase_ : List[Any] = model_class(lowerCAmelCase__ ) lowerCAmelCase_ : str = model.generate(lowerCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] ,lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = jit(model.generate ) lowerCAmelCase_ : int = jit_generate(lowerCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist() ) def UpperCAmelCase_ ( self : Any ) -> Dict: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[str] = self._get_input_ids_and_config() lowerCAmelCase_ : Union[str, Any] = max_length lowerCAmelCase_ : List[str] = 1 lowerCAmelCase_ : int = 8 lowerCAmelCase_ : List[str] = 9 for model_class in self.all_generative_model_classes: lowerCAmelCase_ : Optional[int] = model_class(lowerCAmelCase__ ) lowerCAmelCase_ : Any = model.generate(lowerCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] ,lowerCAmelCase__ ) lowerCAmelCase_ : str = jit(model.generate ) lowerCAmelCase_ : Tuple = jit_generate(lowerCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist() ) def UpperCAmelCase_ ( self : Any ) -> Tuple: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : int = self._get_input_ids_and_config() lowerCAmelCase_ : str = max_length lowerCAmelCase_ : Optional[int] = 2 lowerCAmelCase_ : Dict = 1 lowerCAmelCase_ : List[str] = 8 lowerCAmelCase_ : Any = 9 for model_class in self.all_generative_model_classes: lowerCAmelCase_ : Optional[int] = model_class(lowerCAmelCase__ ) lowerCAmelCase_ : str = model.generate(lowerCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] ,lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = jit(model.generate ) lowerCAmelCase_ : Dict = jit_generate(lowerCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist() ) def UpperCAmelCase_ ( self : int ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[str] = self._get_input_ids_and_config() # pad attention mask on the left lowerCAmelCase_ : List[Any] = attention_mask.at[(0, 0)].set(0 ) lowerCAmelCase_ : Dict = False lowerCAmelCase_ : Any = max_length for model_class in self.all_generative_model_classes: lowerCAmelCase_ : str = model_class(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = model.generate(lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] ,lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = jit(model.generate ) lowerCAmelCase_ : str = jit_generate(lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist() ) def UpperCAmelCase_ ( self : Tuple ) -> List[str]: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = self._get_input_ids_and_config() # pad attention mask on the left lowerCAmelCase_ : List[str] = attention_mask.at[(0, 0)].set(0 ) lowerCAmelCase_ : int = True lowerCAmelCase_ : Union[str, Any] = max_length for model_class in self.all_generative_model_classes: lowerCAmelCase_ : List[str] = model_class(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = model.generate(lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] ,lowerCAmelCase__ ) lowerCAmelCase_ : str = jit(model.generate ) lowerCAmelCase_ : int = jit_generate(lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist() ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[str] = self._get_input_ids_and_config() # pad attention mask on the left lowerCAmelCase_ : Tuple = attention_mask.at[(0, 0)].set(0 ) lowerCAmelCase_ : Union[str, Any] = 2 lowerCAmelCase_ : Optional[Any] = max_length for model_class in self.all_generative_model_classes: lowerCAmelCase_ : Optional[int] = model_class(lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = model.generate(lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ).sequences self.assertEqual(generation_outputs.shape[-1] ,lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = jit(model.generate ) lowerCAmelCase_ : List[Any] = jit_generate(lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ).sequences self.assertListEqual(generation_outputs.tolist() ,jit_generation_outputs.tolist() ) @require_flax class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : int ) -> int: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-bert" ) lowerCAmelCase_ : List[Any] = FlaxAutoModelForCausalLM.from_pretrained("hf-internal-testing/tiny-bert-flax-only" ) lowerCAmelCase_ : List[Any] = "Hello world" lowerCAmelCase_ : Optional[Any] = tokenizer(lowerCAmelCase__ ,return_tensors="np" ).input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(lowerCAmelCase__ ,"do_samples" ): model.generate(lowerCAmelCase__ ,do_samples=lowerCAmelCase__ ) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(lowerCAmelCase__ ,"foo" ): lowerCAmelCase_ : Any = {"foo": "bar"} model.generate(lowerCAmelCase__ ,**lowerCAmelCase__ )
659
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, PNDMScheduler, StableDiffusionLDMaDPipeline, UNetaDConditionModel, ) from diffusers.utils import nightly, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS enable_full_determinism() class __snake_case ( unittest.TestCase ): """simple docstring""" UpperCamelCase_ = StableDiffusionLDMaDPipeline UpperCamelCase_ = TEXT_TO_IMAGE_PARAMS UpperCamelCase_ = TEXT_TO_IMAGE_BATCH_PARAMS UpperCamelCase_ = TEXT_TO_IMAGE_IMAGE_PARAMS def UpperCAmelCase_ ( self : Tuple ) -> str: '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase_ : Optional[Any] = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") ,up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") ,cross_attention_dim=32 ,) lowerCAmelCase_ : Any = DDIMScheduler( beta_start=0.00_085 ,beta_end=0.012 ,beta_schedule="scaled_linear" ,clip_sample=lowerCAmelCase__ ,set_alpha_to_one=lowerCAmelCase__ ,) torch.manual_seed(0 ) lowerCAmelCase_ : str = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=6 ,out_channels=6 ,down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] ,up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] ,latent_channels=4 ,) torch.manual_seed(0 ) lowerCAmelCase_ : Optional[Any] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=10_00 ,) lowerCAmelCase_ : Optional[int] = CLIPTextModel(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) lowerCAmelCase_ : List[Any] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : List[str]=0 ) -> Dict: '''simple docstring''' if str(lowerCAmelCase__ ).startswith("mps" ): lowerCAmelCase_ : Optional[int] = torch.manual_seed(lowerCAmelCase__ ) else: lowerCAmelCase_ : Dict = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : str = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : Any ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Dict = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : List[str] = self.get_dummy_components() lowerCAmelCase_ : Union[str, Any] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Any = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Any = output.rgb, output.depth lowerCAmelCase_ : Dict = rgb[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) lowerCAmelCase_ : Optional[Any] = np.array( [0.37_338_176, 0.70_247, 0.74_203_193, 0.51_643_604, 0.58_256_793, 0.60_932_136, 0.4_181_095, 0.48_355_877, 0.46_535_262] ) lowerCAmelCase_ : Tuple = np.array([103.46_727, 85.812_004, 87.849_236] ) assert np.abs(image_slice_rgb.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(image_slice_depth.flatten() - expected_slice_depth ).max() < 1e-2 def UpperCAmelCase_ ( self : int ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Dict = self.get_dummy_components() lowerCAmelCase_ : List[str] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : str = 3 * [inputs["prompt"]] # forward lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = output.rgb, output.depth lowerCAmelCase_ : str = rgb_slice_a[0, -3:, -3:, -1] lowerCAmelCase_ : List[str] = depth_slice_a[0, -3:, -1] lowerCAmelCase_ : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = 3 * [inputs.pop("prompt" )] lowerCAmelCase_ : str = ldmad_pipe.tokenizer( lowerCAmelCase__ ,padding="max_length" ,max_length=ldmad_pipe.tokenizer.model_max_length ,truncation=lowerCAmelCase__ ,return_tensors="pt" ,) lowerCAmelCase_ : Union[str, Any] = text_inputs["input_ids"].to(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = ldmad_pipe.text_encoder(lowerCAmelCase__ )[0] lowerCAmelCase_ : Optional[int] = prompt_embeds # forward lowerCAmelCase_ : str = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : str = output.rgb, output.depth lowerCAmelCase_ : Optional[Any] = rgb_slice_a[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth_slice_a[0, -3:, -1] assert np.abs(rgb_slice_a.flatten() - rgb_slice_a.flatten() ).max() < 1e-4 assert np.abs(depth_slice_a.flatten() - depth_slice_a.flatten() ).max() < 1e-4 def UpperCAmelCase_ ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : Any = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : Optional[int] = self.get_dummy_components() lowerCAmelCase_ : Dict = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : Any = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = "french fries" lowerCAmelCase_ : Optional[int] = ldmad_pipe(**lowerCAmelCase__ ,negative_prompt=lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = output.rgb, output.depth lowerCAmelCase_ : Any = rgb[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) lowerCAmelCase_ : int = np.array( [0.37_044, 0.71_811_503, 0.7_223_251, 0.48_603_675, 0.5_638_391, 0.6_364_948, 0.42_833_704, 0.4_901_315, 0.47_926_217] ) lowerCAmelCase_ : Union[str, Any] = np.array([107.84_738, 84.62_802, 89.962_135] ) assert np.abs(rgb_slice.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(depth_slice.flatten() - expected_slice_depth ).max() < 1e-2 @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Dict="cpu" ,lowerCAmelCase__ : Union[str, Any]=torch.floataa ,lowerCAmelCase__ : List[str]=0 ) -> int: '''simple docstring''' lowerCAmelCase_ : Any = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = np.random.RandomState(lowerCAmelCase__ ).standard_normal((1, 4, 64, 64) ) lowerCAmelCase_ : Optional[Any] = torch.from_numpy(lowerCAmelCase__ ).to(device=lowerCAmelCase__ ,dtype=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : List[Any] ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ) lowerCAmelCase_ : List[str] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Dict = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Dict = output.rgb, output.depth lowerCAmelCase_ : List[str] = rgb[0, -3:, -3:, -1].flatten() lowerCAmelCase_ : Optional[int] = rgb[0, -3:, -1].flatten() assert rgb.shape == (1, 5_12, 5_12, 3) assert depth.shape == (1, 5_12, 5_12) lowerCAmelCase_ : int = np.array( [0.53_805_465, 0.56_707_305, 0.5_486_515, 0.57_012_236, 0.5_814_511, 0.56_253_487, 0.54_843_014, 0.55_092_263, 0.6_459_706] ) lowerCAmelCase_ : Optional[Any] = np.array( [0.9_263_781, 0.6_678_672, 0.5_486_515, 0.92_202_145, 0.67_831_135, 0.56_253_487, 0.9_241_694, 0.7_551_478, 0.6_459_706] ) assert np.abs(rgb_slice - expected_slice_rgb ).max() < 3e-3 assert np.abs(depth_slice - expected_slice_depth ).max() < 3e-3 @nightly @require_torch_gpu class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Dict="cpu" ,lowerCAmelCase__ : List[str]=torch.floataa ,lowerCAmelCase__ : Optional[int]=0 ) -> int: '''simple docstring''' lowerCAmelCase_ : Dict = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = np.random.RandomState(lowerCAmelCase__ ).standard_normal((1, 4, 64, 64) ) lowerCAmelCase_ : Any = torch.from_numpy(lowerCAmelCase__ ).to(device=lowerCAmelCase__ ,dtype=lowerCAmelCase__ ) lowerCAmelCase_ : int = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 50, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : Dict ) -> int: '''simple docstring''' lowerCAmelCase_ : List[Any] = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ).to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Any = output.rgb, output.depth lowerCAmelCase_ : Dict = 0.495_586 lowerCAmelCase_ : Optional[Any] = 0.33_795_515 lowerCAmelCase_ : Any = 112.48_518 lowerCAmelCase_ : List[Any] = 98.489_746 assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3 def UpperCAmelCase_ ( self : Tuple ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : int = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d-4c" ).to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : str = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Tuple = output.rgb, output.depth lowerCAmelCase_ : List[str] = 0.4_194_127 lowerCAmelCase_ : List[str] = 0.35_375_586 lowerCAmelCase_ : str = 0.5_638_502 lowerCAmelCase_ : Optional[Any] = 0.34_686_103 assert rgb.shape == (1, 5_12, 5_12, 3) assert depth.shape == (1, 5_12, 5_12, 1) assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3
659
1
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Sequence, Value from .base import TaskTemplate @dataclass(frozen=snake_case__ ) class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = field(default='question-answering-extractive' , metadata={'include_in_asdict_even_if_is_default': True} ) UpperCamelCase_ = Features({'question': Value('string' ), 'context': Value('string' )} ) UpperCamelCase_ = Features( { 'answers': Sequence( { 'text': Value('string' ), 'answer_start': Value('int32' ), } ) } ) UpperCamelCase_ = "question" UpperCamelCase_ = "context" UpperCamelCase_ = "answers" @property def UpperCAmelCase_ ( self : str ) -> Dict[str, str]: '''simple docstring''' return {self.question_column: "question", self.context_column: "context", self.answers_column: "answers"}
659
import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) _lowercase = { '''iou_prediction_head.layers.0''': '''iou_prediction_head.proj_in''', '''iou_prediction_head.layers.1''': '''iou_prediction_head.layers.0''', '''iou_prediction_head.layers.2''': '''iou_prediction_head.proj_out''', '''mask_decoder.output_upscaling.0''': '''mask_decoder.upscale_conv1''', '''mask_decoder.output_upscaling.1''': '''mask_decoder.upscale_layer_norm''', '''mask_decoder.output_upscaling.3''': '''mask_decoder.upscale_conv2''', '''mask_downscaling.0''': '''mask_embed.conv1''', '''mask_downscaling.1''': '''mask_embed.layer_norm1''', '''mask_downscaling.3''': '''mask_embed.conv2''', '''mask_downscaling.4''': '''mask_embed.layer_norm2''', '''mask_downscaling.6''': '''mask_embed.conv3''', '''point_embeddings''': '''point_embed''', '''pe_layer.positional_encoding_gaussian_matrix''': '''shared_embedding.positional_embedding''', '''image_encoder''': '''vision_encoder''', '''neck.0''': '''neck.conv1''', '''neck.1''': '''neck.layer_norm1''', '''neck.2''': '''neck.conv2''', '''neck.3''': '''neck.layer_norm2''', '''patch_embed.proj''': '''patch_embed.projection''', '''.norm''': '''.layer_norm''', '''blocks''': '''layers''', } def UpperCamelCase ( snake_case__): lowerCAmelCase_ : int = {} state_dict.pop("pixel_mean" , snake_case__) state_dict.pop("pixel_std" , snake_case__) lowerCAmelCase_ : List[Any] = R".*.output_hypernetworks_mlps.(\d+).layers.(\d+).*" for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: lowerCAmelCase_ : Dict = key.replace(snake_case__ , snake_case__) if re.match(snake_case__ , snake_case__): lowerCAmelCase_ : Any = int(re.match(snake_case__ , snake_case__).group(2)) if layer_nb == 0: lowerCAmelCase_ : List[Any] = key.replace("layers.0" , "proj_in") elif layer_nb == 1: lowerCAmelCase_ : List[Any] = key.replace("layers.1" , "layers.0") elif layer_nb == 2: lowerCAmelCase_ : int = key.replace("layers.2" , "proj_out") lowerCAmelCase_ : int = value lowerCAmelCase_ : Optional[int] = model_state_dict[ "prompt_encoder.shared_embedding.positional_embedding" ] return model_state_dict def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__="ybelkada/segment-anything"): lowerCAmelCase_ : Optional[int] = hf_hub_download(snake_case__ , F'''checkpoints/{model_name}.pth''') if "sam_vit_b" in model_name: lowerCAmelCase_ : Optional[Any] = SamConfig() elif "sam_vit_l" in model_name: lowerCAmelCase_ : Optional[int] = SamVisionConfig( hidden_size=10_24 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) lowerCAmelCase_ : Union[str, Any] = SamConfig( vision_config=snake_case__ , ) elif "sam_vit_h" in model_name: lowerCAmelCase_ : Optional[Any] = SamVisionConfig( hidden_size=12_80 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) lowerCAmelCase_ : Tuple = SamConfig( vision_config=snake_case__ , ) lowerCAmelCase_ : Optional[Any] = torch.load(snake_case__ , map_location="cpu") lowerCAmelCase_ : Union[str, Any] = replace_keys(snake_case__) lowerCAmelCase_ : List[Any] = SamImageProcessor() lowerCAmelCase_ : Any = SamProcessor(image_processor=snake_case__) lowerCAmelCase_ : Any = SamModel(snake_case__) hf_model.load_state_dict(snake_case__) lowerCAmelCase_ : Dict = hf_model.to("cuda") lowerCAmelCase_ : List[str] = "https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png" lowerCAmelCase_ : List[Any] = Image.open(requests.get(snake_case__ , stream=snake_case__).raw).convert("RGB") lowerCAmelCase_ : Optional[int] = [[[4_00, 6_50]]] lowerCAmelCase_ : int = [[1]] lowerCAmelCase_ : Optional[Any] = processor(images=np.array(snake_case__) , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Optional[Any] = hf_model(**snake_case__) lowerCAmelCase_ : Optional[int] = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579_890_251_159_668 lowerCAmelCase_ : Any = processor( images=np.array(snake_case__) , input_points=snake_case__ , input_labels=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Optional[Any] = hf_model(**snake_case__) lowerCAmelCase_ : Union[str, Any] = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_712_603_092_193_604 lowerCAmelCase_ : Tuple = ((75, 2_75, 17_25, 8_50),) lowerCAmelCase_ : Optional[Any] = processor(images=np.array(snake_case__) , input_boxes=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : List[Any] = hf_model(**snake_case__) lowerCAmelCase_ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.8_686_015_605_926_514 # Test with 2 points and 1 image. lowerCAmelCase_ : int = [[[4_00, 6_50], [8_00, 6_50]]] lowerCAmelCase_ : Optional[Any] = [[1, 1]] lowerCAmelCase_ : List[Any] = processor( images=np.array(snake_case__) , input_points=snake_case__ , input_labels=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Tuple = hf_model(**snake_case__) lowerCAmelCase_ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_936_047_792_434_692 if __name__ == "__main__": _lowercase = argparse.ArgumentParser() _lowercase = ['''sam_vit_b_01ec64''', '''sam_vit_h_4b8939''', '''sam_vit_l_0b3195'''] parser.add_argument( '''--model_name''', default='''sam_vit_h_4b8939''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model and processor to the hub after converting''', ) parser.add_argument( '''--model_hub_id''', default='''ybelkada/segment-anything''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) _lowercase = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
659
1
def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): def count_of_possible_combinations(snake_case__) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(target - item) for item in array) return count_of_possible_combinations(snake_case__) def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): def count_of_possible_combinations_with_dp_array( snake_case__ , snake_case__) -> int: if target < 0: return 0 if target == 0: return 1 if dp_array[target] != -1: return dp_array[target] lowerCAmelCase_ : Optional[Any] = sum( count_of_possible_combinations_with_dp_array(target - item , snake_case__) for item in array) lowerCAmelCase_ : Any = answer return answer lowerCAmelCase_ : List[Any] = [-1] * (target + 1) return count_of_possible_combinations_with_dp_array(snake_case__ , snake_case__) def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : Optional[Any] = [0] * (target + 1) lowerCAmelCase_ : Union[str, Any] = 1 for i in range(1 , target + 1): for j in range(snake_case__): if i - array[j] >= 0: dp_array[i] += dp_array[i - array[j]] return dp_array[target] if __name__ == "__main__": import doctest doctest.testmod() _lowercase = 3 _lowercase = 5 _lowercase = [1, 2, 5] print(combination_sum_iv(n, array, target))
659
class __snake_case : """simple docstring""" def __init__( self : Union[str, Any] ,lowerCAmelCase__ : str = "" ,lowerCAmelCase__ : bool = False ) -> None: '''simple docstring''' lowerCAmelCase_ : dict[str, RadixNode] = {} # A node will be a leaf if the tree contains its word lowerCAmelCase_ : Optional[int] = is_leaf lowerCAmelCase_ : List[str] = prefix def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : str ) -> tuple[str, str, str]: '''simple docstring''' lowerCAmelCase_ : List[str] = 0 for q, w in zip(self.prefix ,lowerCAmelCase__ ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : list[str] ) -> None: '''simple docstring''' for word in words: self.insert(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : str ) -> None: '''simple docstring''' if self.prefix == word: lowerCAmelCase_ : Optional[Any] = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: lowerCAmelCase_ : Optional[int] = RadixNode(prefix=lowerCAmelCase__ ,is_leaf=lowerCAmelCase__ ) else: lowerCAmelCase_ : Optional[Any] = self.nodes[word[0]] lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = incoming_node.match( lowerCAmelCase__ ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: lowerCAmelCase_ : Dict = remaining_prefix lowerCAmelCase_ : str = self.nodes[matching_string[0]] lowerCAmelCase_ : Dict = RadixNode(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Any = aux_node if remaining_word == "": lowerCAmelCase_ : Optional[Any] = True else: self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : str ) -> bool: '''simple docstring''' lowerCAmelCase_ : List[str] = self.nodes.get(word[0] ,lowerCAmelCase__ ) if not incoming_node: return False else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : str ) -> bool: '''simple docstring''' lowerCAmelCase_ : int = self.nodes.get(word[0] ,lowerCAmelCase__ ) if not incoming_node: return False else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(lowerCAmelCase__ ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: lowerCAmelCase_ : int = list(self.nodes.values() )[0] lowerCAmelCase_ : List[Any] = merging_node.is_leaf self.prefix += merging_node.prefix lowerCAmelCase_ : int = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: lowerCAmelCase_ : List[str] = False # If there is 1 edge, we merge it with its child else: lowerCAmelCase_ : Union[str, Any] = list(incoming_node.nodes.values() )[0] lowerCAmelCase_ : Optional[int] = merging_node.is_leaf incoming_node.prefix += merging_node.prefix lowerCAmelCase_ : List[str] = merging_node.nodes return True def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : int = 0 ) -> None: '''simple docstring''' if self.prefix != "": print("-" * height ,self.prefix ," (leaf)" if self.is_leaf else "" ) for value in self.nodes.values(): value.print_tree(height + 1 ) def UpperCamelCase ( ): lowerCAmelCase_ : List[Any] = "banana bananas bandana band apple all beast".split() lowerCAmelCase_ : Optional[Any] = RadixNode() root.insert_many(snake_case__) assert all(root.find(snake_case__) for word in words) assert not root.find("bandanas") assert not root.find("apps") root.delete("all") assert not root.find("all") root.delete("banana") assert not root.find("banana") assert root.find("bananas") return True def UpperCamelCase ( ): assert test_trie() def UpperCamelCase ( ): lowerCAmelCase_ : str = RadixNode() lowerCAmelCase_ : str = "banana bananas bandanas bandana band apple all beast".split() root.insert_many(snake_case__) print("Words:" , snake_case__) print("Tree:") root.print_tree() if __name__ == "__main__": main()
659
1
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer _lowercase = logging.get_logger(__name__) _lowercase = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } _lowercase = { '''vocab_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json''' }, '''merges_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt''' }, '''tokenizer_config_file''': { '''facebook/blenderbot_small-90M''': ( '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json''' ) }, } _lowercase = { '''facebook/blenderbot_small-90M''': 512, } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = BlenderbotSmallTokenizer def __init__( self : Optional[int] ,lowerCAmelCase__ : Optional[int]=None ,lowerCAmelCase__ : Union[str, Any]=None ,lowerCAmelCase__ : Any="<|endoftext|>" ,lowerCAmelCase__ : int="<|endoftext|>" ,lowerCAmelCase__ : Optional[Any]="<|endoftext|>" ,lowerCAmelCase__ : Union[str, Any]=False ,lowerCAmelCase__ : Optional[Any]=True ,**lowerCAmelCase__ : Union[str, Any] ,) -> str: '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=lowerCAmelCase__ ,merges=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ,) ,bos_token=lowerCAmelCase__ ,eos_token=lowerCAmelCase__ ,unk_token=lowerCAmelCase__ ,**lowerCAmelCase__ ,) lowerCAmelCase_ : Dict = add_prefix_space def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Tuple=None ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : str = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : Dict = [self.sep_token_id] lowerCAmelCase_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
659
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class __snake_case : """simple docstring""" def __init__( self : Tuple ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Optional[Any]=12 ,lowerCAmelCase__ : Union[str, Any]=7 ,lowerCAmelCase__ : Union[str, Any]=True ,lowerCAmelCase__ : List[str]=True ,lowerCAmelCase__ : Any=True ,lowerCAmelCase__ : Optional[Any]=99 ,lowerCAmelCase__ : List[str]=32 ,lowerCAmelCase__ : Dict=32 ,lowerCAmelCase__ : str=2 ,lowerCAmelCase__ : Optional[int]=4 ,lowerCAmelCase__ : str=37 ,lowerCAmelCase__ : Dict=0.1 ,lowerCAmelCase__ : List[str]=0.1 ,lowerCAmelCase__ : str=5_12 ,lowerCAmelCase__ : Union[str, Any]=0.02 ,lowerCAmelCase__ : Tuple=0 ,lowerCAmelCase__ : str=None ,) -> str: '''simple docstring''' lowerCAmelCase_ : int = parent lowerCAmelCase_ : str = batch_size lowerCAmelCase_ : int = seq_length lowerCAmelCase_ : Union[str, Any] = is_training lowerCAmelCase_ : int = use_input_mask lowerCAmelCase_ : List[Any] = use_labels lowerCAmelCase_ : Dict = vocab_size lowerCAmelCase_ : Union[str, Any] = hidden_size lowerCAmelCase_ : Union[str, Any] = projection_dim lowerCAmelCase_ : List[Any] = num_hidden_layers lowerCAmelCase_ : Any = num_attention_heads lowerCAmelCase_ : List[Any] = intermediate_size lowerCAmelCase_ : Any = dropout lowerCAmelCase_ : Optional[int] = attention_dropout lowerCAmelCase_ : int = max_position_embeddings lowerCAmelCase_ : Optional[int] = initializer_range lowerCAmelCase_ : Any = scope lowerCAmelCase_ : Tuple = bos_token_id def UpperCAmelCase_ ( self : str ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) lowerCAmelCase_ : Dict = None if self.use_input_mask: lowerCAmelCase_ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowerCAmelCase_ : List[Any] = input_mask.numpy() lowerCAmelCase_ , lowerCAmelCase_ : str = input_mask.shape lowerCAmelCase_ : Dict = np.random.randint(1 ,seq_length - 1 ,size=(batch_size,) ) for batch_idx, start_index in enumerate(lowerCAmelCase__ ): lowerCAmelCase_ : Union[str, Any] = 1 lowerCAmelCase_ : Optional[Any] = 0 lowerCAmelCase_ : List[Any] = self.get_config() return config, input_ids, tf.convert_to_tensor(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[str] ) -> str: '''simple docstring''' return BlipTextConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,projection_dim=self.projection_dim ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,dropout=self.dropout ,attention_dropout=self.attention_dropout ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,bos_token_id=self.bos_token_id ,) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Dict ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : List[Any] = TFBlipTextModel(config=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = model(lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,training=lowerCAmelCase__ ) lowerCAmelCase_ : str = model(lowerCAmelCase__ ,training=lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def UpperCAmelCase_ ( self : Optional[int] ) -> int: '''simple docstring''' lowerCAmelCase_ : List[str] = self.prepare_config_and_inputs() lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = config_and_inputs lowerCAmelCase_ : Tuple = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = (TFBlipTextModel,) if is_tf_available() else () UpperCamelCase_ = False UpperCamelCase_ = False UpperCamelCase_ = False def UpperCAmelCase_ ( self : Optional[Any] ) -> str: '''simple docstring''' lowerCAmelCase_ : List[str] = BlipTextModelTester(self ) lowerCAmelCase_ : Tuple = ConfigTester(self ,config_class=lowerCAmelCase__ ,hidden_size=37 ) def UpperCAmelCase_ ( self : str ) -> Any: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase_ ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' pass def UpperCAmelCase_ ( self : Union[str, Any] ) -> Any: '''simple docstring''' pass @unittest.skip(reason="Blip does not use inputs_embeds" ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCAmelCase_ ( self : int ) -> Optional[Any]: '''simple docstring''' pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCAmelCase_ ( self : Dict ) -> Union[str, Any]: '''simple docstring''' pass @slow def UpperCAmelCase_ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Tuple = TFBlipTextModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : str=True ) -> List[Any]: '''simple docstring''' super().test_pt_tf_model_equivalence(allow_missing_keys=lowerCAmelCase__ )
659
1
import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = LEDTokenizer UpperCamelCase_ = LEDTokenizerFast UpperCamelCase_ = True def UpperCAmelCase_ ( self : List[Any] ) -> Optional[int]: '''simple docstring''' super().setUp() lowerCAmelCase_ : Union[str, Any] = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] lowerCAmelCase_ : Tuple = dict(zip(lowerCAmelCase__ ,range(len(lowerCAmelCase__ ) ) ) ) lowerCAmelCase_ : int = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] lowerCAmelCase_ : Union[str, Any] = {"unk_token": "<unk>"} lowerCAmelCase_ : List[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase_ : Any = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file ,"w" ,encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase__ ) + "\n" ) with open(self.merges_file ,"w" ,encoding="utf-8" ) as fp: fp.write("\n".join(lowerCAmelCase__ ) ) def UpperCAmelCase_ ( self : List[Any] ,**lowerCAmelCase__ : int ) -> Tuple: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ,**lowerCAmelCase__ : Optional[int] ) -> List[Any]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCAmelCase__ ) def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : int ) -> List[str]: '''simple docstring''' return "lower newer", "lower newer" @cached_property def UpperCAmelCase_ ( self : int ) -> Union[str, Any]: '''simple docstring''' return LEDTokenizer.from_pretrained("allenai/led-base-16384" ) @cached_property def UpperCAmelCase_ ( self : List[str] ) -> Dict: '''simple docstring''' return LEDTokenizerFast.from_pretrained("allenai/led-base-16384" ) @require_torch def UpperCAmelCase_ ( self : int ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = ["A long paragraph for summarization.", "Another paragraph for summarization."] lowerCAmelCase_ : int = [0, 2_50, 2_51, 1_78_18, 13, 3_91_86, 19_38, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Any = tokenizer(lowerCAmelCase__ ,max_length=len(lowerCAmelCase__ ) ,padding=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual((2, 9) ,batch.input_ids.shape ) self.assertEqual((2, 9) ,batch.attention_mask.shape ) lowerCAmelCase_ : int = batch.input_ids.tolist()[0] self.assertListEqual(lowerCAmelCase__ ,lowerCAmelCase__ ) @require_torch def UpperCAmelCase_ ( self : Dict ) -> Any: '''simple docstring''' lowerCAmelCase_ : int = ["A long paragraph for summarization.", "Another paragraph for summarization."] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[Any] = tokenizer(lowerCAmelCase__ ,padding=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIn("input_ids" ,lowerCAmelCase__ ) self.assertIn("attention_mask" ,lowerCAmelCase__ ) self.assertNotIn("labels" ,lowerCAmelCase__ ) self.assertNotIn("decoder_attention_mask" ,lowerCAmelCase__ ) @require_torch def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : int = [ "Summary of the text.", "Another summary.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[int] = tokenizer(text_target=lowerCAmelCase__ ,max_length=32 ,padding="max_length" ,return_tensors="pt" ) self.assertEqual(32 ,targets["input_ids"].shape[1] ) @require_torch def UpperCAmelCase_ ( self : Tuple ) -> List[str]: '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Tuple = tokenizer( ["I am a small frog" * 10_24, "I am a small frog"] ,padding=lowerCAmelCase__ ,truncation=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual(batch.input_ids.shape ,(2, 51_22) ) @require_torch def UpperCAmelCase_ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Tuple = ["A long paragraph for summarization."] lowerCAmelCase_ : Dict = [ "Summary of the text.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[Any] = tokenizer(lowerCAmelCase__ ,return_tensors="pt" ) lowerCAmelCase_ : Optional[Any] = tokenizer(text_target=lowerCAmelCase__ ,return_tensors="pt" ) lowerCAmelCase_ : List[str] = inputs["input_ids"] lowerCAmelCase_ : Any = targets["input_ids"] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) @require_torch def UpperCAmelCase_ ( self : str ) -> Tuple: '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : str = ["Summary of the text.", "Another summary."] lowerCAmelCase_ : str = [[0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, -1, -1]] lowerCAmelCase_ : List[Any] = tokenizer(lowerCAmelCase__ ,padding=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = [[0] * len(lowerCAmelCase__ ) for x in encoded_output["input_ids"]] lowerCAmelCase_ : Optional[int] = tokenizer.pad(lowerCAmelCase__ ) self.assertSequenceEqual(outputs["global_attention_mask"] ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' pass def UpperCAmelCase_ ( self : str ) -> Union[str, Any]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCAmelCase_ : Dict = self.rust_tokenizer_class.from_pretrained(lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = self.tokenizer_class.from_pretrained(lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Dict = "A, <mask> AllenNLP sentence." lowerCAmelCase_ : Tuple = tokenizer_r.encode_plus(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,return_token_type_ids=lowerCAmelCase__ ) lowerCAmelCase_ : int = tokenizer_p.encode_plus(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,return_token_type_ids=lowerCAmelCase__ ) self.assertEqual(sum(tokens_r["token_type_ids"] ) ,sum(tokens_p["token_type_ids"] ) ) self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) ,sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) ,) lowerCAmelCase_ : Any = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) lowerCAmelCase_ : Union[str, Any] = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) self.assertSequenceEqual(tokens_p["input_ids"] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( lowerCAmelCase__ ,["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( lowerCAmelCase__ ,["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] )
659
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _lowercase = logging.get_logger(__name__) _lowercase = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} # See all LED models at https://huggingface.co/models?filter=LED _lowercase = { '''vocab_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json''', }, '''merges_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json''', }, } _lowercase = { '''allenai/led-base-16384''': 16384, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def UpperCamelCase ( ): lowerCAmelCase_ : Optional[int] = ( list(range(ord("!") , ord("~") + 1)) + list(range(ord("¡") , ord("¬") + 1)) + list(range(ord("®") , ord("ÿ") + 1)) ) lowerCAmelCase_ : List[Any] = bs[:] lowerCAmelCase_ : Optional[int] = 0 for b in range(2**8): if b not in bs: bs.append(snake_case__) cs.append(2**8 + n) n += 1 lowerCAmelCase_ : Tuple = [chr(snake_case__) for n in cs] return dict(zip(snake_case__ , snake_case__)) def UpperCamelCase ( snake_case__): lowerCAmelCase_ : str = set() lowerCAmelCase_ : List[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char)) lowerCAmelCase_ : Union[str, Any] = char return pairs class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = ['input_ids', 'attention_mask'] def __init__( self : int ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Tuple="replace" ,lowerCAmelCase__ : Optional[int]="<s>" ,lowerCAmelCase__ : Optional[int]="</s>" ,lowerCAmelCase__ : Tuple="</s>" ,lowerCAmelCase__ : int="<s>" ,lowerCAmelCase__ : Union[str, Any]="<unk>" ,lowerCAmelCase__ : str="<pad>" ,lowerCAmelCase__ : Tuple="<mask>" ,lowerCAmelCase__ : Optional[int]=False ,**lowerCAmelCase__ : Tuple ,) -> Any: '''simple docstring''' lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else bos_token lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else eos_token lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else sep_token lowerCAmelCase_ : Any = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else cls_token lowerCAmelCase_ : Tuple = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else unk_token lowerCAmelCase_ : Any = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase_ : Optional[int] = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else mask_token super().__init__( errors=lowerCAmelCase__ ,bos_token=lowerCAmelCase__ ,eos_token=lowerCAmelCase__ ,unk_token=lowerCAmelCase__ ,sep_token=lowerCAmelCase__ ,cls_token=lowerCAmelCase__ ,pad_token=lowerCAmelCase__ ,mask_token=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,**lowerCAmelCase__ ,) with open(lowerCAmelCase__ ,encoding="utf-8" ) as vocab_handle: lowerCAmelCase_ : List[str] = json.load(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = {v: k for k, v in self.encoder.items()} lowerCAmelCase_ : Optional[int] = errors # how to handle errors in decoding lowerCAmelCase_ : Optional[int] = bytes_to_unicode() lowerCAmelCase_ : str = {v: k for k, v in self.byte_encoder.items()} with open(lowerCAmelCase__ ,encoding="utf-8" ) as merges_handle: lowerCAmelCase_ : List[str] = merges_handle.read().split("\n" )[1:-1] lowerCAmelCase_ : List[Any] = [tuple(merge.split() ) for merge in bpe_merges] lowerCAmelCase_ : Union[str, Any] = dict(zip(lowerCAmelCase__ ,range(len(lowerCAmelCase__ ) ) ) ) lowerCAmelCase_ : Dict = {} lowerCAmelCase_ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowerCAmelCase_ : Any = re.compile(R"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def UpperCAmelCase_ ( self : Dict ) -> Dict: '''simple docstring''' return len(self.encoder ) def UpperCAmelCase_ ( self : Dict ) -> str: '''simple docstring''' return dict(self.encoder ,**self.added_tokens_encoder ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Dict ) -> Dict: '''simple docstring''' if token in self.cache: return self.cache[token] lowerCAmelCase_ : Union[str, Any] = tuple(lowerCAmelCase__ ) lowerCAmelCase_ : str = get_pairs(lowerCAmelCase__ ) if not pairs: return token while True: lowerCAmelCase_ : Optional[int] = min(lowerCAmelCase__ ,key=lambda lowerCAmelCase__ : self.bpe_ranks.get(lowerCAmelCase__ ,float("inf" ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = bigram lowerCAmelCase_ : Tuple = [] lowerCAmelCase_ : str = 0 while i < len(lowerCAmelCase__ ): try: lowerCAmelCase_ : Union[str, Any] = word.index(lowerCAmelCase__ ,lowerCAmelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase_ : List[str] = j if word[i] == first and i < len(lowerCAmelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase_ : Optional[int] = tuple(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = new_word if len(lowerCAmelCase__ ) == 1: break else: lowerCAmelCase_ : Dict = get_pairs(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = " ".join(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = word return word def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : Dict ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : Any = [] for token in re.findall(self.pat ,lowerCAmelCase__ ): lowerCAmelCase_ : Optional[int] = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCAmelCase__ ).split(" " ) ) return bpe_tokens def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any] ) -> Tuple: '''simple docstring''' return self.encoder.get(lowerCAmelCase__ ,self.encoder.get(self.unk_token ) ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Union[str, Any] ) -> Optional[int]: '''simple docstring''' return self.decoder.get(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[Any] ) -> Any: '''simple docstring''' lowerCAmelCase_ : int = "".join(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" ,errors=self.errors ) return text def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(lowerCAmelCase__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCAmelCase_ : Optional[int] = os.path.join( lowerCAmelCase__ ,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase_ : List[str] = os.path.join( lowerCAmelCase__ ,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(lowerCAmelCase__ ,"w" ,encoding="utf-8" ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=lowerCAmelCase__ ,ensure_ascii=lowerCAmelCase__ ) + "\n" ) lowerCAmelCase_ : Dict = 0 with open(lowerCAmelCase__ ,"w" ,encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda lowerCAmelCase__ : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' " Please check that the tokenizer is not corrupted!" ) lowerCAmelCase_ : List[Any] = token_index writer.write(" ".join(lowerCAmelCase__ ) + "\n" ) index += 1 return vocab_file, merge_file def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase_ : Union[str, Any] = [self.cls_token_id] lowerCAmelCase_ : str = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ,lowerCAmelCase__ : bool = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ ,token_ids_a=lowerCAmelCase__ ,already_has_special_tokens=lowerCAmelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase__ )) + [1] return [1] + ([0] * len(lowerCAmelCase__ )) + [1, 1] + ([0] * len(lowerCAmelCase__ )) + [1] def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = [self.sep_token_id] lowerCAmelCase_ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : Optional[int]=False ,**lowerCAmelCase__ : str ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = kwargs.pop("add_prefix_space" ,self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCAmelCase__ ) > 0 and not text[0].isspace()): lowerCAmelCase_ : List[str] = " " + text return (text, kwargs) def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : Union[Dict[str, EncodedInput], BatchEncoding] ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : Optional[bool] = None ,) -> dict: '''simple docstring''' lowerCAmelCase_ : int = super()._pad( encoded_inputs=lowerCAmelCase__ ,max_length=lowerCAmelCase__ ,padding_strategy=lowerCAmelCase__ ,pad_to_multiple_of=lowerCAmelCase__ ,return_attention_mask=lowerCAmelCase__ ,) # Load from model defaults if return_attention_mask is None: lowerCAmelCase_ : List[Any] = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowerCAmelCase_ : Dict = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowerCAmelCase_ : List[Any] = len(encoded_inputs["global_attention_mask"] ) != len(lowerCAmelCase__ ) if needs_to_be_padded: lowerCAmelCase_ : Union[str, Any] = len(lowerCAmelCase__ ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowerCAmelCase_ : Optional[int] = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": lowerCAmelCase_ : List[Any] = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
659
1
from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo _lowercase = '''\ @misc{wu2016googles, title={Google\'s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation}, author={Yonghui Wu and Mike Schuster and Zhifeng Chen and Quoc V. Le and Mohammad Norouzi and Wolfgang Macherey and Maxim Krikun and Yuan Cao and Qin Gao and Klaus Macherey and Jeff Klingner and Apurva Shah and Melvin Johnson and Xiaobing Liu and Łukasz Kaiser and Stephan Gouws and Yoshikiyo Kato and Taku Kudo and Hideto Kazawa and Keith Stevens and George Kurian and Nishant Patil and Wei Wang and Cliff Young and Jason Smith and Jason Riesa and Alex Rudnick and Oriol Vinyals and Greg Corrado and Macduff Hughes and Jeffrey Dean}, year={2016}, eprint={1609.08144}, archivePrefix={arXiv}, primaryClass={cs.CL} } ''' _lowercase = '''\ The BLEU score has some undesirable properties when used for single sentences, as it was designed to be a corpus measure. We therefore use a slightly different score for our RL experiments which we call the \'GLEU score\'. For the GLEU score, we record all sub-sequences of 1, 2, 3 or 4 tokens in output and target sequence (n-grams). We then compute a recall, which is the ratio of the number of matching n-grams to the number of total n-grams in the target (ground truth) sequence, and a precision, which is the ratio of the number of matching n-grams to the number of total n-grams in the generated output sequence. Then GLEU score is simply the minimum of recall and precision. This GLEU score\'s range is always between 0 (no matches) and 1 (all match) and it is symmetrical when switching output and target. According to our experiments, GLEU score correlates quite well with the BLEU metric on a corpus level but does not have its drawbacks for our per sentence reward objective. ''' _lowercase = '''\ Computes corpus-level Google BLEU (GLEU) score of translated segments against one or more references. Instead of averaging the sentence level GLEU scores (i.e. macro-average precision), Wu et al. (2016) sum up the matching tokens and the max of hypothesis and reference tokens for each sentence, then compute using the aggregate values. Args: predictions (list of str): list of translations to score. Each translation should be tokenized into a list of tokens. references (list of list of str): list of lists of references for each translation. Each reference should be tokenized into a list of tokens. min_len (int): The minimum order of n-gram this function should extract. Defaults to 1. max_len (int): The maximum order of n-gram this function should extract. Defaults to 4. Returns: \'google_bleu\': google_bleu score Examples: Example 1: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results["google_bleu"], 2)) 0.44 Example 2: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\', ... \'heed\', \'the\', \'cat\', \'commands\'] >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\', ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\', ... \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results["google_bleu"], 2)) 0.61 Example 3: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\', ... \'heed\', \'the\', \'cat\', \'commands\'] >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\', ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\', ... \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references, min_len=2) >>> print(round(results["google_bleu"], 2)) 0.53 Example 4: >>> hyp1 = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'which\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'always\', ... \'disobeys\', \'the\', \'commands\', \'of\', \'the\', \'cat\'] >>> ref1a = [\'It\', \'is\', \'the\', \'guiding\', \'principle\', \'which\', ... \'guarantees\', \'the\', \'rubber\', \'duck\', \'forces\', \'never\', ... \'being\', \'under\', \'the\', \'command\', \'of\', \'the\', \'cat\'] >>> ref1b = [\'It\', \'is\', \'a\', \'guide\', \'to\', \'action\', \'that\', ... \'ensures\', \'that\', \'the\', \'rubber\', \'duck\', \'will\', \'never\', ... \'heed\', \'the\', \'cat\', \'commands\'] >>> ref1c = [\'It\', \'is\', \'the\', \'practical\', \'guide\', \'for\', \'the\', ... \'rubber\', \'duck\', \'army\', \'never\', \'to\', \'heed\', \'the\', \'directions\', ... \'of\', \'the\', \'cat\'] >>> hyp2 = [\'he\', \'read\', \'the\', \'book\', \'because\', \'he\', \'was\', ... \'interested\', \'in\', \'world\', \'history\'] >>> ref2a = [\'he\', \'was\', \'interested\', \'in\', \'world\', \'history\', ... \'because\', \'he\', \'read\', \'the\', \'book\'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric("google_bleu") >>> results = google_bleu.compute(predictions=hypotheses,references=list_of_references, min_len=2, max_len=6) >>> print(round(results["google_bleu"], 2)) 0.4 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __snake_case ( datasets.Metric ): """simple docstring""" def UpperCAmelCase_ ( self : List[str] ) -> MetricInfo: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { "predictions": datasets.Sequence(datasets.Value("string" ,id="token" ) ,id="sequence" ), "references": datasets.Sequence( datasets.Sequence(datasets.Value("string" ,id="token" ) ,id="sequence" ) ,id="references" ), } ) ,) def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : List[List[List[str]]] ,lowerCAmelCase__ : List[List[str]] ,lowerCAmelCase__ : int = 1 ,lowerCAmelCase__ : int = 4 ,) -> Dict[str, float]: '''simple docstring''' return { "google_bleu": gleu_score.corpus_gleu( list_of_references=lowerCAmelCase__ ,hypotheses=lowerCAmelCase__ ,min_len=lowerCAmelCase__ ,max_len=lowerCAmelCase__ ) }
659
import os _lowercase = {'''I''': 1, '''V''': 5, '''X''': 10, '''L''': 50, '''C''': 100, '''D''': 500, '''M''': 1000} def UpperCamelCase ( snake_case__): lowerCAmelCase_ : List[str] = 0 lowerCAmelCase_ : Any = 0 while index < len(snake_case__) - 1: lowerCAmelCase_ : Optional[Any] = SYMBOLS[numerals[index]] lowerCAmelCase_ : int = SYMBOLS[numerals[index + 1]] if current_value < next_value: total_value -= current_value else: total_value += current_value index += 1 total_value += SYMBOLS[numerals[index]] return total_value def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Optional[int] = "" lowerCAmelCase_ : Tuple = num // 10_00 numerals += m_count * "M" num %= 10_00 lowerCAmelCase_ : int = num // 1_00 if c_count == 9: numerals += "CM" c_count -= 9 elif c_count == 4: numerals += "CD" c_count -= 4 if c_count >= 5: numerals += "D" c_count -= 5 numerals += c_count * "C" num %= 1_00 lowerCAmelCase_ : int = num // 10 if x_count == 9: numerals += "XC" x_count -= 9 elif x_count == 4: numerals += "XL" x_count -= 4 if x_count >= 5: numerals += "L" x_count -= 5 numerals += x_count * "X" num %= 10 if num == 9: numerals += "IX" num -= 9 elif num == 4: numerals += "IV" num -= 4 if num >= 5: numerals += "V" num -= 5 numerals += num * "I" return numerals def UpperCamelCase ( snake_case__ = "/p089_roman.txt"): lowerCAmelCase_ : int = 0 with open(os.path.dirname(snake_case__) + roman_numerals_filename) as filea: lowerCAmelCase_ : List[Any] = filea.readlines() for line in lines: lowerCAmelCase_ : Any = line.strip() lowerCAmelCase_ : Tuple = parse_roman_numerals(snake_case__) lowerCAmelCase_ : List[Any] = generate_roman_numerals(snake_case__) savings += len(snake_case__) - len(snake_case__) return savings if __name__ == "__main__": print(f"{solution() = }")
659
1
def UpperCamelCase ( snake_case__ = 1 , snake_case__ = 10_00): lowerCAmelCase_ : str = 1 lowerCAmelCase_ : Union[str, Any] = 0 for divide_by_number in range(snake_case__ , digit + 1): lowerCAmelCase_ : list[int] = [] lowerCAmelCase_ : Union[str, Any] = numerator for _ in range(1 , digit + 1): if now_divide in has_been_divided: if longest_list_length < len(snake_case__): lowerCAmelCase_ : List[str] = len(snake_case__) lowerCAmelCase_ : Any = divide_by_number else: has_been_divided.append(snake_case__) lowerCAmelCase_ : List[str] = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
659
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def UpperCamelCase ( ): lowerCAmelCase_ : Dict = HfArgumentParser(snake_case__) lowerCAmelCase_ : Dict = parser.parse_args_into_dataclasses()[0] lowerCAmelCase_ : List[Any] = TensorFlowBenchmark(args=snake_case__) try: lowerCAmelCase_ : str = parser.parse_args_into_dataclasses()[0] except ValueError as e: lowerCAmelCase_ : Optional[Any] = "Arg --no_{0} is no longer used, please use --no-{0} instead." lowerCAmelCase_ : Tuple = " ".join(str(snake_case__).split(" ")[:-1]) lowerCAmelCase_ : List[Any] = "" lowerCAmelCase_ : Optional[Any] = eval(str(snake_case__).split(" ")[-1]) lowerCAmelCase_ : List[Any] = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:]) else: wrong_args.append(snake_case__) if len(snake_case__) > 0: lowerCAmelCase_ : int = full_error_msg + begin_error_msg + str(snake_case__) raise ValueError(snake_case__) benchmark.run() if __name__ == "__main__": main()
659
1
from torch import nn def UpperCamelCase ( snake_case__): if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(F'''Unsupported activation function: {act_fn}''')
659
from collections import defaultdict from pathlib import Path import pandas as pd from rouge_cli import calculate_rouge_path from utils import calculate_rouge _lowercase = [ '''Prosecutor: "No videos were used in the crash investigation" German papers say they saw a cell phone video of the''' ''' final seconds on board Flight 9525. The Germanwings co-pilot says he had a "previous episode of severe''' ''' depression\" German airline confirms it knew of Andreas Lubitz\'s depression years before he took control.''', '''The Palestinian Authority officially becomes the 123rd member of the International Criminal Court. The formal''' ''' accession was marked with a ceremony at The Hague, in the Netherlands. The Palestinians signed the ICC\'s''' ''' founding Rome Statute in January. Israel and the United States opposed the Palestinians\' efforts to join the''' ''' body.''', '''Amnesty International releases its annual report on the death penalty. The report catalogs the use of''' ''' state-sanctioned killing as a punitive measure across the globe. At least 607 people were executed around the''' ''' world in 2014, compared to 778 in 2013. The U.S. remains one of the worst offenders for imposing capital''' ''' punishment.''', ] _lowercase = [ '''Marseille prosecutor says "so far no videos were used in the crash investigation" despite media reports .''' ''' Journalists at Bild and Paris Match are "very confident" the video clip is real, an editor says . Andreas Lubitz''' ''' had informed his Lufthansa training school of an episode of severe depression, airline says .''', '''Membership gives the ICC jurisdiction over alleged crimes committed in Palestinian territories since last June .''' ''' Israel and the United States opposed the move, which could open the door to war crimes investigations against''' ''' Israelis .''', '''Amnesty\'s annual death penalty report catalogs encouraging signs, but setbacks in numbers of those sentenced to''' ''' death . Organization claims that governments around the world are using the threat of terrorism to advance''' ''' executions . The number of executions worldwide has gone down by almost 22% compared with 2013, but death''' ''' sentences up by 28% .''', ] def UpperCamelCase ( ): lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , bootstrap_aggregation=snake_case__ , rouge_keys=["rouge2", "rougeL"]) assert isinstance(snake_case__ , snake_case__) lowerCAmelCase_ : str = calculate_rouge(snake_case__ , snake_case__ , bootstrap_aggregation=snake_case__ , rouge_keys=["rouge2"]) assert ( pd.DataFrame(no_aggregation["rouge2"]).fmeasure.mean() == pd.DataFrame(no_aggregation_just_ra["rouge2"]).fmeasure.mean() ) def UpperCamelCase ( ): lowerCAmelCase_ : str = "rougeLsum" lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=[k])[k] lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=[k])[k] assert score > score_no_sep def UpperCamelCase ( ): lowerCAmelCase_ : int = ["rouge1", "rouge2", "rougeL"] lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=snake_case__) lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=snake_case__) assert score_sep == score_no_sep def UpperCamelCase ( ): lowerCAmelCase_ : List[str] = [ "Her older sister, Margot Frank, died in 1945, a month earlier than previously thought.", "Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports .", ] lowerCAmelCase_ : Dict = [ "Margot Frank, died in 1945, a month earlier than previously thought.", "Prosecutor: \"No videos were used in the crash investigation\" German papers say they saw a cell phone video of" " the final seconds on board Flight 9525.", ] assert calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__) == calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__) def UpperCamelCase ( ): lowerCAmelCase_ : Optional[int] = [ "\" \"a person who has such a video needs to immediately give it to the investigators,\" prosecutor says .<n> \"it is a very disturbing scene,\" editor-in-chief of bild online tells \"erin burnett: outfront\" " ] lowerCAmelCase_ : Any = [ " Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports . Journalists at Bild and Paris Match are \"very confident\" the video clip is real, an editor says . Andreas Lubitz had informed his Lufthansa training school of an episode of severe depression, airline says ." ] lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , rouge_keys=["rougeLsum"] , newline_sep=snake_case__)["rougeLsum"] lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , rouge_keys=["rougeLsum"])["rougeLsum"] assert new_score > prev_score def UpperCamelCase ( ): lowerCAmelCase_ : int = Path("examples/seq2seq/test_data/wmt_en_ro") lowerCAmelCase_ : Dict = calculate_rouge_path(data_dir.joinpath("test.source") , data_dir.joinpath("test.target")) assert isinstance(snake_case__ , snake_case__) lowerCAmelCase_ : Any = calculate_rouge_path( data_dir.joinpath("test.source") , data_dir.joinpath("test.target") , bootstrap_aggregation=snake_case__) assert isinstance(snake_case__ , snake_case__)
659
1
from math import factorial def UpperCamelCase ( snake_case__ , snake_case__): # If either of the conditions are true, the function is being asked # to calculate a factorial of a negative number, which is not possible if n < k or k < 0: raise ValueError("Please enter positive integers for n and k where n >= k") return factorial(snake_case__) // (factorial(snake_case__) * factorial(n - k)) if __name__ == "__main__": print( '''The number of five-card hands possible from a standard''', f"fifty-two card deck is: {combinations(52, 5)}\n", ) print( '''If a class of 40 students must be arranged into groups of''', f"4 for group projects, there are {combinations(40, 4)} ways", '''to arrange them.\n''', ) print( '''If 10 teams are competing in a Formula One race, there''', f"are {combinations(10, 3)} ways that first, second and", '''third place can be awarded.''', )
659
import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = LEDTokenizer UpperCamelCase_ = LEDTokenizerFast UpperCamelCase_ = True def UpperCAmelCase_ ( self : List[Any] ) -> Optional[int]: '''simple docstring''' super().setUp() lowerCAmelCase_ : Union[str, Any] = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] lowerCAmelCase_ : Tuple = dict(zip(lowerCAmelCase__ ,range(len(lowerCAmelCase__ ) ) ) ) lowerCAmelCase_ : int = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] lowerCAmelCase_ : Union[str, Any] = {"unk_token": "<unk>"} lowerCAmelCase_ : List[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase_ : Any = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file ,"w" ,encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase__ ) + "\n" ) with open(self.merges_file ,"w" ,encoding="utf-8" ) as fp: fp.write("\n".join(lowerCAmelCase__ ) ) def UpperCAmelCase_ ( self : List[Any] ,**lowerCAmelCase__ : int ) -> Tuple: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ,**lowerCAmelCase__ : Optional[int] ) -> List[Any]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCAmelCase__ ) def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : int ) -> List[str]: '''simple docstring''' return "lower newer", "lower newer" @cached_property def UpperCAmelCase_ ( self : int ) -> Union[str, Any]: '''simple docstring''' return LEDTokenizer.from_pretrained("allenai/led-base-16384" ) @cached_property def UpperCAmelCase_ ( self : List[str] ) -> Dict: '''simple docstring''' return LEDTokenizerFast.from_pretrained("allenai/led-base-16384" ) @require_torch def UpperCAmelCase_ ( self : int ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = ["A long paragraph for summarization.", "Another paragraph for summarization."] lowerCAmelCase_ : int = [0, 2_50, 2_51, 1_78_18, 13, 3_91_86, 19_38, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Any = tokenizer(lowerCAmelCase__ ,max_length=len(lowerCAmelCase__ ) ,padding=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual((2, 9) ,batch.input_ids.shape ) self.assertEqual((2, 9) ,batch.attention_mask.shape ) lowerCAmelCase_ : int = batch.input_ids.tolist()[0] self.assertListEqual(lowerCAmelCase__ ,lowerCAmelCase__ ) @require_torch def UpperCAmelCase_ ( self : Dict ) -> Any: '''simple docstring''' lowerCAmelCase_ : int = ["A long paragraph for summarization.", "Another paragraph for summarization."] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[Any] = tokenizer(lowerCAmelCase__ ,padding=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIn("input_ids" ,lowerCAmelCase__ ) self.assertIn("attention_mask" ,lowerCAmelCase__ ) self.assertNotIn("labels" ,lowerCAmelCase__ ) self.assertNotIn("decoder_attention_mask" ,lowerCAmelCase__ ) @require_torch def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : int = [ "Summary of the text.", "Another summary.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[int] = tokenizer(text_target=lowerCAmelCase__ ,max_length=32 ,padding="max_length" ,return_tensors="pt" ) self.assertEqual(32 ,targets["input_ids"].shape[1] ) @require_torch def UpperCAmelCase_ ( self : Tuple ) -> List[str]: '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Tuple = tokenizer( ["I am a small frog" * 10_24, "I am a small frog"] ,padding=lowerCAmelCase__ ,truncation=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual(batch.input_ids.shape ,(2, 51_22) ) @require_torch def UpperCAmelCase_ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Tuple = ["A long paragraph for summarization."] lowerCAmelCase_ : Dict = [ "Summary of the text.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[Any] = tokenizer(lowerCAmelCase__ ,return_tensors="pt" ) lowerCAmelCase_ : Optional[Any] = tokenizer(text_target=lowerCAmelCase__ ,return_tensors="pt" ) lowerCAmelCase_ : List[str] = inputs["input_ids"] lowerCAmelCase_ : Any = targets["input_ids"] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) @require_torch def UpperCAmelCase_ ( self : str ) -> Tuple: '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : str = ["Summary of the text.", "Another summary."] lowerCAmelCase_ : str = [[0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, -1, -1]] lowerCAmelCase_ : List[Any] = tokenizer(lowerCAmelCase__ ,padding=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = [[0] * len(lowerCAmelCase__ ) for x in encoded_output["input_ids"]] lowerCAmelCase_ : Optional[int] = tokenizer.pad(lowerCAmelCase__ ) self.assertSequenceEqual(outputs["global_attention_mask"] ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' pass def UpperCAmelCase_ ( self : str ) -> Union[str, Any]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCAmelCase_ : Dict = self.rust_tokenizer_class.from_pretrained(lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = self.tokenizer_class.from_pretrained(lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Dict = "A, <mask> AllenNLP sentence." lowerCAmelCase_ : Tuple = tokenizer_r.encode_plus(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,return_token_type_ids=lowerCAmelCase__ ) lowerCAmelCase_ : int = tokenizer_p.encode_plus(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,return_token_type_ids=lowerCAmelCase__ ) self.assertEqual(sum(tokens_r["token_type_ids"] ) ,sum(tokens_p["token_type_ids"] ) ) self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) ,sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) ,) lowerCAmelCase_ : Any = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) lowerCAmelCase_ : Union[str, Any] = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) self.assertSequenceEqual(tokens_p["input_ids"] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( lowerCAmelCase__ ,["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( lowerCAmelCase__ ,["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] )
659
1
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = LongformerTokenizer UpperCamelCase_ = True UpperCamelCase_ = LongformerTokenizerFast UpperCamelCase_ = True def UpperCAmelCase_ ( self : Optional[int] ) -> Union[str, Any]: '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowerCAmelCase_ : Dict = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] lowerCAmelCase_ : int = dict(zip(lowerCAmelCase__ ,range(len(lowerCAmelCase__ ) ) ) ) lowerCAmelCase_ : Union[str, Any] = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] lowerCAmelCase_ : Tuple = {"unk_token": "<unk>"} lowerCAmelCase_ : Optional[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase_ : Optional[int] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file ,"w" ,encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase__ ) + "\n" ) with open(self.merges_file ,"w" ,encoding="utf-8" ) as fp: fp.write("\n".join(lowerCAmelCase__ ) ) def UpperCAmelCase_ ( self : int ,**lowerCAmelCase__ : str ) -> str: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[str] ,**lowerCAmelCase__ : int ) -> Any: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Dict ,lowerCAmelCase__ : Tuple ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : str = "lower newer" lowerCAmelCase_ : List[str] = "lower newer" return input_text, output_text def UpperCAmelCase_ ( self : List[str] ) -> int: '''simple docstring''' lowerCAmelCase_ : Dict = self.tokenizer_class(self.vocab_file ,self.merges_file ,**self.special_tokens_map ) lowerCAmelCase_ : int = "lower newer" lowerCAmelCase_ : Optional[Any] = ["l", "o", "w", "er", "\u0120", "n", "e", "w", "er"] lowerCAmelCase_ : Any = tokenizer.tokenize(lowerCAmelCase__ ) # , add_prefix_space=True) self.assertListEqual(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Dict = tokens + [tokenizer.unk_token] lowerCAmelCase_ : str = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' lowerCAmelCase_ : List[Any] = self.get_tokenizer() self.assertListEqual(tokenizer.encode("Hello world!" ,add_special_tokens=lowerCAmelCase__ ) ,[0, 3_14_14, 2_32, 3_28, 2] ) self.assertListEqual( tokenizer.encode("Hello world! cécé herlolip 418" ,add_special_tokens=lowerCAmelCase__ ) ,[0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2] ,) @slow def UpperCAmelCase_ ( self : str ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : Any = self.tokenizer_class.from_pretrained("allenai/longformer-base-4096" ) lowerCAmelCase_ : Union[str, Any] = tokenizer.encode("sequence builders" ,add_special_tokens=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = tokenizer.encode("multi-sequence build" ,add_special_tokens=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = tokenizer.encode( "sequence builders" ,add_special_tokens=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = tokenizer.encode( "sequence builders" ,"multi-sequence build" ,add_special_tokens=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = tokenizer.build_inputs_with_special_tokens(lowerCAmelCase__ ,lowerCAmelCase__ ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode def UpperCAmelCase_ ( self : Dict ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : str = self.get_tokenizer() lowerCAmelCase_ : int = "Encode this sequence." lowerCAmelCase_ : str = tokenizer.byte_encoder[" ".encode("utf-8" )[0]] # Testing encoder arguments lowerCAmelCase_ : Any = tokenizer.encode(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ) lowerCAmelCase_ : str = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertNotEqual(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : int = tokenizer.encode(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = tokenizer.convert_ids_to_tokens(encoded[0] )[0] self.assertEqual(lowerCAmelCase__ ,lowerCAmelCase__ ) tokenizer.add_special_tokens({"bos_token": "<s>"} ) lowerCAmelCase_ : str = tokenizer.encode(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = tokenizer.convert_ids_to_tokens(encoded[1] )[0] self.assertNotEqual(lowerCAmelCase__ ,lowerCAmelCase__ ) # Testing spaces after special tokens lowerCAmelCase_ : Union[str, Any] = "<mask>" tokenizer.add_special_tokens( {"mask_token": AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ )} ) # mask token has a left space lowerCAmelCase_ : Union[str, Any] = tokenizer.convert_tokens_to_ids(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = "Encode <mask> sequence" lowerCAmelCase_ : str = "Encode <mask>sequence" lowerCAmelCase_ : Any = tokenizer.encode(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = encoded.index(lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertEqual(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Dict = tokenizer.encode(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = encoded.index(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = tokenizer.convert_ids_to_tokens(encoded[mask_loc + 1] )[0] self.assertNotEqual(lowerCAmelCase__ ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : str ) -> List[Any]: '''simple docstring''' pass def UpperCAmelCase_ ( self : Tuple ) -> Optional[int]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCAmelCase_ : Dict = self.rust_tokenizer_class.from_pretrained(lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = self.tokenizer_class.from_pretrained(lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Dict = "A, <mask> AllenNLP sentence." lowerCAmelCase_ : List[str] = tokenizer_r.encode_plus(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,return_token_type_ids=lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = tokenizer_p.encode_plus(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,return_token_type_ids=lowerCAmelCase__ ) # token_type_ids should put 0 everywhere self.assertEqual(sum(tokens_r["token_type_ids"] ) ,sum(tokens_p["token_type_ids"] ) ) # attention_mask should put 1 everywhere, so sum over length should be 1 self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) ,sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) ,) lowerCAmelCase_ : str = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) lowerCAmelCase_ : Optional[int] = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) # Rust correctly handles the space before the mask while python doesnt self.assertSequenceEqual(tokens_p["input_ids"] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( lowerCAmelCase__ ,["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( lowerCAmelCase__ ,["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) def UpperCAmelCase_ ( self : int ) -> List[Any]: '''simple docstring''' for trim_offsets, add_prefix_space in itertools.product([True, False] ,repeat=2 ): lowerCAmelCase_ : List[str] = self.rust_tokenizer_class.from_pretrained( self.tmpdirname ,use_fast=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = json.loads(tokenizer_r.backend_tokenizer.pre_tokenizer.__getstate__() ) lowerCAmelCase_ : int = json.loads(tokenizer_r.backend_tokenizer.post_processor.__getstate__() ) self.assertEqual(pre_tokenizer_state["add_prefix_space"] ,lowerCAmelCase__ ) self.assertEqual(post_processor_state["add_prefix_space"] ,lowerCAmelCase__ ) self.assertEqual(post_processor_state["trim_offsets"] ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> Dict: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCAmelCase_ : List[str] = "hello" # `hello` is a token in the vocabulary of `pretrained_name` lowerCAmelCase_ : Dict = f'''{text_of_1_token} {text_of_1_token}''' lowerCAmelCase_ : Optional[Any] = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase__ ,use_fast=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = tokenizer_r(lowerCAmelCase__ ,return_offsets_mapping=lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ) self.assertEqual(encoding.offset_mapping[0] ,(0, len(lowerCAmelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] ,(len(lowerCAmelCase__ ) + 1, len(lowerCAmelCase__ ) + 1 + len(lowerCAmelCase__ )) ,) lowerCAmelCase_ : Dict = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase__ ,use_fast=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ) lowerCAmelCase_ : str = tokenizer_r(lowerCAmelCase__ ,return_offsets_mapping=lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ) self.assertEqual(encoding.offset_mapping[0] ,(0, len(lowerCAmelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] ,(len(lowerCAmelCase__ ) + 1, len(lowerCAmelCase__ ) + 1 + len(lowerCAmelCase__ )) ,) lowerCAmelCase_ : Tuple = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase__ ,use_fast=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = tokenizer_r(lowerCAmelCase__ ,return_offsets_mapping=lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ) self.assertEqual(encoding.offset_mapping[0] ,(0, len(lowerCAmelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] ,(len(lowerCAmelCase__ ), len(lowerCAmelCase__ ) + 1 + len(lowerCAmelCase__ )) ,) lowerCAmelCase_ : int = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase__ ,use_fast=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ) lowerCAmelCase_ : Dict = tokenizer_r(lowerCAmelCase__ ,return_offsets_mapping=lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ) self.assertEqual(encoding.offset_mapping[0] ,(0, len(lowerCAmelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] ,(len(lowerCAmelCase__ ), len(lowerCAmelCase__ ) + 1 + len(lowerCAmelCase__ )) ,) lowerCAmelCase_ : str = f''' {text}''' # tokenizer_r = self.rust_tokenizer_class.from_pretrained( # pretrained_name, use_fast=True, add_prefix_space=True, trim_offsets=True # ) # encoding = tokenizer_r(text, return_offsets_mapping=True, add_special_tokens=False) # self.assertEqual(encoding.offset_mapping[0], (1, 1 + len(text_of_1_token))) # self.assertEqual( # encoding.offset_mapping[1], # (1 + len(text_of_1_token) + 1, 1 + len(text_of_1_token) + 1 + len(text_of_1_token)), # ) lowerCAmelCase_ : Dict = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase__ ,use_fast=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ) lowerCAmelCase_ : int = tokenizer_r(lowerCAmelCase__ ,return_offsets_mapping=lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ) self.assertEqual(encoding.offset_mapping[0] ,(1, 1 + len(lowerCAmelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] ,(1 + len(lowerCAmelCase__ ) + 1, 1 + len(lowerCAmelCase__ ) + 1 + len(lowerCAmelCase__ )) ,) lowerCAmelCase_ : str = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase__ ,use_fast=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = tokenizer_r(lowerCAmelCase__ ,return_offsets_mapping=lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ) self.assertEqual(encoding.offset_mapping[0] ,(0, 1 + len(lowerCAmelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] ,(1 + len(lowerCAmelCase__ ), 1 + len(lowerCAmelCase__ ) + 1 + len(lowerCAmelCase__ )) ,) lowerCAmelCase_ : Optional[Any] = self.rust_tokenizer_class.from_pretrained( lowerCAmelCase__ ,use_fast=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = tokenizer_r(lowerCAmelCase__ ,return_offsets_mapping=lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ) self.assertEqual(encoding.offset_mapping[0] ,(0, 1 + len(lowerCAmelCase__ )) ) self.assertEqual( encoding.offset_mapping[1] ,(1 + len(lowerCAmelCase__ ), 1 + len(lowerCAmelCase__ ) + 1 + len(lowerCAmelCase__ )) ,)
659
from ....configuration_utils import PretrainedConfig from ....utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''Visual-Attention-Network/van-base''': ( '''https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json''' ), } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = 'van' def __init__( self : List[str] ,lowerCAmelCase__ : int=2_24 ,lowerCAmelCase__ : Optional[int]=3 ,lowerCAmelCase__ : Dict=[7, 3, 3, 3] ,lowerCAmelCase__ : List[str]=[4, 2, 2, 2] ,lowerCAmelCase__ : Union[str, Any]=[64, 1_28, 3_20, 5_12] ,lowerCAmelCase__ : Union[str, Any]=[3, 3, 12, 3] ,lowerCAmelCase__ : Any=[8, 8, 4, 4] ,lowerCAmelCase__ : Optional[int]="gelu" ,lowerCAmelCase__ : List[str]=0.02 ,lowerCAmelCase__ : Optional[Any]=1e-6 ,lowerCAmelCase__ : Dict=1e-2 ,lowerCAmelCase__ : Union[str, Any]=0.0 ,lowerCAmelCase__ : Optional[Any]=0.0 ,**lowerCAmelCase__ : List[str] ,) -> Tuple: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = image_size lowerCAmelCase_ : List[str] = num_channels lowerCAmelCase_ : str = patch_sizes lowerCAmelCase_ : Optional[Any] = strides lowerCAmelCase_ : List[Any] = hidden_sizes lowerCAmelCase_ : int = depths lowerCAmelCase_ : int = mlp_ratios lowerCAmelCase_ : str = hidden_act lowerCAmelCase_ : List[str] = initializer_range lowerCAmelCase_ : Dict = layer_norm_eps lowerCAmelCase_ : str = layer_scale_init_value lowerCAmelCase_ : Tuple = drop_path_rate lowerCAmelCase_ : Dict = dropout_rate
659
1
def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : str = len(snake_case__) print("The following activities are selected:") # The first activity is always selected lowerCAmelCase_ : Optional[Any] = 0 print(snake_case__ , end=",") # Consider rest of the activities for j in range(snake_case__): # If this activity has start time greater than # or equal to the finish time of previously # selected activity, then select it if start[j] >= finish[i]: print(snake_case__ , end=",") lowerCAmelCase_ : Any = j if __name__ == "__main__": import doctest doctest.testmod() _lowercase = [1, 3, 0, 5, 8, 5] _lowercase = [2, 4, 6, 7, 9, 9] print_max_activities(start, finish)
659
from math import factorial def UpperCamelCase ( snake_case__ , snake_case__): # If either of the conditions are true, the function is being asked # to calculate a factorial of a negative number, which is not possible if n < k or k < 0: raise ValueError("Please enter positive integers for n and k where n >= k") return factorial(snake_case__) // (factorial(snake_case__) * factorial(n - k)) if __name__ == "__main__": print( '''The number of five-card hands possible from a standard''', f"fifty-two card deck is: {combinations(52, 5)}\n", ) print( '''If a class of 40 students must be arranged into groups of''', f"4 for group projects, there are {combinations(40, 4)} ways", '''to arrange them.\n''', ) print( '''If 10 teams are competing in a Formula One race, there''', f"are {combinations(10, 3)} ways that first, second and", '''third place can be awarded.''', )
659
1
import heapq import sys import numpy as np _lowercase = tuple[int, int] class __snake_case : """simple docstring""" def __init__( self : Tuple ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = [] lowerCAmelCase_ : int = set() def UpperCAmelCase_ ( self : str ) -> Union[str, Any]: '''simple docstring''' if not self.empty(): return self.elements[0][0] else: return float("inf" ) def UpperCAmelCase_ ( self : Any ) -> List[Any]: '''simple docstring''' return len(self.elements ) == 0 def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : int ) -> List[Any]: '''simple docstring''' if item not in self.set: heapq.heappush(self.elements ,(priority, item) ) self.set.add(lowerCAmelCase__ ) else: # update # print("update", item) lowerCAmelCase_ : Any = [] ((lowerCAmelCase_) , (lowerCAmelCase_)) : str = heapq.heappop(self.elements ) while x != item: temp.append((pri, x) ) ((lowerCAmelCase_) , (lowerCAmelCase_)) : Dict = heapq.heappop(self.elements ) temp.append((priority, item) ) for pro, xxx in temp: heapq.heappush(self.elements ,(pro, xxx) ) def UpperCAmelCase_ ( self : Dict ,lowerCAmelCase__ : str ) -> Optional[int]: '''simple docstring''' if item in self.set: self.set.remove(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = [] ((lowerCAmelCase_) , (lowerCAmelCase_)) : str = heapq.heappop(self.elements ) while x != item: temp.append((pro, x) ) ((lowerCAmelCase_) , (lowerCAmelCase_)) : int = heapq.heappop(self.elements ) for prito, yyy in temp: heapq.heappush(self.elements ,(prito, yyy) ) def UpperCAmelCase_ ( self : int ) -> List[str]: '''simple docstring''' return self.elements[0][1] def UpperCAmelCase_ ( self : str ) -> str: '''simple docstring''' ((lowerCAmelCase_) , (lowerCAmelCase_)) : Any = heapq.heappop(self.elements ) self.set.remove(lowerCAmelCase__ ) return (priority, item) def UpperCamelCase ( snake_case__ , snake_case__): # euclidean distance lowerCAmelCase_ : Union[str, Any] = np.array(snake_case__) lowerCAmelCase_ : Dict = np.array(snake_case__) return np.linalg.norm(a - b) def UpperCamelCase ( snake_case__ , snake_case__): # integer division by time variable return consistent_heuristic(snake_case__ , snake_case__) // t def UpperCamelCase ( snake_case__ , snake_case__): # manhattan distance return abs(p[0] - goal[0]) + abs(p[1] - goal[1]) def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : Union[str, Any] = g_function[start] + Wa * heuristics[i](snake_case__ , snake_case__) return ans def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : str = np.chararray((n, n)) for i in range(snake_case__): for j in range(snake_case__): lowerCAmelCase_ : Optional[Any] = "*" for i in range(snake_case__): for j in range(snake_case__): if (j, (n - 1) - i) in blocks: lowerCAmelCase_ : Tuple = "#" lowerCAmelCase_ : int = "-" lowerCAmelCase_ : List[str] = back_pointer[goal] while x != start: ((lowerCAmelCase_) , (lowerCAmelCase_)) : Optional[int] = x # print(x) lowerCAmelCase_ : Optional[Any] = "-" lowerCAmelCase_ : Optional[Any] = back_pointer[x] lowerCAmelCase_ : Optional[Any] = "-" for i in range(snake_case__): for j in range(snake_case__): if (i, j) == (0, n - 1): print(grid[i][j] , end=" ") print("<-- End position" , end=" ") else: print(grid[i][j] , end=" ") print() print("^") print("Start position") print() print("# is an obstacle") print("- is the path taken by algorithm") print("PATH TAKEN BY THE ALGORITHM IS:-") lowerCAmelCase_ : Any = back_pointer[goal] while x != start: print(snake_case__ , end=" ") lowerCAmelCase_ : Optional[int] = back_pointer[x] print(snake_case__) sys.exit() def UpperCamelCase ( snake_case__): if p[0] < 0 or p[0] > n - 1: return False if p[1] < 0 or p[1] > n - 1: return False return True def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ): for itera in range(snake_case__): open_list[itera].remove_element(snake_case__) # print("s", s) # print("j", j) ((lowerCAmelCase_) , (lowerCAmelCase_)) : int = s lowerCAmelCase_ : str = (x - 1, y) lowerCAmelCase_ : Any = (x + 1, y) lowerCAmelCase_ : Tuple = (x, y + 1) lowerCAmelCase_ : Optional[int] = (x, y - 1) for neighbours in [left, right, up, down]: if neighbours not in blocks: if valid(snake_case__) and neighbours not in visited: # print("neighbour", neighbours) visited.add(snake_case__) lowerCAmelCase_ : int = -1 lowerCAmelCase_ : List[Any] = float("inf") if valid(snake_case__) and g_function[neighbours] > g_function[s] + 1: lowerCAmelCase_ : Optional[Any] = g_function[s] + 1 lowerCAmelCase_ : Any = s if neighbours not in close_list_anchor: open_list[0].put(snake_case__ , key(snake_case__ , 0 , snake_case__ , snake_case__)) if neighbours not in close_list_inad: for var in range(1 , snake_case__): if key(snake_case__ , snake_case__ , snake_case__ , snake_case__) <= Wa * key( snake_case__ , 0 , snake_case__ , snake_case__): open_list[j].put( snake_case__ , key(snake_case__ , snake_case__ , snake_case__ , snake_case__)) def UpperCamelCase ( ): lowerCAmelCase_ : str = [] for x in range(1 , 5): for y in range(1 , 6): some_list.append((x, y)) for x in range(15 , 20): some_list.append((x, 17)) for x in range(10 , 19): for y in range(1 , 15): some_list.append((x, y)) # L block for x in range(1 , 4): for y in range(12 , 19): some_list.append((x, y)) for x in range(3 , 13): for y in range(16 , 19): some_list.append((x, y)) return some_list _lowercase = {0: consistent_heuristic, 1: heuristic_a, 2: heuristic_a} _lowercase = [ (0, 1), (1, 1), (2, 1), (3, 1), (4, 1), (5, 1), (6, 1), (7, 1), (8, 1), (9, 1), (10, 1), (11, 1), (12, 1), (13, 1), (14, 1), (15, 1), (16, 1), (17, 1), (18, 1), (19, 1), ] _lowercase = make_common_ground() _lowercase = blocks_blk # hyper parameters _lowercase = 1 _lowercase = 1 _lowercase = 20 _lowercase = 3 # one consistent and two other inconsistent # start and end destination _lowercase = (0, 0) _lowercase = (n - 1, n - 1) _lowercase = 1 def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__): lowerCAmelCase_ : List[str] = {start: 0, goal: float("inf")} lowerCAmelCase_ : Tuple = {start: -1, goal: -1} lowerCAmelCase_ : str = [] lowerCAmelCase_ : List[str] = set() for i in range(snake_case__): open_list.append(PriorityQueue()) open_list[i].put(snake_case__ , key(snake_case__ , snake_case__ , snake_case__ , snake_case__)) lowerCAmelCase_ : list[int] = [] lowerCAmelCase_ : list[int] = [] while open_list[0].minkey() < float("inf"): for i in range(1 , snake_case__): # print(open_list[0].minkey(), open_list[i].minkey()) if open_list[i].minkey() <= Wa * open_list[0].minkey(): global t t += 1 if g_function[goal] <= open_list[i].minkey(): if g_function[goal] < float("inf"): do_something(snake_case__ , snake_case__ , snake_case__) else: lowerCAmelCase_ , lowerCAmelCase_ : str = open_list[i].top_show() visited.add(snake_case__) expand_state( snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ) close_list_inad.append(snake_case__) else: if g_function[goal] <= open_list[0].minkey(): if g_function[goal] < float("inf"): do_something(snake_case__ , snake_case__ , snake_case__) else: lowerCAmelCase_ : Tuple = open_list[0].top_show() visited.add(snake_case__) expand_state( snake_case__ , 0 , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , snake_case__ , ) close_list_anchor.append(snake_case__) print("No path found to goal") print() for i in range(n - 1 , -1 , -1): for j in range(snake_case__): if (j, i) in blocks: print("#" , end=" ") elif (j, i) in back_pointer: if (j, i) == (n - 1, n - 1): print("*" , end=" ") else: print("-" , end=" ") else: print("*" , end=" ") if (j, i) == (n - 1, n - 1): print("<-- End position" , end=" ") print() print("^") print("Start position") print() print("# is an obstacle") print("- is the path taken by algorithm") if __name__ == "__main__": multi_a_star(start, goal, n_heuristic)
659
import argparse import json from tqdm import tqdm def UpperCamelCase ( ): lowerCAmelCase_ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--src_path" , type=snake_case__ , default="biencoder-nq-dev.json" , help="Path to raw DPR training data" , ) parser.add_argument( "--evaluation_set" , type=snake_case__ , help="where to store parsed evaluation_set file" , ) parser.add_argument( "--gold_data_path" , type=snake_case__ , help="where to store parsed gold_data_path file" , ) lowerCAmelCase_ : Dict = parser.parse_args() with open(args.src_path , "r") as src_file, open(args.evaluation_set , "w") as eval_file, open( args.gold_data_path , "w") as gold_file: lowerCAmelCase_ : Optional[int] = json.load(snake_case__) for dpr_record in tqdm(snake_case__): lowerCAmelCase_ : str = dpr_record["question"] lowerCAmelCase_ : Dict = [context["title"] for context in dpr_record["positive_ctxs"]] eval_file.write(question + "\n") gold_file.write("\t".join(snake_case__) + "\n") if __name__ == "__main__": main()
659
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore _lowercase = ''' Human: <<task>> Assistant: ''' _lowercase = '''huggingface-tools/default-prompts''' _lowercase = {'''chat''': '''chat_prompt_template.txt''', '''run''': '''run_prompt_template.txt'''} def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__="run"): if prompt_or_repo_id is None: lowerCAmelCase_ : List[Any] = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search("\\s" , snake_case__) is not None: return prompt_or_repo_id lowerCAmelCase_ : Optional[int] = cached_file( snake_case__ , PROMPT_FILES[mode] , repo_type="dataset" , user_agent={"agent": agent_name}) with open(snake_case__ , "r" , encoding="utf-8") as f: return f.read()
659
from collections.abc import Sequence def UpperCamelCase ( snake_case__ = None): if nums is None or not nums: raise ValueError("Input sequence should not be empty") lowerCAmelCase_ : Dict = nums[0] for i in range(1 , len(snake_case__)): lowerCAmelCase_ : Optional[int] = nums[i] lowerCAmelCase_ : Optional[int] = max(snake_case__ , ans + num , snake_case__) return ans if __name__ == "__main__": import doctest doctest.testmod() # Try on a sample input from the user _lowercase = int(input('''Enter number of elements : ''').strip()) _lowercase = list(map(int, input('''\nEnter the numbers : ''').strip().split()))[:n] print(max_subsequence_sum(array))
659
1
from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class __snake_case : """simple docstring""" def __init__( self : Optional[Any] ,lowerCAmelCase__ : Optional[int] ,lowerCAmelCase__ : int=13 ,lowerCAmelCase__ : Tuple=7 ,lowerCAmelCase__ : Any=True ,lowerCAmelCase__ : Any=True ,lowerCAmelCase__ : List[str]=True ,lowerCAmelCase__ : int=True ,lowerCAmelCase__ : int=99 ,lowerCAmelCase__ : int=32 ,lowerCAmelCase__ : Union[str, Any]=2 ,lowerCAmelCase__ : int=4 ,lowerCAmelCase__ : Dict=37 ,lowerCAmelCase__ : Optional[int]="gelu" ,lowerCAmelCase__ : Union[str, Any]=0.1 ,lowerCAmelCase__ : List[Any]=0.1 ,lowerCAmelCase__ : Any=5_12 ,lowerCAmelCase__ : str=16 ,lowerCAmelCase__ : List[Any]=2 ,lowerCAmelCase__ : int=0.02 ,lowerCAmelCase__ : str=3 ,lowerCAmelCase__ : str=4 ,lowerCAmelCase__ : Any=None ,lowerCAmelCase__ : Union[str, Any]=10_00 ,) -> int: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = parent lowerCAmelCase_ : Optional[int] = batch_size lowerCAmelCase_ : Any = seq_length lowerCAmelCase_ : List[str] = is_training lowerCAmelCase_ : str = use_input_mask lowerCAmelCase_ : Optional[int] = use_token_type_ids lowerCAmelCase_ : Tuple = use_labels lowerCAmelCase_ : Tuple = vocab_size lowerCAmelCase_ : List[Any] = hidden_size lowerCAmelCase_ : List[str] = num_hidden_layers lowerCAmelCase_ : Union[str, Any] = num_attention_heads lowerCAmelCase_ : Optional[Any] = intermediate_size lowerCAmelCase_ : Union[str, Any] = hidden_act lowerCAmelCase_ : List[str] = hidden_dropout_prob lowerCAmelCase_ : Optional[int] = attention_probs_dropout_prob lowerCAmelCase_ : Optional[int] = max_position_embeddings lowerCAmelCase_ : Tuple = type_vocab_size lowerCAmelCase_ : Dict = type_sequence_label_size lowerCAmelCase_ : int = initializer_range lowerCAmelCase_ : Tuple = num_labels lowerCAmelCase_ : List[str] = num_choices lowerCAmelCase_ : Dict = scope lowerCAmelCase_ : List[Any] = range_bbox def UpperCAmelCase_ ( self : Dict ) -> str: '''simple docstring''' lowerCAmelCase_ : str = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) # convert bbox to numpy since TF does not support item assignment lowerCAmelCase_ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length, 4] ,self.range_bbox ).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: lowerCAmelCase_ : Optional[Any] = bbox[i, j, 3] lowerCAmelCase_ : List[Any] = bbox[i, j, 1] lowerCAmelCase_ : List[str] = t if bbox[i, j, 2] < bbox[i, j, 0]: lowerCAmelCase_ : str = bbox[i, j, 2] lowerCAmelCase_ : Tuple = bbox[i, j, 0] lowerCAmelCase_ : Dict = t lowerCAmelCase_ : str = tf.convert_to_tensor(lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = None if self.use_input_mask: lowerCAmelCase_ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase_ : Any = None if self.use_token_type_ids: lowerCAmelCase_ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) lowerCAmelCase_ : List[Any] = None lowerCAmelCase_ : Tuple = None lowerCAmelCase_ : Optional[int] = None if self.use_labels: lowerCAmelCase_ : List[str] = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) lowerCAmelCase_ : str = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) lowerCAmelCase_ : Any = ids_tensor([self.batch_size] ,self.num_choices ) lowerCAmelCase_ : Dict = LayoutLMConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_range=self.initializer_range ,) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : Dict ,lowerCAmelCase__ : str ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : int ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : Optional[int] ,lowerCAmelCase__ : Tuple ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : int = TFLayoutLMModel(config=lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = model(lowerCAmelCase__ ,lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,token_type_ids=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = model(lowerCAmelCase__ ,lowerCAmelCase__ ,token_type_ids=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = model(lowerCAmelCase__ ,lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : int ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : Optional[int] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : List[Any] = TFLayoutLMForMaskedLM(config=lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = model(lowerCAmelCase__ ,lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,token_type_ids=lowerCAmelCase__ ,labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Optional[int] ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : Dict ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Tuple ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : List[Any] = self.num_labels lowerCAmelCase_ : List[Any] = TFLayoutLMForSequenceClassification(config=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = model(lowerCAmelCase__ ,lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,token_type_ids=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : Optional[int] ,lowerCAmelCase__ : Optional[int] ,lowerCAmelCase__ : Dict ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Dict ) -> Any: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = self.num_labels lowerCAmelCase_ : Dict = TFLayoutLMForTokenClassification(config=lowerCAmelCase__ ) lowerCAmelCase_ : str = model(lowerCAmelCase__ ,lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,token_type_ids=lowerCAmelCase__ ,labels=lowerCAmelCase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : Dict ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : str ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : int ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : str = TFLayoutLMForQuestionAnswering(config=lowerCAmelCase__ ) lowerCAmelCase_ : Dict = model(lowerCAmelCase__ ,lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,token_type_ids=lowerCAmelCase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def UpperCAmelCase_ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = self.prepare_config_and_inputs() ( ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ( lowerCAmelCase_ ) , ) : Dict = config_and_inputs lowerCAmelCase_ : Optional[Any] = { "input_ids": input_ids, "bbox": bbox, "token_type_ids": token_type_ids, "attention_mask": input_mask, } return config, inputs_dict @require_tf class __snake_case ( snake_case__ , snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = ( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) UpperCamelCase_ = ( { 'feature-extraction': TFLayoutLMModel, 'fill-mask': TFLayoutLMForMaskedLM, 'text-classification': TFLayoutLMForSequenceClassification, 'token-classification': TFLayoutLMForTokenClassification, 'zero-shot': TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) UpperCamelCase_ = False UpperCamelCase_ = True UpperCamelCase_ = 1_0 def UpperCAmelCase_ ( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = TFLayoutLMModelTester(self ) lowerCAmelCase_ : Union[str, Any] = ConfigTester(self ,config_class=lowerCAmelCase__ ,hidden_size=37 ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase_ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[str] ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Any ) -> int: '''simple docstring''' lowerCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCAmelCase__ ) def UpperCAmelCase_ ( self : str ) -> Any: '''simple docstring''' lowerCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> Dict: '''simple docstring''' lowerCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCAmelCase__ ) @slow def UpperCAmelCase_ ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : int = TFLayoutLMModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) @unittest.skip("Onnx compliancy broke with TF 2.10" ) def UpperCAmelCase_ ( self : int ) -> List[Any]: '''simple docstring''' pass def UpperCamelCase ( ): # Here we prepare a batch of 2 sequences to test a LayoutLM forward pass on: # fmt: off lowerCAmelCase_ : Optional[Any] = tf.convert_to_tensor([[1_01,10_19,10_14,10_16,10_37,1_28_49,47_47,10_04,1_42_46,22_78,54_39,45_24,50_02,29_30,21_93,29_30,43_41,32_08,10_05,10_55,21_71,28_48,1_13_00,35_31,1_02],[1_01,40_70,40_34,70_20,10_24,30_58,10_15,10_13,28_61,10_13,60_70,1_92_74,27_72,62_05,2_78_14,1_61_47,1_61_47,43_43,20_47,1_02_83,1_09_69,1_43_89,10_12,23_38,1_02]]) # noqa: E231 lowerCAmelCase_ : List[str] = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],]) # noqa: E231 lowerCAmelCase_ : List[Any] = tf.convert_to_tensor([[[0,0,0,0],[4_23,2_37,4_40,2_51],[4_27,2_72,4_41,2_87],[4_19,1_15,4_37,1_29],[9_61,8_85,9_92,9_12],[2_56,38,3_30,58],[2_56,38,3_30,58],[3_36,42,3_53,57],[3_60,39,4_01,56],[3_60,39,4_01,56],[4_11,39,4_71,59],[4_79,41,5_28,59],[5_33,39,6_30,60],[67,1_13,1_34,1_31],[1_41,1_15,2_09,1_32],[68,1_49,1_33,1_66],[1_41,1_49,1_87,1_64],[1_95,1_48,2_87,1_65],[1_95,1_48,2_87,1_65],[1_95,1_48,2_87,1_65],[2_95,1_48,3_49,1_65],[4_41,1_49,4_92,1_66],[4_97,1_49,5_46,1_64],[64,2_01,1_25,2_18],[10_00,10_00,10_00,10_00]],[[0,0,0,0],[6_62,1_50,7_54,1_66],[6_65,1_99,7_42,2_11],[5_19,2_13,5_54,2_28],[5_19,2_13,5_54,2_28],[1_34,4_33,1_87,4_54],[1_30,4_67,2_04,4_80],[1_30,4_67,2_04,4_80],[1_30,4_67,2_04,4_80],[1_30,4_67,2_04,4_80],[1_30,4_67,2_04,4_80],[3_14,4_69,3_76,4_82],[5_04,6_84,5_82,7_06],[9_41,8_25,9_73,9_00],[9_41,8_25,9_73,9_00],[9_41,8_25,9_73,9_00],[9_41,8_25,9_73,9_00],[6_10,7_49,6_52,7_65],[1_30,6_59,1_68,6_72],[1_76,6_57,2_37,6_72],[2_38,6_57,3_12,6_72],[4_43,6_53,6_28,6_72],[4_43,6_53,6_28,6_72],[7_16,3_01,8_25,3_17],[10_00,10_00,10_00,10_00]]]) # noqa: E231 lowerCAmelCase_ : List[str] = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]]) # noqa: E231 # these are sequence labels (i.e. at the token level) lowerCAmelCase_ : Tuple = tf.convert_to_tensor([[-1_00,10,10,10,9,1,-1_00,7,7,-1_00,7,7,4,2,5,2,8,8,-1_00,-1_00,5,0,3,2,-1_00],[-1_00,12,12,12,-1_00,12,10,-1_00,-1_00,-1_00,-1_00,10,12,9,-1_00,-1_00,-1_00,10,10,10,9,12,-1_00,10,-1_00]]) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class __snake_case ( unittest.TestCase ): """simple docstring""" @slow def UpperCAmelCase_ ( self : Optional[Any] ) -> Dict: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = TFLayoutLMModel.from_pretrained("microsoft/layoutlm-base-uncased" ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Tuple = prepare_layoutlm_batch_inputs() # forward pass lowerCAmelCase_ : Optional[int] = model(input_ids=lowerCAmelCase__ ,bbox=lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,token_type_ids=lowerCAmelCase__ ) # test the sequence output on [0, :3, :3] lowerCAmelCase_ : Any = tf.convert_to_tensor( [[0.1_785, -0.1_947, -0.0_425], [-0.3_254, -0.2_807, 0.2_553], [-0.5_391, -0.3_322, 0.3_364]] ,) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] ,lowerCAmelCase__ ,atol=1e-3 ) ) # test the pooled output on [1, :3] lowerCAmelCase_ : List[Any] = tf.convert_to_tensor([-0.6_580, -0.0_214, 0.8_552] ) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] ,lowerCAmelCase__ ,atol=1e-3 ) ) @slow def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : int = TFLayoutLMForSequenceClassification.from_pretrained("microsoft/layoutlm-base-uncased" ,num_labels=2 ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = prepare_layoutlm_batch_inputs() # forward pass lowerCAmelCase_ : Tuple = model( input_ids=lowerCAmelCase__ ,bbox=lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,token_type_ids=lowerCAmelCase__ ,labels=tf.convert_to_tensor([1, 1] ) ,) # test whether we get a loss as a scalar lowerCAmelCase_ : Tuple = outputs.loss lowerCAmelCase_ : Optional[Any] = (2,) self.assertEqual(loss.shape ,lowerCAmelCase__ ) # test the shape of the logits lowerCAmelCase_ : Any = outputs.logits lowerCAmelCase_ : List[str] = (2, 2) self.assertEqual(logits.shape ,lowerCAmelCase__ ) @slow def UpperCAmelCase_ ( self : Optional[int] ) -> Any: '''simple docstring''' lowerCAmelCase_ : List[str] = TFLayoutLMForTokenClassification.from_pretrained("microsoft/layoutlm-base-uncased" ,num_labels=13 ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : int = prepare_layoutlm_batch_inputs() # forward pass lowerCAmelCase_ : List[Any] = model( input_ids=lowerCAmelCase__ ,bbox=lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,token_type_ids=lowerCAmelCase__ ,labels=lowerCAmelCase__ ) # test the shape of the logits lowerCAmelCase_ : Optional[Any] = outputs.logits lowerCAmelCase_ : Tuple = tf.convert_to_tensor((2, 25, 13) ) self.assertEqual(logits.shape ,lowerCAmelCase__ ) @slow def UpperCAmelCase_ ( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = TFLayoutLMForQuestionAnswering.from_pretrained("microsoft/layoutlm-base-uncased" ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = prepare_layoutlm_batch_inputs() # forward pass lowerCAmelCase_ : List[str] = model(input_ids=lowerCAmelCase__ ,bbox=lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,token_type_ids=lowerCAmelCase__ ) # test the shape of the logits lowerCAmelCase_ : List[str] = tf.convert_to_tensor((2, 25) ) self.assertEqual(outputs.start_logits.shape ,lowerCAmelCase__ ) self.assertEqual(outputs.end_logits.shape ,lowerCAmelCase__ )
659
from typing import TYPE_CHECKING from ....utils import _LazyModule _lowercase = {'''tokenization_tapex''': ['''TapexTokenizer''']} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
659
1
import torch from diffusers import DPMSolverSDEScheduler from diffusers.utils import torch_device from diffusers.utils.testing_utils import require_torchsde from .test_schedulers import SchedulerCommonTest @require_torchsde class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = (DPMSolverSDEScheduler,) UpperCamelCase_ = 1_0 def UpperCAmelCase_ ( self : Optional[Any] ,**lowerCAmelCase__ : Tuple ) -> int: '''simple docstring''' lowerCAmelCase_ : Any = { "num_train_timesteps": 11_00, "beta_start": 0.0_001, "beta_end": 0.02, "beta_schedule": "linear", "noise_sampler_seed": 0, } config.update(**lowerCAmelCase__ ) return config def UpperCAmelCase_ ( self : List[Any] ) -> List[str]: '''simple docstring''' for timesteps in [10, 50, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=lowerCAmelCase__ ) def UpperCAmelCase_ ( self : str ) -> List[Any]: '''simple docstring''' for beta_start, beta_end in zip([0.00_001, 0.0_001, 0.001] ,[0.0_002, 0.002, 0.02] ): self.check_over_configs(beta_start=lowerCAmelCase__ ,beta_end=lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[str] ) -> List[Any]: '''simple docstring''' for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ) -> Any: '''simple docstring''' for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : str = self.scheduler_classes[0] lowerCAmelCase_ : Dict = self.get_scheduler_config() lowerCAmelCase_ : Dict = scheduler_class(**lowerCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase_ : Any = self.dummy_model() lowerCAmelCase_ : List[str] = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase_ : Optional[int] = sample.to(lowerCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase_ : List[str] = scheduler.scale_model_input(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = model(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Dict = output.prev_sample lowerCAmelCase_ : Dict = torch.sum(torch.abs(lowerCAmelCase__ ) ) lowerCAmelCase_ : str = torch.mean(torch.abs(lowerCAmelCase__ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 167.47_821_044_921_875 ) < 1e-2 assert abs(result_mean.item() - 0.2_178_705_964_565_277 ) < 1e-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 171.59_352_111_816_406 ) < 1e-2 assert abs(result_mean.item() - 0.22_342_906_892_299_652 ) < 1e-3 else: assert abs(result_sum.item() - 162.52_383_422_851_562 ) < 1e-2 assert abs(result_mean.item() - 0.211_619_570_851_326 ) < 1e-3 def UpperCAmelCase_ ( self : str ) -> int: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = self.scheduler_classes[0] lowerCAmelCase_ : Optional[Any] = self.get_scheduler_config(prediction_type="v_prediction" ) lowerCAmelCase_ : int = scheduler_class(**lowerCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps ) lowerCAmelCase_ : Tuple = self.dummy_model() lowerCAmelCase_ : int = self.dummy_sample_deter * scheduler.init_noise_sigma lowerCAmelCase_ : Union[str, Any] = sample.to(lowerCAmelCase__ ) for i, t in enumerate(scheduler.timesteps ): lowerCAmelCase_ : Dict = scheduler.scale_model_input(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = model(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : str = output.prev_sample lowerCAmelCase_ : str = torch.sum(torch.abs(lowerCAmelCase__ ) ) lowerCAmelCase_ : str = torch.mean(torch.abs(lowerCAmelCase__ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 124.77_149_200_439_453 ) < 1e-2 assert abs(result_mean.item() - 0.16_226_289_014_816_284 ) < 1e-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 128.1_663_360_595_703 ) < 1e-2 assert abs(result_mean.item() - 0.16_688_326_001_167_297 ) < 1e-3 else: assert abs(result_sum.item() - 119.8_487_548_828_125 ) < 1e-2 assert abs(result_mean.item() - 0.1_560_530_662_536_621 ) < 1e-3 def UpperCAmelCase_ ( self : int ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : List[str] = self.scheduler_classes[0] lowerCAmelCase_ : List[str] = self.get_scheduler_config() lowerCAmelCase_ : Optional[Any] = scheduler_class(**lowerCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps ,device=lowerCAmelCase__ ) lowerCAmelCase_ : Any = self.dummy_model() lowerCAmelCase_ : Optional[int] = self.dummy_sample_deter.to(lowerCAmelCase__ ) * scheduler.init_noise_sigma for t in scheduler.timesteps: lowerCAmelCase_ : List[Any] = scheduler.scale_model_input(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = model(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = output.prev_sample lowerCAmelCase_ : int = torch.sum(torch.abs(lowerCAmelCase__ ) ) lowerCAmelCase_ : Optional[int] = torch.mean(torch.abs(lowerCAmelCase__ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 167.46_957_397_460_938 ) < 1e-2 assert abs(result_mean.item() - 0.21_805_934_607_982_635 ) < 1e-3 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 171.59_353_637_695_312 ) < 1e-2 assert abs(result_mean.item() - 0.22_342_908_382_415_771 ) < 1e-3 else: assert abs(result_sum.item() - 162.52_383_422_851_562 ) < 1e-2 assert abs(result_mean.item() - 0.211_619_570_851_326 ) < 1e-3 def UpperCAmelCase_ ( self : Tuple ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Any = self.scheduler_classes[0] lowerCAmelCase_ : Optional[Any] = self.get_scheduler_config() lowerCAmelCase_ : List[Any] = scheduler_class(**lowerCAmelCase__ ,use_karras_sigmas=lowerCAmelCase__ ) scheduler.set_timesteps(self.num_inference_steps ,device=lowerCAmelCase__ ) lowerCAmelCase_ : str = self.dummy_model() lowerCAmelCase_ : Union[str, Any] = self.dummy_sample_deter.to(lowerCAmelCase__ ) * scheduler.init_noise_sigma lowerCAmelCase_ : str = sample.to(lowerCAmelCase__ ) for t in scheduler.timesteps: lowerCAmelCase_ : str = scheduler.scale_model_input(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = model(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = scheduler.step(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = output.prev_sample lowerCAmelCase_ : int = torch.sum(torch.abs(lowerCAmelCase__ ) ) lowerCAmelCase_ : int = torch.mean(torch.abs(lowerCAmelCase__ ) ) if torch_device in ["mps"]: assert abs(result_sum.item() - 176.66_974_135_742_188 ) < 1e-2 assert abs(result_mean.item() - 0.23_003_872_730_981_811 ) < 1e-2 elif torch_device in ["cuda"]: assert abs(result_sum.item() - 177.63_653_564_453_125 ) < 1e-2 assert abs(result_mean.item() - 0.23_003_872_730_981_811 ) < 1e-2 else: assert abs(result_sum.item() - 170.3_135_223_388_672 ) < 1e-2 assert abs(result_mean.item() - 0.23_003_872_730_981_811 ) < 1e-2
659
import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _lowercase = '''src/diffusers''' _lowercase = '''.''' # This is to make sure the diffusers module imported is the one in the repo. _lowercase = importlib.util.spec_from_file_location( '''diffusers''', os.path.join(DIFFUSERS_PATH, '''__init__.py'''), submodule_search_locations=[DIFFUSERS_PATH], ) _lowercase = spec.loader.load_module() def UpperCamelCase ( snake_case__ , snake_case__): return line.startswith(snake_case__) or len(snake_case__) <= 1 or re.search(R"^\s*\)(\s*->.*:|:)\s*$" , snake_case__) is not None def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Tuple = object_name.split(".") lowerCAmelCase_ : Union[str, Any] = 0 # First let's find the module where our object lives. lowerCAmelCase_ : Union[str, Any] = parts[i] while i < len(snake_case__) and not os.path.isfile(os.path.join(snake_case__ , F'''{module}.py''')): i += 1 if i < len(snake_case__): lowerCAmelCase_ : Dict = os.path.join(snake_case__ , parts[i]) if i >= len(snake_case__): raise ValueError(F'''`object_name` should begin with the name of a module of diffusers but got {object_name}.''') with open(os.path.join(snake_case__ , F'''{module}.py''') , "r" , encoding="utf-8" , newline="\n") as f: lowerCAmelCase_ : Optional[Any] = f.readlines() # Now let's find the class / func in the code! lowerCAmelCase_ : Union[str, Any] = "" lowerCAmelCase_ : int = 0 for name in parts[i + 1 :]: while ( line_index < len(snake_case__) and re.search(RF'''^{indent}(class|def)\s+{name}(\(|\:)''' , lines[line_index]) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(snake_case__): raise ValueError(F''' {object_name} does not match any function or class in {module}.''') # We found the beginning of the class / func, now let's find the end (when the indent diminishes). lowerCAmelCase_ : Union[str, Any] = line_index while line_index < len(snake_case__) and _should_continue(lines[line_index] , snake_case__): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1]) <= 1: line_index -= 1 lowerCAmelCase_ : List[str] = lines[start_index:line_index] return "".join(snake_case__) _lowercase = re.compile(r'''^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)''') _lowercase = re.compile(r'''^\s*(\S+)->(\S+)(\s+.*|$)''') _lowercase = re.compile(r'''<FILL\s+[^>]*>''') def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Any = code.split("\n") lowerCAmelCase_ : Any = 0 while idx < len(snake_case__) and len(lines[idx]) == 0: idx += 1 if idx < len(snake_case__): return re.search(R"^(\s*)\S" , lines[idx]).groups()[0] return "" def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Dict = len(get_indent(snake_case__)) > 0 if has_indent: lowerCAmelCase_ : Dict = F'''class Bla:\n{code}''' lowerCAmelCase_ : Optional[int] = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=1_19 , preview=snake_case__) lowerCAmelCase_ : Optional[Any] = black.format_str(snake_case__ , mode=snake_case__) lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = style_docstrings_in_code(snake_case__) return result[len("class Bla:\n") :] if has_indent else result def UpperCamelCase ( snake_case__ , snake_case__=False): with open(snake_case__ , "r" , encoding="utf-8" , newline="\n") as f: lowerCAmelCase_ : Tuple = f.readlines() lowerCAmelCase_ : Tuple = [] lowerCAmelCase_ : Union[str, Any] = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(snake_case__): lowerCAmelCase_ : Optional[int] = _re_copy_warning.search(lines[line_index]) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = search.groups() lowerCAmelCase_ : int = find_code_in_diffusers(snake_case__) lowerCAmelCase_ : Dict = get_indent(snake_case__) lowerCAmelCase_ : Union[str, Any] = line_index + 1 if indent == theoretical_indent else line_index + 2 lowerCAmelCase_ : str = theoretical_indent lowerCAmelCase_ : Union[str, Any] = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. lowerCAmelCase_ : Optional[int] = True while line_index < len(snake_case__) and should_continue: line_index += 1 if line_index >= len(snake_case__): break lowerCAmelCase_ : Dict = lines[line_index] lowerCAmelCase_ : List[str] = _should_continue(snake_case__ , snake_case__) and re.search(F'''^{indent}# End copy''' , snake_case__) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1]) <= 1: line_index -= 1 lowerCAmelCase_ : Dict = lines[start_index:line_index] lowerCAmelCase_ : Optional[int] = "".join(snake_case__) # Remove any nested `Copied from` comments to avoid circular copies lowerCAmelCase_ : List[Any] = [line for line in theoretical_code.split("\n") if _re_copy_warning.search(snake_case__) is None] lowerCAmelCase_ : Optional[Any] = "\n".join(snake_case__) # Before comparing, use the `replace_pattern` on the original code. if len(snake_case__) > 0: lowerCAmelCase_ : List[str] = replace_pattern.replace("with" , "").split(",") lowerCAmelCase_ : Tuple = [_re_replace_pattern.search(snake_case__) for p in patterns] for pattern in patterns: if pattern is None: continue lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[str] = pattern.groups() lowerCAmelCase_ : int = re.sub(snake_case__ , snake_case__ , snake_case__) if option.strip() == "all-casing": lowerCAmelCase_ : List[str] = re.sub(obja.lower() , obja.lower() , snake_case__) lowerCAmelCase_ : int = re.sub(obja.upper() , obja.upper() , snake_case__) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line lowerCAmelCase_ : List[Any] = blackify(lines[start_index - 1] + theoretical_code) lowerCAmelCase_ : Union[str, Any] = theoretical_code[len(lines[start_index - 1]) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index]) if overwrite: lowerCAmelCase_ : List[Any] = lines[:start_index] + [theoretical_code] + lines[line_index:] lowerCAmelCase_ : Union[str, Any] = start_index + 1 if overwrite and len(snake_case__) > 0: # Warn the user a file has been modified. print(F'''Detected changes, rewriting {filename}.''') with open(snake_case__ , "w" , encoding="utf-8" , newline="\n") as f: f.writelines(snake_case__) return diffs def UpperCamelCase ( snake_case__ = False): lowerCAmelCase_ : Tuple = glob.glob(os.path.join(snake_case__ , "**/*.py") , recursive=snake_case__) lowerCAmelCase_ : int = [] for filename in all_files: lowerCAmelCase_ : Union[str, Any] = is_copy_consistent(snake_case__ , snake_case__) diffs += [F'''- {filename}: copy does not match {d[0]} at line {d[1]}''' for d in new_diffs] if not overwrite and len(snake_case__) > 0: lowerCAmelCase_ : Optional[Any] = "\n".join(snake_case__) raise Exception( "Found the following copy inconsistencies:\n" + diff + "\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.") if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') _lowercase = parser.parse_args() check_copies(args.fix_and_overwrite)
659
1
def UpperCamelCase ( snake_case__): if n_term == "": return [] lowerCAmelCase_ : list = [] for temp in range(int(snake_case__)): series.append(F'''1/{temp + 1}''' if series else "1") return series if __name__ == "__main__": _lowercase = input('''Enter the last number (nth term) of the Harmonic Series''') print('''Formula of Harmonic Series => 1+1/2+1/3 ..... 1/n''') print(harmonic_series(nth_term))
659
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''microsoft/swinv2-tiny-patch4-window8-256''': ( '''https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json''' ), } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = 'swinv2' UpperCamelCase_ = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : List[Any] ,lowerCAmelCase__ : Optional[int]=2_24 ,lowerCAmelCase__ : Dict=4 ,lowerCAmelCase__ : Dict=3 ,lowerCAmelCase__ : List[Any]=96 ,lowerCAmelCase__ : Optional[Any]=[2, 2, 6, 2] ,lowerCAmelCase__ : Optional[Any]=[3, 6, 12, 24] ,lowerCAmelCase__ : Optional[int]=7 ,lowerCAmelCase__ : Dict=4.0 ,lowerCAmelCase__ : Dict=True ,lowerCAmelCase__ : str=0.0 ,lowerCAmelCase__ : Tuple=0.0 ,lowerCAmelCase__ : str=0.1 ,lowerCAmelCase__ : List[str]="gelu" ,lowerCAmelCase__ : Union[str, Any]=False ,lowerCAmelCase__ : Dict=0.02 ,lowerCAmelCase__ : int=1e-5 ,lowerCAmelCase__ : List[str]=32 ,**lowerCAmelCase__ : Tuple ,) -> List[str]: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = image_size lowerCAmelCase_ : List[Any] = patch_size lowerCAmelCase_ : Dict = num_channels lowerCAmelCase_ : Optional[int] = embed_dim lowerCAmelCase_ : Optional[Any] = depths lowerCAmelCase_ : Any = len(lowerCAmelCase__ ) lowerCAmelCase_ : str = num_heads lowerCAmelCase_ : List[str] = window_size lowerCAmelCase_ : List[str] = mlp_ratio lowerCAmelCase_ : Dict = qkv_bias lowerCAmelCase_ : str = hidden_dropout_prob lowerCAmelCase_ : str = attention_probs_dropout_prob lowerCAmelCase_ : Union[str, Any] = drop_path_rate lowerCAmelCase_ : List[Any] = hidden_act lowerCAmelCase_ : Any = use_absolute_embeddings lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : int = initializer_range lowerCAmelCase_ : Union[str, Any] = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase_ : Tuple = int(embed_dim * 2 ** (len(lowerCAmelCase__ ) - 1) ) lowerCAmelCase_ : str = (0, 0, 0, 0)
659
1
import multiprocessing from typing import TYPE_CHECKING, Optional, Union from .. import Dataset, Features, config from ..formatting import query_table from ..packaged_modules.sql.sql import Sql from ..utils import logging from .abc import AbstractDatasetInputStream if TYPE_CHECKING: import sqlitea import sqlalchemy class __snake_case ( snake_case__ ): """simple docstring""" def __init__( self : str ,lowerCAmelCase__ : Union[str, "sqlalchemy.sql.Selectable"] ,lowerCAmelCase__ : Union[str, "sqlalchemy.engine.Connection", "sqlalchemy.engine.Engine", "sqlite3.Connection"] ,lowerCAmelCase__ : Optional[Features] = None ,lowerCAmelCase__ : str = None ,lowerCAmelCase__ : bool = False ,**lowerCAmelCase__ : Union[str, Any] ,) -> Union[str, Any]: '''simple docstring''' super().__init__(features=lowerCAmelCase__ ,cache_dir=lowerCAmelCase__ ,keep_in_memory=lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Dict = Sql( cache_dir=lowerCAmelCase__ ,features=lowerCAmelCase__ ,sql=lowerCAmelCase__ ,con=lowerCAmelCase__ ,**lowerCAmelCase__ ,) def UpperCAmelCase_ ( self : List[str] ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = None lowerCAmelCase_ : List[Any] = None lowerCAmelCase_ : Any = None lowerCAmelCase_ : Any = None self.builder.download_and_prepare( download_config=lowerCAmelCase__ ,download_mode=lowerCAmelCase__ ,verification_mode=lowerCAmelCase__ ,base_path=lowerCAmelCase__ ,) # Build dataset for splits lowerCAmelCase_ : Dict = self.builder.as_dataset( split="train" ,verification_mode=lowerCAmelCase__ ,in_memory=self.keep_in_memory ) return dataset class __snake_case : """simple docstring""" def __init__( self : Optional[Any] ,lowerCAmelCase__ : Dataset ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Union[str, "sqlalchemy.engine.Connection", "sqlalchemy.engine.Engine", "sqlite3.Connection"] ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : Optional[int] = None ,**lowerCAmelCase__ : str ,) -> Any: '''simple docstring''' if num_proc is not None and num_proc <= 0: raise ValueError(f'''num_proc {num_proc} must be an integer > 0.''' ) lowerCAmelCase_ : List[str] = dataset lowerCAmelCase_ : Optional[Any] = name lowerCAmelCase_ : List[Any] = con lowerCAmelCase_ : Tuple = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE lowerCAmelCase_ : Optional[Any] = num_proc lowerCAmelCase_ : str = to_sql_kwargs def UpperCAmelCase_ ( self : List[str] ) -> int: '''simple docstring''' lowerCAmelCase_ : int = self.to_sql_kwargs.pop("sql" ,lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = self.to_sql_kwargs.pop("con" ,lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = self.to_sql_kwargs.pop("index" ,lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = self._write(index=lowerCAmelCase__ ,**self.to_sql_kwargs ) return written def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : List[Any] ) -> Any: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : int = args lowerCAmelCase_ : Dict = {**to_sql_kwargs, "if_exists": "append"} if offset > 0 else to_sql_kwargs lowerCAmelCase_ : str = query_table( table=self.dataset.data ,key=slice(lowerCAmelCase__ ,offset + self.batch_size ) ,indices=self.dataset._indices ,) lowerCAmelCase_ : Dict = batch.to_pandas() lowerCAmelCase_ : str = df.to_sql(self.name ,self.con ,index=lowerCAmelCase__ ,**lowerCAmelCase__ ) return num_rows or len(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : Dict ,**lowerCAmelCase__ : Union[str, Any] ) -> int: '''simple docstring''' lowerCAmelCase_ : int = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 ,len(self.dataset ) ,self.batch_size ) ,unit="ba" ,disable=not logging.is_progress_bar_enabled() ,desc="Creating SQL from Arrow format" ,): written += self._batch_sql((offset, index, to_sql_kwargs) ) else: lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for num_rows in logging.tqdm( pool.imap( self._batch_sql ,[(offset, index, to_sql_kwargs) for offset in range(0 ,lowerCAmelCase__ ,lowerCAmelCase__ )] ,) ,total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size ,unit="ba" ,disable=not logging.is_progress_bar_enabled() ,desc="Creating SQL from Arrow format" ,): written += num_rows return written
659
from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging _lowercase = logging.get_logger(__name__) class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = ['input_features', 'attention_mask'] def __init__( self : Optional[Any] ,lowerCAmelCase__ : Any=80 ,lowerCAmelCase__ : Optional[Any]=1_60_00 ,lowerCAmelCase__ : List[str]=0.0 ,lowerCAmelCase__ : Tuple=10 ,lowerCAmelCase__ : Optional[Any]=25 ,lowerCAmelCase__ : Any="hamming_window" ,lowerCAmelCase__ : List[str]=32_768.0 ,lowerCAmelCase__ : Union[str, Any]=0.97 ,lowerCAmelCase__ : Any=1.0 ,lowerCAmelCase__ : str=True ,lowerCAmelCase__ : int=True ,lowerCAmelCase__ : Tuple=False ,**lowerCAmelCase__ : Optional[int] ,) -> Optional[Any]: '''simple docstring''' super().__init__(feature_size=lowerCAmelCase__ ,sampling_rate=lowerCAmelCase__ ,padding_value=lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = feature_size lowerCAmelCase_ : List[Any] = sampling_rate lowerCAmelCase_ : Union[str, Any] = padding_value lowerCAmelCase_ : str = hop_length lowerCAmelCase_ : str = win_length lowerCAmelCase_ : str = frame_signal_scale lowerCAmelCase_ : Any = preemphasis_coeff lowerCAmelCase_ : Optional[Any] = mel_floor lowerCAmelCase_ : List[str] = normalize_means lowerCAmelCase_ : Optional[Any] = normalize_vars lowerCAmelCase_ : Dict = win_function lowerCAmelCase_ : List[Any] = return_attention_mask lowerCAmelCase_ : Tuple = win_length * sampling_rate // 10_00 lowerCAmelCase_ : str = hop_length * sampling_rate // 10_00 lowerCAmelCase_ : Dict = optimal_fft_length(self.sample_size ) lowerCAmelCase_ : Optional[int] = (self.n_fft // 2) + 1 def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : np.array ) -> np.ndarray: '''simple docstring''' if self.win_function == "hamming_window": lowerCAmelCase_ : int = window_function(window_length=self.sample_size ,name=self.win_function ,periodic=lowerCAmelCase__ ) else: lowerCAmelCase_ : Tuple = window_function(window_length=self.sample_size ,name=self.win_function ) lowerCAmelCase_ : List[str] = mel_filter_bank( num_frequency_bins=self.n_freqs ,num_mel_filters=self.feature_size ,min_frequency=0.0 ,max_frequency=self.sampling_rate / 2.0 ,sampling_rate=self.sampling_rate ,) lowerCAmelCase_ : Any = spectrogram( one_waveform * self.frame_signal_scale ,window=lowerCAmelCase__ ,frame_length=self.sample_size ,hop_length=self.sample_stride ,fft_length=self.n_fft ,center=lowerCAmelCase__ ,preemphasis=self.preemphasis_coeff ,mel_filters=lowerCAmelCase__ ,mel_floor=self.mel_floor ,log_mel="log" ,) return msfc_features.T def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : Tuple ) -> Optional[Any]: '''simple docstring''' if self.normalize_means: lowerCAmelCase_ : Optional[int] = x[:input_length].mean(axis=0 ) lowerCAmelCase_ : List[str] = np.subtract(lowerCAmelCase__ ,lowerCAmelCase__ ) if self.normalize_vars: lowerCAmelCase_ : Optional[Any] = x[:input_length].std(axis=0 ) lowerCAmelCase_ : Tuple = np.divide(lowerCAmelCase__ ,lowerCAmelCase__ ) if input_length < x.shape[0]: lowerCAmelCase_ : int = padding_value # make sure array is in float32 lowerCAmelCase_ : Any = x.astype(np.floataa ) return x def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[np.ndarray] ,lowerCAmelCase__ : Optional[np.ndarray] = None ) -> List[np.ndarray]: '''simple docstring''' lowerCAmelCase_ : List[Any] = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(lowerCAmelCase__ ,lowerCAmelCase__ ,self.padding_value ) for x, n in zip(lowerCAmelCase__ ,lowerCAmelCase__ )] def __call__( self : int ,lowerCAmelCase__ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,lowerCAmelCase__ : Union[bool, str, PaddingStrategy] = False ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : bool = False ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : Optional[bool] = None ,lowerCAmelCase__ : Optional[Union[str, TensorType]] = None ,lowerCAmelCase__ : Optional[int] = None ,**lowerCAmelCase__ : Union[str, Any] ,) -> BatchFeature: '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) lowerCAmelCase_ : List[Any] = isinstance(lowerCAmelCase__ ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase_ : str = is_batched_numpy or ( isinstance(lowerCAmelCase__ ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase_ : Tuple = [np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(lowerCAmelCase__ ,np.ndarray ): lowerCAmelCase_ : int = np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) elif isinstance(lowerCAmelCase__ ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase_ : Union[str, Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase_ : Optional[int] = [raw_speech] # extract fbank features lowerCAmelCase_ : Dict = [self._extract_mfsc_features(lowerCAmelCase__ ) for one_waveform in raw_speech] # convert into correct format for padding lowerCAmelCase_ : int = BatchFeature({"input_features": features} ) lowerCAmelCase_ : Union[str, Any] = self.pad( lowerCAmelCase__ ,padding=lowerCAmelCase__ ,max_length=lowerCAmelCase__ ,truncation=lowerCAmelCase__ ,pad_to_multiple_of=lowerCAmelCase__ ,return_attention_mask=lowerCAmelCase__ ,**lowerCAmelCase__ ,) # make sure list is in array format lowerCAmelCase_ : Optional[Any] = padded_inputs.get("input_features" ) if isinstance(input_features[0] ,lowerCAmelCase__ ): lowerCAmelCase_ : Optional[int] = [np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) for feature in input_features] lowerCAmelCase_ : List[Any] = padded_inputs.get("attention_mask" ) if attention_mask is not None: lowerCAmelCase_ : Dict = [np.asarray(lowerCAmelCase__ ,dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: lowerCAmelCase_ : Dict = ( np.array(lowerCAmelCase__ ,dtype=np.intaa ) if self._get_padding_strategies(lowerCAmelCase__ ,max_length=lowerCAmelCase__ ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) lowerCAmelCase_ : List[str] = self.normalize( padded_inputs["input_features"] ,attention_mask=lowerCAmelCase__ ) if return_tensors is not None: lowerCAmelCase_ : Dict = padded_inputs.convert_to_tensors(lowerCAmelCase__ ) return padded_inputs
659
1
from __future__ import annotations # This is the precision for this function which can be altered. # It is recommended for users to keep this number greater than or equal to 10. _lowercase = 10 def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): for i in range(snake_case__ , snake_case__): if array[i] == target: return i return -1 def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : List[str] = 0 lowerCAmelCase_ : Tuple = len(snake_case__) while left <= right: if right - left < precision: return lin_search(snake_case__ , snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : List[str] = (left + right) // 3 + 1 lowerCAmelCase_ : Tuple = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: lowerCAmelCase_ : str = one_third - 1 elif array[two_third] < target: lowerCAmelCase_ : Any = two_third + 1 else: lowerCAmelCase_ : List[str] = one_third + 1 lowerCAmelCase_ : Tuple = two_third - 1 else: return -1 def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): if left < right: if right - left < precision: return lin_search(snake_case__ , snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : Dict = (left + right) // 3 + 1 lowerCAmelCase_ : List[Any] = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: return rec_ternary_search(snake_case__ , one_third - 1 , snake_case__ , snake_case__) elif array[two_third] < target: return rec_ternary_search(two_third + 1 , snake_case__ , snake_case__ , snake_case__) else: return rec_ternary_search(one_third + 1 , two_third - 1 , snake_case__ , snake_case__) else: return -1 if __name__ == "__main__": import doctest doctest.testmod() _lowercase = input('''Enter numbers separated by comma:\n''').strip() _lowercase = [int(item.strip()) for item in user_input.split(''',''')] assert collection == sorted(collection), f"List must be ordered.\n{collection}." _lowercase = int(input('''Enter the number to be found in the list:\n''').strip()) _lowercase = ite_ternary_search(collection, target) _lowercase = rec_ternary_search(0, len(collection) - 1, collection, target) if resulta != -1: print(f"Iterative search: {target} found at positions: {resulta}") print(f"Recursive search: {target} found at positions: {resulta}") else: print('''Not found''')
659
from __future__ import annotations # This is the precision for this function which can be altered. # It is recommended for users to keep this number greater than or equal to 10. _lowercase = 10 def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): for i in range(snake_case__ , snake_case__): if array[i] == target: return i return -1 def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : List[str] = 0 lowerCAmelCase_ : Tuple = len(snake_case__) while left <= right: if right - left < precision: return lin_search(snake_case__ , snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : List[str] = (left + right) // 3 + 1 lowerCAmelCase_ : Tuple = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: lowerCAmelCase_ : str = one_third - 1 elif array[two_third] < target: lowerCAmelCase_ : Any = two_third + 1 else: lowerCAmelCase_ : List[str] = one_third + 1 lowerCAmelCase_ : Tuple = two_third - 1 else: return -1 def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): if left < right: if right - left < precision: return lin_search(snake_case__ , snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : Dict = (left + right) // 3 + 1 lowerCAmelCase_ : List[Any] = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: return rec_ternary_search(snake_case__ , one_third - 1 , snake_case__ , snake_case__) elif array[two_third] < target: return rec_ternary_search(two_third + 1 , snake_case__ , snake_case__ , snake_case__) else: return rec_ternary_search(one_third + 1 , two_third - 1 , snake_case__ , snake_case__) else: return -1 if __name__ == "__main__": import doctest doctest.testmod() _lowercase = input('''Enter numbers separated by comma:\n''').strip() _lowercase = [int(item.strip()) for item in user_input.split(''',''')] assert collection == sorted(collection), f"List must be ordered.\n{collection}." _lowercase = int(input('''Enter the number to be found in the list:\n''').strip()) _lowercase = ite_ternary_search(collection, target) _lowercase = rec_ternary_search(0, len(collection) - 1, collection, target) if resulta != -1: print(f"Iterative search: {target} found at positions: {resulta}") print(f"Recursive search: {target} found at positions: {resulta}") else: print('''Not found''')
659
1
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def UpperCamelCase ( ): lowerCAmelCase_ : Dict = HfArgumentParser(snake_case__) lowerCAmelCase_ : Dict = parser.parse_args_into_dataclasses()[0] lowerCAmelCase_ : List[Any] = TensorFlowBenchmark(args=snake_case__) try: lowerCAmelCase_ : str = parser.parse_args_into_dataclasses()[0] except ValueError as e: lowerCAmelCase_ : Optional[Any] = "Arg --no_{0} is no longer used, please use --no-{0} instead." lowerCAmelCase_ : Tuple = " ".join(str(snake_case__).split(" ")[:-1]) lowerCAmelCase_ : List[Any] = "" lowerCAmelCase_ : Optional[Any] = eval(str(snake_case__).split(" ")[-1]) lowerCAmelCase_ : List[Any] = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:]) else: wrong_args.append(snake_case__) if len(snake_case__) > 0: lowerCAmelCase_ : int = full_error_msg + begin_error_msg + str(snake_case__) raise ValueError(snake_case__) benchmark.run() if __name__ == "__main__": main()
659
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer _lowercase = logging.get_logger(__name__) _lowercase = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } _lowercase = { '''vocab_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json''' }, '''merges_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt''' }, '''tokenizer_config_file''': { '''facebook/blenderbot_small-90M''': ( '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json''' ) }, } _lowercase = { '''facebook/blenderbot_small-90M''': 512, } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = BlenderbotSmallTokenizer def __init__( self : Optional[int] ,lowerCAmelCase__ : Optional[int]=None ,lowerCAmelCase__ : Union[str, Any]=None ,lowerCAmelCase__ : Any="<|endoftext|>" ,lowerCAmelCase__ : int="<|endoftext|>" ,lowerCAmelCase__ : Optional[Any]="<|endoftext|>" ,lowerCAmelCase__ : Union[str, Any]=False ,lowerCAmelCase__ : Optional[Any]=True ,**lowerCAmelCase__ : Union[str, Any] ,) -> str: '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=lowerCAmelCase__ ,merges=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ,) ,bos_token=lowerCAmelCase__ ,eos_token=lowerCAmelCase__ ,unk_token=lowerCAmelCase__ ,**lowerCAmelCase__ ,) lowerCAmelCase_ : Dict = add_prefix_space def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Tuple=None ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : str = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : Dict = [self.sep_token_id] lowerCAmelCase_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
659
1
import collections import json import math import os import re import time from fnmatch import fnmatch from typing import Dict import requests from slack_sdk import WebClient _lowercase = WebClient(token=os.environ['''CI_SLACK_BOT_TOKEN''']) def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Optional[Any] = test_results.split(" ") lowerCAmelCase_ : List[str] = 0 lowerCAmelCase_ : Optional[int] = 0 # When the output is short enough, the output is surrounded by = signs: "== OUTPUT ==" # When it is too long, those signs are not present. lowerCAmelCase_ : Optional[int] = expressions[-2] if "=" in expressions[-1] else expressions[-1] for i, expression in enumerate(snake_case__): if "failed" in expression: failed += int(expressions[i - 1]) if "passed" in expression: success += int(expressions[i - 1]) return failed, success, time_spent def UpperCamelCase ( snake_case__): lowerCAmelCase_ : int = {} lowerCAmelCase_ : int = None lowerCAmelCase_ : Union[str, Any] = False for line in failures_short_lines.split("\n"): if re.search(R"_ \[doctest\]" , snake_case__): lowerCAmelCase_ : Optional[int] = True lowerCAmelCase_ : List[Any] = line.split(" ")[2] elif in_error and not line.split(" ")[0].isdigit(): lowerCAmelCase_ : List[str] = line lowerCAmelCase_ : Tuple = False return failures class __snake_case : """simple docstring""" def __init__( self : List[str] ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Dict ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : Tuple = title lowerCAmelCase_ : Optional[int] = doc_test_results["time_spent"].split("," )[0] lowerCAmelCase_ : Any = doc_test_results["success"] lowerCAmelCase_ : Any = doc_test_results["failures"] lowerCAmelCase_ : Union[str, Any] = self.n_success + self.n_failures # Failures and success of the modeling tests lowerCAmelCase_ : Optional[Any] = doc_test_results @property def UpperCAmelCase_ ( self : List[Any] ) -> str: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = [self._time_spent] lowerCAmelCase_ : Optional[Any] = 0 for time in time_spent: lowerCAmelCase_ : List[Any] = time.split(":" ) # Time can be formatted as xx:xx:xx, as .xx, or as x.xx if the time spent was less than a minute. if len(lowerCAmelCase__ ) == 1: lowerCAmelCase_ : str = [0, 0, time_parts[0]] lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = int(time_parts[0] ), int(time_parts[1] ), float(time_parts[2] ) total_secs += hours * 36_00 + minutes * 60 + seconds lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = total_secs // 36_00, (total_secs % 36_00) // 60, total_secs % 60 return f'''{int(lowerCAmelCase__ )}h{int(lowerCAmelCase__ )}m{int(lowerCAmelCase__ )}s''' @property def UpperCAmelCase_ ( self : Tuple ) -> Dict: '''simple docstring''' return {"type": "header", "text": {"type": "plain_text", "text": self.title}} @property def UpperCAmelCase_ ( self : List[Any] ) -> Dict: '''simple docstring''' return { "type": "section", "text": { "type": "plain_text", "text": f'''🌞 There were no failures: all {self.n_tests} tests passed. The suite ran in {self.time}.''', "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f'''https://github.com/huggingface/transformers/actions/runs/{os.environ["GITHUB_RUN_ID"]}''', }, } @property def UpperCAmelCase_ ( self : Dict ) -> Dict: '''simple docstring''' return { "type": "section", "text": { "type": "plain_text", "text": ( f'''There were {self.n_failures} failures, out of {self.n_tests} tests.\nThe suite ran in''' f''' {self.time}.''' ), "emoji": True, }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f'''https://github.com/huggingface/transformers/actions/runs/{os.environ["GITHUB_RUN_ID"]}''', }, } @property def UpperCAmelCase_ ( self : List[Any] ) -> Dict: '''simple docstring''' lowerCAmelCase_ : int = 40 lowerCAmelCase_ : List[Any] = {k: v["failed"] for k, v in doc_test_results.items() if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ )} lowerCAmelCase_ : Optional[Any] = "" for category, failures in category_failures.items(): if len(lowerCAmelCase__ ) == 0: continue if report != "": report += "\n\n" report += f'''*{category} failures*:'''.ljust(line_length // 2 ).rjust(line_length // 2 ) + "\n" report += "`" report += "`\n`".join(lowerCAmelCase__ ) report += "`" return { "type": "section", "text": { "type": "mrkdwn", "text": f'''The following examples had failures:\n\n\n{report}\n''', }, } @property def UpperCAmelCase_ ( self : int ) -> str: '''simple docstring''' lowerCAmelCase_ : Any = [self.header] if self.n_failures > 0: blocks.append(self.failures ) if self.n_failures > 0: blocks.extend([self.category_failures] ) if self.n_failures == 0: blocks.append(self.no_failures ) return json.dumps(lowerCAmelCase__ ) @staticmethod def UpperCAmelCase_ ( ) -> str: '''simple docstring''' lowerCAmelCase_ : Optional[int] = [ { "type": "section", "text": { "type": "plain_text", "text": "There was an issue running the tests.", }, "accessory": { "type": "button", "text": {"type": "plain_text", "text": "Check Action results", "emoji": True}, "url": f'''https://github.com/huggingface/transformers/actions/runs/{os.environ["GITHUB_RUN_ID"]}''', }, } ] print("Sending the following payload" ) print(json.dumps({"blocks": json.loads(lowerCAmelCase__ )} ) ) client.chat_postMessage( channel=os.environ["CI_SLACK_CHANNEL_ID_DAILY"] ,text="There was an issue running the tests." ,blocks=lowerCAmelCase__ ,) def UpperCAmelCase_ ( self : Tuple ) -> Tuple: '''simple docstring''' print("Sending the following payload" ) print(json.dumps({"blocks": json.loads(self.payload )} ) ) lowerCAmelCase_ : List[Any] = f'''{self.n_failures} failures out of {self.n_tests} tests,''' if self.n_failures else "All tests passed." lowerCAmelCase_ : Optional[int] = client.chat_postMessage( channel=os.environ["CI_SLACK_CHANNEL_ID_DAILY"] ,blocks=self.payload ,text=lowerCAmelCase__ ,) def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Optional[int] ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : int ) -> Dict: '''simple docstring''' lowerCAmelCase_ : List[Any] = "" for key, value in failures.items(): lowerCAmelCase_ : Tuple = value[:2_00] + " [Truncated]" if len(lowerCAmelCase__ ) > 2_50 else value failures_text += f'''*{key}*\n_{value}_\n\n''' lowerCAmelCase_ : Any = job_name lowerCAmelCase_ : List[Any] = {"type": "section", "text": {"type": "mrkdwn", "text": text}} if job_link is not None: lowerCAmelCase_ : str = { "type": "button", "text": {"type": "plain_text", "text": "GitHub Action job", "emoji": True}, "url": job_link, } return [ {"type": "header", "text": {"type": "plain_text", "text": title.upper(), "emoji": True}}, content, {"type": "section", "text": {"type": "mrkdwn", "text": failures_text}}, ] def UpperCAmelCase_ ( self : Tuple ) -> List[Any]: '''simple docstring''' if self.thread_ts is None: raise ValueError("Can only post reply if a post has been made." ) lowerCAmelCase_ : List[Any] = self.doc_test_results.pop("job_link" ) self.doc_test_results.pop("failures" ) self.doc_test_results.pop("success" ) self.doc_test_results.pop("time_spent" ) lowerCAmelCase_ : Tuple = sorted(self.doc_test_results.items() ,key=lambda lowerCAmelCase__ : t[0] ) for job, job_result in sorted_dict: if len(job_result["failures"] ): lowerCAmelCase_ : Dict = f'''*Num failures* :{len(job_result["failed"] )} \n''' lowerCAmelCase_ : Optional[int] = job_result["failures"] lowerCAmelCase_ : List[Any] = self.get_reply_blocks(lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,text=lowerCAmelCase__ ) print("Sending the following reply" ) print(json.dumps({"blocks": blocks} ) ) client.chat_postMessage( channel=os.environ["CI_SLACK_CHANNEL_ID_DAILY"] ,text=f'''Results for {job}''' ,blocks=lowerCAmelCase__ ,thread_ts=self.thread_ts["ts"] ,) time.sleep(1 ) def UpperCamelCase ( ): lowerCAmelCase_ : int = os.environ["GITHUB_RUN_ID"] lowerCAmelCase_ : str = F'''https://api.github.com/repos/huggingface/transformers/actions/runs/{run_id}/jobs?per_page=100''' lowerCAmelCase_ : Optional[int] = requests.get(snake_case__).json() lowerCAmelCase_ : Union[str, Any] = {} try: jobs.update({job["name"]: job["html_url"] for job in result["jobs"]}) lowerCAmelCase_ : List[Any] = math.ceil((result["total_count"] - 1_00) / 1_00) for i in range(snake_case__): lowerCAmelCase_ : Any = requests.get(url + F'''&page={i + 2}''').json() jobs.update({job["name"]: job["html_url"] for job in result["jobs"]}) return jobs except Exception as e: print("Unknown error, could not fetch links." , snake_case__) return {} def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Optional[Any] = {} if os.path.exists(snake_case__): lowerCAmelCase_ : str = os.listdir(snake_case__) for file in files: try: with open(os.path.join(snake_case__ , snake_case__) , encoding="utf-8") as f: lowerCAmelCase_ : Tuple = f.read() except UnicodeDecodeError as e: raise ValueError(F'''Could not open {os.path.join(snake_case__ , snake_case__)}.''') from e return _artifact def UpperCamelCase ( ): class __snake_case : """simple docstring""" def __init__( self : List[str] ,lowerCAmelCase__ : str ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = name lowerCAmelCase_ : Tuple = [] def __str__( self : Optional[int] ) -> Dict: '''simple docstring''' return self.name def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : str ) -> Dict: '''simple docstring''' self.paths.append({"name": self.name, "path": path} ) lowerCAmelCase_ : Dict[str, Artifact] = {} lowerCAmelCase_ : List[Any] = filter(os.path.isdir , os.listdir()) for directory in directories: lowerCAmelCase_ : Union[str, Any] = directory if artifact_name not in _available_artifacts: lowerCAmelCase_ : int = Artifact(snake_case__) _available_artifacts[artifact_name].add_path(snake_case__) return _available_artifacts if __name__ == "__main__": _lowercase = get_job_links() _lowercase = retrieve_available_artifacts() _lowercase = collections.OrderedDict( [ ('''*.py''', '''API Examples'''), ('''*.md''', '''MD Examples'''), ] ) # This dict will contain all the information relative to each doc test category: # - failed: list of failed tests # - failures: dict in the format 'test': 'error_message' _lowercase = { v: { '''failed''': [], '''failures''': {}, } for v in docs.values() } # Link to the GitHub Action job _lowercase = github_actions_job_links.get('''run_doctests''') _lowercase = available_artifacts['''doc_tests_gpu_test_reports'''].paths[0] _lowercase = retrieve_artifact(artifact_path['''name''']) if "stats" in artifact: _lowercase , _lowercase , _lowercase = handle_test_results(artifact['''stats''']) _lowercase = failed _lowercase = success _lowercase = time_spent[1:-1] + ''', ''' _lowercase = extract_first_line_failure(artifact['''failures_short''']) for line in artifact["summary_short"].split('''\n'''): if re.search('''FAILED''', line): _lowercase = line.replace('''FAILED ''', '''''') _lowercase = line.split()[0].replace('''\n''', '''''') if "::" in line: _lowercase , _lowercase = line.split('''::''') else: _lowercase , _lowercase = line, line for file_regex in docs.keys(): if fnmatch(file_path, file_regex): _lowercase = docs[file_regex] doc_test_results[category]["failed"].append(test) _lowercase = all_failures[test] if test in all_failures else '''N/A''' _lowercase = failure break _lowercase = Message('''🤗 Results of the doc tests.''', doc_test_results) message.post() message.post_reply()
659
from collections.abc import Generator from math import sin def UpperCamelCase ( snake_case__): if len(snake_case__) != 32: raise ValueError("Input must be of length 32") lowerCAmelCase_ : Tuple = b"" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def UpperCamelCase ( snake_case__): if i < 0: raise ValueError("Input must be non-negative") lowerCAmelCase_ : List[str] = format(snake_case__ , "08x")[-8:] lowerCAmelCase_ : Any = b"" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("utf-8") return little_endian_hex def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Union[str, Any] = b"" for char in message: bit_string += format(snake_case__ , "08b").encode("utf-8") lowerCAmelCase_ : Optional[int] = format(len(snake_case__) , "064b").encode("utf-8") # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(snake_case__) % 5_12 != 4_48: bit_string += b"0" bit_string += to_little_endian(start_len[32:]) + to_little_endian(start_len[:32]) return bit_string def UpperCamelCase ( snake_case__): if len(snake_case__) % 5_12 != 0: raise ValueError("Input must have length that's a multiple of 512") for pos in range(0 , len(snake_case__) , 5_12): lowerCAmelCase_ : List[str] = bit_string[pos : pos + 5_12] lowerCAmelCase_ : Union[str, Any] = [] for i in range(0 , 5_12 , 32): block_words.append(int(to_little_endian(block[i : i + 32]) , 2)) yield block_words def UpperCamelCase ( snake_case__): if i < 0: raise ValueError("Input must be non-negative") lowerCAmelCase_ : Dict = format(snake_case__ , "032b") lowerCAmelCase_ : str = "" for c in i_str: new_str += "1" if c == "0" else "0" return int(snake_case__ , 2) def UpperCamelCase ( snake_case__ , snake_case__): return (a + b) % 2**32 def UpperCamelCase ( snake_case__ , snake_case__): if i < 0: raise ValueError("Input must be non-negative") if shift < 0: raise ValueError("Shift must be non-negative") return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Optional[Any] = preprocess(snake_case__) lowerCAmelCase_ : Optional[Any] = [int(2**32 * abs(sin(i + 1))) for i in range(64)] # Starting states lowerCAmelCase_ : List[str] = 0x67_45_23_01 lowerCAmelCase_ : Union[str, Any] = 0xef_cd_ab_89 lowerCAmelCase_ : List[Any] = 0x98_ba_dc_fe lowerCAmelCase_ : Tuple = 0x10_32_54_76 lowerCAmelCase_ : Any = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(snake_case__): lowerCAmelCase_ : Optional[int] = aa lowerCAmelCase_ : List[str] = ba lowerCAmelCase_ : Any = ca lowerCAmelCase_ : Union[str, Any] = da # Hash current chunk for i in range(64): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f lowerCAmelCase_ : Any = d ^ (b & (c ^ d)) lowerCAmelCase_ : Dict = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f lowerCAmelCase_ : Any = c ^ (d & (b ^ c)) lowerCAmelCase_ : List[str] = (5 * i + 1) % 16 elif i <= 47: lowerCAmelCase_ : int = b ^ c ^ d lowerCAmelCase_ : Optional[Any] = (3 * i + 5) % 16 else: lowerCAmelCase_ : List[Any] = c ^ (b | not_aa(snake_case__)) lowerCAmelCase_ : List[Any] = (7 * i) % 16 lowerCAmelCase_ : Optional[Any] = (f + a + added_consts[i] + block_words[g]) % 2**32 lowerCAmelCase_ : Optional[Any] = d lowerCAmelCase_ : Dict = c lowerCAmelCase_ : List[str] = b lowerCAmelCase_ : Any = sum_aa(snake_case__ , left_rotate_aa(snake_case__ , shift_amounts[i])) # Add hashed chunk to running total lowerCAmelCase_ : Dict = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : str = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : Optional[int] = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : int = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : Union[str, Any] = reformat_hex(snake_case__) + reformat_hex(snake_case__) + reformat_hex(snake_case__) + reformat_hex(snake_case__) return digest if __name__ == "__main__": import doctest doctest.testmod()
659
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _lowercase = { '''configuration_bloom''': ['''BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BloomConfig''', '''BloomOnnxConfig'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ['''BloomTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ '''BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BloomForCausalLM''', '''BloomModel''', '''BloomPreTrainedModel''', '''BloomForSequenceClassification''', '''BloomForTokenClassification''', '''BloomForQuestionAnswering''', ] if TYPE_CHECKING: from .configuration_bloom import BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP, BloomConfig, BloomOnnxConfig try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bloom_fast import BloomTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bloom import ( BLOOM_PRETRAINED_MODEL_ARCHIVE_LIST, BloomForCausalLM, BloomForQuestionAnswering, BloomForSequenceClassification, BloomForTokenClassification, BloomModel, BloomPreTrainedModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
659
import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): _lowercase = True from torch.cuda.amp import autocast _lowercase = logging.getLogger(__name__) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Whether to log verbose messages or not.'} , ) UpperCamelCase_ = field( default=2.0 , metadata={'help': 'Maximum temperature for gumbel softmax.'} ) UpperCamelCase_ = field( default=0.5 , metadata={'help': 'Minimum temperature for gumbel softmax.'} ) UpperCamelCase_ = field( default=0.99_99_95 , metadata={'help': 'Decay of gumbel temperature during training.'} ) def UpperCamelCase ( snake_case__ , snake_case__): logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout)] , ) lowerCAmelCase_ : str = logging.WARNING if model_args.verbose_logging: lowerCAmelCase_ : int = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank): lowerCAmelCase_ : Any = logging.INFO logger.setLevel(snake_case__) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The name of the dataset to use (via the datasets library).'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) UpperCamelCase_ = field( default='train' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) UpperCamelCase_ = field( default='validation' , metadata={ 'help': ( 'The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'' ) } , ) UpperCamelCase_ = field( default='file' , metadata={'help': 'Column in the dataset that contains speech file path. Defaults to \'file\''} , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Overwrite the cached preprocessed datasets or not.'} ) UpperCamelCase_ = field( default=1 , metadata={ 'help': 'The percentage of the train set used as validation set in case there\'s no validation split' } , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) UpperCamelCase_ = field( default=20.0 , metadata={'help': 'Filter audio files that are longer than `max_duration_in_seconds` seconds'} ) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = 42 UpperCamelCase_ = 42 UpperCamelCase_ = "longest" UpperCamelCase_ = None UpperCamelCase_ = None def __call__( self : str ,lowerCAmelCase__ : List[Dict[str, Union[List[int], torch.Tensor]]] ) -> Dict[str, torch.Tensor]: '''simple docstring''' lowerCAmelCase_ : Tuple = self.feature_extractor.pad( lowerCAmelCase__ ,max_length=self.max_length ,padding=self.padding ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors="pt" ,) lowerCAmelCase_ : Union[str, Any] = self.model._get_feat_extract_output_lengths(batch["input_values"].shape[-1] ) lowerCAmelCase_ : List[str] = batch["input_values"].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula lowerCAmelCase_ : Tuple = self.model._get_feat_extract_output_lengths(batch["attention_mask"].sum(-1 ) ).to( torch.long ) lowerCAmelCase_ : Optional[Any] = torch.zeros( (batch_size, mask_indices_seq_length) ,dtype=torch.long ,device=batch["input_values"].device ) # these two operations makes sure that all values # before the output lengths indices are attended to lowerCAmelCase_ : Tuple = 1 lowerCAmelCase_ : int = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices lowerCAmelCase_ : str = _compute_mask_indices( (batch_size, mask_indices_seq_length) ,self.model.config.mask_time_prob ,self.model.config.mask_time_length ,attention_mask=lowerCAmelCase__ ,min_masks=2 ,) return batch class __snake_case ( snake_case__ ): """simple docstring""" def __init__( self : List[str] ,*lowerCAmelCase__ : Optional[int] ,lowerCAmelCase__ : Tuple=1 ,lowerCAmelCase__ : Optional[int]=0 ,lowerCAmelCase__ : Optional[Any]=1.0 ,**lowerCAmelCase__ : Any ) -> str: '''simple docstring''' super().__init__(*lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = 0 lowerCAmelCase_ : int = max_gumbel_temp lowerCAmelCase_ : Union[str, Any] = min_gumbel_temp lowerCAmelCase_ : str = gumbel_temp_decay def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : nn.Module ,lowerCAmelCase__ : Dict[str, Union[torch.Tensor, Any]] ) -> torch.Tensor: '''simple docstring''' model.train() lowerCAmelCase_ : str = self._prepare_inputs(lowerCAmelCase__ ) if self.use_amp: with autocast(): lowerCAmelCase_ : List[Any] = self.compute_loss(lowerCAmelCase__ ,lowerCAmelCase__ ) else: lowerCAmelCase_ : List[Any] = self.compute_loss(lowerCAmelCase__ ,lowerCAmelCase__ ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": lowerCAmelCase_ : List[Any] = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": lowerCAmelCase_ : Optional[Any] = loss.sum() / (inputs["mask_time_indices"]).sum() else: raise ValueError(f'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: lowerCAmelCase_ : int = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(lowerCAmelCase__ ).backward() elif self.use_apex: with amp.scale_loss(lowerCAmelCase__ ,self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(lowerCAmelCase__ ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) return loss.detach() def UpperCamelCase ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase_ : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments)) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = parser.parse_args_into_dataclasses() configure_logger(snake_case__ , snake_case__) # Downloading and loading a dataset from the hub. lowerCAmelCase_ : List[str] = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" lowerCAmelCase_ : Any = DatasetDict() lowerCAmelCase_ : Union[str, Any] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : List[str] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" lowerCAmelCase_ : Union[str, Any] = DatasetDict() lowerCAmelCase_ : int = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split="validation" , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : Any = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported lowerCAmelCase_ : Dict = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=snake_case__) def prepare_dataset(snake_case__): # check that all files have the correct sampling rate lowerCAmelCase_ , lowerCAmelCase_ : str = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate) return batch # load audio files into numpy arrays lowerCAmelCase_ : int = datasets.map( snake_case__ , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets["train"].column_names) # filter audio files that are too long lowerCAmelCase_ : int = vectorized_datasets.filter( lambda snake_case__: len(data["speech"]) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate)) def normalize(snake_case__): return feature_extractor(batch["speech"] , sampling_rate=feature_extractor.sampling_rate) # normalize and transform to `BatchFeatures` lowerCAmelCase_ : str = vectorized_datasets.map( snake_case__ , batched=snake_case__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets["train"].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 lowerCAmelCase_ : Optional[Any] = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( "PreTraining is only supported for ``config.do_stable_layer_norm=True`` and" " ``config.feat_extract_norm='layer'") lowerCAmelCase_ : Dict = WavaVecaForPreTraining(snake_case__) lowerCAmelCase_ : int = DataCollatorForWavaVecaPretraining(model=snake_case__ , feature_extractor=snake_case__) lowerCAmelCase_ : List[Any] = WavaVecaPreTrainer( model=snake_case__ , data_collator=snake_case__ , args=snake_case__ , train_dataset=vectorized_datasets["train"] , eval_dataset=vectorized_datasets["validation"] , tokenizer=snake_case__ , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
659
1
from __future__ import annotations from typing import Any class __snake_case : """simple docstring""" def __init__( self : int ,lowerCAmelCase__ : int ,lowerCAmelCase__ : int ,lowerCAmelCase__ : float = 0 ) -> None: '''simple docstring''' lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = row, column lowerCAmelCase_ : str = [[default_value for c in range(lowerCAmelCase__ )] for r in range(lowerCAmelCase__ )] def __str__( self : Tuple ) -> str: '''simple docstring''' lowerCAmelCase_ : Tuple = f'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier lowerCAmelCase_ : str = 0 for row_vector in self.array: for obj in row_vector: lowerCAmelCase_ : Any = max(lowerCAmelCase__ ,len(str(lowerCAmelCase__ ) ) ) lowerCAmelCase_ : Tuple = f'''%{max_element_length}s''' # Make string and return def single_line(lowerCAmelCase__ : list[float] ) -> str: nonlocal string_format_identifier lowerCAmelCase_ : List[Any] = "[" line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowerCAmelCase__ ) for row_vector in self.array ) return s def __repr__( self : Dict ) -> str: '''simple docstring''' return str(self ) def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : tuple[int, int] ) -> bool: '''simple docstring''' if not (isinstance(lowerCAmelCase__ ,(list, tuple) ) and len(lowerCAmelCase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : List[str] ,lowerCAmelCase__ : tuple[int, int] ) -> Any: '''simple docstring''' assert self.validate_indicies(lowerCAmelCase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : List[str] ,lowerCAmelCase__ : tuple[int, int] ,lowerCAmelCase__ : float ) -> None: '''simple docstring''' assert self.validate_indicies(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = value def __add__( self : List[str] ,lowerCAmelCase__ : Matrix ) -> Matrix: '''simple docstring''' assert isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) assert self.row == another.row and self.column == another.column # Add lowerCAmelCase_ : Tuple = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase_ : Union[str, Any] = self[r, c] + another[r, c] return result def __neg__( self : Union[str, Any] ) -> Matrix: '''simple docstring''' lowerCAmelCase_ : Any = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase_ : List[Any] = -self[r, c] return result def __sub__( self : int ,lowerCAmelCase__ : Matrix ) -> Matrix: '''simple docstring''' return self + (-another) def __mul__( self : Any ,lowerCAmelCase__ : int | float | Matrix ) -> Matrix: '''simple docstring''' if isinstance(lowerCAmelCase__ ,(int, float) ): # Scalar multiplication lowerCAmelCase_ : Tuple = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase_ : Tuple = self[r, c] * another return result elif isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ): # Matrix multiplication assert self.column == another.row lowerCAmelCase_ : Union[str, Any] = Matrix(self.row ,another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: lowerCAmelCase_ : int = f'''Unsupported type given for another ({type(lowerCAmelCase__ )})''' raise TypeError(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Dict ) -> Matrix: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = Matrix(self.column ,self.row ) for r in range(self.row ): for c in range(self.column ): lowerCAmelCase_ : List[str] = self[r, c] return result def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Matrix ,lowerCAmelCase__ : Matrix ) -> Any: '''simple docstring''' assert isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) and isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate lowerCAmelCase_ : Optional[int] = v.transpose() lowerCAmelCase_ : int = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def UpperCamelCase ( ): # a^(-1) lowerCAmelCase_ : str = Matrix(3 , 3 , 0) for i in range(3): lowerCAmelCase_ : Optional[Any] = 1 print(F'''a^(-1) is {ainv}''') # u, v lowerCAmelCase_ : Optional[Any] = Matrix(3 , 1 , 0) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = 1, 2, -3 lowerCAmelCase_ : Tuple = Matrix(3 , 1 , 0) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = 4, -2, 5 print(F'''u is {u}''') print(F'''v is {v}''') print(F'''uv^T is {u * v.transpose()}''') # Sherman Morrison print(F'''(a + uv^T)^(-1) is {ainv.sherman_morrison(snake_case__ , snake_case__)}''') def UpperCamelCase ( ): import doctest doctest.testmod() testa()
659
from __future__ import annotations from collections.abc import Callable def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1_00 , ): lowerCAmelCase_ : Any = x_start lowerCAmelCase_ : Optional[Any] = fnc(snake_case__) lowerCAmelCase_ : Union[str, Any] = 0.0 for _ in range(snake_case__): # Approximates small segments of curve as linear and solve # for trapezoidal area lowerCAmelCase_ : Any = (x_end - x_start) / steps + xa lowerCAmelCase_ : Dict = fnc(snake_case__) area += abs(fxa + fxa) * (xa - xa) / 2 # Increment step lowerCAmelCase_ : int = xa lowerCAmelCase_ : str = fxa return area if __name__ == "__main__": def UpperCamelCase ( snake_case__): return x**3 + x**2 print('''f(x) = x^3 + x^2''') print('''The area between the curve, x = -5, x = 5 and the x axis is:''') _lowercase = 10 while i <= 100000: print(f"with {i} steps: {trapezoidal_area(f, -5, 5, i)}") i *= 10
659
1
import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = VideoToVideoSDPipeline UpperCamelCase_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({'video'} ) - {'image', 'width', 'height'} UpperCamelCase_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'video'} ) - {'image'} UpperCamelCase_ = PipelineTesterMixin.required_optional_params - {'latents'} UpperCamelCase_ = False # No `output_type`. UpperCamelCase_ = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> List[str]: '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase_ : Tuple = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "DownBlock3D") ,up_block_types=("UpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D") ,cross_attention_dim=32 ,attention_head_dim=4 ,) lowerCAmelCase_ : Optional[Any] = DDIMScheduler( beta_start=0.00_085 ,beta_end=0.012 ,beta_schedule="scaled_linear" ,clip_sample=lowerCAmelCase__ ,set_alpha_to_one=lowerCAmelCase__ ,) torch.manual_seed(0 ) lowerCAmelCase_ : List[Any] = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=3 ,out_channels=3 ,down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] ,up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] ,latent_channels=4 ,sample_size=1_28 ,) torch.manual_seed(0 ) lowerCAmelCase_ : Optional[int] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=10_00 ,hidden_act="gelu" ,projection_dim=5_12 ,) lowerCAmelCase_ : Union[str, Any] = CLIPTextModel(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) lowerCAmelCase_ : Any = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, } return components def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : List[Any]=0 ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : Dict = floats_tensor((1, 3, 3, 32, 32) ,rng=random.Random(lowerCAmelCase__ ) ).to(lowerCAmelCase__ ) if str(lowerCAmelCase__ ).startswith("mps" ): lowerCAmelCase_ : Any = torch.manual_seed(lowerCAmelCase__ ) else: lowerCAmelCase_ : Optional[Any] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = { "prompt": "A painting of a squirrel eating a burger", "video": video, "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "pt", } return inputs def UpperCAmelCase_ ( self : List[str] ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : List[str] = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : List[Any] = self.get_dummy_components() lowerCAmelCase_ : List[Any] = VideoToVideoSDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = sd_pipe.to(lowerCAmelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = "np" lowerCAmelCase_ : Tuple = sd_pipe(**lowerCAmelCase__ ).frames lowerCAmelCase_ : Tuple = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) lowerCAmelCase_ : Optional[int] = np.array([1_06, 1_17, 1_13, 1_74, 1_37, 1_12, 1_48, 1_51, 1_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() ,reason="XFormers attention is only available with CUDA and `xformers` installed" ,) def UpperCAmelCase_ ( self : Any ) -> str: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCAmelCase__ ,expected_max_diff=5e-3 ) @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def UpperCAmelCase_ ( self : Dict ) -> Dict: '''simple docstring''' pass @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def UpperCAmelCase_ ( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' pass @unittest.skip(reason="`num_images_per_prompt` argument is not supported for this pipeline." ) def UpperCAmelCase_ ( self : Dict ) -> Dict: '''simple docstring''' pass def UpperCAmelCase_ ( self : int ) -> int: '''simple docstring''' return super().test_progress_bar() @slow @skip_mps class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : Optional[int] ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : Tuple = VideoToVideoSDPipeline.from_pretrained("cerspense/zeroscope_v2_XL" ,torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames lowerCAmelCase_ : Any = torch.Generator(device="cpu" ).manual_seed(0 ) lowerCAmelCase_ : Optional[Any] = torch.randn((1, 10, 3, 10_24, 5_76) ,generator=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = video.to("cuda" ) lowerCAmelCase_ : Optional[int] = "Spiderman is surfing" lowerCAmelCase_ : Tuple = pipe(lowerCAmelCase__ ,video=lowerCAmelCase__ ,generator=lowerCAmelCase__ ,num_inference_steps=3 ,output_type="pt" ).frames lowerCAmelCase_ : Optional[int] = np.array([-1.0_458_984, -1.1_279_297, -0.9_663_086, -0.91_503_906, -0.75_097_656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1e-2
659
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, PNDMScheduler, StableDiffusionLDMaDPipeline, UNetaDConditionModel, ) from diffusers.utils import nightly, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS enable_full_determinism() class __snake_case ( unittest.TestCase ): """simple docstring""" UpperCamelCase_ = StableDiffusionLDMaDPipeline UpperCamelCase_ = TEXT_TO_IMAGE_PARAMS UpperCamelCase_ = TEXT_TO_IMAGE_BATCH_PARAMS UpperCamelCase_ = TEXT_TO_IMAGE_IMAGE_PARAMS def UpperCAmelCase_ ( self : Tuple ) -> str: '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase_ : Optional[Any] = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") ,up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") ,cross_attention_dim=32 ,) lowerCAmelCase_ : Any = DDIMScheduler( beta_start=0.00_085 ,beta_end=0.012 ,beta_schedule="scaled_linear" ,clip_sample=lowerCAmelCase__ ,set_alpha_to_one=lowerCAmelCase__ ,) torch.manual_seed(0 ) lowerCAmelCase_ : str = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=6 ,out_channels=6 ,down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] ,up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] ,latent_channels=4 ,) torch.manual_seed(0 ) lowerCAmelCase_ : Optional[Any] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=10_00 ,) lowerCAmelCase_ : Optional[int] = CLIPTextModel(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) lowerCAmelCase_ : List[Any] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : List[str]=0 ) -> Dict: '''simple docstring''' if str(lowerCAmelCase__ ).startswith("mps" ): lowerCAmelCase_ : Optional[int] = torch.manual_seed(lowerCAmelCase__ ) else: lowerCAmelCase_ : Dict = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : str = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : Any ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Dict = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : List[str] = self.get_dummy_components() lowerCAmelCase_ : Union[str, Any] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Any = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Any = output.rgb, output.depth lowerCAmelCase_ : Dict = rgb[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) lowerCAmelCase_ : Optional[Any] = np.array( [0.37_338_176, 0.70_247, 0.74_203_193, 0.51_643_604, 0.58_256_793, 0.60_932_136, 0.4_181_095, 0.48_355_877, 0.46_535_262] ) lowerCAmelCase_ : Tuple = np.array([103.46_727, 85.812_004, 87.849_236] ) assert np.abs(image_slice_rgb.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(image_slice_depth.flatten() - expected_slice_depth ).max() < 1e-2 def UpperCAmelCase_ ( self : int ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Dict = self.get_dummy_components() lowerCAmelCase_ : List[str] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : str = 3 * [inputs["prompt"]] # forward lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = output.rgb, output.depth lowerCAmelCase_ : str = rgb_slice_a[0, -3:, -3:, -1] lowerCAmelCase_ : List[str] = depth_slice_a[0, -3:, -1] lowerCAmelCase_ : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = 3 * [inputs.pop("prompt" )] lowerCAmelCase_ : str = ldmad_pipe.tokenizer( lowerCAmelCase__ ,padding="max_length" ,max_length=ldmad_pipe.tokenizer.model_max_length ,truncation=lowerCAmelCase__ ,return_tensors="pt" ,) lowerCAmelCase_ : Union[str, Any] = text_inputs["input_ids"].to(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = ldmad_pipe.text_encoder(lowerCAmelCase__ )[0] lowerCAmelCase_ : Optional[int] = prompt_embeds # forward lowerCAmelCase_ : str = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : str = output.rgb, output.depth lowerCAmelCase_ : Optional[Any] = rgb_slice_a[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth_slice_a[0, -3:, -1] assert np.abs(rgb_slice_a.flatten() - rgb_slice_a.flatten() ).max() < 1e-4 assert np.abs(depth_slice_a.flatten() - depth_slice_a.flatten() ).max() < 1e-4 def UpperCAmelCase_ ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : Any = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : Optional[int] = self.get_dummy_components() lowerCAmelCase_ : Dict = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : Any = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = "french fries" lowerCAmelCase_ : Optional[int] = ldmad_pipe(**lowerCAmelCase__ ,negative_prompt=lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = output.rgb, output.depth lowerCAmelCase_ : Any = rgb[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) lowerCAmelCase_ : int = np.array( [0.37_044, 0.71_811_503, 0.7_223_251, 0.48_603_675, 0.5_638_391, 0.6_364_948, 0.42_833_704, 0.4_901_315, 0.47_926_217] ) lowerCAmelCase_ : Union[str, Any] = np.array([107.84_738, 84.62_802, 89.962_135] ) assert np.abs(rgb_slice.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(depth_slice.flatten() - expected_slice_depth ).max() < 1e-2 @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Dict="cpu" ,lowerCAmelCase__ : Union[str, Any]=torch.floataa ,lowerCAmelCase__ : List[str]=0 ) -> int: '''simple docstring''' lowerCAmelCase_ : Any = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = np.random.RandomState(lowerCAmelCase__ ).standard_normal((1, 4, 64, 64) ) lowerCAmelCase_ : Optional[Any] = torch.from_numpy(lowerCAmelCase__ ).to(device=lowerCAmelCase__ ,dtype=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : List[Any] ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ) lowerCAmelCase_ : List[str] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Dict = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Dict = output.rgb, output.depth lowerCAmelCase_ : List[str] = rgb[0, -3:, -3:, -1].flatten() lowerCAmelCase_ : Optional[int] = rgb[0, -3:, -1].flatten() assert rgb.shape == (1, 5_12, 5_12, 3) assert depth.shape == (1, 5_12, 5_12) lowerCAmelCase_ : int = np.array( [0.53_805_465, 0.56_707_305, 0.5_486_515, 0.57_012_236, 0.5_814_511, 0.56_253_487, 0.54_843_014, 0.55_092_263, 0.6_459_706] ) lowerCAmelCase_ : Optional[Any] = np.array( [0.9_263_781, 0.6_678_672, 0.5_486_515, 0.92_202_145, 0.67_831_135, 0.56_253_487, 0.9_241_694, 0.7_551_478, 0.6_459_706] ) assert np.abs(rgb_slice - expected_slice_rgb ).max() < 3e-3 assert np.abs(depth_slice - expected_slice_depth ).max() < 3e-3 @nightly @require_torch_gpu class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Dict="cpu" ,lowerCAmelCase__ : List[str]=torch.floataa ,lowerCAmelCase__ : Optional[int]=0 ) -> int: '''simple docstring''' lowerCAmelCase_ : Dict = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = np.random.RandomState(lowerCAmelCase__ ).standard_normal((1, 4, 64, 64) ) lowerCAmelCase_ : Any = torch.from_numpy(lowerCAmelCase__ ).to(device=lowerCAmelCase__ ,dtype=lowerCAmelCase__ ) lowerCAmelCase_ : int = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 50, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : Dict ) -> int: '''simple docstring''' lowerCAmelCase_ : List[Any] = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ).to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Any = output.rgb, output.depth lowerCAmelCase_ : Dict = 0.495_586 lowerCAmelCase_ : Optional[Any] = 0.33_795_515 lowerCAmelCase_ : Any = 112.48_518 lowerCAmelCase_ : List[Any] = 98.489_746 assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3 def UpperCAmelCase_ ( self : Tuple ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : int = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d-4c" ).to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : str = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Tuple = output.rgb, output.depth lowerCAmelCase_ : List[str] = 0.4_194_127 lowerCAmelCase_ : List[str] = 0.35_375_586 lowerCAmelCase_ : str = 0.5_638_502 lowerCAmelCase_ : Optional[Any] = 0.34_686_103 assert rgb.shape == (1, 5_12, 5_12, 3) assert depth.shape == (1, 5_12, 5_12, 1) assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3
659
1
from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING _lowercase = logging.get_logger(__name__) @add_end_docstrings(snake_case__ ) class __snake_case ( snake_case__ ): """simple docstring""" def __init__( self : Optional[int] ,**lowerCAmelCase__ : Union[str, Any] ) -> Dict: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) requires_backends(self ,"vision" ) requires_backends(self ,"torch" ) if self.framework != "pt": raise ValueError(f'''The {self.__class__} is only available in PyTorch.''' ) self.check_model_type(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Dict ,**lowerCAmelCase__ : str ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : Any = {} lowerCAmelCase_ : Optional[Any] = {} lowerCAmelCase_ : Optional[int] = {} # preprocess args if "points_per_batch" in kwargs: lowerCAmelCase_ : List[Any] = kwargs["points_per_batch"] if "points_per_crop" in kwargs: lowerCAmelCase_ : Union[str, Any] = kwargs["points_per_crop"] if "crops_n_layers" in kwargs: lowerCAmelCase_ : str = kwargs["crops_n_layers"] if "crop_overlap_ratio" in kwargs: lowerCAmelCase_ : Union[str, Any] = kwargs["crop_overlap_ratio"] if "crop_n_points_downscale_factor" in kwargs: lowerCAmelCase_ : int = kwargs["crop_n_points_downscale_factor"] # postprocess args if "pred_iou_thresh" in kwargs: lowerCAmelCase_ : Dict = kwargs["pred_iou_thresh"] if "stability_score_offset" in kwargs: lowerCAmelCase_ : str = kwargs["stability_score_offset"] if "mask_threshold" in kwargs: lowerCAmelCase_ : List[Any] = kwargs["mask_threshold"] if "stability_score_thresh" in kwargs: lowerCAmelCase_ : Dict = kwargs["stability_score_thresh"] if "crops_nms_thresh" in kwargs: lowerCAmelCase_ : List[str] = kwargs["crops_nms_thresh"] if "output_rle_mask" in kwargs: lowerCAmelCase_ : List[str] = kwargs["output_rle_mask"] if "output_bboxes_mask" in kwargs: lowerCAmelCase_ : Optional[int] = kwargs["output_bboxes_mask"] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self : str ,lowerCAmelCase__ : Optional[int] ,*lowerCAmelCase__ : Any ,lowerCAmelCase__ : Optional[int]=None ,lowerCAmelCase__ : Dict=None ,**lowerCAmelCase__ : Optional[Any] ) -> Tuple: '''simple docstring''' return super().__call__(lowerCAmelCase__ ,*lowerCAmelCase__ ,num_workers=lowerCAmelCase__ ,batch_size=lowerCAmelCase__ ,**lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Union[str, Any]=64 ,lowerCAmelCase__ : int = 0 ,lowerCAmelCase__ : float = 5_12 / 15_00 ,lowerCAmelCase__ : Optional[int] = 32 ,lowerCAmelCase__ : Optional[int] = 1 ,) -> Dict: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = load_image(lowerCAmelCase__ ) lowerCAmelCase_ : int = self.image_processor.size["longest_edge"] lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = self.image_processor.generate_crop_boxes( lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = self.image_processor(images=lowerCAmelCase__ ,return_tensors="pt" ) with self.device_placement(): if self.framework == "pt": lowerCAmelCase_ : Union[str, Any] = self.get_inference_context() with inference_context(): lowerCAmelCase_ : Any = self._ensure_tensor_on_device(lowerCAmelCase__ ,device=self.device ) lowerCAmelCase_ : Optional[Any] = self.model.get_image_embeddings(model_inputs.pop("pixel_values" ) ) lowerCAmelCase_ : Any = image_embeddings lowerCAmelCase_ : int = grid_points.shape[1] lowerCAmelCase_ : str = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( "Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. " "To return all points at once, set points_per_batch to None" ) for i in range(0 ,lowerCAmelCase__ ,lowerCAmelCase__ ): lowerCAmelCase_ : Any = grid_points[:, i : i + points_per_batch, :, :] lowerCAmelCase_ : List[Any] = input_labels[:, i : i + points_per_batch] lowerCAmelCase_ : List[Any] = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def UpperCAmelCase_ ( self : Dict ,lowerCAmelCase__ : Optional[int] ,lowerCAmelCase__ : str=0.88 ,lowerCAmelCase__ : Dict=0.95 ,lowerCAmelCase__ : int=0 ,lowerCAmelCase__ : Union[str, Any]=1 ,) -> int: '''simple docstring''' lowerCAmelCase_ : str = model_inputs.pop("input_boxes" ) lowerCAmelCase_ : int = model_inputs.pop("is_last" ) lowerCAmelCase_ : Dict = model_inputs.pop("original_sizes" ).tolist() lowerCAmelCase_ : Dict = model_inputs.pop("reshaped_input_sizes" ).tolist() lowerCAmelCase_ : Tuple = self.model(**lowerCAmelCase__ ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks lowerCAmelCase_ : int = model_outputs["pred_masks"] lowerCAmelCase_ : int = self.image_processor.post_process_masks( lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,binarize=lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = model_outputs["iou_scores"] lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = self.image_processor.filter_masks( masks[0] ,iou_scores[0] ,original_sizes[0] ,input_boxes[0] ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : Optional[int]=False ,lowerCAmelCase__ : Optional[int]=False ,lowerCAmelCase__ : Optional[int]=0.7 ,) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = [] lowerCAmelCase_ : int = [] lowerCAmelCase_ : Optional[Any] = [] for model_output in model_outputs: all_scores.append(model_output.pop("iou_scores" ) ) all_masks.extend(model_output.pop("masks" ) ) all_boxes.append(model_output.pop("boxes" ) ) lowerCAmelCase_ : Optional[Any] = torch.cat(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = torch.cat(lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = self.image_processor.post_process_for_mask_generation( lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = defaultdict(lowerCAmelCase__ ) for output in model_outputs: for k, v in output.items(): extra[k].append(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = {} if output_rle_mask: lowerCAmelCase_ : Tuple = rle_mask if output_bboxes_mask: lowerCAmelCase_ : str = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
659
import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) _lowercase = { '''iou_prediction_head.layers.0''': '''iou_prediction_head.proj_in''', '''iou_prediction_head.layers.1''': '''iou_prediction_head.layers.0''', '''iou_prediction_head.layers.2''': '''iou_prediction_head.proj_out''', '''mask_decoder.output_upscaling.0''': '''mask_decoder.upscale_conv1''', '''mask_decoder.output_upscaling.1''': '''mask_decoder.upscale_layer_norm''', '''mask_decoder.output_upscaling.3''': '''mask_decoder.upscale_conv2''', '''mask_downscaling.0''': '''mask_embed.conv1''', '''mask_downscaling.1''': '''mask_embed.layer_norm1''', '''mask_downscaling.3''': '''mask_embed.conv2''', '''mask_downscaling.4''': '''mask_embed.layer_norm2''', '''mask_downscaling.6''': '''mask_embed.conv3''', '''point_embeddings''': '''point_embed''', '''pe_layer.positional_encoding_gaussian_matrix''': '''shared_embedding.positional_embedding''', '''image_encoder''': '''vision_encoder''', '''neck.0''': '''neck.conv1''', '''neck.1''': '''neck.layer_norm1''', '''neck.2''': '''neck.conv2''', '''neck.3''': '''neck.layer_norm2''', '''patch_embed.proj''': '''patch_embed.projection''', '''.norm''': '''.layer_norm''', '''blocks''': '''layers''', } def UpperCamelCase ( snake_case__): lowerCAmelCase_ : int = {} state_dict.pop("pixel_mean" , snake_case__) state_dict.pop("pixel_std" , snake_case__) lowerCAmelCase_ : List[Any] = R".*.output_hypernetworks_mlps.(\d+).layers.(\d+).*" for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: lowerCAmelCase_ : Dict = key.replace(snake_case__ , snake_case__) if re.match(snake_case__ , snake_case__): lowerCAmelCase_ : Any = int(re.match(snake_case__ , snake_case__).group(2)) if layer_nb == 0: lowerCAmelCase_ : List[Any] = key.replace("layers.0" , "proj_in") elif layer_nb == 1: lowerCAmelCase_ : List[Any] = key.replace("layers.1" , "layers.0") elif layer_nb == 2: lowerCAmelCase_ : int = key.replace("layers.2" , "proj_out") lowerCAmelCase_ : int = value lowerCAmelCase_ : Optional[int] = model_state_dict[ "prompt_encoder.shared_embedding.positional_embedding" ] return model_state_dict def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__="ybelkada/segment-anything"): lowerCAmelCase_ : Optional[int] = hf_hub_download(snake_case__ , F'''checkpoints/{model_name}.pth''') if "sam_vit_b" in model_name: lowerCAmelCase_ : Optional[Any] = SamConfig() elif "sam_vit_l" in model_name: lowerCAmelCase_ : Optional[int] = SamVisionConfig( hidden_size=10_24 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) lowerCAmelCase_ : Union[str, Any] = SamConfig( vision_config=snake_case__ , ) elif "sam_vit_h" in model_name: lowerCAmelCase_ : Optional[Any] = SamVisionConfig( hidden_size=12_80 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) lowerCAmelCase_ : Tuple = SamConfig( vision_config=snake_case__ , ) lowerCAmelCase_ : Optional[Any] = torch.load(snake_case__ , map_location="cpu") lowerCAmelCase_ : Union[str, Any] = replace_keys(snake_case__) lowerCAmelCase_ : List[Any] = SamImageProcessor() lowerCAmelCase_ : Any = SamProcessor(image_processor=snake_case__) lowerCAmelCase_ : Any = SamModel(snake_case__) hf_model.load_state_dict(snake_case__) lowerCAmelCase_ : Dict = hf_model.to("cuda") lowerCAmelCase_ : List[str] = "https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png" lowerCAmelCase_ : List[Any] = Image.open(requests.get(snake_case__ , stream=snake_case__).raw).convert("RGB") lowerCAmelCase_ : Optional[int] = [[[4_00, 6_50]]] lowerCAmelCase_ : int = [[1]] lowerCAmelCase_ : Optional[Any] = processor(images=np.array(snake_case__) , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Optional[Any] = hf_model(**snake_case__) lowerCAmelCase_ : Optional[int] = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579_890_251_159_668 lowerCAmelCase_ : Any = processor( images=np.array(snake_case__) , input_points=snake_case__ , input_labels=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Optional[Any] = hf_model(**snake_case__) lowerCAmelCase_ : Union[str, Any] = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_712_603_092_193_604 lowerCAmelCase_ : Tuple = ((75, 2_75, 17_25, 8_50),) lowerCAmelCase_ : Optional[Any] = processor(images=np.array(snake_case__) , input_boxes=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : List[Any] = hf_model(**snake_case__) lowerCAmelCase_ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.8_686_015_605_926_514 # Test with 2 points and 1 image. lowerCAmelCase_ : int = [[[4_00, 6_50], [8_00, 6_50]]] lowerCAmelCase_ : Optional[Any] = [[1, 1]] lowerCAmelCase_ : List[Any] = processor( images=np.array(snake_case__) , input_points=snake_case__ , input_labels=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Tuple = hf_model(**snake_case__) lowerCAmelCase_ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_936_047_792_434_692 if __name__ == "__main__": _lowercase = argparse.ArgumentParser() _lowercase = ['''sam_vit_b_01ec64''', '''sam_vit_h_4b8939''', '''sam_vit_l_0b3195'''] parser.add_argument( '''--model_name''', default='''sam_vit_h_4b8939''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model and processor to the hub after converting''', ) parser.add_argument( '''--model_hub_id''', default='''ybelkada/segment-anything''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) _lowercase = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
659
1
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''asapp/sew-d-tiny-100k''': '''https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json''', # See all SEW-D models at https://huggingface.co/models?filter=sew-d } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = 'sew-d' def __init__( self : Dict ,lowerCAmelCase__ : int=32 ,lowerCAmelCase__ : str=7_68 ,lowerCAmelCase__ : int=12 ,lowerCAmelCase__ : str=12 ,lowerCAmelCase__ : List[Any]=30_72 ,lowerCAmelCase__ : int=2 ,lowerCAmelCase__ : Optional[int]=5_12 ,lowerCAmelCase__ : Tuple=2_56 ,lowerCAmelCase__ : List[Any]=True ,lowerCAmelCase__ : int=True ,lowerCAmelCase__ : Optional[Any]=("p2c", "c2p") ,lowerCAmelCase__ : Tuple="layer_norm" ,lowerCAmelCase__ : Tuple="gelu_python" ,lowerCAmelCase__ : List[str]=0.1 ,lowerCAmelCase__ : Union[str, Any]=0.1 ,lowerCAmelCase__ : Union[str, Any]=0.1 ,lowerCAmelCase__ : Dict=0.0 ,lowerCAmelCase__ : Tuple=0.1 ,lowerCAmelCase__ : Optional[Any]=0.02 ,lowerCAmelCase__ : Dict=1e-7 ,lowerCAmelCase__ : Optional[int]=1e-5 ,lowerCAmelCase__ : Optional[Any]="group" ,lowerCAmelCase__ : Optional[Any]="gelu" ,lowerCAmelCase__ : Union[str, Any]=(64, 1_28, 1_28, 1_28, 1_28, 2_56, 2_56, 2_56, 2_56, 5_12, 5_12, 5_12, 5_12) ,lowerCAmelCase__ : Tuple=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) ,lowerCAmelCase__ : Optional[Any]=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) ,lowerCAmelCase__ : Union[str, Any]=False ,lowerCAmelCase__ : List[str]=1_28 ,lowerCAmelCase__ : int=16 ,lowerCAmelCase__ : Optional[Any]=True ,lowerCAmelCase__ : List[str]=0.05 ,lowerCAmelCase__ : Tuple=10 ,lowerCAmelCase__ : str=2 ,lowerCAmelCase__ : Union[str, Any]=0.0 ,lowerCAmelCase__ : Optional[int]=10 ,lowerCAmelCase__ : Any=0 ,lowerCAmelCase__ : List[Any]="mean" ,lowerCAmelCase__ : int=False ,lowerCAmelCase__ : str=False ,lowerCAmelCase__ : List[str]=2_56 ,lowerCAmelCase__ : Any=0 ,lowerCAmelCase__ : List[Any]=1 ,lowerCAmelCase__ : Any=2 ,**lowerCAmelCase__ : Tuple ,) -> int: '''simple docstring''' super().__init__(**lowerCAmelCase__ ,pad_token_id=lowerCAmelCase__ ,bos_token_id=lowerCAmelCase__ ,eos_token_id=lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = hidden_size lowerCAmelCase_ : str = feat_extract_norm lowerCAmelCase_ : str = feat_extract_activation lowerCAmelCase_ : Union[str, Any] = list(lowerCAmelCase__ ) lowerCAmelCase_ : str = list(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = list(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = conv_bias lowerCAmelCase_ : int = num_conv_pos_embeddings lowerCAmelCase_ : Any = num_conv_pos_embedding_groups lowerCAmelCase_ : Dict = len(self.conv_dim ) lowerCAmelCase_ : str = num_hidden_layers lowerCAmelCase_ : int = intermediate_size lowerCAmelCase_ : Union[str, Any] = squeeze_factor lowerCAmelCase_ : Any = max_position_embeddings lowerCAmelCase_ : Optional[int] = position_buckets lowerCAmelCase_ : Any = share_att_key lowerCAmelCase_ : int = relative_attention lowerCAmelCase_ : List[Any] = norm_rel_ebd lowerCAmelCase_ : Optional[Any] = list(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = hidden_act lowerCAmelCase_ : int = num_attention_heads lowerCAmelCase_ : int = hidden_dropout lowerCAmelCase_ : str = attention_dropout lowerCAmelCase_ : int = activation_dropout lowerCAmelCase_ : Tuple = feat_proj_dropout lowerCAmelCase_ : Any = final_dropout lowerCAmelCase_ : Optional[Any] = layer_norm_eps lowerCAmelCase_ : Optional[Any] = feature_layer_norm_eps lowerCAmelCase_ : List[str] = initializer_range lowerCAmelCase_ : List[Any] = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect." "It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`," f'''but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)''' f'''= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowerCAmelCase_ : Optional[Any] = apply_spec_augment lowerCAmelCase_ : Optional[int] = mask_time_prob lowerCAmelCase_ : List[str] = mask_time_length lowerCAmelCase_ : str = mask_time_min_masks lowerCAmelCase_ : List[str] = mask_feature_prob lowerCAmelCase_ : Optional[Any] = mask_feature_length lowerCAmelCase_ : Union[str, Any] = mask_feature_min_masks # ctc loss lowerCAmelCase_ : int = ctc_loss_reduction lowerCAmelCase_ : List[str] = ctc_zero_infinity # sequence classification lowerCAmelCase_ : str = use_weighted_layer_sum lowerCAmelCase_ : Any = classifier_proj_size @property def UpperCAmelCase_ ( self : Tuple ) -> List[str]: '''simple docstring''' return functools.reduce(operator.mul ,self.conv_stride ,1 )
659
class __snake_case : """simple docstring""" def __init__( self : Union[str, Any] ,lowerCAmelCase__ : str = "" ,lowerCAmelCase__ : bool = False ) -> None: '''simple docstring''' lowerCAmelCase_ : dict[str, RadixNode] = {} # A node will be a leaf if the tree contains its word lowerCAmelCase_ : Optional[int] = is_leaf lowerCAmelCase_ : List[str] = prefix def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : str ) -> tuple[str, str, str]: '''simple docstring''' lowerCAmelCase_ : List[str] = 0 for q, w in zip(self.prefix ,lowerCAmelCase__ ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : list[str] ) -> None: '''simple docstring''' for word in words: self.insert(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : str ) -> None: '''simple docstring''' if self.prefix == word: lowerCAmelCase_ : Optional[Any] = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: lowerCAmelCase_ : Optional[int] = RadixNode(prefix=lowerCAmelCase__ ,is_leaf=lowerCAmelCase__ ) else: lowerCAmelCase_ : Optional[Any] = self.nodes[word[0]] lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = incoming_node.match( lowerCAmelCase__ ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: lowerCAmelCase_ : Dict = remaining_prefix lowerCAmelCase_ : str = self.nodes[matching_string[0]] lowerCAmelCase_ : Dict = RadixNode(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Any = aux_node if remaining_word == "": lowerCAmelCase_ : Optional[Any] = True else: self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : str ) -> bool: '''simple docstring''' lowerCAmelCase_ : List[str] = self.nodes.get(word[0] ,lowerCAmelCase__ ) if not incoming_node: return False else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : str ) -> bool: '''simple docstring''' lowerCAmelCase_ : int = self.nodes.get(word[0] ,lowerCAmelCase__ ) if not incoming_node: return False else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(lowerCAmelCase__ ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: lowerCAmelCase_ : int = list(self.nodes.values() )[0] lowerCAmelCase_ : List[Any] = merging_node.is_leaf self.prefix += merging_node.prefix lowerCAmelCase_ : int = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: lowerCAmelCase_ : List[str] = False # If there is 1 edge, we merge it with its child else: lowerCAmelCase_ : Union[str, Any] = list(incoming_node.nodes.values() )[0] lowerCAmelCase_ : Optional[int] = merging_node.is_leaf incoming_node.prefix += merging_node.prefix lowerCAmelCase_ : List[str] = merging_node.nodes return True def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : int = 0 ) -> None: '''simple docstring''' if self.prefix != "": print("-" * height ,self.prefix ," (leaf)" if self.is_leaf else "" ) for value in self.nodes.values(): value.print_tree(height + 1 ) def UpperCamelCase ( ): lowerCAmelCase_ : List[Any] = "banana bananas bandana band apple all beast".split() lowerCAmelCase_ : Optional[Any] = RadixNode() root.insert_many(snake_case__) assert all(root.find(snake_case__) for word in words) assert not root.find("bandanas") assert not root.find("apps") root.delete("all") assert not root.find("all") root.delete("banana") assert not root.find("banana") assert root.find("bananas") return True def UpperCamelCase ( ): assert test_trie() def UpperCamelCase ( ): lowerCAmelCase_ : str = RadixNode() lowerCAmelCase_ : str = "banana bananas bandanas bandana band apple all beast".split() root.insert_many(snake_case__) print("Words:" , snake_case__) print("Tree:") root.print_tree() if __name__ == "__main__": main()
659
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _lowercase = { '''configuration_m2m_100''': ['''M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''M2M100Config''', '''M2M100OnnxConfig'''], '''tokenization_m2m_100''': ['''M2M100Tokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ '''M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST''', '''M2M100ForConditionalGeneration''', '''M2M100Model''', '''M2M100PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
659
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class __snake_case : """simple docstring""" def __init__( self : Tuple ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Optional[Any]=12 ,lowerCAmelCase__ : Union[str, Any]=7 ,lowerCAmelCase__ : Union[str, Any]=True ,lowerCAmelCase__ : List[str]=True ,lowerCAmelCase__ : Any=True ,lowerCAmelCase__ : Optional[Any]=99 ,lowerCAmelCase__ : List[str]=32 ,lowerCAmelCase__ : Dict=32 ,lowerCAmelCase__ : str=2 ,lowerCAmelCase__ : Optional[int]=4 ,lowerCAmelCase__ : str=37 ,lowerCAmelCase__ : Dict=0.1 ,lowerCAmelCase__ : List[str]=0.1 ,lowerCAmelCase__ : str=5_12 ,lowerCAmelCase__ : Union[str, Any]=0.02 ,lowerCAmelCase__ : Tuple=0 ,lowerCAmelCase__ : str=None ,) -> str: '''simple docstring''' lowerCAmelCase_ : int = parent lowerCAmelCase_ : str = batch_size lowerCAmelCase_ : int = seq_length lowerCAmelCase_ : Union[str, Any] = is_training lowerCAmelCase_ : int = use_input_mask lowerCAmelCase_ : List[Any] = use_labels lowerCAmelCase_ : Dict = vocab_size lowerCAmelCase_ : Union[str, Any] = hidden_size lowerCAmelCase_ : Union[str, Any] = projection_dim lowerCAmelCase_ : List[Any] = num_hidden_layers lowerCAmelCase_ : Any = num_attention_heads lowerCAmelCase_ : List[Any] = intermediate_size lowerCAmelCase_ : Any = dropout lowerCAmelCase_ : Optional[int] = attention_dropout lowerCAmelCase_ : int = max_position_embeddings lowerCAmelCase_ : Optional[int] = initializer_range lowerCAmelCase_ : Any = scope lowerCAmelCase_ : Tuple = bos_token_id def UpperCAmelCase_ ( self : str ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) lowerCAmelCase_ : Dict = None if self.use_input_mask: lowerCAmelCase_ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowerCAmelCase_ : List[Any] = input_mask.numpy() lowerCAmelCase_ , lowerCAmelCase_ : str = input_mask.shape lowerCAmelCase_ : Dict = np.random.randint(1 ,seq_length - 1 ,size=(batch_size,) ) for batch_idx, start_index in enumerate(lowerCAmelCase__ ): lowerCAmelCase_ : Union[str, Any] = 1 lowerCAmelCase_ : Optional[Any] = 0 lowerCAmelCase_ : List[Any] = self.get_config() return config, input_ids, tf.convert_to_tensor(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[str] ) -> str: '''simple docstring''' return BlipTextConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,projection_dim=self.projection_dim ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,dropout=self.dropout ,attention_dropout=self.attention_dropout ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,bos_token_id=self.bos_token_id ,) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Dict ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : List[Any] = TFBlipTextModel(config=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = model(lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,training=lowerCAmelCase__ ) lowerCAmelCase_ : str = model(lowerCAmelCase__ ,training=lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def UpperCAmelCase_ ( self : Optional[int] ) -> int: '''simple docstring''' lowerCAmelCase_ : List[str] = self.prepare_config_and_inputs() lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = config_and_inputs lowerCAmelCase_ : Tuple = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = (TFBlipTextModel,) if is_tf_available() else () UpperCamelCase_ = False UpperCamelCase_ = False UpperCamelCase_ = False def UpperCAmelCase_ ( self : Optional[Any] ) -> str: '''simple docstring''' lowerCAmelCase_ : List[str] = BlipTextModelTester(self ) lowerCAmelCase_ : Tuple = ConfigTester(self ,config_class=lowerCAmelCase__ ,hidden_size=37 ) def UpperCAmelCase_ ( self : str ) -> Any: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase_ ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' pass def UpperCAmelCase_ ( self : Union[str, Any] ) -> Any: '''simple docstring''' pass @unittest.skip(reason="Blip does not use inputs_embeds" ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCAmelCase_ ( self : int ) -> Optional[Any]: '''simple docstring''' pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCAmelCase_ ( self : Dict ) -> Union[str, Any]: '''simple docstring''' pass @slow def UpperCAmelCase_ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Tuple = TFBlipTextModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : str=True ) -> List[Any]: '''simple docstring''' super().test_pt_tf_model_equivalence(allow_missing_keys=lowerCAmelCase__ )
659
1
import math class __snake_case : """simple docstring""" def __init__( self : Dict ,lowerCAmelCase__ : List[Any]=0 ) -> Optional[Any]: # a graph with Node 0,1,...,N-1 '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = n lowerCAmelCase_ : Any = [ [math.inf for j in range(0 ,lowerCAmelCase__ )] for i in range(0 ,lowerCAmelCase__ ) ] # adjacency matrix for weight lowerCAmelCase_ : Optional[Any] = [ [math.inf for j in range(0 ,lowerCAmelCase__ )] for i in range(0 ,lowerCAmelCase__ ) ] # dp[i][j] stores minimum distance from i to j def UpperCAmelCase_ ( self : Dict ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = w def UpperCAmelCase_ ( self : Any ) -> Dict: '''simple docstring''' for k in range(0 ,self.n ): for i in range(0 ,self.n ): for j in range(0 ,self.n ): lowerCAmelCase_ : Optional[Any] = min(self.dp[i][j] ,self.dp[i][k] + self.dp[k][j] ) def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Dict ,lowerCAmelCase__ : str ) -> List[Any]: '''simple docstring''' return self.dp[u][v] if __name__ == "__main__": _lowercase = Graph(5) graph.add_edge(0, 2, 9) graph.add_edge(0, 4, 10) graph.add_edge(1, 3, 5) graph.add_edge(2, 3, 7) graph.add_edge(3, 0, 10) graph.add_edge(3, 1, 2) graph.add_edge(3, 2, 1) graph.add_edge(3, 4, 6) graph.add_edge(4, 1, 3) graph.add_edge(4, 2, 4) graph.add_edge(4, 3, 9) graph.floyd_warshall() graph.show_min(1, 4) graph.show_min(0, 3)
659
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _lowercase = logging.get_logger(__name__) _lowercase = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} # See all LED models at https://huggingface.co/models?filter=LED _lowercase = { '''vocab_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json''', }, '''merges_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json''', }, } _lowercase = { '''allenai/led-base-16384''': 16384, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def UpperCamelCase ( ): lowerCAmelCase_ : Optional[int] = ( list(range(ord("!") , ord("~") + 1)) + list(range(ord("¡") , ord("¬") + 1)) + list(range(ord("®") , ord("ÿ") + 1)) ) lowerCAmelCase_ : List[Any] = bs[:] lowerCAmelCase_ : Optional[int] = 0 for b in range(2**8): if b not in bs: bs.append(snake_case__) cs.append(2**8 + n) n += 1 lowerCAmelCase_ : Tuple = [chr(snake_case__) for n in cs] return dict(zip(snake_case__ , snake_case__)) def UpperCamelCase ( snake_case__): lowerCAmelCase_ : str = set() lowerCAmelCase_ : List[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char)) lowerCAmelCase_ : Union[str, Any] = char return pairs class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = ['input_ids', 'attention_mask'] def __init__( self : int ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Tuple="replace" ,lowerCAmelCase__ : Optional[int]="<s>" ,lowerCAmelCase__ : Optional[int]="</s>" ,lowerCAmelCase__ : Tuple="</s>" ,lowerCAmelCase__ : int="<s>" ,lowerCAmelCase__ : Union[str, Any]="<unk>" ,lowerCAmelCase__ : str="<pad>" ,lowerCAmelCase__ : Tuple="<mask>" ,lowerCAmelCase__ : Optional[int]=False ,**lowerCAmelCase__ : Tuple ,) -> Any: '''simple docstring''' lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else bos_token lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else eos_token lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else sep_token lowerCAmelCase_ : Any = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else cls_token lowerCAmelCase_ : Tuple = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else unk_token lowerCAmelCase_ : Any = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase_ : Optional[int] = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else mask_token super().__init__( errors=lowerCAmelCase__ ,bos_token=lowerCAmelCase__ ,eos_token=lowerCAmelCase__ ,unk_token=lowerCAmelCase__ ,sep_token=lowerCAmelCase__ ,cls_token=lowerCAmelCase__ ,pad_token=lowerCAmelCase__ ,mask_token=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,**lowerCAmelCase__ ,) with open(lowerCAmelCase__ ,encoding="utf-8" ) as vocab_handle: lowerCAmelCase_ : List[str] = json.load(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = {v: k for k, v in self.encoder.items()} lowerCAmelCase_ : Optional[int] = errors # how to handle errors in decoding lowerCAmelCase_ : Optional[int] = bytes_to_unicode() lowerCAmelCase_ : str = {v: k for k, v in self.byte_encoder.items()} with open(lowerCAmelCase__ ,encoding="utf-8" ) as merges_handle: lowerCAmelCase_ : List[str] = merges_handle.read().split("\n" )[1:-1] lowerCAmelCase_ : List[Any] = [tuple(merge.split() ) for merge in bpe_merges] lowerCAmelCase_ : Union[str, Any] = dict(zip(lowerCAmelCase__ ,range(len(lowerCAmelCase__ ) ) ) ) lowerCAmelCase_ : Dict = {} lowerCAmelCase_ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowerCAmelCase_ : Any = re.compile(R"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def UpperCAmelCase_ ( self : Dict ) -> Dict: '''simple docstring''' return len(self.encoder ) def UpperCAmelCase_ ( self : Dict ) -> str: '''simple docstring''' return dict(self.encoder ,**self.added_tokens_encoder ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Dict ) -> Dict: '''simple docstring''' if token in self.cache: return self.cache[token] lowerCAmelCase_ : Union[str, Any] = tuple(lowerCAmelCase__ ) lowerCAmelCase_ : str = get_pairs(lowerCAmelCase__ ) if not pairs: return token while True: lowerCAmelCase_ : Optional[int] = min(lowerCAmelCase__ ,key=lambda lowerCAmelCase__ : self.bpe_ranks.get(lowerCAmelCase__ ,float("inf" ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = bigram lowerCAmelCase_ : Tuple = [] lowerCAmelCase_ : str = 0 while i < len(lowerCAmelCase__ ): try: lowerCAmelCase_ : Union[str, Any] = word.index(lowerCAmelCase__ ,lowerCAmelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase_ : List[str] = j if word[i] == first and i < len(lowerCAmelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase_ : Optional[int] = tuple(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = new_word if len(lowerCAmelCase__ ) == 1: break else: lowerCAmelCase_ : Dict = get_pairs(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = " ".join(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = word return word def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : Dict ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : Any = [] for token in re.findall(self.pat ,lowerCAmelCase__ ): lowerCAmelCase_ : Optional[int] = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCAmelCase__ ).split(" " ) ) return bpe_tokens def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any] ) -> Tuple: '''simple docstring''' return self.encoder.get(lowerCAmelCase__ ,self.encoder.get(self.unk_token ) ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Union[str, Any] ) -> Optional[int]: '''simple docstring''' return self.decoder.get(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[Any] ) -> Any: '''simple docstring''' lowerCAmelCase_ : int = "".join(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" ,errors=self.errors ) return text def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(lowerCAmelCase__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCAmelCase_ : Optional[int] = os.path.join( lowerCAmelCase__ ,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase_ : List[str] = os.path.join( lowerCAmelCase__ ,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(lowerCAmelCase__ ,"w" ,encoding="utf-8" ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=lowerCAmelCase__ ,ensure_ascii=lowerCAmelCase__ ) + "\n" ) lowerCAmelCase_ : Dict = 0 with open(lowerCAmelCase__ ,"w" ,encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda lowerCAmelCase__ : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' " Please check that the tokenizer is not corrupted!" ) lowerCAmelCase_ : List[Any] = token_index writer.write(" ".join(lowerCAmelCase__ ) + "\n" ) index += 1 return vocab_file, merge_file def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase_ : Union[str, Any] = [self.cls_token_id] lowerCAmelCase_ : str = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ,lowerCAmelCase__ : bool = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ ,token_ids_a=lowerCAmelCase__ ,already_has_special_tokens=lowerCAmelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase__ )) + [1] return [1] + ([0] * len(lowerCAmelCase__ )) + [1, 1] + ([0] * len(lowerCAmelCase__ )) + [1] def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = [self.sep_token_id] lowerCAmelCase_ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : Optional[int]=False ,**lowerCAmelCase__ : str ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = kwargs.pop("add_prefix_space" ,self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCAmelCase__ ) > 0 and not text[0].isspace()): lowerCAmelCase_ : List[str] = " " + text return (text, kwargs) def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : Union[Dict[str, EncodedInput], BatchEncoding] ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : Optional[bool] = None ,) -> dict: '''simple docstring''' lowerCAmelCase_ : int = super()._pad( encoded_inputs=lowerCAmelCase__ ,max_length=lowerCAmelCase__ ,padding_strategy=lowerCAmelCase__ ,pad_to_multiple_of=lowerCAmelCase__ ,return_attention_mask=lowerCAmelCase__ ,) # Load from model defaults if return_attention_mask is None: lowerCAmelCase_ : List[Any] = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowerCAmelCase_ : Dict = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowerCAmelCase_ : List[Any] = len(encoded_inputs["global_attention_mask"] ) != len(lowerCAmelCase__ ) if needs_to_be_padded: lowerCAmelCase_ : Union[str, Any] = len(lowerCAmelCase__ ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowerCAmelCase_ : Optional[int] = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": lowerCAmelCase_ : List[Any] = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
659
1
from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = CustomTokenizer pass
659
import os _lowercase = {'''I''': 1, '''V''': 5, '''X''': 10, '''L''': 50, '''C''': 100, '''D''': 500, '''M''': 1000} def UpperCamelCase ( snake_case__): lowerCAmelCase_ : List[str] = 0 lowerCAmelCase_ : Any = 0 while index < len(snake_case__) - 1: lowerCAmelCase_ : Optional[Any] = SYMBOLS[numerals[index]] lowerCAmelCase_ : int = SYMBOLS[numerals[index + 1]] if current_value < next_value: total_value -= current_value else: total_value += current_value index += 1 total_value += SYMBOLS[numerals[index]] return total_value def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Optional[int] = "" lowerCAmelCase_ : Tuple = num // 10_00 numerals += m_count * "M" num %= 10_00 lowerCAmelCase_ : int = num // 1_00 if c_count == 9: numerals += "CM" c_count -= 9 elif c_count == 4: numerals += "CD" c_count -= 4 if c_count >= 5: numerals += "D" c_count -= 5 numerals += c_count * "C" num %= 1_00 lowerCAmelCase_ : int = num // 10 if x_count == 9: numerals += "XC" x_count -= 9 elif x_count == 4: numerals += "XL" x_count -= 4 if x_count >= 5: numerals += "L" x_count -= 5 numerals += x_count * "X" num %= 10 if num == 9: numerals += "IX" num -= 9 elif num == 4: numerals += "IV" num -= 4 if num >= 5: numerals += "V" num -= 5 numerals += num * "I" return numerals def UpperCamelCase ( snake_case__ = "/p089_roman.txt"): lowerCAmelCase_ : int = 0 with open(os.path.dirname(snake_case__) + roman_numerals_filename) as filea: lowerCAmelCase_ : List[Any] = filea.readlines() for line in lines: lowerCAmelCase_ : Any = line.strip() lowerCAmelCase_ : Tuple = parse_roman_numerals(snake_case__) lowerCAmelCase_ : List[Any] = generate_roman_numerals(snake_case__) savings += len(snake_case__) - len(snake_case__) return savings if __name__ == "__main__": print(f"{solution() = }")
659
1
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class __snake_case : """simple docstring""" def __init__( self : Tuple ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Optional[Any]=12 ,lowerCAmelCase__ : Union[str, Any]=7 ,lowerCAmelCase__ : Union[str, Any]=True ,lowerCAmelCase__ : List[str]=True ,lowerCAmelCase__ : Any=True ,lowerCAmelCase__ : Optional[Any]=99 ,lowerCAmelCase__ : List[str]=32 ,lowerCAmelCase__ : Dict=32 ,lowerCAmelCase__ : str=2 ,lowerCAmelCase__ : Optional[int]=4 ,lowerCAmelCase__ : str=37 ,lowerCAmelCase__ : Dict=0.1 ,lowerCAmelCase__ : List[str]=0.1 ,lowerCAmelCase__ : str=5_12 ,lowerCAmelCase__ : Union[str, Any]=0.02 ,lowerCAmelCase__ : Tuple=0 ,lowerCAmelCase__ : str=None ,) -> str: '''simple docstring''' lowerCAmelCase_ : int = parent lowerCAmelCase_ : str = batch_size lowerCAmelCase_ : int = seq_length lowerCAmelCase_ : Union[str, Any] = is_training lowerCAmelCase_ : int = use_input_mask lowerCAmelCase_ : List[Any] = use_labels lowerCAmelCase_ : Dict = vocab_size lowerCAmelCase_ : Union[str, Any] = hidden_size lowerCAmelCase_ : Union[str, Any] = projection_dim lowerCAmelCase_ : List[Any] = num_hidden_layers lowerCAmelCase_ : Any = num_attention_heads lowerCAmelCase_ : List[Any] = intermediate_size lowerCAmelCase_ : Any = dropout lowerCAmelCase_ : Optional[int] = attention_dropout lowerCAmelCase_ : int = max_position_embeddings lowerCAmelCase_ : Optional[int] = initializer_range lowerCAmelCase_ : Any = scope lowerCAmelCase_ : Tuple = bos_token_id def UpperCAmelCase_ ( self : str ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) lowerCAmelCase_ : Dict = None if self.use_input_mask: lowerCAmelCase_ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowerCAmelCase_ : List[Any] = input_mask.numpy() lowerCAmelCase_ , lowerCAmelCase_ : str = input_mask.shape lowerCAmelCase_ : Dict = np.random.randint(1 ,seq_length - 1 ,size=(batch_size,) ) for batch_idx, start_index in enumerate(lowerCAmelCase__ ): lowerCAmelCase_ : Union[str, Any] = 1 lowerCAmelCase_ : Optional[Any] = 0 lowerCAmelCase_ : List[Any] = self.get_config() return config, input_ids, tf.convert_to_tensor(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[str] ) -> str: '''simple docstring''' return BlipTextConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,projection_dim=self.projection_dim ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,dropout=self.dropout ,attention_dropout=self.attention_dropout ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,bos_token_id=self.bos_token_id ,) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Dict ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : List[Any] = TFBlipTextModel(config=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = model(lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,training=lowerCAmelCase__ ) lowerCAmelCase_ : str = model(lowerCAmelCase__ ,training=lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def UpperCAmelCase_ ( self : Optional[int] ) -> int: '''simple docstring''' lowerCAmelCase_ : List[str] = self.prepare_config_and_inputs() lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = config_and_inputs lowerCAmelCase_ : Tuple = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = (TFBlipTextModel,) if is_tf_available() else () UpperCamelCase_ = False UpperCamelCase_ = False UpperCamelCase_ = False def UpperCAmelCase_ ( self : Optional[Any] ) -> str: '''simple docstring''' lowerCAmelCase_ : List[str] = BlipTextModelTester(self ) lowerCAmelCase_ : Tuple = ConfigTester(self ,config_class=lowerCAmelCase__ ,hidden_size=37 ) def UpperCAmelCase_ ( self : str ) -> Any: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase_ ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' pass def UpperCAmelCase_ ( self : Union[str, Any] ) -> Any: '''simple docstring''' pass @unittest.skip(reason="Blip does not use inputs_embeds" ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCAmelCase_ ( self : int ) -> Optional[Any]: '''simple docstring''' pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCAmelCase_ ( self : Dict ) -> Union[str, Any]: '''simple docstring''' pass @slow def UpperCAmelCase_ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Tuple = TFBlipTextModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : str=True ) -> List[Any]: '''simple docstring''' super().test_pt_tf_model_equivalence(allow_missing_keys=lowerCAmelCase__ )
659
from transformers import HfArgumentParser, TensorFlowBenchmark, TensorFlowBenchmarkArguments def UpperCamelCase ( ): lowerCAmelCase_ : Dict = HfArgumentParser(snake_case__) lowerCAmelCase_ : Dict = parser.parse_args_into_dataclasses()[0] lowerCAmelCase_ : List[Any] = TensorFlowBenchmark(args=snake_case__) try: lowerCAmelCase_ : str = parser.parse_args_into_dataclasses()[0] except ValueError as e: lowerCAmelCase_ : Optional[Any] = "Arg --no_{0} is no longer used, please use --no-{0} instead." lowerCAmelCase_ : Tuple = " ".join(str(snake_case__).split(" ")[:-1]) lowerCAmelCase_ : List[Any] = "" lowerCAmelCase_ : Optional[Any] = eval(str(snake_case__).split(" ")[-1]) lowerCAmelCase_ : List[Any] = [] for arg in depreciated_args: # arg[2:] removes '--' if arg[2:] in TensorFlowBenchmark.deprecated_args: # arg[5:] removes '--no_' full_error_msg += arg_error_msg.format(arg[5:]) else: wrong_args.append(snake_case__) if len(snake_case__) > 0: lowerCAmelCase_ : int = full_error_msg + begin_error_msg + str(snake_case__) raise ValueError(snake_case__) benchmark.run() if __name__ == "__main__": main()
659
1
def __lowercase ( snake_case ): """simple docstring""" __magic_name__ :str = current_set.copy() for row_index, row in enumerate(snake_case ): __magic_name__ :Tuple = row[0] for column_index, column in enumerate(snake_case ): if magnitude == 0: __magic_name__ :Optional[Any] = column continue __magic_name__ :Union[str, Any] = column / magnitude # Subtract to cancel term __magic_name__ :List[str] = current_set[0] __magic_name__ :List[Any] = [first_row] __magic_name__ :Any = current_set[1::] for row in current_set: __magic_name__ :str = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(snake_case ) continue for column_index in range(len(snake_case ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(snake_case ) # Create next recursion iteration set if len(final_set[0] ) != 3: __magic_name__ :Tuple = final_set[0] __magic_name__ :List[Any] = [] __magic_name__ :int = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) __magic_name__ :int = simplify(snake_case ) for i in range(len(snake_case ) ): resultant[i].insert(0, current_first_column[i] ) resultant.insert(0, snake_case ) __magic_name__ :Any = resultant return final_set def __lowercase ( snake_case ): """simple docstring""" if len(snake_case ) == 0: raise IndexError('''solve_simultaneous() requires n lists of length n+1''' ) __magic_name__ :int = len(snake_case ) + 1 if any(len(snake_case ) != _length for item in equations ): raise IndexError('''solve_simultaneous() requires n lists of length n+1''' ) for row in equations: if any(not isinstance(snake_case, (int, float) ) for column in row ): raise ValueError('''solve_simultaneous() requires lists of integers''' ) if len(snake_case ) == 1: return [equations[0][-1] / equations[0][0]] __magic_name__ :List[Any] = equations.copy() if any(0 in row for row in data_set ): __magic_name__ :Any = data_set.copy() __magic_name__ :List[str] = [] for row_index, row in enumerate(snake_case ): if 0 not in row: __magic_name__ :Any = data_set.pop(snake_case ) break if not full_row: raise ValueError('''solve_simultaneous() requires at least 1 full equation''' ) data_set.insert(0, snake_case ) __magic_name__ :Tuple = data_set.copy() __magic_name__ :Optional[Any] = simplify(snake_case ) __magic_name__ :Optional[Any] = simplified[::-1] __magic_name__ :list = [] for row in simplified: __magic_name__ :int = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue __magic_name__ :Union[str, Any] = row.copy()[: len(snake_case ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(snake_case ) == 0: solutions.append(0 ) continue __magic_name__ :Dict = temp_row[1::] __magic_name__ :Any = temp_row[::-1] for column_index, column in enumerate(snake_case ): current_solution -= column * solutions[column_index] solutions.append(snake_case ) __magic_name__ :List[Any] = [] for item in solutions: final.append(float(round(snake_case, 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() SCREAMING_SNAKE_CASE__ : List[str] = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
0
from collections import defaultdict from pathlib import Path import pandas as pd from rouge_cli import calculate_rouge_path from utils import calculate_rouge _lowercase = [ '''Prosecutor: "No videos were used in the crash investigation" German papers say they saw a cell phone video of the''' ''' final seconds on board Flight 9525. The Germanwings co-pilot says he had a "previous episode of severe''' ''' depression\" German airline confirms it knew of Andreas Lubitz\'s depression years before he took control.''', '''The Palestinian Authority officially becomes the 123rd member of the International Criminal Court. The formal''' ''' accession was marked with a ceremony at The Hague, in the Netherlands. The Palestinians signed the ICC\'s''' ''' founding Rome Statute in January. Israel and the United States opposed the Palestinians\' efforts to join the''' ''' body.''', '''Amnesty International releases its annual report on the death penalty. The report catalogs the use of''' ''' state-sanctioned killing as a punitive measure across the globe. At least 607 people were executed around the''' ''' world in 2014, compared to 778 in 2013. The U.S. remains one of the worst offenders for imposing capital''' ''' punishment.''', ] _lowercase = [ '''Marseille prosecutor says "so far no videos were used in the crash investigation" despite media reports .''' ''' Journalists at Bild and Paris Match are "very confident" the video clip is real, an editor says . Andreas Lubitz''' ''' had informed his Lufthansa training school of an episode of severe depression, airline says .''', '''Membership gives the ICC jurisdiction over alleged crimes committed in Palestinian territories since last June .''' ''' Israel and the United States opposed the move, which could open the door to war crimes investigations against''' ''' Israelis .''', '''Amnesty\'s annual death penalty report catalogs encouraging signs, but setbacks in numbers of those sentenced to''' ''' death . Organization claims that governments around the world are using the threat of terrorism to advance''' ''' executions . The number of executions worldwide has gone down by almost 22% compared with 2013, but death''' ''' sentences up by 28% .''', ] def UpperCamelCase ( ): lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , bootstrap_aggregation=snake_case__ , rouge_keys=["rouge2", "rougeL"]) assert isinstance(snake_case__ , snake_case__) lowerCAmelCase_ : str = calculate_rouge(snake_case__ , snake_case__ , bootstrap_aggregation=snake_case__ , rouge_keys=["rouge2"]) assert ( pd.DataFrame(no_aggregation["rouge2"]).fmeasure.mean() == pd.DataFrame(no_aggregation_just_ra["rouge2"]).fmeasure.mean() ) def UpperCamelCase ( ): lowerCAmelCase_ : str = "rougeLsum" lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=[k])[k] lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=[k])[k] assert score > score_no_sep def UpperCamelCase ( ): lowerCAmelCase_ : int = ["rouge1", "rouge2", "rougeL"] lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=snake_case__) lowerCAmelCase_ : List[Any] = calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__ , rouge_keys=snake_case__) assert score_sep == score_no_sep def UpperCamelCase ( ): lowerCAmelCase_ : List[str] = [ "Her older sister, Margot Frank, died in 1945, a month earlier than previously thought.", "Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports .", ] lowerCAmelCase_ : Dict = [ "Margot Frank, died in 1945, a month earlier than previously thought.", "Prosecutor: \"No videos were used in the crash investigation\" German papers say they saw a cell phone video of" " the final seconds on board Flight 9525.", ] assert calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__) == calculate_rouge(snake_case__ , snake_case__ , newline_sep=snake_case__) def UpperCamelCase ( ): lowerCAmelCase_ : Optional[int] = [ "\" \"a person who has such a video needs to immediately give it to the investigators,\" prosecutor says .<n> \"it is a very disturbing scene,\" editor-in-chief of bild online tells \"erin burnett: outfront\" " ] lowerCAmelCase_ : Any = [ " Marseille prosecutor says \"so far no videos were used in the crash investigation\" despite media reports . Journalists at Bild and Paris Match are \"very confident\" the video clip is real, an editor says . Andreas Lubitz had informed his Lufthansa training school of an episode of severe depression, airline says ." ] lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , rouge_keys=["rougeLsum"] , newline_sep=snake_case__)["rougeLsum"] lowerCAmelCase_ : Any = calculate_rouge(snake_case__ , snake_case__ , rouge_keys=["rougeLsum"])["rougeLsum"] assert new_score > prev_score def UpperCamelCase ( ): lowerCAmelCase_ : int = Path("examples/seq2seq/test_data/wmt_en_ro") lowerCAmelCase_ : Dict = calculate_rouge_path(data_dir.joinpath("test.source") , data_dir.joinpath("test.target")) assert isinstance(snake_case__ , snake_case__) lowerCAmelCase_ : Any = calculate_rouge_path( data_dir.joinpath("test.source") , data_dir.joinpath("test.target") , bootstrap_aggregation=snake_case__) assert isinstance(snake_case__ , snake_case__)
659
0
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def _A ( _lowercase ) -> Tuple: """simple docstring""" monkeypatch.setattr('datasets.utils.deprecation_utils._emitted_deprecation_warnings' , set() ) @pytest.fixture def _A ( _lowercase ) -> int: """simple docstring""" class __lowerCamelCase : def __init__( self: Optional[Any],A_: Tuple ): '''simple docstring''' __UpperCamelCase = metric_id class __lowerCamelCase : _lowercase = [MetricMock(_a ) for metric_id in ["""accuracy""", """mse""", """precision""", """codeparrot/apps_metric"""]] def snake_case_ ( self: Any ): '''simple docstring''' return self._metrics monkeypatch.setattr('datasets.inspect.huggingface_hub' , HfhMock() ) @pytest.mark.parametrize( 'func, args' , [(load_metric, ('metrics/mse',)), (list_metrics, ()), (inspect_metric, ('metrics/mse', 'tmp_path'))] ) def _A ( _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ) -> str: """simple docstring""" if "tmp_path" in args: __UpperCamelCase = tuple(arg if arg != 'tmp_path' else tmp_path for arg in args ) with pytest.warns(_lowercase , match='https://huggingface.co/docs/evaluate' ): func(*_lowercase )
1
import json import os import unittest from transformers import BatchEncoding, LEDTokenizer, LEDTokenizerFast from transformers.models.led.tokenization_led import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = LEDTokenizer UpperCamelCase_ = LEDTokenizerFast UpperCamelCase_ = True def UpperCAmelCase_ ( self : List[Any] ) -> Optional[int]: '''simple docstring''' super().setUp() lowerCAmelCase_ : Union[str, Any] = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "\u0120", "\u0120l", "\u0120n", "\u0120lo", "\u0120low", "er", "\u0120lowest", "\u0120newer", "\u0120wider", "<unk>", ] lowerCAmelCase_ : Tuple = dict(zip(lowerCAmelCase__ ,range(len(lowerCAmelCase__ ) ) ) ) lowerCAmelCase_ : int = ["#version: 0.2", "\u0120 l", "\u0120l o", "\u0120lo w", "e r", ""] lowerCAmelCase_ : Union[str, Any] = {"unk_token": "<unk>"} lowerCAmelCase_ : List[Any] = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase_ : Any = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file ,"w" ,encoding="utf-8" ) as fp: fp.write(json.dumps(lowerCAmelCase__ ) + "\n" ) with open(self.merges_file ,"w" ,encoding="utf-8" ) as fp: fp.write("\n".join(lowerCAmelCase__ ) ) def UpperCAmelCase_ ( self : List[Any] ,**lowerCAmelCase__ : int ) -> Tuple: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ,**lowerCAmelCase__ : Optional[int] ) -> List[Any]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return self.rust_tokenizer_class.from_pretrained(self.tmpdirname ,**lowerCAmelCase__ ) def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : int ) -> List[str]: '''simple docstring''' return "lower newer", "lower newer" @cached_property def UpperCAmelCase_ ( self : int ) -> Union[str, Any]: '''simple docstring''' return LEDTokenizer.from_pretrained("allenai/led-base-16384" ) @cached_property def UpperCAmelCase_ ( self : List[str] ) -> Dict: '''simple docstring''' return LEDTokenizerFast.from_pretrained("allenai/led-base-16384" ) @require_torch def UpperCAmelCase_ ( self : int ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Union[str, Any] = ["A long paragraph for summarization.", "Another paragraph for summarization."] lowerCAmelCase_ : int = [0, 2_50, 2_51, 1_78_18, 13, 3_91_86, 19_38, 4, 2] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Any = tokenizer(lowerCAmelCase__ ,max_length=len(lowerCAmelCase__ ) ,padding=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual((2, 9) ,batch.input_ids.shape ) self.assertEqual((2, 9) ,batch.attention_mask.shape ) lowerCAmelCase_ : int = batch.input_ids.tolist()[0] self.assertListEqual(lowerCAmelCase__ ,lowerCAmelCase__ ) @require_torch def UpperCAmelCase_ ( self : Dict ) -> Any: '''simple docstring''' lowerCAmelCase_ : int = ["A long paragraph for summarization.", "Another paragraph for summarization."] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[Any] = tokenizer(lowerCAmelCase__ ,padding=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIn("input_ids" ,lowerCAmelCase__ ) self.assertIn("attention_mask" ,lowerCAmelCase__ ) self.assertNotIn("labels" ,lowerCAmelCase__ ) self.assertNotIn("decoder_attention_mask" ,lowerCAmelCase__ ) @require_torch def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : int = [ "Summary of the text.", "Another summary.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[int] = tokenizer(text_target=lowerCAmelCase__ ,max_length=32 ,padding="max_length" ,return_tensors="pt" ) self.assertEqual(32 ,targets["input_ids"].shape[1] ) @require_torch def UpperCAmelCase_ ( self : Tuple ) -> List[str]: '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Tuple = tokenizer( ["I am a small frog" * 10_24, "I am a small frog"] ,padding=lowerCAmelCase__ ,truncation=lowerCAmelCase__ ,return_tensors="pt" ) self.assertIsInstance(lowerCAmelCase__ ,lowerCAmelCase__ ) self.assertEqual(batch.input_ids.shape ,(2, 51_22) ) @require_torch def UpperCAmelCase_ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Tuple = ["A long paragraph for summarization."] lowerCAmelCase_ : Dict = [ "Summary of the text.", ] for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : Optional[Any] = tokenizer(lowerCAmelCase__ ,return_tensors="pt" ) lowerCAmelCase_ : Optional[Any] = tokenizer(text_target=lowerCAmelCase__ ,return_tensors="pt" ) lowerCAmelCase_ : List[str] = inputs["input_ids"] lowerCAmelCase_ : Any = targets["input_ids"] self.assertTrue((input_ids[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((labels[:, 0] == tokenizer.bos_token_id).all().item() ) self.assertTrue((input_ids[:, -1] == tokenizer.eos_token_id).all().item() ) self.assertTrue((labels[:, -1] == tokenizer.eos_token_id).all().item() ) @require_torch def UpperCAmelCase_ ( self : str ) -> Tuple: '''simple docstring''' for tokenizer in [self.default_tokenizer, self.default_tokenizer_fast]: lowerCAmelCase_ : str = ["Summary of the text.", "Another summary."] lowerCAmelCase_ : str = [[0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, -1, -1]] lowerCAmelCase_ : List[Any] = tokenizer(lowerCAmelCase__ ,padding=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = [[0] * len(lowerCAmelCase__ ) for x in encoded_output["input_ids"]] lowerCAmelCase_ : Optional[int] = tokenizer.pad(lowerCAmelCase__ ) self.assertSequenceEqual(outputs["global_attention_mask"] ,lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Dict: '''simple docstring''' pass def UpperCAmelCase_ ( self : str ) -> Union[str, Any]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowerCAmelCase_ : Dict = self.rust_tokenizer_class.from_pretrained(lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = self.tokenizer_class.from_pretrained(lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Dict = "A, <mask> AllenNLP sentence." lowerCAmelCase_ : Tuple = tokenizer_r.encode_plus(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,return_token_type_ids=lowerCAmelCase__ ) lowerCAmelCase_ : int = tokenizer_p.encode_plus(lowerCAmelCase__ ,add_special_tokens=lowerCAmelCase__ ,return_token_type_ids=lowerCAmelCase__ ) self.assertEqual(sum(tokens_r["token_type_ids"] ) ,sum(tokens_p["token_type_ids"] ) ) self.assertEqual( sum(tokens_r["attention_mask"] ) / len(tokens_r["attention_mask"] ) ,sum(tokens_p["attention_mask"] ) / len(tokens_p["attention_mask"] ) ,) lowerCAmelCase_ : Any = tokenizer_r.convert_ids_to_tokens(tokens_r["input_ids"] ) lowerCAmelCase_ : Union[str, Any] = tokenizer_p.convert_ids_to_tokens(tokens_p["input_ids"] ) self.assertSequenceEqual(tokens_p["input_ids"] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual(tokens_r["input_ids"] ,[0, 2_50, 6, 5_02_64, 38_23, 4_87, 2_19_92, 36_45, 4, 2] ) self.assertSequenceEqual( lowerCAmelCase__ ,["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] ) self.assertSequenceEqual( lowerCAmelCase__ ,["<s>", "A", ",", "<mask>", "ĠAllen", "N", "LP", "Ġsentence", ".", "</s>"] )
659
0
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class lowerCamelCase__ ( datasets.BeamBasedBuilder): """simple docstring""" def snake_case_ ( self : List[Any] ) -> List[str]: return datasets.DatasetInfo( features=datasets.Features({'''content''': datasets.Value('''string''' )} ) , supervised_keys=__lowerCAmelCase , ) def snake_case_ ( self : Optional[Any] , __lowerCAmelCase : List[str] , __lowerCAmelCase : Dict ) -> List[Any]: return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''examples''': get_test_dummy_examples()} )] def snake_case_ ( self : List[Any] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Tuple ) -> List[str]: import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(__lowerCAmelCase ) class lowerCamelCase__ ( datasets.BeamBasedBuilder): """simple docstring""" def snake_case_ ( self : Tuple ) -> int: return datasets.DatasetInfo( features=datasets.Features({'''a''': datasets.Sequence({'''b''': datasets.Value('''string''' )} )} ) , supervised_keys=__lowerCAmelCase , ) def snake_case_ ( self : List[str] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Optional[int] ) -> Dict: return [ datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''examples''': get_test_nested_examples()} ) ] def snake_case_ ( self : Optional[int] , __lowerCAmelCase : Tuple , __lowerCAmelCase : Any ) -> str: import apache_beam as beam return pipeline | "Load Examples" >> beam.Create(__lowerCAmelCase ) def SCREAMING_SNAKE_CASE_ ( ) -> Any: return [(i, {"content": content}) for i, content in enumerate(['''foo''', '''bar''', '''foobar'''] )] def SCREAMING_SNAKE_CASE_ ( ) -> Tuple: return [(i, {"a": {"b": [content]}}) for i, content in enumerate(['''foo''', '''bar''', '''foobar'''] )] class lowerCamelCase__ ( _A): """simple docstring""" @require_beam def snake_case_ ( self : Union[str, Any] ) -> List[str]: _A = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _A = DummyBeamDataset(cache_dir=__lowerCAmelCase , beam_runner='''DirectRunner''' ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(__lowerCAmelCase , builder.name , '''default''' , '''0.0.0''' , f'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({'''content''': datasets.Value('''string''' )} ) ) _A = builder.as_dataset() self.assertEqual(dset['''train'''].num_rows , __lowerCAmelCase ) self.assertEqual(dset['''train'''].info.splits['''train'''].num_examples , __lowerCAmelCase ) self.assertDictEqual(dset['''train'''][0] , get_test_dummy_examples()[0][1] ) self.assertDictEqual( dset['''train'''][expected_num_examples - 1] , get_test_dummy_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(__lowerCAmelCase , builder.name , '''default''' , '''0.0.0''' , '''dataset_info.json''' ) ) ) del dset @require_beam def snake_case_ ( self : int ) -> str: import apache_beam as beam _A = beam.io.parquetio.WriteToParquet _A = len(get_test_dummy_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _A = DummyBeamDataset(cache_dir=__lowerCAmelCase , beam_runner='''DirectRunner''' ) with patch('''apache_beam.io.parquetio.WriteToParquet''' ) as write_parquet_mock: _A = partial(__lowerCAmelCase , num_shards=2 ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join( __lowerCAmelCase , builder.name , '''default''' , '''0.0.0''' , f'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertTrue( os.path.exists( os.path.join( __lowerCAmelCase , builder.name , '''default''' , '''0.0.0''' , f'''{builder.name}-train-00000-of-00002.arrow''' ) ) ) self.assertDictEqual(builder.info.features , datasets.Features({'''content''': datasets.Value('''string''' )} ) ) _A = builder.as_dataset() self.assertEqual(dset['''train'''].num_rows , __lowerCAmelCase ) self.assertEqual(dset['''train'''].info.splits['''train'''].num_examples , __lowerCAmelCase ) # Order is not preserved when sharding, so we just check that all the elements are there self.assertListEqual(sorted(dset['''train''']['''content'''] ) , sorted(['''foo''', '''bar''', '''foobar'''] ) ) self.assertTrue( os.path.exists(os.path.join(__lowerCAmelCase , builder.name , '''default''' , '''0.0.0''' , '''dataset_info.json''' ) ) ) del dset @require_beam def snake_case_ ( self : Optional[Any] ) -> Optional[int]: with tempfile.TemporaryDirectory() as tmp_cache_dir: _A = DummyBeamDataset(cache_dir=__lowerCAmelCase ) self.assertRaises(datasets.builder.MissingBeamOptions , builder.download_and_prepare ) @require_beam def snake_case_ ( self : Any ) -> int: _A = len(get_test_nested_examples() ) with tempfile.TemporaryDirectory() as tmp_cache_dir: _A = NestedBeamDataset(cache_dir=__lowerCAmelCase , beam_runner='''DirectRunner''' ) builder.download_and_prepare() self.assertTrue( os.path.exists( os.path.join(__lowerCAmelCase , builder.name , '''default''' , '''0.0.0''' , f'''{builder.name}-train.arrow''' ) ) ) self.assertDictEqual( builder.info.features , datasets.Features({'''a''': datasets.Sequence({'''b''': datasets.Value('''string''' )} )} ) ) _A = builder.as_dataset() self.assertEqual(dset['''train'''].num_rows , __lowerCAmelCase ) self.assertEqual(dset['''train'''].info.splits['''train'''].num_examples , __lowerCAmelCase ) self.assertDictEqual(dset['''train'''][0] , get_test_nested_examples()[0][1] ) self.assertDictEqual( dset['''train'''][expected_num_examples - 1] , get_test_nested_examples()[expected_num_examples - 1][1] ) self.assertTrue( os.path.exists(os.path.join(__lowerCAmelCase , builder.name , '''default''' , '''0.0.0''' , '''dataset_info.json''' ) ) ) del dset
2
from ....configuration_utils import PretrainedConfig from ....utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''Visual-Attention-Network/van-base''': ( '''https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json''' ), } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = 'van' def __init__( self : List[str] ,lowerCAmelCase__ : int=2_24 ,lowerCAmelCase__ : Optional[int]=3 ,lowerCAmelCase__ : Dict=[7, 3, 3, 3] ,lowerCAmelCase__ : List[str]=[4, 2, 2, 2] ,lowerCAmelCase__ : Union[str, Any]=[64, 1_28, 3_20, 5_12] ,lowerCAmelCase__ : Union[str, Any]=[3, 3, 12, 3] ,lowerCAmelCase__ : Any=[8, 8, 4, 4] ,lowerCAmelCase__ : Optional[int]="gelu" ,lowerCAmelCase__ : List[str]=0.02 ,lowerCAmelCase__ : Optional[Any]=1e-6 ,lowerCAmelCase__ : Dict=1e-2 ,lowerCAmelCase__ : Union[str, Any]=0.0 ,lowerCAmelCase__ : Optional[Any]=0.0 ,**lowerCAmelCase__ : List[str] ,) -> Tuple: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = image_size lowerCAmelCase_ : List[str] = num_channels lowerCAmelCase_ : str = patch_sizes lowerCAmelCase_ : Optional[Any] = strides lowerCAmelCase_ : List[Any] = hidden_sizes lowerCAmelCase_ : int = depths lowerCAmelCase_ : int = mlp_ratios lowerCAmelCase_ : str = hidden_act lowerCAmelCase_ : List[str] = initializer_range lowerCAmelCase_ : Dict = layer_norm_eps lowerCAmelCase_ : str = layer_scale_init_value lowerCAmelCase_ : Tuple = drop_path_rate lowerCAmelCase_ : Dict = dropout_rate
659
0
'''simple docstring''' def A_( A : int): UpperCamelCase = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(27)) print(perfect_cube(4))
3
from math import factorial def UpperCamelCase ( snake_case__ , snake_case__): # If either of the conditions are true, the function is being asked # to calculate a factorial of a negative number, which is not possible if n < k or k < 0: raise ValueError("Please enter positive integers for n and k where n >= k") return factorial(snake_case__) // (factorial(snake_case__) * factorial(n - k)) if __name__ == "__main__": print( '''The number of five-card hands possible from a standard''', f"fifty-two card deck is: {combinations(52, 5)}\n", ) print( '''If a class of 40 students must be arranged into groups of''', f"4 for group projects, there are {combinations(40, 4)} ways", '''to arrange them.\n''', ) print( '''If 10 teams are competing in a Formula One race, there''', f"are {combinations(10, 3)} ways that first, second and", '''third place can be awarded.''', )
659
0
"""simple docstring""" import inspect import os import torch from transformers import AutoModel from transformers.testing_utils import mockenv_context from transformers.trainer_utils import set_seed import accelerate from accelerate.accelerator import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils.testing import ( AccelerateTestCase, TempDirTestCase, execute_subprocess_async, require_cuda, require_fsdp, require_multi_gpu, slow, ) from accelerate.utils.constants import ( FSDP_AUTO_WRAP_POLICY, FSDP_BACKWARD_PREFETCH, FSDP_SHARDING_STRATEGY, FSDP_STATE_DICT_TYPE, ) from accelerate.utils.dataclasses import FullyShardedDataParallelPlugin from accelerate.utils.other import patch_environment set_seed(42) __UpperCamelCase : Optional[int] = '''bert-base-cased''' __UpperCamelCase : List[Any] = '''fp16''' __UpperCamelCase : Any = '''bf16''' __UpperCamelCase : Dict = [FPaa, BFaa] @require_fsdp @require_cuda class a ( a__ ): def UpperCamelCase__ ( self ): """simple docstring""" super().setUp() lowerCAmelCase = dict( ACCELERATE_USE_FSDP='true' , MASTER_ADDR='localhost' , MASTER_PORT='10999' , RANK='0' , LOCAL_RANK='0' , WORLD_SIZE='1' , ) def UpperCamelCase__ ( self ): """simple docstring""" from torch.distributed.fsdp.fully_sharded_data_parallel import ShardingStrategy for i, strategy in enumerate(_snake_case ): lowerCAmelCase = self.dist_env.copy() lowerCAmelCase = F'{i + 1}' lowerCAmelCase = strategy with mockenv_context(**_snake_case ): lowerCAmelCase = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.sharding_strategy , ShardingStrategy(i + 1 ) ) def UpperCamelCase__ ( self ): """simple docstring""" from torch.distributed.fsdp.fully_sharded_data_parallel import BackwardPrefetch for i, prefetch_policy in enumerate(_snake_case ): lowerCAmelCase = self.dist_env.copy() lowerCAmelCase = prefetch_policy with mockenv_context(**_snake_case ): lowerCAmelCase = FullyShardedDataParallelPlugin() if prefetch_policy == "NO_PREFETCH": self.assertIsNone(fsdp_plugin.backward_prefetch ) else: self.assertEqual(fsdp_plugin.backward_prefetch , BackwardPrefetch(i + 1 ) ) def UpperCamelCase__ ( self ): """simple docstring""" from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType for i, state_dict_type in enumerate(_snake_case ): lowerCAmelCase = self.dist_env.copy() lowerCAmelCase = state_dict_type with mockenv_context(**_snake_case ): lowerCAmelCase = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.state_dict_type , StateDictType(i + 1 ) ) if state_dict_type == "FULL_STATE_DICT": self.assertTrue(fsdp_plugin.state_dict_config.offload_to_cpu ) self.assertTrue(fsdp_plugin.state_dict_config.ranka_only ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = AutoModel.from_pretrained(_snake_case ) for policy in FSDP_AUTO_WRAP_POLICY: lowerCAmelCase = self.dist_env.copy() lowerCAmelCase = policy if policy == "TRANSFORMER_BASED_WRAP": lowerCAmelCase = 'BertLayer' elif policy == "SIZE_BASED_WRAP": lowerCAmelCase = '2000' with mockenv_context(**_snake_case ): lowerCAmelCase = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(_snake_case ) if policy == "NO_WRAP": self.assertIsNone(fsdp_plugin.auto_wrap_policy ) else: self.assertIsNotNone(fsdp_plugin.auto_wrap_policy ) lowerCAmelCase = self.dist_env.copy() lowerCAmelCase = 'TRANSFORMER_BASED_WRAP' lowerCAmelCase = 'T5Layer' with mockenv_context(**_snake_case ): lowerCAmelCase = FullyShardedDataParallelPlugin() with self.assertRaises(_snake_case ) as cm: fsdp_plugin.set_auto_wrap_policy(_snake_case ) self.assertTrue('Could not find the transformer layer class to wrap in the model.' in str(cm.exception ) ) lowerCAmelCase = self.dist_env.copy() lowerCAmelCase = 'SIZE_BASED_WRAP' lowerCAmelCase = '0' with mockenv_context(**_snake_case ): lowerCAmelCase = FullyShardedDataParallelPlugin() fsdp_plugin.set_auto_wrap_policy(_snake_case ) self.assertIsNone(fsdp_plugin.auto_wrap_policy ) def UpperCamelCase__ ( self ): """simple docstring""" from torch.distributed.fsdp.fully_sharded_data_parallel import MixedPrecision from torch.distributed.fsdp.sharded_grad_scaler import ShardedGradScaler for mp_dtype in dtypes: lowerCAmelCase = self.dist_env.copy() lowerCAmelCase = mp_dtype with mockenv_context(**_snake_case ): lowerCAmelCase = Accelerator() if mp_dtype == "fp16": lowerCAmelCase = torch.floataa elif mp_dtype == "bf16": lowerCAmelCase = torch.bfloataa lowerCAmelCase = MixedPrecision(param_dtype=_snake_case , reduce_dtype=_snake_case , buffer_dtype=_snake_case ) self.assertEqual(accelerator.state.fsdp_plugin.mixed_precision_policy , _snake_case ) if mp_dtype == FPaa: self.assertTrue(isinstance(accelerator.scaler , _snake_case ) ) elif mp_dtype == BFaa: self.assertIsNone(accelerator.scaler ) AcceleratorState._reset_state(_snake_case ) def UpperCamelCase__ ( self ): """simple docstring""" from torch.distributed.fsdp.fully_sharded_data_parallel import CPUOffload for flag in [True, False]: lowerCAmelCase = self.dist_env.copy() lowerCAmelCase = str(_snake_case ).lower() with mockenv_context(**_snake_case ): lowerCAmelCase = FullyShardedDataParallelPlugin() self.assertEqual(fsdp_plugin.cpu_offload , CPUOffload(offload_params=_snake_case ) ) @require_fsdp @require_multi_gpu @slow class a ( a__ ): def UpperCamelCase__ ( self ): """simple docstring""" super().setUp() lowerCAmelCase = 0.82 lowerCAmelCase = [ 'fsdp_shard_grad_op_transformer_based_wrap', 'fsdp_full_shard_transformer_based_wrap', ] lowerCAmelCase = { 'multi_gpu_fp16': 32_00, 'fsdp_shard_grad_op_transformer_based_wrap_fp16': 20_00, 'fsdp_full_shard_transformer_based_wrap_fp16': 19_00, # Disabling below test as it overwhelms the RAM memory usage # on CI self-hosted runner leading to tests getting killed. # "fsdp_full_shard_cpu_offload_transformer_based_wrap_fp32": 1500, # fp16 was leading to indefinite hang } lowerCAmelCase = 1_60 lowerCAmelCase = 1_60 lowerCAmelCase = inspect.getfile(accelerate.test_utils ) lowerCAmelCase = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ['scripts', 'external_deps'] ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = os.path.join(self.test_scripts_folder , 'test_performance.py' ) lowerCAmelCase = ['accelerate', 'launch', '--num_processes=2', '--num_machines=1', '--machine_rank=0', '--use_fsdp'] for config in self.performance_configs: lowerCAmelCase = cmd.copy() for i, strategy in enumerate(_snake_case ): if strategy.lower() in config: cmd_config.append(F'--fsdp_sharding_strategy={i+1}' ) break if "fp32" in config: cmd_config.append('--mixed_precision=no' ) else: cmd_config.append('--mixed_precision=fp16' ) if "cpu_offload" in config: cmd_config.append('--fsdp_offload_params=True' ) for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in config: cmd_config.append(F'--fsdp_auto_wrap_policy={policy}' ) break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append('--fsdp_transformer_layer_cls_to_wrap=BertLayer' ) elif policy == "SIZE_BASED_WRAP": cmd_config.append('--fsdp_min_num_params=2000' ) cmd_config.extend( [ self.test_file_path, F'--output_dir={self.tmpdir}', F'--performance_lower_bound={self.performance_lower_bound}', ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_snake_case , env=os.environ.copy() ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = os.path.join(self.test_scripts_folder , 'test_checkpointing.py' ) lowerCAmelCase = [ 'accelerate', 'launch', '--num_processes=2', '--num_machines=1', '--machine_rank=0', '--use_fsdp', '--mixed_precision=fp16', '--fsdp_transformer_layer_cls_to_wrap=BertLayer', ] for i, strategy in enumerate(_snake_case ): lowerCAmelCase = cmd.copy() cmd_config.append(F'--fsdp_sharding_strategy={i+1}' ) if strategy != "FULL_SHARD": continue lowerCAmelCase = len(_snake_case ) for state_dict_type in FSDP_STATE_DICT_TYPE: lowerCAmelCase = cmd_config[:state_dict_config_index] cmd_config.append(F'--fsdp_state_dict_type={state_dict_type}' ) cmd_config.extend( [ self.test_file_path, F'--output_dir={self.tmpdir}', '--partial_train_epoch=1', ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_snake_case , env=os.environ.copy() ) lowerCAmelCase = cmd_config[:-1] lowerCAmelCase = os.path.join(self.tmpdir , 'epoch_0' ) cmd_config.extend( [ F'--resume_from_checkpoint={resume_from_checkpoint}', ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_snake_case , env=os.environ.copy() ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = os.path.join(self.test_scripts_folder , 'test_peak_memory_usage.py' ) lowerCAmelCase = [ 'accelerate', 'launch', '--num_processes=2', '--num_machines=1', '--machine_rank=0', ] for spec, peak_mem_upper_bound in self.peak_memory_usage_upper_bound.items(): lowerCAmelCase = cmd.copy() if "fp16" in spec: cmd_config.extend(['--mixed_precision=fp16'] ) else: cmd_config.extend(['--mixed_precision=no'] ) if "multi_gpu" in spec: continue else: cmd_config.extend(['--use_fsdp'] ) for i, strategy in enumerate(_snake_case ): if strategy.lower() in spec: cmd_config.append(F'--fsdp_sharding_strategy={i+1}' ) break if "cpu_offload" in spec: cmd_config.append('--fsdp_offload_params=True' ) for policy in FSDP_AUTO_WRAP_POLICY: if policy.lower() in spec: cmd_config.append(F'--fsdp_auto_wrap_policy={policy}' ) break if policy == "TRANSFORMER_BASED_WRAP": cmd_config.append('--fsdp_transformer_layer_cls_to_wrap=BertLayer' ) elif policy == "SIZE_BASED_WRAP": cmd_config.append('--fsdp_min_num_params=2000' ) cmd_config.extend( [ self.test_file_path, F'--output_dir={self.tmpdir}', F'--peak_memory_upper_bound={peak_mem_upper_bound}', F'--n_train={self.n_train}', F'--n_val={self.n_val}', ] ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(_snake_case , env=os.environ.copy() )
4
import argparse import json from tqdm import tqdm def UpperCamelCase ( ): lowerCAmelCase_ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( "--src_path" , type=snake_case__ , default="biencoder-nq-dev.json" , help="Path to raw DPR training data" , ) parser.add_argument( "--evaluation_set" , type=snake_case__ , help="where to store parsed evaluation_set file" , ) parser.add_argument( "--gold_data_path" , type=snake_case__ , help="where to store parsed gold_data_path file" , ) lowerCAmelCase_ : Dict = parser.parse_args() with open(args.src_path , "r") as src_file, open(args.evaluation_set , "w") as eval_file, open( args.gold_data_path , "w") as gold_file: lowerCAmelCase_ : Optional[int] = json.load(snake_case__) for dpr_record in tqdm(snake_case__): lowerCAmelCase_ : str = dpr_record["question"] lowerCAmelCase_ : Dict = [context["title"] for context in dpr_record["positive_ctxs"]] eval_file.write(question + "\n") gold_file.write("\t".join(snake_case__) + "\n") if __name__ == "__main__": main()
659
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { """facebook/data2vec-text-base""": """https://huggingface.co/data2vec/resolve/main/config.json""", } class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' _lowercase : List[Any] = '''data2vec-text''' def __init__( self , _lowercase=30_522 , _lowercase=768 , _lowercase=12 , _lowercase=12 , _lowercase=3_072 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=512 , _lowercase=2 , _lowercase=0.02 , _lowercase=1e-12 , _lowercase=1 , _lowercase=0 , _lowercase=2 , _lowercase="absolute" , _lowercase=True , _lowercase=None , **_lowercase , ): """simple docstring""" super().__init__(pad_token_id=_lowercase , bos_token_id=_lowercase , eos_token_id=_lowercase , **_lowercase ) _lowerCAmelCase = vocab_size _lowerCAmelCase = hidden_size _lowerCAmelCase = num_hidden_layers _lowerCAmelCase = num_attention_heads _lowerCAmelCase = hidden_act _lowerCAmelCase = intermediate_size _lowerCAmelCase = hidden_dropout_prob _lowerCAmelCase = attention_probs_dropout_prob _lowerCAmelCase = max_position_embeddings _lowerCAmelCase = type_vocab_size _lowerCAmelCase = initializer_range _lowerCAmelCase = layer_norm_eps _lowerCAmelCase = position_embedding_type _lowerCAmelCase = use_cache _lowerCAmelCase = classifier_dropout class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def _lowercase ( self ): """simple docstring""" if self.task == "multiple-choice": _lowerCAmelCase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: _lowerCAmelCase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
5
from collections.abc import Sequence def UpperCamelCase ( snake_case__ = None): if nums is None or not nums: raise ValueError("Input sequence should not be empty") lowerCAmelCase_ : Dict = nums[0] for i in range(1 , len(snake_case__)): lowerCAmelCase_ : Optional[int] = nums[i] lowerCAmelCase_ : Optional[int] = max(snake_case__ , ans + num , snake_case__) return ans if __name__ == "__main__": import doctest doctest.testmod() # Try on a sample input from the user _lowercase = int(input('''Enter number of elements : ''').strip()) _lowercase = list(map(int, input('''\nEnter the numbers : ''').strip().split()))[:n] print(max_subsequence_sum(array))
659
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Tuple=None ): if subparsers is not None: SCREAMING_SNAKE_CASE__ = subparsers.add_parser("""test""" ) else: SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser("""Accelerate test command""" ) parser.add_argument( """--config_file""" , default=UpperCamelCase__ , help=( """The path to use to store the config file. Will default to a file named default_config.yaml in the cache """ """location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have """ """such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed """ """with 'huggingface'.""" ) , ) if subparsers is not None: parser.set_defaults(func=UpperCamelCase__ ) return parser def SCREAMING_SNAKE_CASE__ ( UpperCamelCase__: Optional[Any] ): SCREAMING_SNAKE_CASE__ = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["""test_utils""", """scripts""", """test_script.py"""] ) if args.config_file is None: SCREAMING_SNAKE_CASE__ = script_name else: SCREAMING_SNAKE_CASE__ = f'''--config_file={args.config_file} {script_name}''' SCREAMING_SNAKE_CASE__ = ["""accelerate-launch"""] + test_args.split() SCREAMING_SNAKE_CASE__ = execute_subprocess_async(UpperCamelCase__ , env=os.environ.copy() ) if result.returncode == 0: print("""Test is a success! You are ready for your distributed training!""" ) def SCREAMING_SNAKE_CASE__ ( ): SCREAMING_SNAKE_CASE__ = test_command_parser() SCREAMING_SNAKE_CASE__ = parser.parse_args() test_command(UpperCamelCase__ ) if __name__ == "__main__": main()
6
from typing import TYPE_CHECKING from ....utils import _LazyModule _lowercase = {'''tokenization_tapex''': ['''TapexTokenizer''']} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys _lowercase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
659
0
"""simple docstring""" import json import os import torch from diffusers import UNetaDModel os.makedirs('''hub/hopper-medium-v2/unet/hor32''', exist_ok=True) os.makedirs('''hub/hopper-medium-v2/unet/hor128''', exist_ok=True) os.makedirs('''hub/hopper-medium-v2/value_function''', exist_ok=True) def _snake_case ( _snake_case : int ) -> Any: '''simple docstring''' if hor == 1_28: _A = ('DownResnetBlock1D', 'DownResnetBlock1D', 'DownResnetBlock1D') _A = (32, 1_28, 2_56) _A = ('UpResnetBlock1D', 'UpResnetBlock1D') elif hor == 32: _A = ('DownResnetBlock1D', 'DownResnetBlock1D', 'DownResnetBlock1D', 'DownResnetBlock1D') _A = (32, 64, 1_28, 2_56) _A = ('UpResnetBlock1D', 'UpResnetBlock1D', 'UpResnetBlock1D') _A = torch.load(F'''/Users/bglickenhaus/Documents/diffuser/temporal_unet-hopper-mediumv2-hor{hor}.torch''' ) _A = model.state_dict() _A = { 'down_block_types': down_block_types, 'block_out_channels': block_out_channels, 'up_block_types': up_block_types, 'layers_per_block': 1, 'use_timestep_embedding': True, 'out_block_type': 'OutConv1DBlock', 'norm_num_groups': 8, 'downsample_each_block': False, 'in_channels': 14, 'out_channels': 14, 'extra_in_channels': 0, 'time_embedding_type': 'positional', 'flip_sin_to_cos': False, 'freq_shift': 1, 'sample_size': 6_55_36, 'mid_block_type': 'MidResTemporalBlock1D', 'act_fn': 'mish', } _A = UNetaDModel(**_snake_case ) print(F'''length of state dict: {len(state_dict.keys() )}''' ) print(F'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) _A = dict(zip(model.state_dict().keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): _A = state_dict.pop(_snake_case ) hf_value_function.load_state_dict(_snake_case ) torch.save(hf_value_function.state_dict() , F'''hub/hopper-medium-v2/unet/hor{hor}/diffusion_pytorch_model.bin''' ) with open(F'''hub/hopper-medium-v2/unet/hor{hor}/config.json''' , 'w' ) as f: json.dump(_snake_case , _snake_case ) def _snake_case ( ) -> List[str]: '''simple docstring''' _A = { 'in_channels': 14, 'down_block_types': ('DownResnetBlock1D', 'DownResnetBlock1D', 'DownResnetBlock1D', 'DownResnetBlock1D'), 'up_block_types': (), 'out_block_type': 'ValueFunction', 'mid_block_type': 'ValueFunctionMidBlock1D', 'block_out_channels': (32, 64, 1_28, 2_56), 'layers_per_block': 1, 'downsample_each_block': True, 'sample_size': 6_55_36, 'out_channels': 14, 'extra_in_channels': 0, 'time_embedding_type': 'positional', 'use_timestep_embedding': True, 'flip_sin_to_cos': False, 'freq_shift': 1, 'norm_num_groups': 8, 'act_fn': 'mish', } _A = torch.load('/Users/bglickenhaus/Documents/diffuser/value_function-hopper-mediumv2-hor32.torch' ) _A = model _A = UNetaDModel(**_snake_case ) print(F'''length of state dict: {len(state_dict.keys() )}''' ) print(F'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) _A = dict(zip(state_dict.keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): _A = state_dict.pop(_snake_case ) hf_value_function.load_state_dict(_snake_case ) torch.save(hf_value_function.state_dict() , 'hub/hopper-medium-v2/value_function/diffusion_pytorch_model.bin' ) with open('hub/hopper-medium-v2/value_function/config.json' , 'w' ) as f: json.dump(_snake_case , _snake_case ) if __name__ == "__main__": unet(32) # unet(128) value_function()
7
import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _lowercase = '''src/diffusers''' _lowercase = '''.''' # This is to make sure the diffusers module imported is the one in the repo. _lowercase = importlib.util.spec_from_file_location( '''diffusers''', os.path.join(DIFFUSERS_PATH, '''__init__.py'''), submodule_search_locations=[DIFFUSERS_PATH], ) _lowercase = spec.loader.load_module() def UpperCamelCase ( snake_case__ , snake_case__): return line.startswith(snake_case__) or len(snake_case__) <= 1 or re.search(R"^\s*\)(\s*->.*:|:)\s*$" , snake_case__) is not None def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Tuple = object_name.split(".") lowerCAmelCase_ : Union[str, Any] = 0 # First let's find the module where our object lives. lowerCAmelCase_ : Union[str, Any] = parts[i] while i < len(snake_case__) and not os.path.isfile(os.path.join(snake_case__ , F'''{module}.py''')): i += 1 if i < len(snake_case__): lowerCAmelCase_ : Dict = os.path.join(snake_case__ , parts[i]) if i >= len(snake_case__): raise ValueError(F'''`object_name` should begin with the name of a module of diffusers but got {object_name}.''') with open(os.path.join(snake_case__ , F'''{module}.py''') , "r" , encoding="utf-8" , newline="\n") as f: lowerCAmelCase_ : Optional[Any] = f.readlines() # Now let's find the class / func in the code! lowerCAmelCase_ : Union[str, Any] = "" lowerCAmelCase_ : int = 0 for name in parts[i + 1 :]: while ( line_index < len(snake_case__) and re.search(RF'''^{indent}(class|def)\s+{name}(\(|\:)''' , lines[line_index]) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(snake_case__): raise ValueError(F''' {object_name} does not match any function or class in {module}.''') # We found the beginning of the class / func, now let's find the end (when the indent diminishes). lowerCAmelCase_ : Union[str, Any] = line_index while line_index < len(snake_case__) and _should_continue(lines[line_index] , snake_case__): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1]) <= 1: line_index -= 1 lowerCAmelCase_ : List[str] = lines[start_index:line_index] return "".join(snake_case__) _lowercase = re.compile(r'''^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)''') _lowercase = re.compile(r'''^\s*(\S+)->(\S+)(\s+.*|$)''') _lowercase = re.compile(r'''<FILL\s+[^>]*>''') def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Any = code.split("\n") lowerCAmelCase_ : Any = 0 while idx < len(snake_case__) and len(lines[idx]) == 0: idx += 1 if idx < len(snake_case__): return re.search(R"^(\s*)\S" , lines[idx]).groups()[0] return "" def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Dict = len(get_indent(snake_case__)) > 0 if has_indent: lowerCAmelCase_ : Dict = F'''class Bla:\n{code}''' lowerCAmelCase_ : Optional[int] = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=1_19 , preview=snake_case__) lowerCAmelCase_ : Optional[Any] = black.format_str(snake_case__ , mode=snake_case__) lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = style_docstrings_in_code(snake_case__) return result[len("class Bla:\n") :] if has_indent else result def UpperCamelCase ( snake_case__ , snake_case__=False): with open(snake_case__ , "r" , encoding="utf-8" , newline="\n") as f: lowerCAmelCase_ : Tuple = f.readlines() lowerCAmelCase_ : Tuple = [] lowerCAmelCase_ : Union[str, Any] = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(snake_case__): lowerCAmelCase_ : Optional[int] = _re_copy_warning.search(lines[line_index]) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : str = search.groups() lowerCAmelCase_ : int = find_code_in_diffusers(snake_case__) lowerCAmelCase_ : Dict = get_indent(snake_case__) lowerCAmelCase_ : Union[str, Any] = line_index + 1 if indent == theoretical_indent else line_index + 2 lowerCAmelCase_ : str = theoretical_indent lowerCAmelCase_ : Union[str, Any] = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. lowerCAmelCase_ : Optional[int] = True while line_index < len(snake_case__) and should_continue: line_index += 1 if line_index >= len(snake_case__): break lowerCAmelCase_ : Dict = lines[line_index] lowerCAmelCase_ : List[str] = _should_continue(snake_case__ , snake_case__) and re.search(F'''^{indent}# End copy''' , snake_case__) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1]) <= 1: line_index -= 1 lowerCAmelCase_ : Dict = lines[start_index:line_index] lowerCAmelCase_ : Optional[int] = "".join(snake_case__) # Remove any nested `Copied from` comments to avoid circular copies lowerCAmelCase_ : List[Any] = [line for line in theoretical_code.split("\n") if _re_copy_warning.search(snake_case__) is None] lowerCAmelCase_ : Optional[Any] = "\n".join(snake_case__) # Before comparing, use the `replace_pattern` on the original code. if len(snake_case__) > 0: lowerCAmelCase_ : List[str] = replace_pattern.replace("with" , "").split(",") lowerCAmelCase_ : Tuple = [_re_replace_pattern.search(snake_case__) for p in patterns] for pattern in patterns: if pattern is None: continue lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[str] = pattern.groups() lowerCAmelCase_ : int = re.sub(snake_case__ , snake_case__ , snake_case__) if option.strip() == "all-casing": lowerCAmelCase_ : List[str] = re.sub(obja.lower() , obja.lower() , snake_case__) lowerCAmelCase_ : int = re.sub(obja.upper() , obja.upper() , snake_case__) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line lowerCAmelCase_ : List[Any] = blackify(lines[start_index - 1] + theoretical_code) lowerCAmelCase_ : Union[str, Any] = theoretical_code[len(lines[start_index - 1]) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index]) if overwrite: lowerCAmelCase_ : List[Any] = lines[:start_index] + [theoretical_code] + lines[line_index:] lowerCAmelCase_ : Union[str, Any] = start_index + 1 if overwrite and len(snake_case__) > 0: # Warn the user a file has been modified. print(F'''Detected changes, rewriting {filename}.''') with open(snake_case__ , "w" , encoding="utf-8" , newline="\n") as f: f.writelines(snake_case__) return diffs def UpperCamelCase ( snake_case__ = False): lowerCAmelCase_ : Tuple = glob.glob(os.path.join(snake_case__ , "**/*.py") , recursive=snake_case__) lowerCAmelCase_ : int = [] for filename in all_files: lowerCAmelCase_ : Union[str, Any] = is_copy_consistent(snake_case__ , snake_case__) diffs += [F'''- {filename}: copy does not match {d[0]} at line {d[1]}''' for d in new_diffs] if not overwrite and len(snake_case__) > 0: lowerCAmelCase_ : Optional[Any] = "\n".join(snake_case__) raise Exception( "Found the following copy inconsistencies:\n" + diff + "\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.") if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') _lowercase = parser.parse_args() check_copies(args.fix_and_overwrite)
659
0
'''simple docstring''' def _lowerCAmelCase ( __snake_case : list ) -> list: __A : Dict = False while is_sorted is False: # Until all the indices are traversed keep looping __A : int = True for i in range(0 , len(__snake_case ) - 1 , 2 ): # iterating over all even indices if input_list[i] > input_list[i + 1]: __A ,__A : List[Any] = input_list[i + 1], input_list[i] # swapping if elements not in order __A : int = False for i in range(1 , len(__snake_case ) - 1 , 2 ): # iterating over all odd indices if input_list[i] > input_list[i + 1]: __A ,__A : Optional[int] = input_list[i + 1], input_list[i] # swapping if elements not in order __A : str = False return input_list if __name__ == "__main__": print('''Enter list to be sorted''') lowercase__ : Dict = [int(x) for x in input().split()] # inputing elements of the list in one line lowercase__ : Union[str, Any] = odd_even_sort(input_list) print('''The sorted list is''') print(sorted_list)
8
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''microsoft/swinv2-tiny-patch4-window8-256''': ( '''https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/config.json''' ), } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = 'swinv2' UpperCamelCase_ = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : List[Any] ,lowerCAmelCase__ : Optional[int]=2_24 ,lowerCAmelCase__ : Dict=4 ,lowerCAmelCase__ : Dict=3 ,lowerCAmelCase__ : List[Any]=96 ,lowerCAmelCase__ : Optional[Any]=[2, 2, 6, 2] ,lowerCAmelCase__ : Optional[Any]=[3, 6, 12, 24] ,lowerCAmelCase__ : Optional[int]=7 ,lowerCAmelCase__ : Dict=4.0 ,lowerCAmelCase__ : Dict=True ,lowerCAmelCase__ : str=0.0 ,lowerCAmelCase__ : Tuple=0.0 ,lowerCAmelCase__ : str=0.1 ,lowerCAmelCase__ : List[str]="gelu" ,lowerCAmelCase__ : Union[str, Any]=False ,lowerCAmelCase__ : Dict=0.02 ,lowerCAmelCase__ : int=1e-5 ,lowerCAmelCase__ : List[str]=32 ,**lowerCAmelCase__ : Tuple ,) -> List[str]: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = image_size lowerCAmelCase_ : List[Any] = patch_size lowerCAmelCase_ : Dict = num_channels lowerCAmelCase_ : Optional[int] = embed_dim lowerCAmelCase_ : Optional[Any] = depths lowerCAmelCase_ : Any = len(lowerCAmelCase__ ) lowerCAmelCase_ : str = num_heads lowerCAmelCase_ : List[str] = window_size lowerCAmelCase_ : List[str] = mlp_ratio lowerCAmelCase_ : Dict = qkv_bias lowerCAmelCase_ : str = hidden_dropout_prob lowerCAmelCase_ : str = attention_probs_dropout_prob lowerCAmelCase_ : Union[str, Any] = drop_path_rate lowerCAmelCase_ : List[Any] = hidden_act lowerCAmelCase_ : Any = use_absolute_embeddings lowerCAmelCase_ : List[str] = layer_norm_eps lowerCAmelCase_ : int = initializer_range lowerCAmelCase_ : Union[str, Any] = encoder_stride # we set the hidden_size attribute in order to make Swinv2 work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model lowerCAmelCase_ : Tuple = int(embed_dim * 2 ** (len(lowerCAmelCase__ ) - 1) ) lowerCAmelCase_ : str = (0, 0, 0, 0)
659
0
import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def A ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) -> Tuple: A__ = s.rsplit(__UpperCamelCase , __UpperCamelCase ) return new.join(__UpperCamelCase ) def A ( __UpperCamelCase ) -> Dict: # encoder.embeddings are double copied in original FLAVA return sum(param.float().sum() if 'encoder.embeddings' not in key else 0 for key, param in state_dict.items() ) def A ( __UpperCamelCase ) -> List[str]: A__ = {} A__ = ['group_1', 'group_2', 'group_3', 'group_4'] for key, value in state_dict.items(): for group_key in group_keys: if group_key in key: A__ = key.replace(f'''{group_key}.''' , f'''{group_key}.group.''' ) if "res_path" in key: A__ = key.replace('res_path.' , 'res_path.path.' ) if key.endswith('.w' ): A__ = rreplace(__UpperCamelCase , '.w' , '.weight' , 1 ) if key.endswith('.b' ): A__ = rreplace(__UpperCamelCase , '.b' , '.bias' , 1 ) A__ = value.float() return upgrade @torch.no_grad() def A ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , __UpperCamelCase=True ) -> Optional[Any]: from dall_e import Encoder A__ = Encoder() if os.path.exists(__UpperCamelCase ): A__ = torch.load(__UpperCamelCase ) else: A__ = torch.hub.load_state_dict_from_url(__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ): A__ = ckpt.state_dict() encoder.load_state_dict(__UpperCamelCase ) if config_path is not None: A__ = FlavaImageCodebookConfig.from_pretrained(__UpperCamelCase ) else: A__ = FlavaImageCodebookConfig() A__ = FlavaImageCodebook(__UpperCamelCase ).eval() A__ = encoder.state_dict() A__ = upgrade_state_dict(__UpperCamelCase ) hf_model.load_state_dict(__UpperCamelCase ) A__ = hf_model.state_dict() A__ = count_parameters(__UpperCamelCase ) A__ = count_parameters(__UpperCamelCase ) assert torch.allclose(__UpperCamelCase , __UpperCamelCase , atol=1E-3 ) if save_checkpoint: hf_model.save_pretrained(__UpperCamelCase ) else: return hf_state_dict if __name__ == "__main__": SCREAMING_SNAKE_CASE__ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to flava checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') SCREAMING_SNAKE_CASE__ = parser.parse_args() convert_dalle_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
9
from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging _lowercase = logging.get_logger(__name__) class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = ['input_features', 'attention_mask'] def __init__( self : Optional[Any] ,lowerCAmelCase__ : Any=80 ,lowerCAmelCase__ : Optional[Any]=1_60_00 ,lowerCAmelCase__ : List[str]=0.0 ,lowerCAmelCase__ : Tuple=10 ,lowerCAmelCase__ : Optional[Any]=25 ,lowerCAmelCase__ : Any="hamming_window" ,lowerCAmelCase__ : List[str]=32_768.0 ,lowerCAmelCase__ : Union[str, Any]=0.97 ,lowerCAmelCase__ : Any=1.0 ,lowerCAmelCase__ : str=True ,lowerCAmelCase__ : int=True ,lowerCAmelCase__ : Tuple=False ,**lowerCAmelCase__ : Optional[int] ,) -> Optional[Any]: '''simple docstring''' super().__init__(feature_size=lowerCAmelCase__ ,sampling_rate=lowerCAmelCase__ ,padding_value=lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = feature_size lowerCAmelCase_ : List[Any] = sampling_rate lowerCAmelCase_ : Union[str, Any] = padding_value lowerCAmelCase_ : str = hop_length lowerCAmelCase_ : str = win_length lowerCAmelCase_ : str = frame_signal_scale lowerCAmelCase_ : Any = preemphasis_coeff lowerCAmelCase_ : Optional[Any] = mel_floor lowerCAmelCase_ : List[str] = normalize_means lowerCAmelCase_ : Optional[Any] = normalize_vars lowerCAmelCase_ : Dict = win_function lowerCAmelCase_ : List[Any] = return_attention_mask lowerCAmelCase_ : Tuple = win_length * sampling_rate // 10_00 lowerCAmelCase_ : str = hop_length * sampling_rate // 10_00 lowerCAmelCase_ : Dict = optimal_fft_length(self.sample_size ) lowerCAmelCase_ : Optional[int] = (self.n_fft // 2) + 1 def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : np.array ) -> np.ndarray: '''simple docstring''' if self.win_function == "hamming_window": lowerCAmelCase_ : int = window_function(window_length=self.sample_size ,name=self.win_function ,periodic=lowerCAmelCase__ ) else: lowerCAmelCase_ : Tuple = window_function(window_length=self.sample_size ,name=self.win_function ) lowerCAmelCase_ : List[str] = mel_filter_bank( num_frequency_bins=self.n_freqs ,num_mel_filters=self.feature_size ,min_frequency=0.0 ,max_frequency=self.sampling_rate / 2.0 ,sampling_rate=self.sampling_rate ,) lowerCAmelCase_ : Any = spectrogram( one_waveform * self.frame_signal_scale ,window=lowerCAmelCase__ ,frame_length=self.sample_size ,hop_length=self.sample_stride ,fft_length=self.n_fft ,center=lowerCAmelCase__ ,preemphasis=self.preemphasis_coeff ,mel_filters=lowerCAmelCase__ ,mel_floor=self.mel_floor ,log_mel="log" ,) return msfc_features.T def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : Optional[Any] ,lowerCAmelCase__ : Tuple ) -> Optional[Any]: '''simple docstring''' if self.normalize_means: lowerCAmelCase_ : Optional[int] = x[:input_length].mean(axis=0 ) lowerCAmelCase_ : List[str] = np.subtract(lowerCAmelCase__ ,lowerCAmelCase__ ) if self.normalize_vars: lowerCAmelCase_ : Optional[Any] = x[:input_length].std(axis=0 ) lowerCAmelCase_ : Tuple = np.divide(lowerCAmelCase__ ,lowerCAmelCase__ ) if input_length < x.shape[0]: lowerCAmelCase_ : int = padding_value # make sure array is in float32 lowerCAmelCase_ : Any = x.astype(np.floataa ) return x def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[np.ndarray] ,lowerCAmelCase__ : Optional[np.ndarray] = None ) -> List[np.ndarray]: '''simple docstring''' lowerCAmelCase_ : List[Any] = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(lowerCAmelCase__ ,lowerCAmelCase__ ,self.padding_value ) for x, n in zip(lowerCAmelCase__ ,lowerCAmelCase__ )] def __call__( self : int ,lowerCAmelCase__ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] ,lowerCAmelCase__ : Union[bool, str, PaddingStrategy] = False ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : bool = False ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : Optional[bool] = None ,lowerCAmelCase__ : Optional[Union[str, TensorType]] = None ,lowerCAmelCase__ : Optional[int] = None ,**lowerCAmelCase__ : Union[str, Any] ,) -> BatchFeature: '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) lowerCAmelCase_ : List[Any] = isinstance(lowerCAmelCase__ ,np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) lowerCAmelCase_ : str = is_batched_numpy or ( isinstance(lowerCAmelCase__ ,(list, tuple) ) and (isinstance(raw_speech[0] ,(np.ndarray, tuple, list) )) ) if is_batched: lowerCAmelCase_ : Tuple = [np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(lowerCAmelCase__ ,np.ndarray ): lowerCAmelCase_ : int = np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) elif isinstance(lowerCAmelCase__ ,np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): lowerCAmelCase_ : Union[str, Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: lowerCAmelCase_ : Optional[int] = [raw_speech] # extract fbank features lowerCAmelCase_ : Dict = [self._extract_mfsc_features(lowerCAmelCase__ ) for one_waveform in raw_speech] # convert into correct format for padding lowerCAmelCase_ : int = BatchFeature({"input_features": features} ) lowerCAmelCase_ : Union[str, Any] = self.pad( lowerCAmelCase__ ,padding=lowerCAmelCase__ ,max_length=lowerCAmelCase__ ,truncation=lowerCAmelCase__ ,pad_to_multiple_of=lowerCAmelCase__ ,return_attention_mask=lowerCAmelCase__ ,**lowerCAmelCase__ ,) # make sure list is in array format lowerCAmelCase_ : Optional[Any] = padded_inputs.get("input_features" ) if isinstance(input_features[0] ,lowerCAmelCase__ ): lowerCAmelCase_ : Optional[int] = [np.asarray(lowerCAmelCase__ ,dtype=np.floataa ) for feature in input_features] lowerCAmelCase_ : List[Any] = padded_inputs.get("attention_mask" ) if attention_mask is not None: lowerCAmelCase_ : Dict = [np.asarray(lowerCAmelCase__ ,dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: lowerCAmelCase_ : Dict = ( np.array(lowerCAmelCase__ ,dtype=np.intaa ) if self._get_padding_strategies(lowerCAmelCase__ ,max_length=lowerCAmelCase__ ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) lowerCAmelCase_ : List[str] = self.normalize( padded_inputs["input_features"] ,attention_mask=lowerCAmelCase__ ) if return_tensors is not None: lowerCAmelCase_ : Dict = padded_inputs.convert_to_tensors(lowerCAmelCase__ ) return padded_inputs
659
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCAmelCase = { "configuration_rembert": ["REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "RemBertConfig", "RemBertOnnxConfig"] } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = ["RemBertTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = ["RemBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = [ "REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "RemBertForCausalLM", "RemBertForMaskedLM", "RemBertForMultipleChoice", "RemBertForQuestionAnswering", "RemBertForSequenceClassification", "RemBertForTokenClassification", "RemBertLayer", "RemBertModel", "RemBertPreTrainedModel", "load_tf_weights_in_rembert", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = [ "TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFRemBertForCausalLM", "TFRemBertForMaskedLM", "TFRemBertForMultipleChoice", "TFRemBertForQuestionAnswering", "TFRemBertForSequenceClassification", "TFRemBertForTokenClassification", "TFRemBertLayer", "TFRemBertModel", "TFRemBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_rembert import REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RemBertConfig, RemBertOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert import RemBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert_fast import RemBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rembert import ( REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, RemBertForCausalLM, RemBertForMaskedLM, RemBertForMultipleChoice, RemBertForQuestionAnswering, RemBertForSequenceClassification, RemBertForTokenClassification, RemBertLayer, RemBertModel, RemBertPreTrainedModel, load_tf_weights_in_rembert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rembert import ( TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFRemBertForCausalLM, TFRemBertForMaskedLM, TFRemBertForMultipleChoice, TFRemBertForQuestionAnswering, TFRemBertForSequenceClassification, TFRemBertForTokenClassification, TFRemBertLayer, TFRemBertModel, TFRemBertPreTrainedModel, ) else: import sys _lowerCAmelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
10
from __future__ import annotations # This is the precision for this function which can be altered. # It is recommended for users to keep this number greater than or equal to 10. _lowercase = 10 def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): for i in range(snake_case__ , snake_case__): if array[i] == target: return i return -1 def UpperCamelCase ( snake_case__ , snake_case__): lowerCAmelCase_ : List[str] = 0 lowerCAmelCase_ : Tuple = len(snake_case__) while left <= right: if right - left < precision: return lin_search(snake_case__ , snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : List[str] = (left + right) // 3 + 1 lowerCAmelCase_ : Tuple = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: lowerCAmelCase_ : str = one_third - 1 elif array[two_third] < target: lowerCAmelCase_ : Any = two_third + 1 else: lowerCAmelCase_ : List[str] = one_third + 1 lowerCAmelCase_ : Tuple = two_third - 1 else: return -1 def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__): if left < right: if right - left < precision: return lin_search(snake_case__ , snake_case__ , snake_case__ , snake_case__) lowerCAmelCase_ : Dict = (left + right) // 3 + 1 lowerCAmelCase_ : List[Any] = 2 * (left + right) // 3 + 1 if array[one_third] == target: return one_third elif array[two_third] == target: return two_third elif target < array[one_third]: return rec_ternary_search(snake_case__ , one_third - 1 , snake_case__ , snake_case__) elif array[two_third] < target: return rec_ternary_search(two_third + 1 , snake_case__ , snake_case__ , snake_case__) else: return rec_ternary_search(one_third + 1 , two_third - 1 , snake_case__ , snake_case__) else: return -1 if __name__ == "__main__": import doctest doctest.testmod() _lowercase = input('''Enter numbers separated by comma:\n''').strip() _lowercase = [int(item.strip()) for item in user_input.split(''',''')] assert collection == sorted(collection), f"List must be ordered.\n{collection}." _lowercase = int(input('''Enter the number to be found in the list:\n''').strip()) _lowercase = ite_ternary_search(collection, target) _lowercase = rec_ternary_search(0, len(collection) - 1, collection, target) if resulta != -1: print(f"Iterative search: {target} found at positions: {resulta}") print(f"Recursive search: {target} found at positions: {resulta}") else: print('''Not found''')
659
0
'''simple docstring''' import os import re import shutil import sys import tempfile import unittest import black lowercase_ = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If BertLMPredictionHead is changed in modeling_bert.py, this code needs to be manually updated. lowercase_ = " def __init__(self, config):\n super().__init__()\n self.transform = BertPredictionHeadTransform(config)\n\n # The output weights are the same as the input embeddings, but there is\n # an output-only bias for each token.\n self.decoder = nn.Linear(config.hidden_size, config.vocab_size, bias=False)\n\n self.bias = nn.Parameter(torch.zeros(config.vocab_size))\n\n # Need a link between the two variables so that the bias is correctly resized with `resize_token_embeddings`\n self.decoder.bias = self.bias\n\n def forward(self, hidden_states):\n hidden_states = self.transform(hidden_states)\n hidden_states = self.decoder(hidden_states)\n return hidden_states\n" class __A ( unittest.TestCase ): '''simple docstring''' def a__ (self ) -> Optional[int]: """simple docstring""" _a = tempfile.mkdtemp() os.makedirs(os.path.join(self.transformer_dir , '''models/bert/''' ) ) _a = self.transformer_dir shutil.copy( os.path.join(A , '''src/transformers/models/bert/modeling_bert.py''' ) , os.path.join(self.transformer_dir , '''models/bert/modeling_bert.py''' ) , ) def a__ (self ) -> Union[str, Any]: """simple docstring""" _a = '''src/transformers''' shutil.rmtree(self.transformer_dir ) def a__ (self , A , A , A , A=None ) -> Dict: """simple docstring""" _a = comment + f'''\nclass {class_name}(nn.Module):\n''' + class_code if overwrite_result is not None: _a = comment + f'''\nclass {class_name}(nn.Module):\n''' + overwrite_result _a = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 ) _a = black.format_str(A , mode=A ) _a = os.path.join(self.transformer_dir , '''new_code.py''' ) with open(A , '''w''' , newline='''\n''' ) as f: f.write(A ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(A ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=A ) with open(A , '''r''' ) as f: self.assertTrue(f.read() , A ) def a__ (self ) -> List[str]: """simple docstring""" _a = check_copies.find_code_in_transformers('''models.bert.modeling_bert.BertLMPredictionHead''' ) self.assertEqual(A , A ) def a__ (self ) -> Any: """simple docstring""" self.check_copy_consistency( '''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead''' , '''BertLMPredictionHead''' , REFERENCE_CODE + '''\n''' , ) # With no empty line at the end self.check_copy_consistency( '''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead''' , '''BertLMPredictionHead''' , A , ) # Copy consistency with rename self.check_copy_consistency( '''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel''' , '''TestModelLMPredictionHead''' , re.sub('''Bert''' , '''TestModel''' , A ) , ) # Copy consistency with a really long name _a = '''TestModelWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason''' self.check_copy_consistency( f'''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->{long_class_name}''' , f'''{long_class_name}LMPredictionHead''' , re.sub('''Bert''' , A , A ) , ) # Copy consistency with overwrite self.check_copy_consistency( '''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel''' , '''TestModelLMPredictionHead''' , A , overwrite_result=re.sub('''Bert''' , '''TestModel''' , A ) , ) def a__ (self ) -> int: """simple docstring""" _a = check_copies.LOCALIZED_READMES['''README_zh-hans.md'''] _a = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the''' ''' Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for''' ''' Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong''' ''' Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.\n1.''' ''' **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (from HuggingFace),''' ''' released together with the paper [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and''' ''' lighter](https://arxiv.org/abs/1910.01108) by Victor Sanh, Lysandre Debut and Thomas Wolf. The same''' ''' method has been applied to compress GPT2 into''' ''' [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into''' ''' [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),''' ''' Multilingual BERT into''' ''' [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German''' ''' version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)**''' ''' (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders''' ''' as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang''' ''' Luong, Quoc V. Le, Christopher D. Manning.''' ) _a = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the''' ''' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of''' ''' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian''' ''' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n''' ) _a = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the''' ''' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of''' ''' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian''' ''' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n1.''' ''' **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (来自 HuggingFace) 伴随论文''' ''' [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and''' ''' lighter](https://arxiv.org/abs/1910.01108) 由 Victor Sanh, Lysandre Debut and Thomas Wolf 发布。 The same''' ''' method has been applied to compress GPT2 into''' ''' [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into''' ''' [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),''' ''' Multilingual BERT into''' ''' [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German''' ''' version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)** (来自''' ''' Google Research/Stanford University) 伴随论文 [ELECTRA: Pre-training text encoders as discriminators rather''' ''' than generators](https://arxiv.org/abs/2003.10555) 由 Kevin Clark, Minh-Thang Luong, Quoc V. Le,''' ''' Christopher D. Manning 发布。\n''' ) _a , _a = check_copies.convert_to_localized_md( A , A , localized_readme['''format_model_list'''] ) self.assertFalse(A ) self.assertEqual(A , A ) _a , _a = check_copies.convert_to_localized_md( A , A , localized_readme['''format_model_list'''] ) # Check whether the number of models is equal to README.md after conversion. self.assertTrue(A ) _a = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the''' ''' Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for''' ''' Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong''' ''' Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.''' ) _a = ( '''1. **[ALBERT](https://huggingface.co/transformers/main/model_doc/albert.html)** (来自 Google Research and''' ''' the Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of''' ''' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian''' ''' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n''' ) _a = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the''' ''' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of''' ''' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian''' ''' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n''' ) _a , _a = check_copies.convert_to_localized_md( A , A , localized_readme['''format_model_list'''] ) # Check if the model link is synchronized. self.assertEqual(A , A )
11
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer _lowercase = logging.get_logger(__name__) _lowercase = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } _lowercase = { '''vocab_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json''' }, '''merges_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt''' }, '''tokenizer_config_file''': { '''facebook/blenderbot_small-90M''': ( '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json''' ) }, } _lowercase = { '''facebook/blenderbot_small-90M''': 512, } class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = BlenderbotSmallTokenizer def __init__( self : Optional[int] ,lowerCAmelCase__ : Optional[int]=None ,lowerCAmelCase__ : Union[str, Any]=None ,lowerCAmelCase__ : Any="<|endoftext|>" ,lowerCAmelCase__ : int="<|endoftext|>" ,lowerCAmelCase__ : Optional[Any]="<|endoftext|>" ,lowerCAmelCase__ : Union[str, Any]=False ,lowerCAmelCase__ : Optional[Any]=True ,**lowerCAmelCase__ : Union[str, Any] ,) -> str: '''simple docstring''' super().__init__( ByteLevelBPETokenizer( vocab=lowerCAmelCase__ ,merges=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,trim_offsets=lowerCAmelCase__ ,) ,bos_token=lowerCAmelCase__ ,eos_token=lowerCAmelCase__ ,unk_token=lowerCAmelCase__ ,**lowerCAmelCase__ ,) lowerCAmelCase_ : Dict = add_prefix_space def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Tuple=None ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : str = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : Dict = [self.sep_token_id] lowerCAmelCase_ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
659
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ : Union[str, Any] = { """configuration_chinese_clip""": [ """CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ChineseCLIPConfig""", """ChineseCLIPOnnxConfig""", """ChineseCLIPTextConfig""", """ChineseCLIPVisionConfig""", ], """processing_chinese_clip""": ["""ChineseCLIPProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : Optional[Any] = ["""ChineseCLIPFeatureExtractor"""] lowerCamelCase__ : Optional[int] = ["""ChineseCLIPImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : List[str] = [ """CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """ChineseCLIPModel""", """ChineseCLIPPreTrainedModel""", """ChineseCLIPTextModel""", """ChineseCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_chinese_clip import ( CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, ChineseCLIPConfig, ChineseCLIPOnnxConfig, ChineseCLIPTextConfig, ChineseCLIPVisionConfig, ) from .processing_chinese_clip import ChineseCLIPProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_chinese_clip import ChineseCLIPFeatureExtractor, ChineseCLIPImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_chinese_clip import ( CHINESE_CLIP_PRETRAINED_MODEL_ARCHIVE_LIST, ChineseCLIPModel, ChineseCLIPPreTrainedModel, ChineseCLIPTextModel, ChineseCLIPVisionModel, ) else: import sys lowerCamelCase__ : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
12
from collections.abc import Generator from math import sin def UpperCamelCase ( snake_case__): if len(snake_case__) != 32: raise ValueError("Input must be of length 32") lowerCAmelCase_ : Tuple = b"" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def UpperCamelCase ( snake_case__): if i < 0: raise ValueError("Input must be non-negative") lowerCAmelCase_ : List[str] = format(snake_case__ , "08x")[-8:] lowerCAmelCase_ : Any = b"" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("utf-8") return little_endian_hex def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Union[str, Any] = b"" for char in message: bit_string += format(snake_case__ , "08b").encode("utf-8") lowerCAmelCase_ : Optional[int] = format(len(snake_case__) , "064b").encode("utf-8") # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(snake_case__) % 5_12 != 4_48: bit_string += b"0" bit_string += to_little_endian(start_len[32:]) + to_little_endian(start_len[:32]) return bit_string def UpperCamelCase ( snake_case__): if len(snake_case__) % 5_12 != 0: raise ValueError("Input must have length that's a multiple of 512") for pos in range(0 , len(snake_case__) , 5_12): lowerCAmelCase_ : List[str] = bit_string[pos : pos + 5_12] lowerCAmelCase_ : Union[str, Any] = [] for i in range(0 , 5_12 , 32): block_words.append(int(to_little_endian(block[i : i + 32]) , 2)) yield block_words def UpperCamelCase ( snake_case__): if i < 0: raise ValueError("Input must be non-negative") lowerCAmelCase_ : Dict = format(snake_case__ , "032b") lowerCAmelCase_ : str = "" for c in i_str: new_str += "1" if c == "0" else "0" return int(snake_case__ , 2) def UpperCamelCase ( snake_case__ , snake_case__): return (a + b) % 2**32 def UpperCamelCase ( snake_case__ , snake_case__): if i < 0: raise ValueError("Input must be non-negative") if shift < 0: raise ValueError("Shift must be non-negative") return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def UpperCamelCase ( snake_case__): lowerCAmelCase_ : Optional[Any] = preprocess(snake_case__) lowerCAmelCase_ : Optional[Any] = [int(2**32 * abs(sin(i + 1))) for i in range(64)] # Starting states lowerCAmelCase_ : List[str] = 0x67_45_23_01 lowerCAmelCase_ : Union[str, Any] = 0xef_cd_ab_89 lowerCAmelCase_ : List[Any] = 0x98_ba_dc_fe lowerCAmelCase_ : Tuple = 0x10_32_54_76 lowerCAmelCase_ : Any = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(snake_case__): lowerCAmelCase_ : Optional[int] = aa lowerCAmelCase_ : List[str] = ba lowerCAmelCase_ : Any = ca lowerCAmelCase_ : Union[str, Any] = da # Hash current chunk for i in range(64): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f lowerCAmelCase_ : Any = d ^ (b & (c ^ d)) lowerCAmelCase_ : Dict = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f lowerCAmelCase_ : Any = c ^ (d & (b ^ c)) lowerCAmelCase_ : List[str] = (5 * i + 1) % 16 elif i <= 47: lowerCAmelCase_ : int = b ^ c ^ d lowerCAmelCase_ : Optional[Any] = (3 * i + 5) % 16 else: lowerCAmelCase_ : List[Any] = c ^ (b | not_aa(snake_case__)) lowerCAmelCase_ : List[Any] = (7 * i) % 16 lowerCAmelCase_ : Optional[Any] = (f + a + added_consts[i] + block_words[g]) % 2**32 lowerCAmelCase_ : Optional[Any] = d lowerCAmelCase_ : Dict = c lowerCAmelCase_ : List[str] = b lowerCAmelCase_ : Any = sum_aa(snake_case__ , left_rotate_aa(snake_case__ , shift_amounts[i])) # Add hashed chunk to running total lowerCAmelCase_ : Dict = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : str = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : Optional[int] = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : int = sum_aa(snake_case__ , snake_case__) lowerCAmelCase_ : Union[str, Any] = reformat_hex(snake_case__) + reformat_hex(snake_case__) + reformat_hex(snake_case__) + reformat_hex(snake_case__) return digest if __name__ == "__main__": import doctest doctest.testmod()
659
0
'''simple docstring''' from __future__ import annotations from collections import Counter from random import random class UpperCAmelCase_ : """simple docstring""" def __init__( self ) -> int: __lowerCamelCase : List[str] = {} def lowercase_ ( self , SCREAMING_SNAKE_CASE_ ) -> None: __lowerCamelCase : Optional[int] = {} def lowercase_ ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> None: if nodea not in self.connections: self.add_node(SCREAMING_SNAKE_CASE_ ) if nodea not in self.connections: self.add_node(SCREAMING_SNAKE_CASE_ ) __lowerCamelCase : Optional[int] = probability def lowercase_ ( self ) -> list[str]: return list(self.connections ) def lowercase_ ( self , SCREAMING_SNAKE_CASE_ ) -> str: __lowerCamelCase : Optional[Any] = 0 __lowerCamelCase : Optional[Any] = random() for dest in self.connections[node]: current_probability += self.connections[node][dest] if current_probability > random_value: return dest return "" def UpperCAmelCase__ ( UpperCAmelCase_ : str , UpperCAmelCase_ : list[tuple[str, str, float]] , UpperCAmelCase_ : int ) -> dict[str, int]: __lowerCamelCase : Any = MarkovChainGraphUndirectedUnweighted() for nodea, nodea, probability in transitions: graph.add_transition_probability(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) __lowerCamelCase : Optional[int] = Counter(graph.get_nodes() ) __lowerCamelCase : Optional[int] = start for _ in range(UpperCAmelCase_ ): __lowerCamelCase : List[Any] = graph.transition(UpperCAmelCase_ ) visited[node] += 1 return visited if __name__ == "__main__": import doctest doctest.testmod()
13
import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): _lowercase = True from torch.cuda.amp import autocast _lowercase = logging.getLogger(__name__) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Whether to log verbose messages or not.'} , ) UpperCamelCase_ = field( default=2.0 , metadata={'help': 'Maximum temperature for gumbel softmax.'} ) UpperCamelCase_ = field( default=0.5 , metadata={'help': 'Minimum temperature for gumbel softmax.'} ) UpperCamelCase_ = field( default=0.99_99_95 , metadata={'help': 'Decay of gumbel temperature during training.'} ) def UpperCamelCase ( snake_case__ , snake_case__): logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout)] , ) lowerCAmelCase_ : str = logging.WARNING if model_args.verbose_logging: lowerCAmelCase_ : int = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank): lowerCAmelCase_ : Any = logging.INFO logger.setLevel(snake_case__) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The name of the dataset to use (via the datasets library).'} ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) UpperCamelCase_ = field( default='train' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) UpperCamelCase_ = field( default='validation' , metadata={ 'help': ( 'The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'' ) } , ) UpperCamelCase_ = field( default='file' , metadata={'help': 'Column in the dataset that contains speech file path. Defaults to \'file\''} , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'Overwrite the cached preprocessed datasets or not.'} ) UpperCamelCase_ = field( default=1 , metadata={ 'help': 'The percentage of the train set used as validation set in case there\'s no validation split' } , ) UpperCamelCase_ = field( default=snake_case__ , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) UpperCamelCase_ = field( default=20.0 , metadata={'help': 'Filter audio files that are longer than `max_duration_in_seconds` seconds'} ) @dataclass class __snake_case : """simple docstring""" UpperCamelCase_ = 42 UpperCamelCase_ = 42 UpperCamelCase_ = "longest" UpperCamelCase_ = None UpperCamelCase_ = None def __call__( self : str ,lowerCAmelCase__ : List[Dict[str, Union[List[int], torch.Tensor]]] ) -> Dict[str, torch.Tensor]: '''simple docstring''' lowerCAmelCase_ : Tuple = self.feature_extractor.pad( lowerCAmelCase__ ,max_length=self.max_length ,padding=self.padding ,pad_to_multiple_of=self.pad_to_multiple_of ,return_tensors="pt" ,) lowerCAmelCase_ : Union[str, Any] = self.model._get_feat_extract_output_lengths(batch["input_values"].shape[-1] ) lowerCAmelCase_ : List[str] = batch["input_values"].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula lowerCAmelCase_ : Tuple = self.model._get_feat_extract_output_lengths(batch["attention_mask"].sum(-1 ) ).to( torch.long ) lowerCAmelCase_ : Optional[Any] = torch.zeros( (batch_size, mask_indices_seq_length) ,dtype=torch.long ,device=batch["input_values"].device ) # these two operations makes sure that all values # before the output lengths indices are attended to lowerCAmelCase_ : Tuple = 1 lowerCAmelCase_ : int = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices lowerCAmelCase_ : str = _compute_mask_indices( (batch_size, mask_indices_seq_length) ,self.model.config.mask_time_prob ,self.model.config.mask_time_length ,attention_mask=lowerCAmelCase__ ,min_masks=2 ,) return batch class __snake_case ( snake_case__ ): """simple docstring""" def __init__( self : List[str] ,*lowerCAmelCase__ : Optional[int] ,lowerCAmelCase__ : Tuple=1 ,lowerCAmelCase__ : Optional[int]=0 ,lowerCAmelCase__ : Optional[Any]=1.0 ,**lowerCAmelCase__ : Any ) -> str: '''simple docstring''' super().__init__(*lowerCAmelCase__ ,**lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = 0 lowerCAmelCase_ : int = max_gumbel_temp lowerCAmelCase_ : Union[str, Any] = min_gumbel_temp lowerCAmelCase_ : str = gumbel_temp_decay def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : nn.Module ,lowerCAmelCase__ : Dict[str, Union[torch.Tensor, Any]] ) -> torch.Tensor: '''simple docstring''' model.train() lowerCAmelCase_ : str = self._prepare_inputs(lowerCAmelCase__ ) if self.use_amp: with autocast(): lowerCAmelCase_ : List[Any] = self.compute_loss(lowerCAmelCase__ ,lowerCAmelCase__ ) else: lowerCAmelCase_ : List[Any] = self.compute_loss(lowerCAmelCase__ ,lowerCAmelCase__ ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": lowerCAmelCase_ : List[Any] = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": lowerCAmelCase_ : Optional[Any] = loss.sum() / (inputs["mask_time_indices"]).sum() else: raise ValueError(f'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: lowerCAmelCase_ : int = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(lowerCAmelCase__ ).backward() elif self.use_apex: with amp.scale_loss(lowerCAmelCase__ ,self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(lowerCAmelCase__ ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step ,self.min_gumbel_temp ) ) return loss.detach() def UpperCamelCase ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowerCAmelCase_ : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments)) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = parser.parse_args_into_dataclasses() configure_logger(snake_case__ , snake_case__) # Downloading and loading a dataset from the hub. lowerCAmelCase_ : List[str] = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" lowerCAmelCase_ : Any = DatasetDict() lowerCAmelCase_ : Union[str, Any] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : List[str] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" lowerCAmelCase_ : Union[str, Any] = DatasetDict() lowerCAmelCase_ : int = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split="validation" , cache_dir=model_args.cache_dir , ) lowerCAmelCase_ : Any = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported lowerCAmelCase_ : Dict = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=snake_case__) def prepare_dataset(snake_case__): # check that all files have the correct sampling rate lowerCAmelCase_ , lowerCAmelCase_ : str = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate) return batch # load audio files into numpy arrays lowerCAmelCase_ : int = datasets.map( snake_case__ , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets["train"].column_names) # filter audio files that are too long lowerCAmelCase_ : int = vectorized_datasets.filter( lambda snake_case__: len(data["speech"]) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate)) def normalize(snake_case__): return feature_extractor(batch["speech"] , sampling_rate=feature_extractor.sampling_rate) # normalize and transform to `BatchFeatures` lowerCAmelCase_ : str = vectorized_datasets.map( snake_case__ , batched=snake_case__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets["train"].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 lowerCAmelCase_ : Optional[Any] = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( "PreTraining is only supported for ``config.do_stable_layer_norm=True`` and" " ``config.feat_extract_norm='layer'") lowerCAmelCase_ : Dict = WavaVecaForPreTraining(snake_case__) lowerCAmelCase_ : int = DataCollatorForWavaVecaPretraining(model=snake_case__ , feature_extractor=snake_case__) lowerCAmelCase_ : List[Any] = WavaVecaPreTrainer( model=snake_case__ , data_collator=snake_case__ , args=snake_case__ , train_dataset=vectorized_datasets["train"] , eval_dataset=vectorized_datasets["validation"] , tokenizer=snake_case__ , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
659
0
import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.text import TextDatasetReader from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def __UpperCAmelCase ( __a : str ,__a : Optional[int] ) -> str: """simple docstring""" assert isinstance(__a ,__a ) assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' ,[False, True] ) def __UpperCAmelCase ( __a : Any ,__a : Union[str, Any] ,__a : int ) -> Tuple: """simple docstring""" _a : str = tmp_path / '''cache''' _a : List[Any] = {'''text''': '''string'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _a : List[Any] = TextDatasetReader(__a ,cache_dir=__a ,keep_in_memory=__a ).read() _check_text_dataset(__a ,__a ) @pytest.mark.parametrize( '''features''' ,[ None, {'''text''': '''string'''}, {'''text''': '''int32'''}, {'''text''': '''float32'''}, ] ,) def __UpperCAmelCase ( __a : Optional[int] ,__a : Optional[Any] ,__a : int ) -> List[str]: """simple docstring""" _a : Dict = tmp_path / '''cache''' _a : Optional[Any] = {'''text''': '''string'''} _a : Optional[Any] = features.copy() if features else default_expected_features _a : Tuple = ( Features({feature: Value(__a ) for feature, dtype in features.items()} ) if features is not None else None ) _a : List[str] = TextDatasetReader(__a ,features=__a ,cache_dir=__a ).read() _check_text_dataset(__a ,__a ) @pytest.mark.parametrize('''split''' ,[None, NamedSplit('''train''' ), '''train''', '''test'''] ) def __UpperCAmelCase ( __a : str ,__a : int ,__a : List[str] ) -> List[Any]: """simple docstring""" _a : Any = tmp_path / '''cache''' _a : List[str] = {'''text''': '''string'''} _a : str = TextDatasetReader(__a ,cache_dir=__a ,split=__a ).read() _check_text_dataset(__a ,__a ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('''path_type''' ,[str, list] ) def __UpperCAmelCase ( __a : int ,__a : int ,__a : str ) -> Optional[Any]: """simple docstring""" if issubclass(__a ,__a ): _a : Tuple = text_path elif issubclass(__a ,__a ): _a : Tuple = [text_path] _a : Union[str, Any] = tmp_path / '''cache''' _a : Any = {'''text''': '''string'''} _a : int = TextDatasetReader(__a ,cache_dir=__a ).read() _check_text_dataset(__a ,__a ) def __UpperCAmelCase ( __a : Tuple ,__a : Union[str, Any] ,__a : Optional[int]=("train",) ) -> List[Any]: """simple docstring""" assert isinstance(__a ,__a ) for split in splits: _a : Union[str, Any] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 1 assert dataset.column_names == ["text"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' ,[False, True] ) def __UpperCAmelCase ( __a : int ,__a : int ,__a : Optional[Any] ) -> Any: """simple docstring""" _a : List[str] = tmp_path / '''cache''' _a : List[str] = {'''text''': '''string'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _a : Union[str, Any] = TextDatasetReader({'''train''': text_path} ,cache_dir=__a ,keep_in_memory=__a ).read() _check_text_datasetdict(__a ,__a ) @pytest.mark.parametrize( '''features''' ,[ None, {'''text''': '''string'''}, {'''text''': '''int32'''}, {'''text''': '''float32'''}, ] ,) def __UpperCAmelCase ( __a : Dict ,__a : List[str] ,__a : Optional[int] ) -> Tuple: """simple docstring""" _a : Optional[Any] = tmp_path / '''cache''' # CSV file loses col_1 string dtype information: default now is "int64" instead of "string" _a : Optional[Any] = {'''text''': '''string'''} _a : str = features.copy() if features else default_expected_features _a : Dict = ( Features({feature: Value(__a ) for feature, dtype in features.items()} ) if features is not None else None ) _a : Union[str, Any] = TextDatasetReader({'''train''': text_path} ,features=__a ,cache_dir=__a ).read() _check_text_datasetdict(__a ,__a ) @pytest.mark.parametrize('''split''' ,[None, NamedSplit('''train''' ), '''train''', '''test'''] ) def __UpperCAmelCase ( __a : int ,__a : int ,__a : Dict ) -> Dict: """simple docstring""" if split: _a : Union[str, Any] = {split: text_path} else: _a : Any = '''train''' _a : Any = {'''train''': text_path, '''test''': text_path} _a : List[str] = tmp_path / '''cache''' _a : Union[str, Any] = {'''text''': '''string'''} _a : Tuple = TextDatasetReader(__a ,cache_dir=__a ).read() _check_text_datasetdict(__a ,__a ,splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() )
14
from __future__ import annotations from collections.abc import Callable def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__ = 1_00 , ): lowerCAmelCase_ : Any = x_start lowerCAmelCase_ : Optional[Any] = fnc(snake_case__) lowerCAmelCase_ : Union[str, Any] = 0.0 for _ in range(snake_case__): # Approximates small segments of curve as linear and solve # for trapezoidal area lowerCAmelCase_ : Any = (x_end - x_start) / steps + xa lowerCAmelCase_ : Dict = fnc(snake_case__) area += abs(fxa + fxa) * (xa - xa) / 2 # Increment step lowerCAmelCase_ : int = xa lowerCAmelCase_ : str = fxa return area if __name__ == "__main__": def UpperCamelCase ( snake_case__): return x**3 + x**2 print('''f(x) = x^3 + x^2''') print('''The area between the curve, x = -5, x = 5 and the x axis is:''') _lowercase = 10 while i <= 100000: print(f"with {i} steps: {trapezoidal_area(f, -5, 5, i)}") i *= 10
659
0
from __future__ import annotations def UpperCamelCase ( __magic_name__ : list[int] ) -> list[int]: # This function is recursive """simple docstring""" lowercase__ = len(__magic_name__ ) # If the array contains only one element, we return it (it's the stop condition of # recursion) if array_length <= 1: return array # Else lowercase__ = array[0] lowercase__ = False lowercase__ = 1 lowercase__ = [] while not is_found and i < array_length: if array[i] < pivot: lowercase__ = True lowercase__ = [element for element in array[i:] if element >= array[i]] lowercase__ = longest_subsequence(__magic_name__ ) if len(__magic_name__ ) > len(__magic_name__ ): lowercase__ = temp_array else: i += 1 lowercase__ = [element for element in array[1:] if element >= pivot] lowercase__ = [pivot, *longest_subsequence(__magic_name__ )] if len(__magic_name__ ) > len(__magic_name__ ): return temp_array else: return longest_subseq if __name__ == "__main__": import doctest doctest.testmod()
15
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, PNDMScheduler, StableDiffusionLDMaDPipeline, UNetaDConditionModel, ) from diffusers.utils import nightly, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS enable_full_determinism() class __snake_case ( unittest.TestCase ): """simple docstring""" UpperCamelCase_ = StableDiffusionLDMaDPipeline UpperCamelCase_ = TEXT_TO_IMAGE_PARAMS UpperCamelCase_ = TEXT_TO_IMAGE_BATCH_PARAMS UpperCamelCase_ = TEXT_TO_IMAGE_IMAGE_PARAMS def UpperCAmelCase_ ( self : Tuple ) -> str: '''simple docstring''' torch.manual_seed(0 ) lowerCAmelCase_ : Optional[Any] = UNetaDConditionModel( block_out_channels=(32, 64) ,layers_per_block=2 ,sample_size=32 ,in_channels=4 ,out_channels=4 ,down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") ,up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") ,cross_attention_dim=32 ,) lowerCAmelCase_ : Any = DDIMScheduler( beta_start=0.00_085 ,beta_end=0.012 ,beta_schedule="scaled_linear" ,clip_sample=lowerCAmelCase__ ,set_alpha_to_one=lowerCAmelCase__ ,) torch.manual_seed(0 ) lowerCAmelCase_ : str = AutoencoderKL( block_out_channels=[32, 64] ,in_channels=6 ,out_channels=6 ,down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] ,up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] ,latent_channels=4 ,) torch.manual_seed(0 ) lowerCAmelCase_ : Optional[Any] = CLIPTextConfig( bos_token_id=0 ,eos_token_id=2 ,hidden_size=32 ,intermediate_size=37 ,layer_norm_eps=1e-0_5 ,num_attention_heads=4 ,num_hidden_layers=5 ,pad_token_id=1 ,vocab_size=10_00 ,) lowerCAmelCase_ : Optional[int] = CLIPTextModel(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) lowerCAmelCase_ : List[Any] = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : List[Any] ,lowerCAmelCase__ : List[str]=0 ) -> Dict: '''simple docstring''' if str(lowerCAmelCase__ ).startswith("mps" ): lowerCAmelCase_ : Optional[int] = torch.manual_seed(lowerCAmelCase__ ) else: lowerCAmelCase_ : Dict = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : str = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : Any ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Dict = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : List[str] = self.get_dummy_components() lowerCAmelCase_ : Union[str, Any] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Any = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Any = output.rgb, output.depth lowerCAmelCase_ : Dict = rgb[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) lowerCAmelCase_ : Optional[Any] = np.array( [0.37_338_176, 0.70_247, 0.74_203_193, 0.51_643_604, 0.58_256_793, 0.60_932_136, 0.4_181_095, 0.48_355_877, 0.46_535_262] ) lowerCAmelCase_ : Tuple = np.array([103.46_727, 85.812_004, 87.849_236] ) assert np.abs(image_slice_rgb.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(image_slice_depth.flatten() - expected_slice_depth ).max() < 1e-2 def UpperCAmelCase_ ( self : int ) -> Optional[int]: '''simple docstring''' lowerCAmelCase_ : Dict = self.get_dummy_components() lowerCAmelCase_ : List[str] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : str = 3 * [inputs["prompt"]] # forward lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = output.rgb, output.depth lowerCAmelCase_ : str = rgb_slice_a[0, -3:, -3:, -1] lowerCAmelCase_ : List[str] = depth_slice_a[0, -3:, -1] lowerCAmelCase_ : Union[str, Any] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = 3 * [inputs.pop("prompt" )] lowerCAmelCase_ : str = ldmad_pipe.tokenizer( lowerCAmelCase__ ,padding="max_length" ,max_length=ldmad_pipe.tokenizer.model_max_length ,truncation=lowerCAmelCase__ ,return_tensors="pt" ,) lowerCAmelCase_ : Union[str, Any] = text_inputs["input_ids"].to(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = ldmad_pipe.text_encoder(lowerCAmelCase__ )[0] lowerCAmelCase_ : Optional[int] = prompt_embeds # forward lowerCAmelCase_ : str = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : str = output.rgb, output.depth lowerCAmelCase_ : Optional[Any] = rgb_slice_a[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth_slice_a[0, -3:, -1] assert np.abs(rgb_slice_a.flatten() - rgb_slice_a.flatten() ).max() < 1e-4 assert np.abs(depth_slice_a.flatten() - depth_slice_a.flatten() ).max() < 1e-4 def UpperCAmelCase_ ( self : Union[str, Any] ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : Any = "cpu" # ensure determinism for the device-dependent torch.Generator lowerCAmelCase_ : Optional[int] = self.get_dummy_components() lowerCAmelCase_ : Dict = PNDMScheduler(skip_prk_steps=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = StableDiffusionLDMaDPipeline(**lowerCAmelCase__ ) lowerCAmelCase_ : Any = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = self.get_dummy_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : List[Any] = "french fries" lowerCAmelCase_ : Optional[int] = ldmad_pipe(**lowerCAmelCase__ ,negative_prompt=lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Union[str, Any] = output.rgb, output.depth lowerCAmelCase_ : Any = rgb[0, -3:, -3:, -1] lowerCAmelCase_ : Tuple = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) lowerCAmelCase_ : int = np.array( [0.37_044, 0.71_811_503, 0.7_223_251, 0.48_603_675, 0.5_638_391, 0.6_364_948, 0.42_833_704, 0.4_901_315, 0.47_926_217] ) lowerCAmelCase_ : Union[str, Any] = np.array([107.84_738, 84.62_802, 89.962_135] ) assert np.abs(rgb_slice.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(depth_slice.flatten() - expected_slice_depth ).max() < 1e-2 @slow @require_torch_gpu class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Dict="cpu" ,lowerCAmelCase__ : Union[str, Any]=torch.floataa ,lowerCAmelCase__ : List[str]=0 ) -> int: '''simple docstring''' lowerCAmelCase_ : Any = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = np.random.RandomState(lowerCAmelCase__ ).standard_normal((1, 4, 64, 64) ) lowerCAmelCase_ : Optional[Any] = torch.from_numpy(lowerCAmelCase__ ).to(device=lowerCAmelCase__ ,dtype=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : List[Any] ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : Optional[Any] = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ) lowerCAmelCase_ : List[str] = ldmad_pipe.to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Dict = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : List[str] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Dict = output.rgb, output.depth lowerCAmelCase_ : List[str] = rgb[0, -3:, -3:, -1].flatten() lowerCAmelCase_ : Optional[int] = rgb[0, -3:, -1].flatten() assert rgb.shape == (1, 5_12, 5_12, 3) assert depth.shape == (1, 5_12, 5_12) lowerCAmelCase_ : int = np.array( [0.53_805_465, 0.56_707_305, 0.5_486_515, 0.57_012_236, 0.5_814_511, 0.56_253_487, 0.54_843_014, 0.55_092_263, 0.6_459_706] ) lowerCAmelCase_ : Optional[Any] = np.array( [0.9_263_781, 0.6_678_672, 0.5_486_515, 0.92_202_145, 0.67_831_135, 0.56_253_487, 0.9_241_694, 0.7_551_478, 0.6_459_706] ) assert np.abs(rgb_slice - expected_slice_rgb ).max() < 3e-3 assert np.abs(depth_slice - expected_slice_depth ).max() < 3e-3 @nightly @require_torch_gpu class __snake_case ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase_ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Dict="cpu" ,lowerCAmelCase__ : List[str]=torch.floataa ,lowerCAmelCase__ : Optional[int]=0 ) -> int: '''simple docstring''' lowerCAmelCase_ : Dict = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = np.random.RandomState(lowerCAmelCase__ ).standard_normal((1, 4, 64, 64) ) lowerCAmelCase_ : Any = torch.from_numpy(lowerCAmelCase__ ).to(device=lowerCAmelCase__ ,dtype=lowerCAmelCase__ ) lowerCAmelCase_ : int = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 50, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self : Dict ) -> int: '''simple docstring''' lowerCAmelCase_ : List[Any] = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ).to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Union[str, Any] = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Any = output.rgb, output.depth lowerCAmelCase_ : Dict = 0.495_586 lowerCAmelCase_ : Optional[Any] = 0.33_795_515 lowerCAmelCase_ : Any = 112.48_518 lowerCAmelCase_ : List[Any] = 98.489_746 assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3 def UpperCAmelCase_ ( self : Tuple ) -> List[str]: '''simple docstring''' lowerCAmelCase_ : int = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d-4c" ).to(lowerCAmelCase__ ) ldmad_pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) lowerCAmelCase_ : str = self.get_inputs(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = ldmad_pipe(**lowerCAmelCase__ ) lowerCAmelCase_ , lowerCAmelCase_ : Tuple = output.rgb, output.depth lowerCAmelCase_ : List[str] = 0.4_194_127 lowerCAmelCase_ : List[str] = 0.35_375_586 lowerCAmelCase_ : str = 0.5_638_502 lowerCAmelCase_ : Optional[Any] = 0.34_686_103 assert rgb.shape == (1, 5_12, 5_12, 3) assert depth.shape == (1, 5_12, 5_12, 1) assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3
659
0
def __a ( A__ : str , A__ : str = " " ): SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = 0 for index, char in enumerate(A__ ): if char == separator: split_words.append(string[last_index:index] ) SCREAMING_SNAKE_CASE = index + 1 elif index + 1 == len(A__ ): split_words.append(string[last_index : index + 1] ) return split_words if __name__ == "__main__": from doctest import testmod testmod()
16
import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) _lowercase = { '''iou_prediction_head.layers.0''': '''iou_prediction_head.proj_in''', '''iou_prediction_head.layers.1''': '''iou_prediction_head.layers.0''', '''iou_prediction_head.layers.2''': '''iou_prediction_head.proj_out''', '''mask_decoder.output_upscaling.0''': '''mask_decoder.upscale_conv1''', '''mask_decoder.output_upscaling.1''': '''mask_decoder.upscale_layer_norm''', '''mask_decoder.output_upscaling.3''': '''mask_decoder.upscale_conv2''', '''mask_downscaling.0''': '''mask_embed.conv1''', '''mask_downscaling.1''': '''mask_embed.layer_norm1''', '''mask_downscaling.3''': '''mask_embed.conv2''', '''mask_downscaling.4''': '''mask_embed.layer_norm2''', '''mask_downscaling.6''': '''mask_embed.conv3''', '''point_embeddings''': '''point_embed''', '''pe_layer.positional_encoding_gaussian_matrix''': '''shared_embedding.positional_embedding''', '''image_encoder''': '''vision_encoder''', '''neck.0''': '''neck.conv1''', '''neck.1''': '''neck.layer_norm1''', '''neck.2''': '''neck.conv2''', '''neck.3''': '''neck.layer_norm2''', '''patch_embed.proj''': '''patch_embed.projection''', '''.norm''': '''.layer_norm''', '''blocks''': '''layers''', } def UpperCamelCase ( snake_case__): lowerCAmelCase_ : int = {} state_dict.pop("pixel_mean" , snake_case__) state_dict.pop("pixel_std" , snake_case__) lowerCAmelCase_ : List[Any] = R".*.output_hypernetworks_mlps.(\d+).layers.(\d+).*" for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: lowerCAmelCase_ : Dict = key.replace(snake_case__ , snake_case__) if re.match(snake_case__ , snake_case__): lowerCAmelCase_ : Any = int(re.match(snake_case__ , snake_case__).group(2)) if layer_nb == 0: lowerCAmelCase_ : List[Any] = key.replace("layers.0" , "proj_in") elif layer_nb == 1: lowerCAmelCase_ : List[Any] = key.replace("layers.1" , "layers.0") elif layer_nb == 2: lowerCAmelCase_ : int = key.replace("layers.2" , "proj_out") lowerCAmelCase_ : int = value lowerCAmelCase_ : Optional[int] = model_state_dict[ "prompt_encoder.shared_embedding.positional_embedding" ] return model_state_dict def UpperCamelCase ( snake_case__ , snake_case__ , snake_case__ , snake_case__="ybelkada/segment-anything"): lowerCAmelCase_ : Optional[int] = hf_hub_download(snake_case__ , F'''checkpoints/{model_name}.pth''') if "sam_vit_b" in model_name: lowerCAmelCase_ : Optional[Any] = SamConfig() elif "sam_vit_l" in model_name: lowerCAmelCase_ : Optional[int] = SamVisionConfig( hidden_size=10_24 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) lowerCAmelCase_ : Union[str, Any] = SamConfig( vision_config=snake_case__ , ) elif "sam_vit_h" in model_name: lowerCAmelCase_ : Optional[Any] = SamVisionConfig( hidden_size=12_80 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) lowerCAmelCase_ : Tuple = SamConfig( vision_config=snake_case__ , ) lowerCAmelCase_ : Optional[Any] = torch.load(snake_case__ , map_location="cpu") lowerCAmelCase_ : Union[str, Any] = replace_keys(snake_case__) lowerCAmelCase_ : List[Any] = SamImageProcessor() lowerCAmelCase_ : Any = SamProcessor(image_processor=snake_case__) lowerCAmelCase_ : Any = SamModel(snake_case__) hf_model.load_state_dict(snake_case__) lowerCAmelCase_ : Dict = hf_model.to("cuda") lowerCAmelCase_ : List[str] = "https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png" lowerCAmelCase_ : List[Any] = Image.open(requests.get(snake_case__ , stream=snake_case__).raw).convert("RGB") lowerCAmelCase_ : Optional[int] = [[[4_00, 6_50]]] lowerCAmelCase_ : int = [[1]] lowerCAmelCase_ : Optional[Any] = processor(images=np.array(snake_case__) , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Optional[Any] = hf_model(**snake_case__) lowerCAmelCase_ : Optional[int] = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.579_890_251_159_668 lowerCAmelCase_ : Any = processor( images=np.array(snake_case__) , input_points=snake_case__ , input_labels=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Optional[Any] = hf_model(**snake_case__) lowerCAmelCase_ : Union[str, Any] = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_712_603_092_193_604 lowerCAmelCase_ : Tuple = ((75, 2_75, 17_25, 8_50),) lowerCAmelCase_ : Optional[Any] = processor(images=np.array(snake_case__) , input_boxes=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : List[Any] = hf_model(**snake_case__) lowerCAmelCase_ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.8_686_015_605_926_514 # Test with 2 points and 1 image. lowerCAmelCase_ : int = [[[4_00, 6_50], [8_00, 6_50]]] lowerCAmelCase_ : Optional[Any] = [[1, 1]] lowerCAmelCase_ : List[Any] = processor( images=np.array(snake_case__) , input_points=snake_case__ , input_labels=snake_case__ , return_tensors="pt").to("cuda") with torch.no_grad(): lowerCAmelCase_ : Tuple = hf_model(**snake_case__) lowerCAmelCase_ : str = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_936_047_792_434_692 if __name__ == "__main__": _lowercase = argparse.ArgumentParser() _lowercase = ['''sam_vit_b_01ec64''', '''sam_vit_h_4b8939''', '''sam_vit_l_0b3195'''] parser.add_argument( '''--model_name''', default='''sam_vit_h_4b8939''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model and processor to the hub after converting''', ) parser.add_argument( '''--model_hub_id''', default='''ybelkada/segment-anything''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) _lowercase = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
659
0
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionInstructPixaPixPipeline, UNetaDConditionModel, ) from diffusers.image_processor import VaeImageProcessor from diffusers.utils import floats_tensor, load_image, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class lowerCamelCase_ ( _lowercase , _lowercase , _lowercase , unittest.TestCase ): _lowercase : List[Any] = StableDiffusionInstructPixaPixPipeline _lowercase : Tuple = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width''', '''cross_attention_kwargs'''} _lowercase : Union[str, Any] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS _lowercase : Dict = IMAGE_TO_IMAGE_IMAGE_PARAMS _lowercase : Dict = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowerCAmelCase_ ( self : Any ): torch.manual_seed(0 ) __A : Tuple = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=8 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=32 , ) __A : List[Any] = PNDMScheduler(skip_prk_steps=__A ) torch.manual_seed(0 ) __A : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) __A : List[str] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) __A : Union[str, Any] = CLIPTextModel(__A ) __A : List[str] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) __A : Optional[Any] = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """safety_checker""": None, """feature_extractor""": None, } return components def lowerCAmelCase_ ( self : str , __A : Union[str, Any] , __A : Union[str, Any]=0 ): __A : int = floats_tensor((1, 3, 32, 32) , rng=random.Random(__A ) ).to(__A ) __A : Dict = image.cpu().permute(0 , 2 , 3 , 1 )[0] __A : Dict = Image.fromarray(np.uinta(__A ) ).convert("""RGB""" ) if str(__A ).startswith("""mps""" ): __A : int = torch.manual_seed(__A ) else: __A : str = torch.Generator(device=__A ).manual_seed(__A ) __A : Union[str, Any] = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """image_guidance_scale""": 1, """output_type""": """numpy""", } return inputs def lowerCAmelCase_ ( self : List[str] ): __A : int = """cpu""" # ensure determinism for the device-dependent torch.Generator __A : str = self.get_dummy_components() __A : List[str] = StableDiffusionInstructPixaPixPipeline(**__A ) __A : List[str] = sd_pipe.to(__A ) sd_pipe.set_progress_bar_config(disable=__A ) __A : Tuple = self.get_dummy_inputs(__A ) __A : Union[str, Any] = sd_pipe(**__A ).images __A : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __A : List[str] = np.array([0.7_5_2_6, 0.3_7_5_0, 0.4_5_4_7, 0.6_1_1_7, 0.5_8_6_6, 0.5_0_1_6, 0.4_3_2_7, 0.5_6_4_2, 0.4_8_1_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowerCAmelCase_ ( self : Optional[Any] ): __A : int = """cpu""" # ensure determinism for the device-dependent torch.Generator __A : Optional[int] = self.get_dummy_components() __A : Any = StableDiffusionInstructPixaPixPipeline(**__A ) __A : str = sd_pipe.to(__A ) sd_pipe.set_progress_bar_config(disable=__A ) __A : Dict = self.get_dummy_inputs(__A ) __A : List[Any] = """french fries""" __A : List[str] = sd_pipe(**__A , negative_prompt=__A ) __A : Union[str, Any] = output.images __A : int = image[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __A : Optional[Any] = np.array([0.7_5_1_1, 0.3_6_4_2, 0.4_5_5_3, 0.6_2_3_6, 0.5_7_9_7, 0.5_0_1_3, 0.4_3_4_3, 0.5_6_1_1, 0.4_8_3_1] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowerCAmelCase_ ( self : Any ): __A : int = """cpu""" # ensure determinism for the device-dependent torch.Generator __A : Dict = self.get_dummy_components() __A : List[str] = StableDiffusionInstructPixaPixPipeline(**__A ) __A : Tuple = sd_pipe.to(__A ) sd_pipe.set_progress_bar_config(disable=__A ) __A : Optional[int] = self.get_dummy_inputs(__A ) __A : int = [inputs["""prompt"""]] * 2 __A : str = np.array(inputs["""image"""] ).astype(np.floataa ) / 2_5_5.0 __A : Optional[Any] = torch.from_numpy(__A ).unsqueeze(0 ).to(__A ) __A : Dict = image / 2 + 0.5 __A : List[Any] = image.permute(0 , 3 , 1 , 2 ) __A : Optional[int] = image.repeat(2 , 1 , 1 , 1 ) __A : List[str] = sd_pipe(**__A ).images __A : List[str] = image[-1, -3:, -3:, -1] assert image.shape == (2, 32, 32, 3) __A : Dict = np.array([0.5_8_1_2, 0.5_7_4_8, 0.5_2_2_2, 0.5_9_0_8, 0.5_6_9_5, 0.7_1_7_4, 0.6_8_0_4, 0.5_5_2_3, 0.5_5_7_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowerCAmelCase_ ( self : Tuple ): __A : Union[str, Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator __A : Dict = self.get_dummy_components() __A : str = EulerAncestralDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="""scaled_linear""" ) __A : List[str] = StableDiffusionInstructPixaPixPipeline(**__A ) __A : Optional[int] = sd_pipe.to(__A ) sd_pipe.set_progress_bar_config(disable=__A ) __A : Optional[Any] = self.get_dummy_inputs(__A ) __A : str = sd_pipe(**__A ).images __A : List[str] = image[0, -3:, -3:, -1] __A : List[Any] = [round(__A , 4 ) for x in image_slice.flatten().tolist()] print(""",""".join([str(__A ) for x in slice] ) ) assert image.shape == (1, 32, 32, 3) __A : Optional[int] = np.array([0.7_4_1_7, 0.3_8_4_2, 0.4_7_3_2, 0.5_7_7_6, 0.5_8_9_1, 0.5_1_3_9, 0.4_0_5_2, 0.5_6_7_3, 0.4_9_8_6] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def lowerCAmelCase_ ( self : int ): super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) def lowerCAmelCase_ ( self : Optional[int] ): __A : Any = self.get_dummy_components() __A : Any = StableDiffusionInstructPixaPixPipeline(**__A ) __A : Any = VaeImageProcessor(do_resize=__A , do_normalize=__A ) __A : Tuple = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) __A : int = pipe(**self.get_dummy_inputs_by_type(__A , input_image_type="""pt""" ) )[0] __A : List[Any] = components["""vae"""] __A : Union[str, Any] = self.get_dummy_inputs_by_type(__A , input_image_type="""pt""" ) for image_param in self.image_latents_params: if image_param in inputs.keys(): __A : Optional[Any] = vae.encode(inputs[image_param] ).latent_dist.mode() __A : Tuple = pipe(**__A )[0] __A : Tuple = np.abs(out - out_latents_inputs ).max() self.assertLess(__A , 1e-4 , """passing latents as image input generate different result from passing image""" ) @slow @require_torch_gpu class lowerCamelCase_ ( unittest.TestCase ): def lowerCAmelCase_ ( self : Dict ): super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCAmelCase_ ( self : List[Any] , __A : Union[str, Any]=0 ): __A : Tuple = torch.manual_seed(__A ) __A : Optional[int] = load_image( """https://huggingface.co/datasets/diffusers/test-arrays/resolve/main/stable_diffusion_pix2pix/example.jpg""" ) __A : Dict = { """prompt""": """turn him into a cyborg""", """image""": image, """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 7.5, """image_guidance_scale""": 1.0, """output_type""": """numpy""", } return inputs def lowerCAmelCase_ ( self : Any ): __A : List[str] = StableDiffusionInstructPixaPixPipeline.from_pretrained( """timbrooks/instruct-pix2pix""" , safety_checker=__A ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) pipe.enable_attention_slicing() __A : Dict = self.get_inputs() __A : List[Any] = pipe(**__A ).images __A : List[Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __A : str = np.array([0.5_9_0_2, 0.6_0_1_5, 0.6_0_2_7, 0.5_9_8_3, 0.6_0_9_2, 0.6_0_6_1, 0.5_7_6_5, 0.5_7_8_5, 0.5_5_5_5] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowerCAmelCase_ ( self : int ): __A : Tuple = StableDiffusionInstructPixaPixPipeline.from_pretrained( """timbrooks/instruct-pix2pix""" , safety_checker=__A ) __A : Dict = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) pipe.enable_attention_slicing() __A : Union[str, Any] = self.get_inputs() __A : str = pipe(**__A ).images __A : Optional[Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __A : Dict = np.array([0.6_5_7_8, 0.6_8_1_7, 0.6_9_7_2, 0.6_7_6_1, 0.6_8_5_6, 0.6_9_1_6, 0.6_4_2_8, 0.6_5_1_6, 0.6_3_0_1] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowerCAmelCase_ ( self : List[str] ): __A : Optional[int] = StableDiffusionInstructPixaPixPipeline.from_pretrained( """timbrooks/instruct-pix2pix""" , safety_checker=__A ) __A : List[Any] = DDIMScheduler.from_config(pipe.scheduler.config ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) pipe.enable_attention_slicing() __A : Tuple = self.get_inputs() __A : List[Any] = pipe(**__A ).images __A : Optional[int] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 512, 3) __A : Optional[Any] = np.array([0.3_8_2_8, 0.3_8_3_4, 0.3_8_1_8, 0.3_7_9_2, 0.3_8_6_5, 0.3_7_5_2, 0.3_7_9_2, 0.3_8_4_7, 0.3_7_5_3] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def lowerCAmelCase_ ( self : int ): __A : List[Any] = 0 def callback_fn(__A : int , __A : int , __A : torch.FloatTensor ) -> None: __A : str = True nonlocal number_of_steps number_of_steps += 1 if step == 1: __A : Dict = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) __A : List[str] = latents[0, -3:, -3:, -1] __A : List[Any] = np.array([-0.2_4_6_3, -0.4_6_4_4, -0.9_7_5_6, 1.5_1_7_6, 1.4_4_1_4, 0.7_8_6_6, 0.9_8_9_7, 0.8_5_2_1, 0.7_9_8_3] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: __A : List[Any] = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 64) __A : Union[str, Any] = latents[0, -3:, -3:, -1] __A : List[Any] = np.array([-0.2_6_4_4, -0.4_6_2_6, -0.9_6_5_3, 1.5_1_7_6, 1.4_5_5_1, 0.7_6_8_6, 0.9_8_0_5, 0.8_4_5_2, 0.8_1_1_5] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 __A : Optional[int] = False __A : Dict = StableDiffusionInstructPixaPixPipeline.from_pretrained( """timbrooks/instruct-pix2pix""" , safety_checker=__A , torch_dtype=torch.floataa ) __A : Optional[Any] = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) pipe.enable_attention_slicing() __A : Dict = self.get_inputs() pipe(**__A , callback=__A , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def lowerCAmelCase_ ( self : str ): torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __A : Optional[Any] = StableDiffusionInstructPixaPixPipeline.from_pretrained( """timbrooks/instruct-pix2pix""" , safety_checker=__A , torch_dtype=torch.floataa ) __A : Optional[int] = pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() __A : List[Any] = self.get_inputs() __A : str = pipe(**__A ) __A : Union[str, Any] = torch.cuda.max_memory_allocated() # make sure that less than 2.2 GB is allocated assert mem_bytes < 2.2 * 10**9 def lowerCAmelCase_ ( self : Optional[int] ): __A : Optional[Any] = self.get_inputs() # resize to resolution that is divisible by 8 but not 16 or 32 __A : int = inputs["""image"""].resize((504, 504) ) __A : Dict = """timbrooks/instruct-pix2pix""" __A : str = StableDiffusionInstructPixaPixPipeline.from_pretrained( __A , safety_checker=__A , ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) pipe.enable_attention_slicing() __A : List[str] = pipe(**__A ) __A : Union[str, Any] = output.images[0] __A : Dict = image[255:258, 383:386, -1] assert image.shape == (504, 504, 3) __A : Tuple = np.array([0.2_7_2_6, 0.2_5_2_9, 0.2_6_6_4, 0.2_6_5_5, 0.2_6_4_1, 0.2_6_4_2, 0.2_5_9_1, 0.2_6_4_9, 0.2_5_9_0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5e-3
17
class __snake_case : """simple docstring""" def __init__( self : Union[str, Any] ,lowerCAmelCase__ : str = "" ,lowerCAmelCase__ : bool = False ) -> None: '''simple docstring''' lowerCAmelCase_ : dict[str, RadixNode] = {} # A node will be a leaf if the tree contains its word lowerCAmelCase_ : Optional[int] = is_leaf lowerCAmelCase_ : List[str] = prefix def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : str ) -> tuple[str, str, str]: '''simple docstring''' lowerCAmelCase_ : List[str] = 0 for q, w in zip(self.prefix ,lowerCAmelCase__ ): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : list[str] ) -> None: '''simple docstring''' for word in words: self.insert(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : str ) -> None: '''simple docstring''' if self.prefix == word: lowerCAmelCase_ : Optional[Any] = True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: lowerCAmelCase_ : Optional[int] = RadixNode(prefix=lowerCAmelCase__ ,is_leaf=lowerCAmelCase__ ) else: lowerCAmelCase_ : Optional[Any] = self.nodes[word[0]] lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Any = incoming_node.match( lowerCAmelCase__ ) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: lowerCAmelCase_ : Dict = remaining_prefix lowerCAmelCase_ : str = self.nodes[matching_string[0]] lowerCAmelCase_ : Dict = RadixNode(lowerCAmelCase__ ,lowerCAmelCase__ ) lowerCAmelCase_ : Any = aux_node if remaining_word == "": lowerCAmelCase_ : Optional[Any] = True else: self.nodes[matching_string[0]].insert(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : str ) -> bool: '''simple docstring''' lowerCAmelCase_ : List[str] = self.nodes.get(word[0] ,lowerCAmelCase__ ) if not incoming_node: return False else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Optional[int] = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : str ) -> bool: '''simple docstring''' lowerCAmelCase_ : int = self.nodes.get(word[0] ,lowerCAmelCase__ ) if not incoming_node: return False else: lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : List[Any] = incoming_node.match( lowerCAmelCase__ ) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(lowerCAmelCase__ ) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes ) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes ) == 1 and not self.is_leaf: lowerCAmelCase_ : int = list(self.nodes.values() )[0] lowerCAmelCase_ : List[Any] = merging_node.is_leaf self.prefix += merging_node.prefix lowerCAmelCase_ : int = merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes ) > 1: lowerCAmelCase_ : List[str] = False # If there is 1 edge, we merge it with its child else: lowerCAmelCase_ : Union[str, Any] = list(incoming_node.nodes.values() )[0] lowerCAmelCase_ : Optional[int] = merging_node.is_leaf incoming_node.prefix += merging_node.prefix lowerCAmelCase_ : List[str] = merging_node.nodes return True def UpperCAmelCase_ ( self : int ,lowerCAmelCase__ : int = 0 ) -> None: '''simple docstring''' if self.prefix != "": print("-" * height ,self.prefix ," (leaf)" if self.is_leaf else "" ) for value in self.nodes.values(): value.print_tree(height + 1 ) def UpperCamelCase ( ): lowerCAmelCase_ : List[Any] = "banana bananas bandana band apple all beast".split() lowerCAmelCase_ : Optional[Any] = RadixNode() root.insert_many(snake_case__) assert all(root.find(snake_case__) for word in words) assert not root.find("bandanas") assert not root.find("apps") root.delete("all") assert not root.find("all") root.delete("banana") assert not root.find("banana") assert root.find("bananas") return True def UpperCamelCase ( ): assert test_trie() def UpperCamelCase ( ): lowerCAmelCase_ : str = RadixNode() lowerCAmelCase_ : str = "banana bananas bandanas bandana band apple all beast".split() root.insert_many(snake_case__) print("Words:" , snake_case__) print("Tree:") root.print_tree() if __name__ == "__main__": main()
659
0
'''simple docstring''' import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartForSequenceClassification, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, TapexTokenizer, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version from transformers.utils.versions import require_version # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("4.17.0.dev0") require_version("datasets>=1.8.0", "To fix: pip install -r examples/pytorch/text-classification/requirements.txt") _SCREAMING_SNAKE_CASE = logging.getLogger(__name__) @dataclass class lowerCAmelCase_ : __lowerCamelCase : Optional[str] = field( default="tab_fact" ,metadata={"help": "The name of the dataset to use (via the datasets library)."} ) __lowerCamelCase : Optional[str] = field( default="tab_fact" ,metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} ,) __lowerCamelCase : int = field( default=1_024 ,metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } ,) __lowerCamelCase : bool = field( default=__magic_name__ ,metadata={"help": "Overwrite the cached preprocessed datasets or not."} ) __lowerCamelCase : bool = field( default=__magic_name__ ,metadata={ "help": ( "Whether to pad all samples to `max_seq_length`. " "If False, will pad the samples dynamically when batching to the maximum length in the batch." ) } ,) __lowerCamelCase : Optional[int] = field( default=__magic_name__ ,metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } ,) __lowerCamelCase : Optional[int] = field( default=__magic_name__ ,metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } ,) __lowerCamelCase : Optional[int] = field( default=__magic_name__ ,metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of prediction examples to this " "value if set." ) } ,) __lowerCamelCase : Optional[str] = field( default=__magic_name__ ,metadata={"help": "A csv or a json file containing the training data."} ) __lowerCamelCase : Optional[str] = field( default=__magic_name__ ,metadata={"help": "A csv or a json file containing the validation data."} ) __lowerCamelCase : Optional[str] = field(default=__magic_name__ ,metadata={"help": "A csv or a json file containing the test data."} ) def _snake_case ( self ) -> Optional[Any]: if self.dataset_name is not None: pass elif self.train_file is None or self.validation_file is None: raise ValueError("Need either a GLUE task, a training/validation file or a dataset name." ) else: _lowerCAmelCase = self.train_file.split("." )[-1] assert train_extension in ["csv", "json"], "`train_file` should be a csv or a json file." _lowerCAmelCase = self.validation_file.split("." )[-1] assert ( validation_extension == train_extension ), "`validation_file` should have the same extension (csv or json) as `train_file`." @dataclass class lowerCAmelCase_ : __lowerCamelCase : str = field( default=__magic_name__ ,metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) __lowerCamelCase : Optional[str] = field( default=__magic_name__ ,metadata={"help": "Pretrained config name or path if not the same as model_name"} ) __lowerCamelCase : Optional[str] = field( default=__magic_name__ ,metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) __lowerCamelCase : Optional[str] = field( default=__magic_name__ ,metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} ,) __lowerCamelCase : bool = field( default=__magic_name__ ,metadata={"help": "Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."} ,) __lowerCamelCase : str = field( default="main" ,metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} ,) __lowerCamelCase : bool = field( default=__magic_name__ ,metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } ,) def __a(): '''simple docstring''' _lowerCAmelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase = parser.parse_args_into_dataclasses() # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) _lowerCAmelCase = training_args.get_process_log_level() logger.setLevel(SCREAMING_SNAKE_CASE_ ) datasets.utils.logging.set_verbosity(SCREAMING_SNAKE_CASE_ ) transformers.utils.logging.set_verbosity(SCREAMING_SNAKE_CASE_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Detecting last checkpoint. _lowerCAmelCase = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _lowerCAmelCase = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON training and evaluation files (see below) # or specify a GLUE benchmark task (the dataset will be downloaded automatically from the datasets Hub). # # For JSON files, this script will use the `question` column for the input question and `table` column for the corresponding table. # # If the CSVs/JSONs contain only one non-label column, the script does single sentence classification on this # single column. You can easily tweak this behavior (see below) # # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.dataset_name is not None: # Downloading and loading a dataset from the hub. _lowerCAmelCase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) else: # Loading a dataset from your local files. # CSV/JSON training and evaluation files are needed. _lowerCAmelCase = {"train": data_args.train_file, "validation": data_args.validation_file} # Get the test dataset: you can provide your own CSV/JSON test file (see below) # when you use `do_predict` without specifying a GLUE benchmark task. if training_args.do_predict: if data_args.test_file is not None: _lowerCAmelCase = data_args.train_file.split("." )[-1] _lowerCAmelCase = data_args.test_file.split("." )[-1] assert ( test_extension == train_extension ), "`test_file` should have the same extension (csv or json) as `train_file`." _lowerCAmelCase = data_args.test_file else: raise ValueError("Need either a GLUE task or a test file for `do_predict`." ) for key in data_files.keys(): logger.info(F'''load a local file for {key}: {data_files[key]}''' ) if data_args.train_file.endswith(".csv" ): # Loading a dataset from local csv files _lowerCAmelCase = load_dataset("csv" , data_files=SCREAMING_SNAKE_CASE_ , cache_dir=model_args.cache_dir ) else: # Loading a dataset from local json files _lowerCAmelCase = load_dataset("json" , data_files=SCREAMING_SNAKE_CASE_ , cache_dir=model_args.cache_dir ) # See more about loading any type of standard or custom dataset at # https://huggingface.co/docs/datasets/loading_datasets.html. # Labels _lowerCAmelCase = raw_datasets["train"].features["label"].names _lowerCAmelCase = len(SCREAMING_SNAKE_CASE_ ) # Load pretrained model and tokenizer # # In distributed training, the .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _lowerCAmelCase = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=SCREAMING_SNAKE_CASE_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # load tapex tokenizer _lowerCAmelCase = TapexTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , add_prefix_space=SCREAMING_SNAKE_CASE_ , ) _lowerCAmelCase = BartForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=SCREAMING_SNAKE_CASE_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Padding strategy if data_args.pad_to_max_length: _lowerCAmelCase = "max_length" else: # We will pad later, dynamically at batch creation, to the max sequence length in each batch _lowerCAmelCase = False # Some models have set the order of the labels to use, so let's make sure we do use it. _lowerCAmelCase = {"Refused": 0, "Entailed": 1} _lowerCAmelCase = {0: "Refused", 1: "Entailed"} if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F'''The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the''' F'''model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.''' ) _lowerCAmelCase = min(data_args.max_seq_length , tokenizer.model_max_length ) def preprocess_tabfact_function(SCREAMING_SNAKE_CASE_ : List[Any] ): # Tokenize the texts def _convert_table_text_to_pandas(SCREAMING_SNAKE_CASE_ : int ): _lowerCAmelCase = [_table_row.split("#" ) for _table_row in _table_text.strip("\n" ).split("\n" )] _lowerCAmelCase = pd.DataFrame.from_records(_table_content[1:] , columns=_table_content[0] ) return _table_pd _lowerCAmelCase = examples["statement"] _lowerCAmelCase = list(map(_convert_table_text_to_pandas , examples["table_text"] ) ) _lowerCAmelCase = tokenizer(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , padding=SCREAMING_SNAKE_CASE_ , max_length=SCREAMING_SNAKE_CASE_ , truncation=SCREAMING_SNAKE_CASE_ ) _lowerCAmelCase = examples["label"] return result with training_args.main_process_first(desc="dataset map pre-processing" ): _lowerCAmelCase = raw_datasets.map( SCREAMING_SNAKE_CASE_ , batched=SCREAMING_SNAKE_CASE_ , load_from_cache_file=not data_args.overwrite_cache , desc="Running tokenizer on dataset" , ) if training_args.do_train: if "train" not in raw_datasets: raise ValueError("--do_train requires a train dataset" ) _lowerCAmelCase = raw_datasets["train"] if data_args.max_train_samples is not None: _lowerCAmelCase = train_dataset.select(range(data_args.max_train_samples ) ) if training_args.do_eval: if "validation" not in raw_datasets and "validation_matched" not in raw_datasets: raise ValueError("--do_eval requires a validation dataset" ) _lowerCAmelCase = raw_datasets["validation"] if data_args.max_eval_samples is not None: _lowerCAmelCase = eval_dataset.select(range(data_args.max_eval_samples ) ) if training_args.do_predict or data_args.test_file is not None: if "test" not in raw_datasets and "test_matched" not in raw_datasets: raise ValueError("--do_predict requires a test dataset" ) _lowerCAmelCase = raw_datasets["test"] if data_args.max_predict_samples is not None: _lowerCAmelCase = predict_dataset.select(range(data_args.max_predict_samples ) ) # Log a few random samples from the training set: if training_args.do_train: for index in random.sample(range(len(SCREAMING_SNAKE_CASE_ ) ) , 3 ): logger.info(F'''Sample {index} of the training set: {train_dataset[index]}.''' ) # You can define your custom compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(SCREAMING_SNAKE_CASE_ : EvalPrediction ): _lowerCAmelCase = p.predictions[0] if isinstance(p.predictions , SCREAMING_SNAKE_CASE_ ) else p.predictions _lowerCAmelCase = np.argmax(SCREAMING_SNAKE_CASE_ , axis=1 ) return {"accuracy": (preds == p.label_ids).astype(np.floataa ).mean().item()} # Data collator will default to DataCollatorWithPadding, so we change it if we already did the padding. if data_args.pad_to_max_length: _lowerCAmelCase = default_data_collator elif training_args.fpaa: _lowerCAmelCase = DataCollatorWithPadding(SCREAMING_SNAKE_CASE_ , pad_to_multiple_of=8 ) else: _lowerCAmelCase = None # Initialize our Trainer _lowerCAmelCase = Trainer( model=SCREAMING_SNAKE_CASE_ , args=SCREAMING_SNAKE_CASE_ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , compute_metrics=SCREAMING_SNAKE_CASE_ , tokenizer=SCREAMING_SNAKE_CASE_ , data_collator=SCREAMING_SNAKE_CASE_ , ) # Training if training_args.do_train: _lowerCAmelCase = None if training_args.resume_from_checkpoint is not None: _lowerCAmelCase = training_args.resume_from_checkpoint elif last_checkpoint is not None: _lowerCAmelCase = last_checkpoint _lowerCAmelCase = trainer.train(resume_from_checkpoint=SCREAMING_SNAKE_CASE_ ) _lowerCAmelCase = train_result.metrics _lowerCAmelCase = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(SCREAMING_SNAKE_CASE_ ) ) _lowerCAmelCase = min(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) ) trainer.save_model() # Saves the tokenizer too for easy upload trainer.log_metrics("train" , SCREAMING_SNAKE_CASE_ ) trainer.save_metrics("train" , SCREAMING_SNAKE_CASE_ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info("*** Evaluate ***" ) _lowerCAmelCase = trainer.evaluate(eval_dataset=SCREAMING_SNAKE_CASE_ ) _lowerCAmelCase = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(SCREAMING_SNAKE_CASE_ ) _lowerCAmelCase = min(SCREAMING_SNAKE_CASE_ , len(SCREAMING_SNAKE_CASE_ ) ) trainer.log_metrics("eval" , SCREAMING_SNAKE_CASE_ ) trainer.save_metrics("eval" , SCREAMING_SNAKE_CASE_ ) if training_args.do_predict: logger.info("*** Predict ***" ) # Removing the `label` columns because it contains -1 and Trainer won't like that. _lowerCAmelCase = predict_dataset.remove_columns("label" ) _lowerCAmelCase = trainer.predict(SCREAMING_SNAKE_CASE_ , metric_key_prefix="predict" ).predictions _lowerCAmelCase = np.argmax(SCREAMING_SNAKE_CASE_ , axis=1 ) _lowerCAmelCase = os.path.join(training_args.output_dir , "predict_results_tabfact.txt" ) if trainer.is_world_process_zero(): with open(SCREAMING_SNAKE_CASE_ , "w" ) as writer: logger.info("***** Predict Results *****" ) writer.write("index\tprediction\n" ) for index, item in enumerate(SCREAMING_SNAKE_CASE_ ): _lowerCAmelCase = label_list[item] writer.write(F'''{index}\t{item}\n''' ) _lowerCAmelCase = {"finetuned_from": model_args.model_name_or_path, "tasks": "text-classification"} if training_args.push_to_hub: trainer.push_to_hub(**SCREAMING_SNAKE_CASE_ ) else: trainer.create_model_card(**SCREAMING_SNAKE_CASE_ ) def __a(SCREAMING_SNAKE_CASE_ : str ): '''simple docstring''' main() if __name__ == "__main__": main()
18
from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class __snake_case : """simple docstring""" def __init__( self : Tuple ,lowerCAmelCase__ : List[str] ,lowerCAmelCase__ : Optional[Any]=12 ,lowerCAmelCase__ : Union[str, Any]=7 ,lowerCAmelCase__ : Union[str, Any]=True ,lowerCAmelCase__ : List[str]=True ,lowerCAmelCase__ : Any=True ,lowerCAmelCase__ : Optional[Any]=99 ,lowerCAmelCase__ : List[str]=32 ,lowerCAmelCase__ : Dict=32 ,lowerCAmelCase__ : str=2 ,lowerCAmelCase__ : Optional[int]=4 ,lowerCAmelCase__ : str=37 ,lowerCAmelCase__ : Dict=0.1 ,lowerCAmelCase__ : List[str]=0.1 ,lowerCAmelCase__ : str=5_12 ,lowerCAmelCase__ : Union[str, Any]=0.02 ,lowerCAmelCase__ : Tuple=0 ,lowerCAmelCase__ : str=None ,) -> str: '''simple docstring''' lowerCAmelCase_ : int = parent lowerCAmelCase_ : str = batch_size lowerCAmelCase_ : int = seq_length lowerCAmelCase_ : Union[str, Any] = is_training lowerCAmelCase_ : int = use_input_mask lowerCAmelCase_ : List[Any] = use_labels lowerCAmelCase_ : Dict = vocab_size lowerCAmelCase_ : Union[str, Any] = hidden_size lowerCAmelCase_ : Union[str, Any] = projection_dim lowerCAmelCase_ : List[Any] = num_hidden_layers lowerCAmelCase_ : Any = num_attention_heads lowerCAmelCase_ : List[Any] = intermediate_size lowerCAmelCase_ : Any = dropout lowerCAmelCase_ : Optional[int] = attention_dropout lowerCAmelCase_ : int = max_position_embeddings lowerCAmelCase_ : Optional[int] = initializer_range lowerCAmelCase_ : Any = scope lowerCAmelCase_ : Tuple = bos_token_id def UpperCAmelCase_ ( self : str ) -> Tuple: '''simple docstring''' lowerCAmelCase_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) lowerCAmelCase_ : Dict = None if self.use_input_mask: lowerCAmelCase_ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowerCAmelCase_ : List[Any] = input_mask.numpy() lowerCAmelCase_ , lowerCAmelCase_ : str = input_mask.shape lowerCAmelCase_ : Dict = np.random.randint(1 ,seq_length - 1 ,size=(batch_size,) ) for batch_idx, start_index in enumerate(lowerCAmelCase__ ): lowerCAmelCase_ : Union[str, Any] = 1 lowerCAmelCase_ : Optional[Any] = 0 lowerCAmelCase_ : List[Any] = self.get_config() return config, input_ids, tf.convert_to_tensor(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[str] ) -> str: '''simple docstring''' return BlipTextConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,projection_dim=self.projection_dim ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,dropout=self.dropout ,attention_dropout=self.attention_dropout ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,bos_token_id=self.bos_token_id ,) def UpperCAmelCase_ ( self : Optional[Any] ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Dict ) -> List[Any]: '''simple docstring''' lowerCAmelCase_ : List[Any] = TFBlipTextModel(config=lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = model(lowerCAmelCase__ ,attention_mask=lowerCAmelCase__ ,training=lowerCAmelCase__ ) lowerCAmelCase_ : str = model(lowerCAmelCase__ ,training=lowerCAmelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def UpperCAmelCase_ ( self : Optional[int] ) -> int: '''simple docstring''' lowerCAmelCase_ : List[str] = self.prepare_config_and_inputs() lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ : Dict = config_and_inputs lowerCAmelCase_ : Tuple = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class __snake_case ( snake_case__ , unittest.TestCase ): """simple docstring""" UpperCamelCase_ = (TFBlipTextModel,) if is_tf_available() else () UpperCamelCase_ = False UpperCamelCase_ = False UpperCamelCase_ = False def UpperCAmelCase_ ( self : Optional[Any] ) -> str: '''simple docstring''' lowerCAmelCase_ : List[str] = BlipTextModelTester(self ) lowerCAmelCase_ : Tuple = ConfigTester(self ,config_class=lowerCAmelCase__ ,hidden_size=37 ) def UpperCAmelCase_ ( self : str ) -> Any: '''simple docstring''' self.config_tester.run_common_tests() def UpperCAmelCase_ ( self : List[Any] ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' pass def UpperCAmelCase_ ( self : Union[str, Any] ) -> Any: '''simple docstring''' pass @unittest.skip(reason="Blip does not use inputs_embeds" ) def UpperCAmelCase_ ( self : Union[str, Any] ) -> Optional[int]: '''simple docstring''' pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCAmelCase_ ( self : int ) -> Optional[Any]: '''simple docstring''' pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCAmelCase_ ( self : Dict ) -> Union[str, Any]: '''simple docstring''' pass @slow def UpperCAmelCase_ ( self : Tuple ) -> Optional[Any]: '''simple docstring''' for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase_ : Tuple = TFBlipTextModel.from_pretrained(lowerCAmelCase__ ) self.assertIsNotNone(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : Any ,lowerCAmelCase__ : str=True ) -> List[Any]: '''simple docstring''' super().test_pt_tf_model_equivalence(allow_missing_keys=lowerCAmelCase__ )
659
0
"""simple docstring""" import argparse import torch from torch import nn from transformers import MaMaaaConfig, MaMaaaForConditionalGeneration def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(__snake_case, __snake_case ) def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase , _UpperCamelCase = emb.weight.shape _UpperCamelCase = nn.Linear(__snake_case, __snake_case, bias=__snake_case ) _UpperCamelCase = emb.weight.data return lin_layer def lowerCamelCase__ ( __snake_case ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = torch.load(__snake_case, map_location='''cpu''' ) _UpperCamelCase = mam_aaa['''args'''] or mam_aaa['''cfg''']['''model'''] _UpperCamelCase = mam_aaa['''model'''] remove_ignore_keys_(__snake_case ) _UpperCamelCase = state_dict['''encoder.embed_tokens.weight'''].shape[0] _UpperCamelCase = MaMaaaConfig( vocab_size=__snake_case, max_position_embeddings=10_24, encoder_layers=args.encoder_layers, decoder_layers=args.decoder_layers, encoder_attention_heads=args.encoder_attention_heads, decoder_attention_heads=args.decoder_attention_heads, encoder_ffn_dim=args.encoder_ffn_embed_dim, decoder_ffn_dim=args.decoder_ffn_embed_dim, d_model=args.encoder_embed_dim, encoder_layerdrop=args.encoder_layerdrop, decoder_layerdrop=args.decoder_layerdrop, dropout=args.dropout, attention_dropout=args.attention_dropout, activation_dropout=args.activation_dropout, activation_function='''relu''', ) _UpperCamelCase = state_dict['''decoder.embed_tokens.weight'''] _UpperCamelCase = MaMaaaForConditionalGeneration(__snake_case ) model.model.load_state_dict(__snake_case, strict=__snake_case ) _UpperCamelCase = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument("""fairseq_path""", type=str, help="""path to a model.pt on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") _a = parser.parse_args() _a = convert_fairseq_mamaaa_checkpoint_from_disk(args.fairseq_pathß) model.save_pretrained(args.pytorch_dump_folder_path)
19
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _lowercase = logging.get_logger(__name__) _lowercase = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt'''} # See all LED models at https://huggingface.co/models?filter=LED _lowercase = { '''vocab_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json''', }, '''merges_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''allenai/led-base-16384''': '''https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json''', }, } _lowercase = { '''allenai/led-base-16384''': 16384, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def UpperCamelCase ( ): lowerCAmelCase_ : Optional[int] = ( list(range(ord("!") , ord("~") + 1)) + list(range(ord("¡") , ord("¬") + 1)) + list(range(ord("®") , ord("ÿ") + 1)) ) lowerCAmelCase_ : List[Any] = bs[:] lowerCAmelCase_ : Optional[int] = 0 for b in range(2**8): if b not in bs: bs.append(snake_case__) cs.append(2**8 + n) n += 1 lowerCAmelCase_ : Tuple = [chr(snake_case__) for n in cs] return dict(zip(snake_case__ , snake_case__)) def UpperCamelCase ( snake_case__): lowerCAmelCase_ : str = set() lowerCAmelCase_ : List[Any] = word[0] for char in word[1:]: pairs.add((prev_char, char)) lowerCAmelCase_ : Union[str, Any] = char return pairs class __snake_case ( snake_case__ ): """simple docstring""" UpperCamelCase_ = VOCAB_FILES_NAMES UpperCamelCase_ = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase_ = ['input_ids', 'attention_mask'] def __init__( self : int ,lowerCAmelCase__ : Tuple ,lowerCAmelCase__ : Any ,lowerCAmelCase__ : Tuple="replace" ,lowerCAmelCase__ : Optional[int]="<s>" ,lowerCAmelCase__ : Optional[int]="</s>" ,lowerCAmelCase__ : Tuple="</s>" ,lowerCAmelCase__ : int="<s>" ,lowerCAmelCase__ : Union[str, Any]="<unk>" ,lowerCAmelCase__ : str="<pad>" ,lowerCAmelCase__ : Tuple="<mask>" ,lowerCAmelCase__ : Optional[int]=False ,**lowerCAmelCase__ : Tuple ,) -> Any: '''simple docstring''' lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else bos_token lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else eos_token lowerCAmelCase_ : int = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else sep_token lowerCAmelCase_ : Any = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else cls_token lowerCAmelCase_ : Tuple = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else unk_token lowerCAmelCase_ : Any = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase_ : Optional[int] = AddedToken(lowerCAmelCase__ ,lstrip=lowerCAmelCase__ ,rstrip=lowerCAmelCase__ ) if isinstance(lowerCAmelCase__ ,lowerCAmelCase__ ) else mask_token super().__init__( errors=lowerCAmelCase__ ,bos_token=lowerCAmelCase__ ,eos_token=lowerCAmelCase__ ,unk_token=lowerCAmelCase__ ,sep_token=lowerCAmelCase__ ,cls_token=lowerCAmelCase__ ,pad_token=lowerCAmelCase__ ,mask_token=lowerCAmelCase__ ,add_prefix_space=lowerCAmelCase__ ,**lowerCAmelCase__ ,) with open(lowerCAmelCase__ ,encoding="utf-8" ) as vocab_handle: lowerCAmelCase_ : List[str] = json.load(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[int] = {v: k for k, v in self.encoder.items()} lowerCAmelCase_ : Optional[int] = errors # how to handle errors in decoding lowerCAmelCase_ : Optional[int] = bytes_to_unicode() lowerCAmelCase_ : str = {v: k for k, v in self.byte_encoder.items()} with open(lowerCAmelCase__ ,encoding="utf-8" ) as merges_handle: lowerCAmelCase_ : List[str] = merges_handle.read().split("\n" )[1:-1] lowerCAmelCase_ : List[Any] = [tuple(merge.split() ) for merge in bpe_merges] lowerCAmelCase_ : Union[str, Any] = dict(zip(lowerCAmelCase__ ,range(len(lowerCAmelCase__ ) ) ) ) lowerCAmelCase_ : Dict = {} lowerCAmelCase_ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowerCAmelCase_ : Any = re.compile(R"'s|'t|'re|'ve|'m|'ll|'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+" ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def UpperCAmelCase_ ( self : Dict ) -> Dict: '''simple docstring''' return len(self.encoder ) def UpperCAmelCase_ ( self : Dict ) -> str: '''simple docstring''' return dict(self.encoder ,**self.added_tokens_encoder ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Dict ) -> Dict: '''simple docstring''' if token in self.cache: return self.cache[token] lowerCAmelCase_ : Union[str, Any] = tuple(lowerCAmelCase__ ) lowerCAmelCase_ : str = get_pairs(lowerCAmelCase__ ) if not pairs: return token while True: lowerCAmelCase_ : Optional[int] = min(lowerCAmelCase__ ,key=lambda lowerCAmelCase__ : self.bpe_ranks.get(lowerCAmelCase__ ,float("inf" ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase_ , lowerCAmelCase_ : Optional[Any] = bigram lowerCAmelCase_ : Tuple = [] lowerCAmelCase_ : str = 0 while i < len(lowerCAmelCase__ ): try: lowerCAmelCase_ : Union[str, Any] = word.index(lowerCAmelCase__ ,lowerCAmelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase_ : List[str] = j if word[i] == first and i < len(lowerCAmelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase_ : Optional[int] = tuple(lowerCAmelCase__ ) lowerCAmelCase_ : Tuple = new_word if len(lowerCAmelCase__ ) == 1: break else: lowerCAmelCase_ : Dict = get_pairs(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = " ".join(lowerCAmelCase__ ) lowerCAmelCase_ : Optional[Any] = word return word def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : Dict ) -> Optional[Any]: '''simple docstring''' lowerCAmelCase_ : Any = [] for token in re.findall(self.pat ,lowerCAmelCase__ ): lowerCAmelCase_ : Optional[int] = "".join( self.byte_encoder[b] for b in token.encode("utf-8" ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCAmelCase__ ).split(" " ) ) return bpe_tokens def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any] ) -> Tuple: '''simple docstring''' return self.encoder.get(lowerCAmelCase__ ,self.encoder.get(self.unk_token ) ) def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : Union[str, Any] ) -> Optional[int]: '''simple docstring''' return self.decoder.get(lowerCAmelCase__ ) def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[Any] ) -> Any: '''simple docstring''' lowerCAmelCase_ : int = "".join(lowerCAmelCase__ ) lowerCAmelCase_ : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode("utf-8" ,errors=self.errors ) return text def UpperCAmelCase_ ( self : Tuple ,lowerCAmelCase__ : str ,lowerCAmelCase__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' if not os.path.isdir(lowerCAmelCase__ ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowerCAmelCase_ : Optional[int] = os.path.join( lowerCAmelCase__ ,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) lowerCAmelCase_ : List[str] = os.path.join( lowerCAmelCase__ ,(filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(lowerCAmelCase__ ,"w" ,encoding="utf-8" ) as f: f.write(json.dumps(self.encoder ,indent=2 ,sort_keys=lowerCAmelCase__ ,ensure_ascii=lowerCAmelCase__ ) + "\n" ) lowerCAmelCase_ : Dict = 0 with open(lowerCAmelCase__ ,"w" ,encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() ,key=lambda lowerCAmelCase__ : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' " Please check that the tokenizer is not corrupted!" ) lowerCAmelCase_ : List[Any] = token_index writer.write(" ".join(lowerCAmelCase__ ) + "\n" ) index += 1 return vocab_file, merge_file def UpperCAmelCase_ ( self : str ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase_ : Union[str, Any] = [self.cls_token_id] lowerCAmelCase_ : str = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ,lowerCAmelCase__ : bool = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ ,token_ids_a=lowerCAmelCase__ ,already_has_special_tokens=lowerCAmelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCAmelCase__ )) + [1] return [1] + ([0] * len(lowerCAmelCase__ )) + [1, 1] + ([0] * len(lowerCAmelCase__ )) + [1] def UpperCAmelCase_ ( self : List[Any] ,lowerCAmelCase__ : List[int] ,lowerCAmelCase__ : Optional[List[int]] = None ) -> List[int]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = [self.sep_token_id] lowerCAmelCase_ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCAmelCase_ ( self : Union[str, Any] ,lowerCAmelCase__ : Union[str, Any] ,lowerCAmelCase__ : Optional[int]=False ,**lowerCAmelCase__ : str ) -> Union[str, Any]: '''simple docstring''' lowerCAmelCase_ : Optional[int] = kwargs.pop("add_prefix_space" ,self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCAmelCase__ ) > 0 and not text[0].isspace()): lowerCAmelCase_ : List[str] = " " + text return (text, kwargs) def UpperCAmelCase_ ( self : List[str] ,lowerCAmelCase__ : Union[Dict[str, EncodedInput], BatchEncoding] ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : PaddingStrategy = PaddingStrategy.DO_NOT_PAD ,lowerCAmelCase__ : Optional[int] = None ,lowerCAmelCase__ : Optional[bool] = None ,) -> dict: '''simple docstring''' lowerCAmelCase_ : int = super()._pad( encoded_inputs=lowerCAmelCase__ ,max_length=lowerCAmelCase__ ,padding_strategy=lowerCAmelCase__ ,pad_to_multiple_of=lowerCAmelCase__ ,return_attention_mask=lowerCAmelCase__ ,) # Load from model defaults if return_attention_mask is None: lowerCAmelCase_ : List[Any] = "attention_mask" in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowerCAmelCase_ : Dict = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowerCAmelCase_ : List[Any] = len(encoded_inputs["global_attention_mask"] ) != len(lowerCAmelCase__ ) if needs_to_be_padded: lowerCAmelCase_ : Union[str, Any] = len(lowerCAmelCase__ ) - len(encoded_inputs["global_attention_mask"] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowerCAmelCase_ : Optional[int] = ( encoded_inputs["global_attention_mask"] + [-1] * difference ) elif self.padding_side == "left": lowerCAmelCase_ : List[Any] = [-1] * difference + encoded_inputs[ "global_attention_mask" ] else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return encoded_inputs
659
0