code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch lowercase__ = '''sshleifer/bart-tiny-random''' lowercase__ = '''patrickvonplaten/t5-tiny-random''' @require_torch class UpperCAmelCase_ ( unittest.TestCase ): """simple docstring""" @cached_property def _lowercase ( self ): return AutoConfig.from_pretrained(UpperCAmelCase_ ) def _lowercase ( self ): snake_case_ , *snake_case_ = create_student_by_copying_alternating_layers(UpperCAmelCase_ , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.num_hidden_layers , 1 ) def _lowercase ( self ): snake_case_ , *snake_case_ = create_student_by_copying_alternating_layers(UpperCAmelCase_ , tempfile.mkdtemp() , e=1 , d=UpperCAmelCase_ ) def _lowercase ( self ): snake_case_ , *snake_case_ = create_student_by_copying_alternating_layers(UpperCAmelCase_ , tempfile.mkdtemp() , e=1 , d=UpperCAmelCase_ ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , self.teacher_config.encoder_layers ) def _lowercase ( self ): snake_case_ , *snake_case_ = create_student_by_copying_alternating_layers(UpperCAmelCase_ , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , 1 ) def _lowercase ( self ): with self.assertRaises(UpperCAmelCase_ ): create_student_by_copying_alternating_layers(UpperCAmelCase_ , tempfile.mkdtemp() , e=UpperCAmelCase_ , d=UpperCAmelCase_ )
508
'''simple docstring''' def __snake_case ( lowercase : int ): if n == 1 or not isinstance(lowercase , lowercase ): return 0 elif n == 2: return 1 else: snake_case_ = [0, 1] for i in range(2 , n + 1 ): sequence.append(sequence[i - 1] + sequence[i - 2] ) return sequence[n] def __snake_case ( lowercase : int ): snake_case_ = 0 snake_case_ = 2 while digits < n: index += 1 snake_case_ = len(str(fibonacci(lowercase ) ) ) return index def __snake_case ( lowercase : int = 1_000 ): return fibonacci_digits_index(lowercase ) if __name__ == "__main__": print(solution(int(str(input()).strip())))
508
1
from dataclasses import dataclass from typing import Optional, Tuple import torch from torch import nn from transformers import RobertaPreTrainedModel, XLMRobertaConfig, XLMRobertaModel from transformers.utils import ModelOutput @dataclass class A (SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCamelCase : Optional[torch.FloatTensor] = None __lowerCamelCase : torch.FloatTensor = None __lowerCamelCase : Optional[Tuple[torch.FloatTensor]] = None __lowerCamelCase : Optional[Tuple[torch.FloatTensor]] = None class A (SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : str , __lowerCAmelCase : List[str]=1 , __lowerCAmelCase : int=0 , __lowerCAmelCase : int=2 , __lowerCAmelCase : Optional[int]=5_12 , __lowerCAmelCase : Tuple="cls" , __lowerCAmelCase : List[str]=False , __lowerCAmelCase : Dict=True , **__lowerCAmelCase : Tuple , ) -> List[Any]: """simple docstring""" super().__init__(pad_token_id=__lowerCAmelCase , bos_token_id=__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase ) A__ = project_dim A__ = pooler_fn A__ = learn_encoder A__ = use_attention_mask class A (SCREAMING_SNAKE_CASE ): '''simple docstring''' __lowerCamelCase : Dict = [R'''pooler''', R'''logit_scale'''] __lowerCamelCase : Optional[int] = [R'''position_ids''', R'''predictions.decoder.bias'''] __lowerCamelCase : Dict = '''roberta''' __lowerCamelCase : List[Any] = RobertaSeriesConfig def __init__( self : Tuple , __lowerCAmelCase : Optional[int] ) -> Any: """simple docstring""" super().__init__(__lowerCAmelCase ) A__ = XLMRobertaModel(__lowerCAmelCase ) A__ = nn.Linear(config.hidden_size , config.project_dim ) A__ = getattr(__lowerCAmelCase , """has_pre_transformation""" , __lowerCAmelCase ) if self.has_pre_transformation: A__ = nn.Linear(config.hidden_size , config.project_dim ) A__ = nn.LayerNorm(config.hidden_size , eps=config.layer_norm_eps ) self.post_init() def a_ ( self : Any , __lowerCAmelCase : Optional[torch.Tensor] = None , __lowerCAmelCase : Optional[torch.Tensor] = None , __lowerCAmelCase : Optional[torch.Tensor] = None , __lowerCAmelCase : Optional[torch.Tensor] = None , __lowerCAmelCase : Optional[torch.Tensor] = None , __lowerCAmelCase : Optional[torch.Tensor] = None , __lowerCAmelCase : Optional[torch.Tensor] = None , __lowerCAmelCase : Optional[torch.Tensor] = None , __lowerCAmelCase : Optional[bool] = None , __lowerCAmelCase : Optional[bool] = None , __lowerCAmelCase : Optional[bool] = None , ) -> Tuple: """simple docstring""" A__ = return_dict if return_dict is not None else self.config.use_return_dict A__ = self.base_model( input_ids=__lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , position_ids=__lowerCAmelCase , head_mask=__lowerCAmelCase , inputs_embeds=__lowerCAmelCase , encoder_hidden_states=__lowerCAmelCase , encoder_attention_mask=__lowerCAmelCase , output_attentions=__lowerCAmelCase , output_hidden_states=True if self.has_pre_transformation else output_hidden_states , return_dict=__lowerCAmelCase , ) if self.has_pre_transformation: A__ = outputs["""hidden_states"""][-2] A__ = self.pre_LN(__lowerCAmelCase ) A__ = self.transformation_pre(__lowerCAmelCase ) return TransformationModelOutput( projection_state=__lowerCAmelCase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , ) else: A__ = self.transformation(outputs.last_hidden_state ) return TransformationModelOutput( projection_state=__lowerCAmelCase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
706
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from accelerate import PartialState from accelerate.utils.operations import broadcast, gather, gather_object, pad_across_processes, reduce def __lowerCamelCase ( __a :Dict ) -> List[Any]: """simple docstring""" return (torch.arange(state.num_processes ) + 1.0 + (state.num_processes * state.process_index)).to(state.device ) def __lowerCamelCase ( __a :str ) -> int: """simple docstring""" A__ = create_tensor(__a ) A__ = gather(__a ) assert gathered_tensor.tolist() == list(range(1 , state.num_processes**2 + 1 ) ) def __lowerCamelCase ( __a :Optional[Any] ) -> Optional[int]: """simple docstring""" A__ = [state.process_index] A__ = gather_object(__a ) assert len(__a ) == state.num_processes, F'{gathered_obj}, {len(__a )} != {state.num_processes}' assert gathered_obj == list(range(state.num_processes ) ), F'{gathered_obj} != {list(range(state.num_processes ) )}' def __lowerCamelCase ( __a :Optional[int] ) -> Dict: """simple docstring""" A__ = create_tensor(__a ) A__ = broadcast(__a ) assert broadcasted_tensor.shape == torch.Size([state.num_processes] ) assert broadcasted_tensor.tolist() == list(range(1 , state.num_processes + 1 ) ) def __lowerCamelCase ( __a :List[str] ) -> Tuple: """simple docstring""" if state.is_main_process: A__ = torch.arange(state.num_processes + 1 ).to(state.device ) else: A__ = torch.arange(state.num_processes ).to(state.device ) A__ = pad_across_processes(__a ) assert padded_tensor.shape == torch.Size([state.num_processes + 1] ) if not state.is_main_process: assert padded_tensor.tolist() == list(range(0 , state.num_processes ) ) + [0] def __lowerCamelCase ( __a :Optional[int] ) -> Tuple: """simple docstring""" if state.num_processes != 2: return A__ = create_tensor(__a ) A__ = reduce(__a , """sum""" ) A__ = torch.tensor([4.0, 6] ).to(state.device ) assert torch.allclose(__a , __a ), F'{reduced_tensor} != {truth_tensor}' def __lowerCamelCase ( __a :str ) -> List[str]: """simple docstring""" if state.num_processes != 2: return A__ = create_tensor(__a ) A__ = reduce(__a , """mean""" ) A__ = torch.tensor([2.0, 3] ).to(state.device ) assert torch.allclose(__a , __a ), F'{reduced_tensor} != {truth_tensor}' def __lowerCamelCase ( __a :List[Any] ) -> Union[str, Any]: """simple docstring""" main() def __lowerCamelCase ( ) -> List[str]: """simple docstring""" A__ = PartialState() state.print(F'State: {state}' ) state.print("""testing gather""" ) test_gather(__a ) state.print("""testing gather_object""" ) test_gather_object(__a ) state.print("""testing broadcast""" ) test_broadcast(__a ) state.print("""testing pad_across_processes""" ) test_pad_across_processes(__a ) state.print("""testing reduce_sum""" ) test_reduce_sum(__a ) state.print("""testing reduce_mean""" ) test_reduce_mean(__a ) if __name__ == "__main__": main()
247
0
import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) lowerCamelCase : Optional[int] = logging.getLogger() def _SCREAMING_SNAKE_CASE ( ): '''simple docstring''' lowerCamelCase_ = argparse.ArgumentParser() parser.add_argument('-f' ) lowerCamelCase_ = parser.parse_args() return args.f class A( _UpperCamelCase ): '''simple docstring''' def a__ ( self : List[Any] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = logging.StreamHandler(sys.stdout ) logger.addHandler(A_ ) def a__ ( self : Union[str, Any] , A_ : Optional[Any] ) -> Union[str, Any]: """simple docstring""" lowerCamelCase_ = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , 'run_glue_deebert.py' ) with patch.object(A_ , 'argv' , A_ ): lowerCamelCase_ = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(A_ , 0.666 ) @slow @require_torch_non_multi_gpu def a__ ( self : str ) -> List[str]: """simple docstring""" lowerCamelCase_ = "\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n ".split() self.run_and_check(A_ ) lowerCamelCase_ = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(A_ ) lowerCamelCase_ = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(A_ )
70
import math class snake_case__: """simple docstring""" def __init__( self : int , SCREAMING_SNAKE_CASE : List[Any]=0 ): # a graph with Node 0,1,...,N-1 lowercase__ : Dict = n lowercase__ : List[Any] = [ [math.inf for j in range(0 , SCREAMING_SNAKE_CASE )] for i in range(0 , SCREAMING_SNAKE_CASE ) ] # adjacency matrix for weight lowercase__ : Optional[int] = [ [math.inf for j in range(0 , SCREAMING_SNAKE_CASE )] for i in range(0 , SCREAMING_SNAKE_CASE ) ] # dp[i][j] stores minimum distance from i to j def snake_case ( self : int , SCREAMING_SNAKE_CASE : Any , SCREAMING_SNAKE_CASE : List[str] , SCREAMING_SNAKE_CASE : Optional[Any] ): lowercase__ : Optional[Any] = w def snake_case ( self : int ): for k in range(0 , self.n ): for i in range(0 , self.n ): for j in range(0 , self.n ): lowercase__ : Union[str, Any] = min(self.dp[i][j] , self.dp[i][k] + self.dp[k][j] ) def snake_case ( self : List[str] , SCREAMING_SNAKE_CASE : str , SCREAMING_SNAKE_CASE : Any ): return self.dp[u][v] if __name__ == "__main__": lowerCAmelCase__ = Graph(5) graph.add_edge(0, 2, 9) graph.add_edge(0, 4, 1_0) graph.add_edge(1, 3, 5) graph.add_edge(2, 3, 7) graph.add_edge(3, 0, 1_0) graph.add_edge(3, 1, 2) graph.add_edge(3, 2, 1) graph.add_edge(3, 4, 6) graph.add_edge(4, 1, 3) graph.add_edge(4, 2, 4) graph.add_edge(4, 3, 9) graph.floyd_warshall() graph.show_min(1, 4) graph.show_min(0, 3)
496
0
import re def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> Dict: return [char.split() for char in re.split(R'[^ a-z A-Z 0-9 \s]' , str_ )] def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> Optional[int]: __lowerCamelCase : int = split_input(str_ ) return "".join( [''.join([char.capitalize() for char in sub_str] ) for sub_str in string_split] ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> List[Any]: try: __lowerCamelCase : Tuple = split_input(__snake_case ) if upper: __lowerCamelCase : List[Any] = ''.join( [ separator.join([char.upper() for char in sub_str] ) for sub_str in string_split ] ) else: __lowerCamelCase : List[Any] = ''.join( [ separator.join([char.lower() for char in sub_str] ) for sub_str in string_split ] ) return res_str except IndexError: return "not valid string" def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> List[str]: return to_simple_case(__snake_case ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> str: try: __lowerCamelCase : int = to_simple_case(__snake_case ) return res_str[0].lower() + res_str[1:] except IndexError: return "not valid string" def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> int: return to_complex_case(__snake_case , __snake_case , '_' ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> Any: return to_complex_case(__snake_case , __snake_case , '-' ) if __name__ == "__main__": __import__("""doctest""").testmod()
705
def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> float: _validate_point(lowerCamelCase__ ) _validate_point(lowerCamelCase__ ) if len(lowerCamelCase__ ) != len(lowerCamelCase__ ): raise ValueError('Both points must be in the same n-dimensional space' ) return float(sum(abs(a - b ) for a, b in zip(lowerCamelCase__ , lowerCamelCase__ ) ) ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> None: if point: if isinstance(lowerCamelCase__ , lowerCamelCase__ ): for item in point: if not isinstance(lowerCamelCase__ , (int, float) ): __lowerCamelCase : int = ( 'Expected a list of numbers as input, found ' F"{type(lowerCamelCase__ ).__name__}" ) raise TypeError(lowerCamelCase__ ) else: __lowerCamelCase : Optional[Any] = F"Expected a list of numbers as input, found {type(lowerCamelCase__ ).__name__}" raise TypeError(lowerCamelCase__ ) else: raise ValueError('Missing an input' ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> float: _validate_point(lowerCamelCase__ ) _validate_point(lowerCamelCase__ ) if len(lowerCamelCase__ ) != len(lowerCamelCase__ ): raise ValueError('Both points must be in the same n-dimensional space' ) return float(sum(abs(x - y ) for x, y in zip(lowerCamelCase__ , lowerCamelCase__ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
337
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import convert_to_rgb, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL A : Optional[int] = logging.get_logger(__name__) class lowerCAmelCase ( UpperCAmelCase__ ): '''simple docstring''' A = ['pixel_values'] def __init__( self :Optional[int] , lowerCamelCase_ :bool = True , lowerCamelCase_ :Dict[str, int] = None , lowerCamelCase_ :PILImageResampling = PILImageResampling.BICUBIC , lowerCamelCase_ :bool = True , lowerCamelCase_ :Union[int, float] = 1 / 2_5_5 , lowerCamelCase_ :bool = True , lowerCamelCase_ :Optional[Union[float, List[float]]] = None , lowerCamelCase_ :Optional[Union[float, List[float]]] = None , lowerCamelCase_ :bool = True , **lowerCamelCase_ :Optional[int] , ) -> List[str]: """simple docstring""" super().__init__(**__lowercase ) UpperCamelCase__ = size if size is not None else {"height": 3_8_4, "width": 3_8_4} UpperCamelCase__ = get_size_dict(__lowercase , default_to_square=__lowercase ) UpperCamelCase__ = do_resize UpperCamelCase__ = size UpperCamelCase__ = resample UpperCamelCase__ = do_rescale UpperCamelCase__ = rescale_factor UpperCamelCase__ = do_normalize UpperCamelCase__ = image_mean if image_mean is not None else OPENAI_CLIP_MEAN UpperCamelCase__ = image_std if image_std is not None else OPENAI_CLIP_STD UpperCamelCase__ = do_convert_rgb def lowerCamelCase__ ( self :Optional[int] , lowerCamelCase_ :np.ndarray , lowerCamelCase_ :Dict[str, int] , lowerCamelCase_ :PILImageResampling = PILImageResampling.BICUBIC , lowerCamelCase_ :Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase_ :int , ) -> Optional[Any]: """simple docstring""" UpperCamelCase__ = get_size_dict(__lowercase , default_to_square=__lowercase ) if "height" not in size or "width" not in size: raise ValueError(f'The `size` dictionary must contain the keys `height` and `width`. Got {size.keys()}' ) UpperCamelCase__ = (size["height"], size["width"]) return resize(__lowercase , size=__lowercase , resample=__lowercase , data_format=__lowercase , **__lowercase ) def lowerCamelCase__ ( self :Tuple , lowerCamelCase_ :np.ndarray , lowerCamelCase_ :Union[int, float] , lowerCamelCase_ :Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase_ :Dict , ) -> List[str]: """simple docstring""" return rescale(__lowercase , scale=__lowercase , data_format=__lowercase , **__lowercase ) def lowerCamelCase__ ( self :Union[str, Any] , lowerCamelCase_ :np.ndarray , lowerCamelCase_ :Union[float, List[float]] , lowerCamelCase_ :Union[float, List[float]] , lowerCamelCase_ :Optional[Union[str, ChannelDimension]] = None , **lowerCamelCase_ :Union[str, Any] , ) -> str: """simple docstring""" return normalize(__lowercase , mean=__lowercase , std=__lowercase , data_format=__lowercase , **__lowercase ) def lowerCamelCase__ ( self :Union[str, Any] , lowerCamelCase_ :ImageInput , lowerCamelCase_ :Optional[bool] = None , lowerCamelCase_ :Optional[Dict[str, int]] = None , lowerCamelCase_ :PILImageResampling = None , lowerCamelCase_ :Optional[bool] = None , lowerCamelCase_ :Optional[float] = None , lowerCamelCase_ :Optional[bool] = None , lowerCamelCase_ :Optional[Union[float, List[float]]] = None , lowerCamelCase_ :Optional[Union[float, List[float]]] = None , lowerCamelCase_ :Optional[Union[str, TensorType]] = None , lowerCamelCase_ :bool = None , lowerCamelCase_ :ChannelDimension = ChannelDimension.FIRST , **lowerCamelCase_ :Union[str, Any] , ) -> int: """simple docstring""" UpperCamelCase__ = do_resize if do_resize is not None else self.do_resize UpperCamelCase__ = resample if resample is not None else self.resample UpperCamelCase__ = do_rescale if do_rescale is not None else self.do_rescale UpperCamelCase__ = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCamelCase__ = do_normalize if do_normalize is not None else self.do_normalize UpperCamelCase__ = image_mean if image_mean is not None else self.image_mean UpperCamelCase__ = image_std if image_std is not None else self.image_std UpperCamelCase__ = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb UpperCamelCase__ = size if size is not None else self.size UpperCamelCase__ = get_size_dict(__lowercase , default_to_square=__lowercase ) UpperCamelCase__ = make_list_of_images(__lowercase ) if not valid_images(__lowercase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # PIL RGBA images are converted to RGB if do_convert_rgb: UpperCamelCase__ = [convert_to_rgb(__lowercase ) for image in images] # All transformations expect numpy arrays. UpperCamelCase__ = [to_numpy_array(__lowercase ) for image in images] if do_resize: UpperCamelCase__ = [self.resize(image=__lowercase , size=__lowercase , resample=__lowercase ) for image in images] if do_rescale: UpperCamelCase__ = [self.rescale(image=__lowercase , scale=__lowercase ) for image in images] if do_normalize: UpperCamelCase__ = [self.normalize(image=__lowercase , mean=__lowercase , std=__lowercase ) for image in images] UpperCamelCase__ = [to_channel_dimension_format(__lowercase , __lowercase ) for image in images] UpperCamelCase__ = BatchFeature(data={"pixel_values": images} , tensor_type=__lowercase ) return encoded_outputs
516
from binascii import hexlify from hashlib import shaaaa from os import urandom # RFC 3526 - More Modular Exponential (MODP) Diffie-Hellman groups for # Internet Key Exchange (IKE) https://tools.ietf.org/html/rfc3526 lowercase__ : Dict = { # 1536-bit 5: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA237327FFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 2048-bit 14: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B" + "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9" + "DE2BCBF6955817183995497CEA956AE515D2261898FA0510" + "15728E5A8AACAA68FFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 3072-bit 15: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B" + "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9" + "DE2BCBF6955817183995497CEA956AE515D2261898FA0510" + "15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64" + "ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7" + "ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B" + "F12FFA06D98A0864D87602733EC86A64521F2B18177B200C" + "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31" + "43DB5BFCE0FD108E4B82D120A93AD2CAFFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 4096-bit 16: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B" + "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9" + "DE2BCBF6955817183995497CEA956AE515D2261898FA0510" + "15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64" + "ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7" + "ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B" + "F12FFA06D98A0864D87602733EC86A64521F2B18177B200C" + "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31" + "43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7" + "88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA" + "2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6" + "287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED" + "1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9" + "93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934063199" + "FFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 6144-bit 17: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD129024E08" + "8A67CC74020BBEA63B139B22514A08798E3404DDEF9519B3CD3A431B" + "302B0A6DF25F14374FE1356D6D51C245E485B576625E7EC6F44C42E9" + "A637ED6B0BFF5CB6F406B7EDEE386BFB5A899FA5AE9F24117C4B1FE6" + "49286651ECE45B3DC2007CB8A163BF0598DA48361C55D39A69163FA8" + "FD24CF5F83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3BE39E772C" + "180E86039B2783A2EC07A28FB5C55DF06F4C52C9DE2BCBF695581718" + "3995497CEA956AE515D2261898FA051015728E5A8AAAC42DAD33170D" + "04507A33A85521ABDF1CBA64ECFB850458DBEF0A8AEA71575D060C7D" + "B3970F85A6E1E4C7ABF5AE8CDB0933D71E8C94E04A25619DCEE3D226" + "1AD2EE6BF12FFA06D98A0864D87602733EC86A64521F2B18177B200C" + "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB3143DB5BFC" + "E0FD108E4B82D120A92108011A723C12A787E6D788719A10BDBA5B26" + "99C327186AF4E23C1A946834B6150BDA2583E9CA2AD44CE8DBBBC2DB" + "04DE8EF92E8EFC141FBECAA6287C59474E6BC05D99B2964FA090C3A2" + "233BA186515BE7ED1F612970CEE2D7AFB81BDD762170481CD0069127" + "D5B05AA993B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492" + "36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BDF8FF9406" + "AD9E530EE5DB382F413001AEB06A53ED9027D831179727B0865A8918" + "DA3EDBEBCF9B14ED44CE6CBACED4BB1BDB7F1447E6CC254B33205151" + "2BD7AF426FB8F401378CD2BF5983CA01C64B92ECF032EA15D1721D03" + "F482D7CE6E74FEF6D55E702F46980C82B5A84031900B1C9E59E7C97F" + "BEC7E8F323A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA" + "CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE32806A1D58B" + "B7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55CDA56C9EC2EF29632" + "387FE8D76E3C0468043E8F663F4860EE12BF2D5B0B7474D6E694F91E" + "6DCC4024FFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, # 8192-bit 18: { "prime": int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1" + "29024E088A67CC74020BBEA63B139B22514A08798E3404DD" + "EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245" + "E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED" + "EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D" + "C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F" + "83655D23DCA3AD961C62F356208552BB9ED529077096966D" + "670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B" + "E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9" + "DE2BCBF6955817183995497CEA956AE515D2261898FA0510" + "15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64" + "ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7" + "ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B" + "F12FFA06D98A0864D87602733EC86A64521F2B18177B200C" + "BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31" + "43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7" + "88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA" + "2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6" + "287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED" + "1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9" + "93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492" + "36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BD" + "F8FF9406AD9E530EE5DB382F413001AEB06A53ED9027D831" + "179727B0865A8918DA3EDBEBCF9B14ED44CE6CBACED4BB1B" + "DB7F1447E6CC254B332051512BD7AF426FB8F401378CD2BF" + "5983CA01C64B92ECF032EA15D1721D03F482D7CE6E74FEF6" + "D55E702F46980C82B5A84031900B1C9E59E7C97FBEC7E8F3" + "23A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA" + "CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE328" + "06A1D58BB7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55C" + "DA56C9EC2EF29632387FE8D76E3C0468043E8F663F4860EE" + "12BF2D5B0B7474D6E694F91E6DBE115974A3926F12FEE5E4" + "38777CB6A932DF8CD8BEC4D073B931BA3BC832B68D9DD300" + "741FA7BF8AFC47ED2576F6936BA424663AAB639C5AE4F568" + "3423B4742BF1C978238F16CBE39D652DE3FDB8BEFC848AD9" + "22222E04A4037C0713EB57A81A23F0C73473FC646CEA306B" + "4BCBC8862F8385DDFA9D4B7FA2C087E879683303ED5BDD3A" + "062B3CF5B3A278A66D2A13F83F44F82DDF310EE074AB6A36" + "4597E899A0255DC164F31CC50846851DF9AB48195DED7EA1" + "B1D510BD7EE74D73FAF36BC31ECFA268359046F4EB879F92" + "4009438B481C6CD7889A002ED5EE382BC9190DA6FC026E47" + "9558E4475677E9AA9E3050E2765694DFC81F56E880B96E71" + "60C980DD98EDD3DFFFFFFFFFFFFFFFFF", base=16, ), "generator": 2, }, } class UpperCAmelCase : '''simple docstring''' def __init__( self : Tuple , __lowercase : int = 14 ): """simple docstring""" if group not in primes: raise ValueError("Unsupported Group" ) snake_case_ = primes[group]["prime"] snake_case_ = primes[group]["generator"] snake_case_ = int(hexlify(urandom(32 ) ) , base=16 ) def snake_case__ ( self : Tuple ): """simple docstring""" return hex(self.__private_key )[2:] def snake_case__ ( self : Optional[int] ): """simple docstring""" snake_case_ = pow(self.generator , self.__private_key , self.prime ) return hex(__lowercase )[2:] def snake_case__ ( self : int , __lowercase : int ): """simple docstring""" return ( 2 <= key <= self.prime - 2 and pow(__lowercase , (self.prime - 1) // 2 , self.prime ) == 1 ) def snake_case__ ( self : Union[str, Any] , __lowercase : str ): """simple docstring""" snake_case_ = int(__lowercase , base=16 ) if not self.is_valid_public_key(__lowercase ): raise ValueError("Invalid public key" ) snake_case_ = pow(__lowercase , self.__private_key , self.prime ) return shaaaa(str(__lowercase ).encode() ).hexdigest() @staticmethod def snake_case__ ( __lowercase : int , __lowercase : int ): """simple docstring""" return ( 2 <= remote_public_key_str <= prime - 2 and pow(__lowercase , (prime - 1) // 2 , __lowercase ) == 1 ) @staticmethod def snake_case__ ( __lowercase : str , __lowercase : str , __lowercase : int = 14 ): """simple docstring""" snake_case_ = int(__lowercase , base=16 ) snake_case_ = int(__lowercase , base=16 ) snake_case_ = primes[group]["prime"] if not DiffieHellman.is_valid_public_key_static(__lowercase , __lowercase ): raise ValueError("Invalid public key" ) snake_case_ = pow(__lowercase , __lowercase , __lowercase ) return shaaaa(str(__lowercase ).encode() ).hexdigest() if __name__ == "__main__": import doctest doctest.testmod()
376
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowerCAmelCase = {'''configuration_xglm''': ['''XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XGLMConfig''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = ['''XGLMTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = ['''XGLMTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ '''XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XGLMForCausalLM''', '''XGLMModel''', '''XGLMPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ '''FlaxXGLMForCausalLM''', '''FlaxXGLMModel''', '''FlaxXGLMPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase = [ '''TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXGLMForCausalLM''', '''TFXGLMModel''', '''TFXGLMPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xglm import XGLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XGLMConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm import XGLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xglm_fast import XGLMTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xglm import XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, XGLMForCausalLM, XGLMModel, XGLMPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xglm import FlaxXGLMForCausalLM, FlaxXGLMModel, FlaxXGLMPreTrainedModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, TFXGLMPreTrainedModel, ) else: import sys __lowerCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
717
from collections import OrderedDict from typing import Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...feature_extraction_utils import FeatureExtractionMixin from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType, logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { '''deepmind/language-perceiver''': '''https://huggingface.co/deepmind/language-perceiver/resolve/main/config.json''', # See all Perceiver models at https://huggingface.co/models?filter=perceiver } class __a ( __UpperCamelCase ): __lowercase : List[str] = 'perceiver' def __init__( self , lowerCAmelCase__=256 , lowerCAmelCase__=1_280 , lowerCAmelCase__=768 , lowerCAmelCase__=1 , lowerCAmelCase__=26 , lowerCAmelCase__=8 , lowerCAmelCase__=8 , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__="kv" , lowerCAmelCase__=1 , lowerCAmelCase__=1 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.0_2 , lowerCAmelCase__=1E-12 , lowerCAmelCase__=True , lowerCAmelCase__=262 , lowerCAmelCase__=2_048 , lowerCAmelCase__=56 , lowerCAmelCase__=[368, 496] , lowerCAmelCase__=16 , lowerCAmelCase__=1_920 , lowerCAmelCase__=16 , lowerCAmelCase__=[1, 16, 224, 224] , **lowerCAmelCase__ , ) -> Dict: '''simple docstring''' super().__init__(**lowerCAmelCase__ ) lowercase__: Tuple = num_latents lowercase__: Dict = d_latents lowercase__: Tuple = d_model lowercase__: Tuple = num_blocks lowercase__: List[Any] = num_self_attends_per_block lowercase__: Any = num_self_attention_heads lowercase__: Optional[int] = num_cross_attention_heads lowercase__: Dict = qk_channels lowercase__: Any = v_channels lowercase__: Dict = cross_attention_shape_for_attention lowercase__: int = self_attention_widening_factor lowercase__: Tuple = cross_attention_widening_factor lowercase__: Optional[Any] = hidden_act lowercase__: Union[str, Any] = attention_probs_dropout_prob lowercase__: str = initializer_range lowercase__: str = layer_norm_eps lowercase__: List[Any] = use_query_residual # masked language modeling attributes lowercase__: List[Any] = vocab_size lowercase__: str = max_position_embeddings # image classification attributes lowercase__: Optional[Any] = image_size # flow attributes lowercase__: int = train_size # multimodal autoencoding attributes lowercase__: Dict = num_frames lowercase__: int = audio_samples_per_frame lowercase__: Optional[int] = samples_per_patch lowercase__: Tuple = output_shape class __a ( __UpperCamelCase ): @property def SCREAMING_SNAKE_CASE__ ( self ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": lowercase__: Any = {0: 'batch', 1: 'choice', 2: 'sequence'} else: lowercase__: List[Any] = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('inputs', dynamic_axis), ('attention_mask', dynamic_axis), ] ) @property def SCREAMING_SNAKE_CASE__ ( self ) -> float: '''simple docstring''' return 1E-4 def SCREAMING_SNAKE_CASE__ ( self , lowerCAmelCase__ , lowerCAmelCase__ = -1 , lowerCAmelCase__ = -1 , lowerCAmelCase__ = -1 , lowerCAmelCase__ = False , lowerCAmelCase__ = None , lowerCAmelCase__ = 3 , lowerCAmelCase__ = 40 , lowerCAmelCase__ = 40 , ) -> Mapping[str, Any]: '''simple docstring''' # copied from `transformers.onnx.config.OnnxConfig` and slightly altered/simplified if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase__: Dict = compute_effective_axis_dimension( lowerCAmelCase__ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX lowercase__: int = preprocessor.num_special_tokens_to_add(lowerCAmelCase__ ) lowercase__: Optional[int] = compute_effective_axis_dimension( lowerCAmelCase__ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=lowerCAmelCase__ ) # Generate dummy inputs according to compute batch and sequence lowercase__: str = [' '.join(['a'] ) * seq_length] * batch_size lowercase__: Optional[int] = dict(preprocessor(lowerCAmelCase__ , return_tensors=lowerCAmelCase__ ) ) lowercase__: str = inputs.pop('input_ids' ) return inputs elif isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) and preprocessor.model_input_names[0] == "pixel_values": # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX lowercase__: str = compute_effective_axis_dimension(lowerCAmelCase__ , fixed_dimension=OnnxConfig.default_fixed_batch ) lowercase__: str = self._generate_dummy_images(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) lowercase__: str = dict(preprocessor(images=lowerCAmelCase__ , return_tensors=lowerCAmelCase__ ) ) lowercase__: Union[str, Any] = inputs.pop('pixel_values' ) return inputs else: raise ValueError( 'Unable to generate dummy inputs for the model. Please provide a tokenizer or a preprocessor.' )
335
0
"""simple docstring""" from typing import Union import fire import torch from tqdm import tqdm def UpperCAmelCase ( A__: str , A__: str = "cpu" , A__: Union[str, None] = None ) -> None: __lowerCamelCase : Tuple = torch.load(SCREAMING_SNAKE_CASE_ , map_location=SCREAMING_SNAKE_CASE_ ) for k, v in tqdm(state_dict.items() ): if not isinstance(SCREAMING_SNAKE_CASE_ , torch.Tensor ): raise TypeError('FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin' ) __lowerCamelCase : Optional[Any] = v.half() if save_path is None: # overwrite src_path __lowerCamelCase : int = src_path torch.save(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if __name__ == "__main__": fire.Fire(convert)
594
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import _LazyModule snake_case_ = {'tokenization_byt5': ['ByT5Tokenizer']} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys snake_case_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
421
0
'''simple docstring''' import os from bleurt import score # From: git+https://github.com/google-research/bleurt.git import datasets lowercase = datasets.logging.get_logger(__name__) lowercase = '''\\n@inproceedings{bleurt,\n title={BLEURT: Learning Robust Metrics for Text Generation},\n author={Thibault Sellam and Dipanjan Das and Ankur P. Parikh},\n booktitle={ACL},\n year={2020},\n url={https://arxiv.org/abs/2004.04696}\n}\n''' lowercase = '''\\nBLEURT a learnt evaluation metric for Natural Language Generation. It is built using multiple phases of transfer learning starting from a pretrained BERT model (Devlin et al. 2018)\nand then employing another pre-training phrase using synthetic data. Finally it is trained on WMT human annotations. You may run BLEURT out-of-the-box or fine-tune\nit for your specific application (the latter is expected to perform better).\n\nSee the project\'s README at https://github.com/google-research/bleurt#readme for more information.\n''' lowercase = '''\nBLEURT score.\n\nArgs:\n `predictions` (list of str): prediction/candidate sentences\n `references` (list of str): reference sentences\n `checkpoint` BLEURT checkpoint. Will default to BLEURT-tiny if None.\n\nReturns:\n \'scores\': List of scores.\nExamples:\n\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> bleurt = datasets.load_metric("bleurt")\n >>> results = bleurt.compute(predictions=predictions, references=references)\n >>> print([round(v, 2) for v in results["scores"]])\n [1.03, 1.04]\n''' lowercase = { '''bleurt-tiny-128''': '''https://storage.googleapis.com/bleurt-oss/bleurt-tiny-128.zip''', '''bleurt-tiny-512''': '''https://storage.googleapis.com/bleurt-oss/bleurt-tiny-512.zip''', '''bleurt-base-128''': '''https://storage.googleapis.com/bleurt-oss/bleurt-base-128.zip''', '''bleurt-base-512''': '''https://storage.googleapis.com/bleurt-oss/bleurt-base-512.zip''', '''bleurt-large-128''': '''https://storage.googleapis.com/bleurt-oss/bleurt-large-128.zip''', '''bleurt-large-512''': '''https://storage.googleapis.com/bleurt-oss/bleurt-large-512.zip''', '''BLEURT-20-D3''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D3.zip''', '''BLEURT-20-D6''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D6.zip''', '''BLEURT-20-D12''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20-D12.zip''', '''BLEURT-20''': '''https://storage.googleapis.com/bleurt-oss-21/BLEURT-20.zip''', } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class UpperCAmelCase ( datasets.Metric): '''simple docstring''' def lowercase_ ( self) -> Dict: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://github.com/google-research/bleurt" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence"), "references": datasets.Value("string" , id="sequence"), }) , codebase_urls=["https://github.com/google-research/bleurt"] , reference_urls=["https://github.com/google-research/bleurt", "https://arxiv.org/abs/2004.04696"] , ) def lowercase_ ( self , lowerCAmelCase_) -> List[Any]: """simple docstring""" if self.config_name == "default": logger.warning( "Using default BLEURT-Base checkpoint for sequence maximum length 128. " "You can use a bigger model for better results with e.g.: datasets.load_metric(\'bleurt\', \'bleurt-large-512\').") a_ ='bleurt-base-128' if self.config_name.lower() in CHECKPOINT_URLS: a_ =self.config_name.lower() elif self.config_name.upper() in CHECKPOINT_URLS: a_ =self.config_name.upper() else: raise KeyError( f"""{self.config_name} model not found. You should supply the name of a model checkpoint for bleurt in {CHECKPOINT_URLS.keys()}""") # download the model checkpoint specified by self.config_name and set up the scorer a_ =dl_manager.download_and_extract(CHECKPOINT_URLS[checkpoint_name]) a_ =score.BleurtScorer(os.path.join(UpperCamelCase__ , UpperCamelCase__)) def lowercase_ ( self , lowerCAmelCase_ , lowerCAmelCase_) -> Any: """simple docstring""" a_ =self.scorer.score(references=UpperCamelCase__ , candidates=UpperCamelCase__) return {"scores": scores}
708
'''simple docstring''' from collections.abc import Generator def UpperCAmelCase_ ( ): '''simple docstring''' a_ , a_ =0, 1 while True: a_ , a_ =b, a + b yield b def UpperCAmelCase_ ( lowercase__ = 1_0_0_0 ): '''simple docstring''' a_ =1 a_ =fibonacci_generator() while len(str(next(lowercase__ ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
41
0
"""simple docstring""" import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def lowercase ( a__ : List[str] ) -> str: _UpperCamelCase = [] for line in lines: _UpperCamelCase = re.sub(R'''#.*''' , '''''' , a__ ) # remove comments if line: filtered_lines.append(a__ ) _UpperCamelCase = '''\n'''.join(a__ ) # Make a hash from all this code _UpperCamelCase = full_str.encode('''utf-8''' ) return shaaaa(a__ ).hexdigest() # get importable module names and hash for caching UpperCAmelCase = { """csv""": (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), """json""": (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), """pandas""": (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), """parquet""": (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), """arrow""": (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), """text""": (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), """imagefolder""": (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), """audiofolder""": (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions UpperCAmelCase = { """.csv""": ("""csv""", {}), """.tsv""": ("""csv""", {"""sep""": """\t"""}), """.json""": ("""json""", {}), """.jsonl""": ("""json""", {}), """.parquet""": ("""parquet""", {}), """.arrow""": ("""arrow""", {}), """.txt""": ("""text""", {}), } _EXTENSION_TO_MODULE.update({ext: ("""imagefolder""", {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ("""imagefolder""", {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ("""audiofolder""", {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ("""audiofolder""", {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) UpperCAmelCase = {"""imagefolder""", """audiofolder"""} # Used to filter data files based on extensions given a module name UpperCAmelCase = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append(""".zip""") _MODULE_TO_EXTENSIONS["audiofolder"].append(""".zip""")
420
"""simple docstring""" from math import factorial, pi def lowercase ( a__ : float , a__ : int = 30 ) -> float: if not isinstance(a__ , (int, float) ): raise ValueError('''maclaurin_sin() requires either an int or float for theta''' ) if not isinstance(a__ , a__ ) or accuracy <= 0: raise ValueError('''maclaurin_sin() requires a positive int for accuracy''' ) _UpperCamelCase = float(a__ ) _UpperCamelCase = theta // (2 * pi) theta -= 2 * div * pi return sum( (-1) ** r * theta ** (2 * r + 1) / factorial(2 * r + 1 ) for r in range(a__ ) ) def lowercase ( a__ : float , a__ : int = 30 ) -> float: if not isinstance(a__ , (int, float) ): raise ValueError('''maclaurin_cos() requires either an int or float for theta''' ) if not isinstance(a__ , a__ ) or accuracy <= 0: raise ValueError('''maclaurin_cos() requires a positive int for accuracy''' ) _UpperCamelCase = float(a__ ) _UpperCamelCase = theta // (2 * pi) theta -= 2 * div * pi return sum((-1) ** r * theta ** (2 * r) / factorial(2 * r ) for r in range(a__ ) ) if __name__ == "__main__": import doctest doctest.testmod() print(maclaurin_sin(10)) print(maclaurin_sin(-10)) print(maclaurin_sin(10, 15)) print(maclaurin_sin(-10, 15)) print(maclaurin_cos(5)) print(maclaurin_cos(-5)) print(maclaurin_cos(10, 15)) print(maclaurin_cos(-10, 15))
420
1
'''simple docstring''' import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class snake_case ( unittest.TestCase ): """simple docstring""" @slow def snake_case ( self ): """simple docstring""" lowerCamelCase_ = AutoImageProcessor.from_pretrained("microsoft/dit-base-finetuned-rvlcdip" ) lowerCamelCase_ = AutoModelForImageClassification.from_pretrained("microsoft/dit-base-finetuned-rvlcdip" ) model.to(UpperCamelCase ) from datasets import load_dataset lowerCamelCase_ = load_dataset("nielsr/rvlcdip-demo" ) lowerCamelCase_ = dataset["train"][0]["image"].convert("RGB" ) lowerCamelCase_ = image_processor(UpperCamelCase , return_tensors="pt" ).to(UpperCamelCase ) # forward pass with torch.no_grad(): lowerCamelCase_ = model(**UpperCamelCase ) lowerCamelCase_ = outputs.logits lowerCamelCase_ = torch.Size((1, 16) ) self.assertEqual(logits.shape , UpperCamelCase ) lowerCamelCase_ = torch.tensor( [-0.4_158, -0.4_092, -0.4_347] , device=UpperCamelCase , dtype=torch.float , ) self.assertTrue(torch.allclose(logits[0, :3] , UpperCamelCase , atol=1e-4 ) )
701
'''simple docstring''' def __snake_case ( UpperCAmelCase_ : int = 100 ): lowerCamelCase_ = n * (n + 1) * (2 * n + 1) / 6 lowerCamelCase_ = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) if __name__ == "__main__": print(f'''{solution() = }''')
445
0
'''simple docstring''' import argparse import random import joblib import numpy as np import torch from igf.igf import ( SecondaryLearner, collect_objective_set, compute_perplexity, generate_datasets, load_gpta, recopy_gpta, set_seed, train_secondary_learner, ) from torch.utils.data import DataLoader, RandomSampler from transformers import GPTaLMHeadModel def a_ ( lowerCamelCase : Any=32 , lowerCamelCase : Optional[int]=10 , lowerCamelCase : Any=100 , lowerCamelCase : Union[str, Any]=1026 , lowerCamelCase : Dict=True , lowerCamelCase : Optional[int]="data/tokenized_stories_train_wikitext103.jbl" , lowerCamelCase : Optional[int]="igf_context_pairs.jbl" , ): set_seed(3 ) # generate train_data and objective_set lowerCAmelCase , lowerCAmelCase = generate_datasets( lowerCamelCase , lowerCamelCase , number=lowerCamelCase , min_len=1026 , trim=lowerCamelCase ) # keeps model same across runs set_seed(4 ) # model, lm_optimizer, lm_scheduler = recopy_gpt2(model, device, max_steps) # store original model weights # can we train on GPU? lowerCAmelCase = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu' ) # load pretrained model lowerCAmelCase = load_gpta('gpt2' ).to(lowerCamelCase ) print('computing perplexity on objective set' ) lowerCAmelCase = compute_perplexity(lowerCamelCase , lowerCamelCase , lowerCamelCase ).item() print('perplexity on objective set:' , lowerCamelCase ) # collect igf pairs and save to file demo.jbl collect_objective_set(lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) # clean up, delete model and data we don't need anymore del model, train_data, objective_set torch.cuda.empty_cache() def a_ ( lowerCamelCase : Dict , lowerCamelCase : Optional[int]=15 , lowerCamelCase : Optional[Any]=128 , lowerCamelCase : List[Any]=100 , lowerCamelCase : Tuple="igf_model.pt" , ): set_seed(42 ) # Load pre-trained model lowerCAmelCase = GPTaLMHeadModel.from_pretrained('gpt2' ) # Initialize secondary learner to use embedding weights of model lowerCAmelCase = SecondaryLearner(lowerCamelCase ) # Train secondary learner lowerCAmelCase = train_secondary_learner( lowerCamelCase , lowerCamelCase , max_epochs=lowerCamelCase , batch_size=lowerCamelCase , eval_freq=100 , igf_model_path=lowerCamelCase , ) del model, secondary_learner_train_data torch.cuda.empty_cache() return secondary_learner def a_ ( lowerCamelCase : List[str] , lowerCamelCase : str , lowerCamelCase : Tuple , lowerCamelCase : Optional[Any]=32 , lowerCamelCase : str=1000 , lowerCamelCase : int=16 , lowerCamelCase : Optional[int]=1.0 , lowerCamelCase : Tuple=recopy_gpta , lowerCamelCase : Optional[int]=None , lowerCamelCase : List[str]=10 , lowerCamelCase : Tuple="gpt2_finetuned.pt" , ): lowerCAmelCase = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu' ) lowerCAmelCase = RandomSampler(lowerCamelCase ) lowerCAmelCase = DataLoader(lowerCamelCase , sampler=lowerCamelCase ) lowerCAmelCase = max_steps // (len(lowerCamelCase )) + 1 lowerCAmelCase = 0 lowerCAmelCase = torch.zeros((1, context_len) , dtype=torch.long , device=lowerCamelCase ) lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = recopy_model(lowerCamelCase , lowerCamelCase , lowerCamelCase ) model.train() if secondary_learner is not None: secondary_learner.to(lowerCamelCase ) secondary_learner.eval() lowerCAmelCase = [] lowerCAmelCase = 0 lowerCAmelCase = [] lowerCAmelCase = [] # Compute the performance of the transformer model at the beginning lowerCAmelCase = compute_perplexity(lowerCamelCase , lowerCamelCase , lowerCamelCase ) test_perps.append(lowerCamelCase ) print('Test perplexity, step' , lowerCamelCase , ':' , lowerCamelCase ) for epoch in range(int(lowerCamelCase ) ): for step, example in enumerate(lowerCamelCase ): torch.cuda.empty_cache() lowerCAmelCase = random.randint(0 , example.size(2 ) - context_len - 1 ) lowerCAmelCase = example[0, 0, start : start + context_len] lm_optimizer.zero_grad() lowerCAmelCase = model(lowerCamelCase , labels=lowerCamelCase ) lowerCAmelCase = True if secondary_learner is not None: lowerCAmelCase = secondary_learner.forward( torch.tensor(lowerCamelCase , dtype=torch.long , device=lowerCamelCase ).unsqueeze(0 ) )[0].item() observed_qs.append(float(lowerCamelCase ) ) # Here we implement the simple non-constant threshold for the predicted IG(X) value # We will decay the selectivity of our secondary learner filter from # 1 standard deviation above average to 1 below average after 10 batches. if global_step == 10: lowerCAmelCase = -1 if predicted_q < threshold: lowerCAmelCase = False # If we passed the filter, add the context to the batch! if do_backprop: contexts.append(np.array(context.cpu() ) ) lowerCAmelCase = outputs[0] lm_loss.backward() examples += 1 del outputs # Once the batch is filled with enough contexts, backprop on the batch. if examples == batch_size: torch.cuda.empty_cache() lowerCAmelCase = 0 # Do LM backprop torch.nn.utils.clip_grad_norm_(model.parameters() , 3.0 ) lm_optimizer.step() lm_scheduler.step() # Update learning rate schedule global_step += 1 # Compute the performance of the transformer model at this batch if global_step % eval_interval == 0: lowerCAmelCase = compute_perplexity(lowerCamelCase , lowerCamelCase , lowerCamelCase ) test_perps.append(lowerCamelCase ) print('Test perplexity, step' , lowerCamelCase , ':' , lowerCamelCase ) # Break out of the loop after 60 batches if max_steps > 0 and global_step > 60: break if max_steps > 0 and global_step > 60: break # save finetuned transformer model torch.save(model.state_dict() , lowerCamelCase ) torch.cuda.empty_cache() # Do some cleaning up so we can reinitialize for the next run of this function del lm_optimizer del lm_scheduler return model def a_ ( ): lowerCAmelCase = argparse.ArgumentParser(description='Fine-tune a transformer model with IGF on a language modeling task' ) # Required parameters parser.add_argument( '--data_dir' , default=lowerCamelCase , type=lowerCamelCase , required=lowerCamelCase , help='The input data dir. Should contain data files for WikiText.' , ) parser.add_argument( '--model_name_or_path' , default=lowerCamelCase , type=lowerCamelCase , required=lowerCamelCase , help='Path to pretrained model or model identifier from huggingface.co/models' , ) parser.add_argument( '--data_file' , type=lowerCamelCase , default=lowerCamelCase , help=( 'A jbl file containing tokenized data which can be split as objective dataset, ' 'train_dataset and test_dataset.' ) , ) parser.add_argument( '--igf_data_file' , type=lowerCamelCase , default=lowerCamelCase , help='A jbl file containing the context and information gain pairs to train secondary learner.' , ) parser.add_argument( '--output_dir' , default=lowerCamelCase , type=lowerCamelCase , required=lowerCamelCase , help='The output directory where the final fine-tuned model is stored.' , ) parser.add_argument( '--tokenizer_name' , default=lowerCamelCase , type=lowerCamelCase , help='Pretrained tokenizer name or path if not the same as model_name' , ) parser.add_argument('--seed' , type=lowerCamelCase , default=lowerCamelCase , help='A seed for reproducible training.' ) parser.add_argument( '--context_len' , default=32 , type=lowerCamelCase , help=( 'The maximum total input sequence length after tokenization. Sequences longer ' 'than this will be truncated, sequences shorter will be padded.' ) , ) parser.add_argument( '--size_objective_set' , default=100 , type=lowerCamelCase , help='number of articles that are long enough to be used as our objective set' , ) parser.add_argument( '--eval_freq' , default=100 , type=lowerCamelCase , help='secondary model evaluation is triggered at eval_freq' ) parser.add_argument('--max_steps' , default=1000 , type=lowerCamelCase , help='To calculate training epochs' ) parser.add_argument( '--secondary_learner_batch_size' , default=128 , type=lowerCamelCase , help='batch size of training data for secondary learner' , ) parser.add_argument( '--batch_size' , default=16 , type=lowerCamelCase , help='batch size of training data of language model(gpt2) ' ) parser.add_argument( '--eval_interval' , default=10 , type=lowerCamelCase , help=( 'decay the selectivity of our secondary learner filter from' '1 standard deviation above average to 1 below average after 10 batches' ) , ) parser.add_argument( '--number' , default=100 , type=lowerCamelCase , help='The number of examples split to be used as objective_set/test_data' ) parser.add_argument( '--min_len' , default=1026 , type=lowerCamelCase , help='The minimum length of the article to be used as objective set' ) parser.add_argument( '--secondary_learner_max_epochs' , default=15 , type=lowerCamelCase , help='number of epochs to train secondary learner' ) parser.add_argument('--trim' , default=lowerCamelCase , type=lowerCamelCase , help='truncate the example if it exceeds context length' ) parser.add_argument( '--threshold' , default=1.0 , type=lowerCamelCase , help=( 'The threshold value used by secondary learner to filter the train_data and allow only' ' informative data as input to the model' ) , ) parser.add_argument('--finetuned_model_name' , default='gpt2_finetuned.pt' , type=lowerCamelCase , help='finetuned_model_name' ) parser.add_argument( '--recopy_model' , default=lowerCamelCase , type=lowerCamelCase , help='Reset the model to the original pretrained GPT-2 weights after each iteration' , ) # function calls # Collecting *n* pairs of context and information gain(X, IG(X)) for training the secondary learner generate_n_pairs( context_len=32 , max_steps=10 , size_objective_set=100 , min_len=1026 , trim=lowerCamelCase , data_file='data/tokenized_stories_train_wikitext103.jbl' , igf_data_file='igf_context_pairs.jbl' , ) # Load train data for secondary learner lowerCAmelCase = joblib.load('data/IGF_values.jbl' ) # Train secondary learner lowerCAmelCase = training_secondary_learner( lowerCamelCase , secondary_learner_max_epochs=15 , secondary_learner_batch_size=128 , eval_freq=100 , igf_model_path='igf_model.pt' , ) # load pretrained gpt2 model lowerCAmelCase = GPTaLMHeadModel.from_pretrained('gpt2' ) set_seed(42 ) # Generate train and test data to train and evaluate gpt2 model lowerCAmelCase , lowerCAmelCase = generate_datasets( context_len=32 , file='data/tokenized_stories_train_wikitext103.jbl' , number=100 , min_len=1026 , trim=lowerCamelCase ) # fine-tuning of the gpt2 model using igf (Information Gain Filtration) finetune( lowerCamelCase , lowerCamelCase , lowerCamelCase , context_len=32 , max_steps=1000 , batch_size=16 , threshold=1.0 , recopy_model=lowerCamelCase , secondary_learner=lowerCamelCase , eval_interval=10 , finetuned_model_name='gpt2_finetuned.pt' , ) if __name__ == "__main__": main()
133
'''simple docstring''' import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets __snake_case ="""\ @inproceedings{pillutla-etal:mauve:neurips2021, title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers}, author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid}, booktitle = {NeurIPS}, year = {2021} } """ __snake_case ="""\ MAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure. MAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences. For details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021). This metrics is a wrapper around the official implementation of MAUVE: https://github.com/krishnap25/mauve """ __snake_case =""" Calculates MAUVE scores between two lists of generated text and reference text. Args: predictions: list of generated text to score. Each predictions should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. Optional Args: num_buckets: the size of the histogram to quantize P and Q. Options: 'auto' (default) or an integer pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1 kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9 kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5 kmeans_max_iter: maximum number of k-means iterations. Default 500 featurize_model_name: name of the model from which features are obtained. Default 'gpt2-large' Use one of ['gpt2', 'gpt2-medium', 'gpt2-large', 'gpt2-xl']. device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU max_text_length: maximum number of tokens to consider. Default 1024 divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25 mauve_scaling_factor: \"c\" from the paper. Default 5. verbose: If True (default), print running time updates seed: random seed to initialize k-means cluster assignments. Returns: mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer, frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer, divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve, p_hist: a discrete distribution, which is a quantized version of the text distribution p_text, q_hist: same as above, but with q_text. Examples: >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest >>> import datasets >>> mauve = datasets.load_metric('mauve') >>> predictions = [\"hello there\", \"general kenobi\"] >>> references = [\"hello there\", \"general kenobi\"] >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP >>> print(out.mauve) # doctest: +SKIP 1.0 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase_ ( datasets.Metric ): def __UpperCAmelCase ( self : Optional[int] ) -> Any: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='https://github.com/krishnap25/mauve' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/krishnap25/mauve'] , reference_urls=[ 'https://arxiv.org/abs/2102.01454', 'https://github.com/krishnap25/mauve', ] , ) def __UpperCAmelCase ( self : List[Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : Union[str, Any]=None , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : Dict="auto" , UpperCAmelCase__ : Union[str, Any]=-1 , UpperCAmelCase__ : int=0.9 , UpperCAmelCase__ : Any=5 , UpperCAmelCase__ : Optional[int]=5_0_0 , UpperCAmelCase__ : List[str]="gpt2-large" , UpperCAmelCase__ : Any=-1 , UpperCAmelCase__ : int=1_0_2_4 , UpperCAmelCase__ : Union[str, Any]=2_5 , UpperCAmelCase__ : Tuple=5 , UpperCAmelCase__ : Tuple=True , UpperCAmelCase__ : Union[str, Any]=2_5 , ) -> Tuple: lowerCAmelCase = compute_mauve( p_text=UpperCAmelCase__ , q_text=UpperCAmelCase__ , p_features=UpperCAmelCase__ , q_features=UpperCAmelCase__ , p_tokens=UpperCAmelCase__ , q_tokens=UpperCAmelCase__ , num_buckets=UpperCAmelCase__ , pca_max_data=UpperCAmelCase__ , kmeans_explained_var=UpperCAmelCase__ , kmeans_num_redo=UpperCAmelCase__ , kmeans_max_iter=UpperCAmelCase__ , featurize_model_name=UpperCAmelCase__ , device_id=UpperCAmelCase__ , max_text_length=UpperCAmelCase__ , divergence_curve_discretization_size=UpperCAmelCase__ , mauve_scaling_factor=UpperCAmelCase__ , verbose=UpperCAmelCase__ , seed=UpperCAmelCase__ , ) return out
133
1
import numpy as np from cva import COLOR_BGR2GRAY, cvtColor, imread from numpy import array, uinta from PIL import Image from digital_image_processing import change_contrast as cc from digital_image_processing import convert_to_negative as cn from digital_image_processing import sepia as sp from digital_image_processing.dithering import burkes as bs from digital_image_processing.edge_detection import canny from digital_image_processing.filters import convolve as conv from digital_image_processing.filters import gaussian_filter as gg from digital_image_processing.filters import local_binary_pattern as lbp from digital_image_processing.filters import median_filter as med from digital_image_processing.filters import sobel_filter as sob from digital_image_processing.resize import resize as rs __lowerCAmelCase : Union[str, Any] = imread(R'digital_image_processing/image_data/lena_small.jpg') __lowerCAmelCase : str = cvtColor(img, COLOR_BGR2GRAY) def a_ ()-> List[str]: snake_case: Union[str, Any] = cn.convert_to_negative(_lowerCAmelCase ) # assert negative_img array for at least one True assert negative_img.any() def a_ ()-> Tuple: with Image.open("""digital_image_processing/image_data/lena_small.jpg""" ) as img: # Work around assertion for response assert str(cc.change_contrast(_lowerCAmelCase , 110 ) ).startswith( """<PIL.Image.Image image mode=RGB size=100x100 at""" ) def a_ ()-> Union[str, Any]: snake_case: Any = canny.gen_gaussian_kernel(9 , sigma=1.4 ) # Assert ambiguous array assert resp.all() def a_ ()-> str: snake_case: Any = imread("""digital_image_processing/image_data/lena_small.jpg""" , 0 ) # assert ambiguous array for all == True assert canny_img.all() snake_case: str = canny.canny(_lowerCAmelCase ) # assert canny array for at least one True assert canny_array.any() def a_ ()-> Optional[Any]: assert gg.gaussian_filter(_lowerCAmelCase , 5 , sigma=0.9 ).all() def a_ ()-> Optional[int]: snake_case: List[str] = array([[0.25, 0.5, 0.25], [0.5, -3, 0.5], [0.25, 0.5, 0.25]] ) snake_case: Dict = conv.img_convolve(_lowerCAmelCase , _lowerCAmelCase ).astype(_lowerCAmelCase ) assert res.any() def a_ ()-> str: assert med.median_filter(_lowerCAmelCase , 3 ).any() def a_ ()-> Any: snake_case: List[Any] = sob.sobel_filter(_lowerCAmelCase ) assert grad.any() and theta.any() def a_ ()-> List[Any]: snake_case: List[Any] = sp.make_sepia(_lowerCAmelCase , 20 ) assert sepia.all() def a_ (_lowerCAmelCase : str = "digital_image_processing/image_data/lena_small.jpg" )-> Tuple: snake_case: str = bs.Burkes(imread(_lowerCAmelCase , 1 ) , 120 ) burkes.process() assert burkes.output_img.any() def a_ (_lowerCAmelCase : Tuple = "digital_image_processing/image_data/lena_small.jpg" , )-> Any: snake_case: List[str] = rs.NearestNeighbour(imread(_lowerCAmelCase , 1 ) , 400 , 200 ) nn.process() assert nn.output.any() def a_ ()-> str: snake_case: Optional[int] = "digital_image_processing/image_data/lena.jpg" # Reading the image and converting it to grayscale. snake_case: Tuple = imread(_lowerCAmelCase , 0 ) # Test for get_neighbors_pixel function() return not None snake_case: Union[str, Any] = 0 snake_case: Any = 0 snake_case: Dict = image[x_coordinate][y_coordinate] snake_case: Optional[Any] = lbp.get_neighbors_pixel( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) assert neighbors_pixels is not None # Test for local_binary_pattern function() # Create a numpy array as the same height and width of read image snake_case: List[str] = np.zeros((image.shape[0], image.shape[1]) ) # Iterating through the image and calculating the local binary pattern value # for each pixel. for i in range(0 , image.shape[0] ): for j in range(0 , image.shape[1] ): snake_case: Optional[int] = lbp.local_binary_value(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) assert lbp_image.any()
719
import argparse import torch from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_from_original_stable_diffusion_ckpt if __name__ == "__main__": __lowerCAmelCase : Any = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str, required=True, help='Path to the checkpoint to convert.' ) # !wget https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml parser.add_argument( '--original_config_file', default=None, type=str, help='The YAML config file corresponding to the original architecture.', ) parser.add_argument( '--num_in_channels', default=None, type=int, help='The number of input channels. If `None` number of input channels will be automatically inferred.', ) parser.add_argument( '--scheduler_type', default='pndm', type=str, help='Type of scheduler to use. Should be one of [\'pndm\', \'lms\', \'ddim\', \'euler\', \'euler-ancestral\', \'dpm\']', ) parser.add_argument( '--pipeline_type', default=None, type=str, help=( 'The pipeline type. One of \'FrozenOpenCLIPEmbedder\', \'FrozenCLIPEmbedder\', \'PaintByExample\'' '. If `None` pipeline will be automatically inferred.' ), ) parser.add_argument( '--image_size', default=None, type=int, help=( 'The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2' ' Base. Use 768 for Stable Diffusion v2.' ), ) parser.add_argument( '--prediction_type', default=None, type=str, help=( 'The prediction type that the model was trained on. Use \'epsilon\' for Stable Diffusion v1.X and Stable' ' Diffusion v2 Base. Use \'v_prediction\' for Stable Diffusion v2.' ), ) parser.add_argument( '--extract_ema', action='store_true', help=( 'Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights' ' or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield' ' higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.' ), ) parser.add_argument( '--upcast_attention', action='store_true', help=( 'Whether the attention computation should always be upcasted. This is necessary when running stable' ' diffusion 2.1.' ), ) parser.add_argument( '--from_safetensors', action='store_true', help='If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.', ) parser.add_argument( '--to_safetensors', action='store_true', help='Whether to store pipeline in safetensors format or not.', ) parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the output model.') parser.add_argument('--device', type=str, help='Device to use (e.g. cpu, cuda:0, cuda:1, etc.)') parser.add_argument( '--stable_unclip', type=str, default=None, required=False, help='Set if this is a stable unCLIP model. One of \'txt2img\' or \'img2img\'.', ) parser.add_argument( '--stable_unclip_prior', type=str, default=None, required=False, help='Set if this is a stable unCLIP txt2img model. Selects which prior to use. If `--stable_unclip` is set to `txt2img`, the karlo prior (https://huggingface.co/kakaobrain/karlo-v1-alpha/tree/main/prior) is selected by default.', ) parser.add_argument( '--clip_stats_path', type=str, help='Path to the clip stats file. Only required if the stable unclip model\'s config specifies `model.params.noise_aug_config.params.clip_stats_path`.', required=False, ) parser.add_argument( '--controlnet', action='store_true', default=None, help='Set flag if this is a controlnet checkpoint.' ) parser.add_argument('--half', action='store_true', help='Save weights in half precision.') parser.add_argument( '--vae_path', type=str, default=None, required=False, help='Set to a path, hub id to an already converted vae to not convert it again.', ) __lowerCAmelCase : str = parser.parse_args() __lowerCAmelCase : Union[str, Any] = download_from_original_stable_diffusion_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, prediction_type=args.prediction_type, model_type=args.pipeline_type, extract_ema=args.extract_ema, scheduler_type=args.scheduler_type, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, stable_unclip=args.stable_unclip, stable_unclip_prior=args.stable_unclip_prior, clip_stats_path=args.clip_stats_path, controlnet=args.controlnet, vae_path=args.vae_path, ) if args.half: pipe.to(torch_dtype=torch.floataa) if args.controlnet: # only save the controlnet model pipe.controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors) else: pipe.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
164
0
"""simple docstring""" from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { """EleutherAI/gpt-j-6B""": """https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json""", # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class lowerCAmelCase__ ( lowercase ): '''simple docstring''' lowerCamelCase__ = """gptj""" lowerCamelCase__ = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase=50400 , lowercase=2048 , lowercase=4096 , lowercase=28 , lowercase=16 , lowercase=64 , lowercase=None , lowercase="gelu_new" , lowercase=0.0 , lowercase=0.0 , lowercase=0.0 , lowercase=1E-5 , lowercase=0.02 , lowercase=True , lowercase=50256 , lowercase=50256 , lowercase=False , **lowercase , ): _lowerCamelCase : Dict = vocab_size _lowerCamelCase : str = n_positions _lowerCamelCase : Dict = n_embd _lowerCamelCase : Union[str, Any] = n_layer _lowerCamelCase : Union[str, Any] = n_head _lowerCamelCase : Tuple = n_inner _lowerCamelCase : Dict = rotary_dim _lowerCamelCase : List[str] = activation_function _lowerCamelCase : Any = resid_pdrop _lowerCamelCase : Optional[int] = embd_pdrop _lowerCamelCase : Union[str, Any] = attn_pdrop _lowerCamelCase : Dict = layer_norm_epsilon _lowerCamelCase : List[Any] = initializer_range _lowerCamelCase : Optional[int] = use_cache _lowerCamelCase : Any = bos_token_id _lowerCamelCase : Optional[Any] = eos_token_id super().__init__( bos_token_id=lowercase , eos_token_id=lowercase , tie_word_embeddings=lowercase , **lowercase ) class lowerCAmelCase__ ( lowercase ): '''simple docstring''' def __init__( self , lowercase , lowercase = "default" , lowercase = None , lowercase = False , ): super().__init__(lowercase , task=lowercase , patching_specs=lowercase , use_past=lowercase ) if not getattr(self._config , 'pad_token_id' , lowercase ): # TODO: how to do that better? _lowerCamelCase : Union[str, Any] = 0 @property def A_ ( self ): _lowerCamelCase : Union[str, Any] = OrderedDict({'input_ids': {0: 'batch', 1: 'sequence'}} ) if self.use_past: self.fill_with_past_key_values_(lowercase , direction='inputs' ) _lowerCamelCase : List[str] = {0: 'batch', 1: 'past_sequence + sequence'} else: _lowerCamelCase : Union[str, Any] = {0: 'batch', 1: 'sequence'} return common_inputs @property def A_ ( self ): return self._config.n_layer @property def A_ ( self ): return self._config.n_head def A_ ( self , lowercase , lowercase = -1 , lowercase = -1 , lowercase = False , lowercase = None , ): _lowerCamelCase : List[str] = super(lowercase , self ).generate_dummy_inputs( lowercase , batch_size=lowercase , seq_length=lowercase , is_pair=lowercase , framework=lowercase ) # We need to order the input in the way they appears in the forward() _lowerCamelCase : List[Any] = OrderedDict({'input_ids': common_inputs['input_ids']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('Cannot generate dummy past_keys inputs without PyTorch installed.' ) else: import torch _lowerCamelCase, _lowerCamelCase : int = common_inputs['input_ids'].shape # Not using the same length for past_key_values _lowerCamelCase : Optional[int] = seqlen + 2 _lowerCamelCase : str = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _lowerCamelCase : Dict = [ (torch.zeros(lowercase ), torch.zeros(lowercase )) for _ in range(self.num_layers ) ] _lowerCamelCase : List[str] = common_inputs['attention_mask'] if self.use_past: _lowerCamelCase : Dict = ordered_inputs['attention_mask'].dtype _lowerCamelCase : Tuple = torch.cat( [ordered_inputs['attention_mask'], torch.ones(lowercase , lowercase , dtype=lowercase )] , dim=1 ) return ordered_inputs @property def A_ ( self ): return 13
630
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowerCAmelCase__ ( lowercase, lowercase, lowercase ): '''simple docstring''' lowerCamelCase__ = [R"""h\.\d+\.attn\.bias""", R"""h\.\d+\.attn\.masked_bias"""] @register_to_config def __init__( self , lowercase , lowercase , lowercase = None , lowercase = 50257 , lowercase = 1024 , lowercase = 768 , lowercase = 12 , lowercase = 12 , lowercase = None , lowercase = "gelu_new" , lowercase = 0.1 , lowercase = 0.1 , lowercase = 0.1 , lowercase = 1E-5 , lowercase = 0.02 , lowercase = True , lowercase = True , lowercase = False , lowercase = False , ): super().__init__() _lowerCamelCase : Tuple = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F'''`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and''' F''' `n_embd`: {n_embd} are not equal.''' ) _lowerCamelCase : Optional[Any] = prefix_inner_dim _lowerCamelCase : Tuple = prefix_hidden_dim _lowerCamelCase : Dict = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCamelCase : Tuple = ( nn.Linear(self.prefix_hidden_dim , lowercase ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCamelCase : Any = GPTaConfig( vocab_size=lowercase , n_positions=lowercase , n_embd=lowercase , n_layer=lowercase , n_head=lowercase , n_inner=lowercase , activation_function=lowercase , resid_pdrop=lowercase , embd_pdrop=lowercase , attn_pdrop=lowercase , layer_norm_epsilon=lowercase , initializer_range=lowercase , scale_attn_weights=lowercase , use_cache=lowercase , scale_attn_by_inverse_layer_idx=lowercase , reorder_and_upcast_attn=lowercase , ) _lowerCamelCase : Dict = GPTaLMHeadModel(lowercase ) def A_ ( self , lowercase , lowercase , lowercase = None , lowercase = None , ): _lowerCamelCase : List[Any] = self.transformer.transformer.wte(lowercase ) _lowerCamelCase : str = self.encode_prefix(lowercase ) _lowerCamelCase : int = self.decode_prefix(lowercase ) _lowerCamelCase : Optional[int] = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: _lowerCamelCase : Any = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) _lowerCamelCase : int = torch.cat((dummy_token, input_ids) , dim=1 ) _lowerCamelCase : List[Any] = self.transformer(inputs_embeds=lowercase , labels=lowercase , attention_mask=lowercase ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def A_ ( self , lowercase , lowercase ): return torch.zeros(lowercase , self.prefix_length , dtype=torch.intaa , device=lowercase ) def A_ ( self , lowercase ): return self.encode_prefix(lowercase ) @torch.no_grad() def A_ ( self , lowercase , lowercase , lowercase ): _lowerCamelCase : Dict = torch.split(lowercase , 1 , dim=0 ) _lowerCamelCase : Any = [] _lowerCamelCase : str = [] for feature in features: _lowerCamelCase : Any = self.decode_prefix(feature.to(lowercase ) ) # back to the clip feature # Only support beam search for now _lowerCamelCase, _lowerCamelCase : Dict = self.generate_beam( input_embeds=lowercase , device=lowercase , eos_token_id=lowercase ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) _lowerCamelCase : List[Any] = torch.stack(lowercase ) _lowerCamelCase : Optional[int] = torch.stack(lowercase ) return generated_tokens, generated_seq_lengths @torch.no_grad() def A_ ( self , lowercase=None , lowercase=None , lowercase=None , lowercase = 5 , lowercase = 67 , lowercase = 1.0 , lowercase = None , ): _lowerCamelCase : Any = eos_token_id _lowerCamelCase : List[str] = None _lowerCamelCase : List[Any] = None _lowerCamelCase : List[str] = torch.ones(lowercase , device=lowercase , dtype=torch.int ) _lowerCamelCase : Optional[int] = torch.zeros(lowercase , device=lowercase , dtype=torch.bool ) if input_embeds is not None: _lowerCamelCase : Optional[int] = input_embeds else: _lowerCamelCase : Union[str, Any] = self.transformer.transformer.wte(lowercase ) for i in range(lowercase ): _lowerCamelCase : Any = self.transformer(inputs_embeds=lowercase ) _lowerCamelCase : Optional[Any] = outputs.logits _lowerCamelCase : List[Any] = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) _lowerCamelCase : Optional[int] = logits.softmax(-1 ).log() if scores is None: _lowerCamelCase, _lowerCamelCase : Union[str, Any] = logits.topk(lowercase , -1 ) _lowerCamelCase : Optional[int] = generated.expand(lowercase , *generated.shape[1:] ) _lowerCamelCase, _lowerCamelCase : Tuple = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: _lowerCamelCase : List[str] = next_tokens else: _lowerCamelCase : Tuple = tokens.expand(lowercase , *tokens.shape[1:] ) _lowerCamelCase : int = torch.cat((tokens, next_tokens) , dim=1 ) else: _lowerCamelCase : Union[str, Any] = -float(np.inf ) _lowerCamelCase : Union[str, Any] = 0 _lowerCamelCase : List[Any] = scores[:, None] + logits seq_lengths[~is_stopped] += 1 _lowerCamelCase : List[Any] = scores_sum / seq_lengths[:, None] _lowerCamelCase, _lowerCamelCase : Optional[Any] = scores_sum_average.view(-1 ).topk(lowercase , -1 ) _lowerCamelCase : Union[str, Any] = next_tokens // scores_sum.shape[1] _lowerCamelCase : Optional[Any] = seq_lengths[next_tokens_source] _lowerCamelCase : Dict = next_tokens % scores_sum.shape[1] _lowerCamelCase : Optional[int] = next_tokens.unsqueeze(1 ) _lowerCamelCase : Tuple = tokens[next_tokens_source] _lowerCamelCase : str = torch.cat((tokens, next_tokens) , dim=1 ) _lowerCamelCase : Optional[int] = generated[next_tokens_source] _lowerCamelCase : Dict = scores_sum_average * seq_lengths _lowerCamelCase : List[Any] = is_stopped[next_tokens_source] _lowerCamelCase : Union[str, Any] = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) _lowerCamelCase : Union[str, Any] = torch.cat((generated, next_token_embed) , dim=1 ) _lowerCamelCase : Optional[Any] = is_stopped + next_tokens.eq(lowercase ).squeeze() if is_stopped.all(): break _lowerCamelCase : Optional[int] = scores / seq_lengths _lowerCamelCase : Optional[int] = scores.argsort(descending=lowercase ) # tokens tensors are already padded to max_seq_length _lowerCamelCase : List[Any] = [tokens[i] for i in order] _lowerCamelCase : Dict = torch.stack(lowercase , dim=0 ) _lowerCamelCase : Optional[int] = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
630
1
import math import sys def lowerCAmelCase__ ( lowerCamelCase_ : str): '''simple docstring''' lowerCAmelCase__ : Tuple = '''''' try: with open(lowerCamelCase_ ,'''rb''') as binary_file: lowerCAmelCase__ : Optional[Any] = binary_file.read() for dat in data: lowerCAmelCase__ : Optional[Any] = f"""{dat:08b}""" result += curr_byte return result except OSError: print('''File not accessible''') sys.exit() def lowerCAmelCase__ ( lowerCamelCase_ : str): '''simple docstring''' lowerCAmelCase__ : str = {'''0''': '''0''', '''1''': '''1'''} lowerCAmelCase__ : Union[str, Any] = '''''', '''''' lowerCAmelCase__ : Tuple = len(lowerCamelCase_) for i in range(len(lowerCamelCase_)): curr_string += data_bits[i] if curr_string not in lexicon: continue lowerCAmelCase__ : Any = lexicon[curr_string] result += last_match_id lowerCAmelCase__ : str = last_match_id + '''0''' if math.loga(lowerCamelCase_).is_integer(): lowerCAmelCase__ : Union[str, Any] = {} for curr_key in list(lowerCamelCase_): lowerCAmelCase__ : Optional[Any] = lexicon.pop(lowerCamelCase_) lowerCAmelCase__ : Any = new_lex lowerCAmelCase__ : List[Any] = last_match_id + '''1''' index += 1 lowerCAmelCase__ : Tuple = '''''' return result def lowerCAmelCase__ ( lowerCamelCase_ : str ,lowerCamelCase_ : str): '''simple docstring''' lowerCAmelCase__ : Optional[Any] = 8 try: with open(lowerCamelCase_ ,'''wb''') as opened_file: lowerCAmelCase__ : Union[str, Any] = [ to_write[i : i + byte_length] for i in range(0 ,len(lowerCamelCase_) ,lowerCamelCase_) ] if len(result_byte_array[-1]) % byte_length == 0: result_byte_array.append('''10000000''') else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1]) - 1 ) for elem in result_byte_array[:-1]: opened_file.write(int(lowerCamelCase_ ,2).to_bytes(1 ,byteorder='''big''')) except OSError: print('''File not accessible''') sys.exit() def lowerCAmelCase__ ( lowerCamelCase_ : str): '''simple docstring''' lowerCAmelCase__ : str = 0 for letter in data_bits: if letter == "1": break counter += 1 lowerCAmelCase__ : Any = data_bits[counter:] lowerCAmelCase__ : Tuple = data_bits[counter + 1 :] return data_bits def lowerCAmelCase__ ( lowerCamelCase_ : str ,lowerCamelCase_ : str): '''simple docstring''' lowerCAmelCase__ : List[str] = read_file_binary(lowerCamelCase_) lowerCAmelCase__ : Any = remove_prefix(lowerCamelCase_) lowerCAmelCase__ : Any = decompress_data(lowerCamelCase_) write_file_binary(lowerCamelCase_ ,lowerCamelCase_) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
716
def lowerCAmelCase__ ( lowerCamelCase_ : Any ,lowerCamelCase_ : Optional[Any]): '''simple docstring''' lowerCAmelCase__ : str = [0 for i in range(r + 1)] # nc0 = 1 lowerCAmelCase__ : Tuple = 1 for i in range(1 ,n + 1): # to compute current row from previous row. lowerCAmelCase__ : Optional[int] = min(lowerCamelCase_ ,lowerCamelCase_) while j > 0: c[j] += c[j - 1] j -= 1 return c[r] print(binomial_coefficient(n=1_0, r=5))
90
0
'''simple docstring''' from random import randint from tempfile import TemporaryFile import numpy as np def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Optional[int] ,_UpperCAmelCase : Optional[int] ,_UpperCAmelCase : Optional[Any] ) -> str: _a : Union[str, Any] =0 if start < end: _a : List[str] =randint(_UpperCAmelCase ,_UpperCAmelCase ) _a : str =a[end] _a : int =a[pivot] _a : Dict =temp _a , _a : str =_in_place_partition(_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ) count += _in_place_quick_sort(_UpperCAmelCase ,_UpperCAmelCase ,p - 1 ) count += _in_place_quick_sort(_UpperCAmelCase ,p + 1 ,_UpperCAmelCase ) return count def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : Dict ,_UpperCAmelCase : Any ,_UpperCAmelCase : Optional[Any] ) -> str: _a : List[str] =0 _a : str =randint(_UpperCAmelCase ,_UpperCAmelCase ) _a : Any =a[end] _a : List[str] =a[pivot] _a : Optional[Any] =temp _a : Optional[int] =start - 1 for index in range(_UpperCAmelCase ,_UpperCAmelCase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value _a : Optional[int] =new_pivot_index + 1 _a : Tuple =a[new_pivot_index] _a : Tuple =a[index] _a : Optional[int] =temp _a : Union[str, Any] =a[new_pivot_index + 1] _a : str =a[end] _a : Dict =temp return new_pivot_index + 1, count A__: List[Any] = TemporaryFile() A__: List[Any] = 100 # 1000 elements are to be sorted A__ , A__: str = 0, 1 # mean and standard deviation A__: str = np.random.normal(mu, sigma, p) np.save(outfile, X) print('''The array is''') print(X) outfile.seek(0) # using the same array A__: Dict = np.load(outfile) A__: Dict = len(M) - 1 A__: Union[str, Any] = _in_place_quick_sort(M, 0, r) print( '''No of Comparisons for 100 elements selected from a standard normal distribution''' '''is :''' ) print(z)
694
'''simple docstring''' def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : int = 4000000 ) -> int: _a : Optional[Any] =[] _a , _a : Union[str, Any] =0, 1 while b <= n: if b % 2 == 0: even_fibs.append(_UpperCAmelCase ) _a , _a : Optional[Any] =b, a + b return sum(_UpperCAmelCase ) if __name__ == "__main__": print(F"{solution() = }")
694
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase = logging.get_logger(__name__) def _A ( lowerCAmelCase_ : Any ): """simple docstring""" if "resnet-50" in model_name: lowerCAmelCase__ = ResNetConfig.from_pretrained("microsoft/resnet-50" ) elif "resnet-101" in model_name: lowerCAmelCase__ = ResNetConfig.from_pretrained("microsoft/resnet-101" ) else: raise ValueError("Model name should include either resnet50 or resnet101" ) lowerCAmelCase__ = DetrConfig(use_timm_backbone=lowerCAmelCase_ , backbone_config=lowerCAmelCase_ ) # set label attributes lowerCAmelCase__ = "panoptic" in model_name if is_panoptic: lowerCAmelCase__ = 250 else: lowerCAmelCase__ = 91 lowerCAmelCase__ = "huggingface/label-files" lowerCAmelCase__ = "coco-detection-id2label.json" lowerCAmelCase__ = json.load(open(hf_hub_download(lowerCAmelCase_ , lowerCAmelCase_ , repo_type="dataset" ) , "r" ) ) lowerCAmelCase__ = {int(lowerCAmelCase_ ): v for k, v in idalabel.items()} lowerCAmelCase__ = idalabel lowerCAmelCase__ = {v: k for k, v in idalabel.items()} return config, is_panoptic def _A ( lowerCAmelCase_ : List[Any] ): """simple docstring""" lowerCAmelCase__ = [] # stem # fmt: off rename_keys.append(("backbone.0.body.conv1.weight", "backbone.conv_encoder.model.embedder.embedder.convolution.weight") ) rename_keys.append(("backbone.0.body.bn1.weight", "backbone.conv_encoder.model.embedder.embedder.normalization.weight") ) rename_keys.append(("backbone.0.body.bn1.bias", "backbone.conv_encoder.model.embedder.embedder.normalization.bias") ) rename_keys.append(("backbone.0.body.bn1.running_mean", "backbone.conv_encoder.model.embedder.embedder.normalization.running_mean") ) rename_keys.append(("backbone.0.body.bn1.running_var", "backbone.conv_encoder.model.embedder.embedder.normalization.running_var") ) # stages for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): # shortcut if layer_idx == 0: rename_keys.append( ( F'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight', F'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight', ) ) rename_keys.append( ( F'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight', F'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight', ) ) rename_keys.append( ( F'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias', F'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias', ) ) rename_keys.append( ( F'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean', F'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean', ) ) rename_keys.append( ( F'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var', F'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var', ) ) # 3 convs for i in range(3 ): rename_keys.append( ( F'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight', F'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight', ) ) rename_keys.append( ( F'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight', F'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight', ) ) rename_keys.append( ( F'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias', F'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias', ) ) rename_keys.append( ( F'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean', F'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean', ) ) rename_keys.append( ( F'backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var', F'backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var', ) ) # fmt: on for i in range(config.encoder_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( F'transformer.encoder.layers.{i}.self_attn.out_proj.weight', F'encoder.layers.{i}.self_attn.out_proj.weight', ) ) rename_keys.append( (F'transformer.encoder.layers.{i}.self_attn.out_proj.bias', F'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.weight', F'encoder.layers.{i}.fc1.weight') ) rename_keys.append((F'transformer.encoder.layers.{i}.linear1.bias', F'encoder.layers.{i}.fc1.bias') ) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.weight', F'encoder.layers.{i}.fc2.weight') ) rename_keys.append((F'transformer.encoder.layers.{i}.linear2.bias', F'encoder.layers.{i}.fc2.bias') ) rename_keys.append( (F'transformer.encoder.layers.{i}.norm1.weight', F'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append( (F'transformer.encoder.layers.{i}.norm1.bias', F'encoder.layers.{i}.self_attn_layer_norm.bias') ) rename_keys.append( (F'transformer.encoder.layers.{i}.norm2.weight', F'encoder.layers.{i}.final_layer_norm.weight') ) rename_keys.append((F'transformer.encoder.layers.{i}.norm2.bias', F'encoder.layers.{i}.final_layer_norm.bias') ) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( F'transformer.decoder.layers.{i}.self_attn.out_proj.weight', F'decoder.layers.{i}.self_attn.out_proj.weight', ) ) rename_keys.append( (F'transformer.decoder.layers.{i}.self_attn.out_proj.bias', F'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( F'transformer.decoder.layers.{i}.multihead_attn.out_proj.weight', F'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( F'transformer.decoder.layers.{i}.multihead_attn.out_proj.bias', F'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.weight', F'decoder.layers.{i}.fc1.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.linear1.bias', F'decoder.layers.{i}.fc1.bias') ) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.weight', F'decoder.layers.{i}.fc2.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.linear2.bias', F'decoder.layers.{i}.fc2.bias') ) rename_keys.append( (F'transformer.decoder.layers.{i}.norm1.weight', F'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.norm1.bias', F'decoder.layers.{i}.self_attn_layer_norm.bias') ) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.weight', F'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (F'transformer.decoder.layers.{i}.norm2.bias', F'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append( (F'transformer.decoder.layers.{i}.norm3.weight', F'decoder.layers.{i}.final_layer_norm.weight') ) rename_keys.append((F'transformer.decoder.layers.{i}.norm3.bias', F'decoder.layers.{i}.final_layer_norm.bias') ) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("input_proj.weight", "input_projection.weight"), ("input_proj.bias", "input_projection.bias"), ("query_embed.weight", "query_position_embeddings.weight"), ("transformer.decoder.norm.weight", "decoder.layernorm.weight"), ("transformer.decoder.norm.bias", "decoder.layernorm.bias"), ("class_embed.weight", "class_labels_classifier.weight"), ("class_embed.bias", "class_labels_classifier.bias"), ("bbox_embed.layers.0.weight", "bbox_predictor.layers.0.weight"), ("bbox_embed.layers.0.bias", "bbox_predictor.layers.0.bias"), ("bbox_embed.layers.1.weight", "bbox_predictor.layers.1.weight"), ("bbox_embed.layers.1.bias", "bbox_predictor.layers.1.bias"), ("bbox_embed.layers.2.weight", "bbox_predictor.layers.2.weight"), ("bbox_embed.layers.2.bias", "bbox_predictor.layers.2.bias"), ] ) return rename_keys def _A ( lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Any , lowerCAmelCase_ : Optional[Any] ): """simple docstring""" lowerCAmelCase__ = state_dict.pop(lowerCAmelCase_ ) lowerCAmelCase__ = val def _A ( lowerCAmelCase_ : Any , lowerCAmelCase_ : Dict=False ): """simple docstring""" lowerCAmelCase__ = "" if is_panoptic: lowerCAmelCase__ = "detr." # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) lowerCAmelCase__ = state_dict.pop(F'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight' ) lowerCAmelCase__ = state_dict.pop(F'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict lowerCAmelCase__ = in_proj_weight[:256, :] lowerCAmelCase__ = in_proj_bias[:256] lowerCAmelCase__ = in_proj_weight[256:512, :] lowerCAmelCase__ = in_proj_bias[256:512] lowerCAmelCase__ = in_proj_weight[-256:, :] lowerCAmelCase__ = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention lowerCAmelCase__ = state_dict.pop(F'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight' ) lowerCAmelCase__ = state_dict.pop(F'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict lowerCAmelCase__ = in_proj_weight[:256, :] lowerCAmelCase__ = in_proj_bias[:256] lowerCAmelCase__ = in_proj_weight[256:512, :] lowerCAmelCase__ = in_proj_bias[256:512] lowerCAmelCase__ = in_proj_weight[-256:, :] lowerCAmelCase__ = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention lowerCAmelCase__ = state_dict.pop( F'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight' ) lowerCAmelCase__ = state_dict.pop(F'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias' ) # next, add query, keys and values (in that order) of cross-attention to the state dict lowerCAmelCase__ = in_proj_weight_cross_attn[:256, :] lowerCAmelCase__ = in_proj_bias_cross_attn[:256] lowerCAmelCase__ = in_proj_weight_cross_attn[256:512, :] lowerCAmelCase__ = in_proj_bias_cross_attn[256:512] lowerCAmelCase__ = in_proj_weight_cross_attn[-256:, :] lowerCAmelCase__ = in_proj_bias_cross_attn[-256:] def _A ( ): """simple docstring""" lowerCAmelCase__ = "http://images.cocodataset.org/val2017/000000039769.jpg" lowerCAmelCase__ = Image.open(requests.get(lowerCAmelCase_ , stream=lowerCAmelCase_ ).raw ) return im @torch.no_grad() def _A ( lowerCAmelCase_ : Any , lowerCAmelCase_ : str=None , lowerCAmelCase_ : int=False ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ = get_detr_config(lowerCAmelCase_ ) # load original model from torch hub lowerCAmelCase__ = { "detr-resnet-50": "detr_resnet50", "detr-resnet-101": "detr_resnet101", } logger.info(F'Converting model {model_name}...' ) lowerCAmelCase__ = torch.hub.load("facebookresearch/detr" , model_name_to_original_name[model_name] , pretrained=lowerCAmelCase_ ).eval() lowerCAmelCase__ = detr.state_dict() # rename keys for src, dest in create_rename_keys(lowerCAmelCase_ ): if is_panoptic: lowerCAmelCase__ = "detr." + src rename_key(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) # query, key and value matrices need special treatment read_in_q_k_v(lowerCAmelCase_ , is_panoptic=lowerCAmelCase_ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them lowerCAmelCase__ = "detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("detr" ) and not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ) ): lowerCAmelCase__ = state_dict.pop(lowerCAmelCase_ ) lowerCAmelCase__ = val elif "class_labels_classifier" in key or "bbox_predictor" in key: lowerCAmelCase__ = state_dict.pop(lowerCAmelCase_ ) lowerCAmelCase__ = val elif key.startswith("bbox_attention" ) or key.startswith("mask_head" ): continue else: lowerCAmelCase__ = state_dict.pop(lowerCAmelCase_ ) lowerCAmelCase__ = val else: if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): lowerCAmelCase__ = state_dict.pop(lowerCAmelCase_ ) lowerCAmelCase__ = val # finally, create HuggingFace model and load state dict lowerCAmelCase__ = DetrForSegmentation(lowerCAmelCase_ ) if is_panoptic else DetrForObjectDetection(lowerCAmelCase_ ) model.load_state_dict(lowerCAmelCase_ ) model.eval() # verify our conversion on an image lowerCAmelCase__ = "coco_panoptic" if is_panoptic else "coco_detection" lowerCAmelCase__ = DetrImageProcessor(format=lowerCAmelCase_ ) lowerCAmelCase__ = processor(images=prepare_img() , return_tensors="pt" ) lowerCAmelCase__ = encoding["pixel_values"] lowerCAmelCase__ = detr(lowerCAmelCase_ ) lowerCAmelCase__ = model(lowerCAmelCase_ ) assert torch.allclose(outputs.logits , original_outputs["pred_logits"] , atol=1E-3 ) assert torch.allclose(outputs.pred_boxes , original_outputs["pred_boxes"] , atol=1E-3 ) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["pred_masks"] , atol=1E-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(F'Saving PyTorch model and image processor to {pytorch_dump_folder_path}...' ) Path(lowerCAmelCase_ ).mkdir(exist_ok=lowerCAmelCase_ ) model.save_pretrained(lowerCAmelCase_ ) processor.save_pretrained(lowerCAmelCase_ ) if push_to_hub: # Upload model and image processor to the hub logger.info("Uploading PyTorch model and image processor to the hub..." ) model.push_to_hub(F'nielsr/{model_name}' ) processor.push_to_hub(F'nielsr/{model_name}' ) if __name__ == "__main__": UpperCamelCase = argparse.ArgumentParser() parser.add_argument( '--model_name', default='detr-resnet-50', type=str, choices=['detr-resnet-50', 'detr-resnet-101'], help='Name of the DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument('--push_to_hub', action='store_true', help='Whether to push the model to the hub or not.') UpperCamelCase = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
703
from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def _A ( ): """simple docstring""" lowerCAmelCase__ = [randint(-1000 , 1000 ) for i in range(10 )] lowerCAmelCase__ = randint(-5000 , 5000 ) return (arr, r) UpperCamelCase = make_dataset() def _A ( lowerCAmelCase_ : list[int] , lowerCAmelCase_ : int ): """simple docstring""" for triplet in permutations(lowerCAmelCase_ , 3 ): if sum(lowerCAmelCase_ ) == target: return tuple(sorted(lowerCAmelCase_ ) ) return (0, 0, 0) def _A ( lowerCAmelCase_ : list[int] , lowerCAmelCase_ : int ): """simple docstring""" arr.sort() lowerCAmelCase__ = len(lowerCAmelCase_ ) for i in range(n - 1 ): lowerCAmelCase__ , lowerCAmelCase__ = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def _A ( ): """simple docstring""" lowerCAmelCase__ = "\nfrom __main__ import dataset, triplet_sum1, triplet_sum2\n" lowerCAmelCase__ = "\ntriplet_sum1(*dataset)\n" lowerCAmelCase__ = "\ntriplet_sum2(*dataset)\n" lowerCAmelCase__ = repeat(setup=lowerCAmelCase_ , stmt=lowerCAmelCase_ , repeat=5 , number=1_0000 ) lowerCAmelCase__ = repeat(setup=lowerCAmelCase_ , stmt=lowerCAmelCase_ , repeat=5 , number=1_0000 ) return (min(lowerCAmelCase_ ), min(lowerCAmelCase_ )) if __name__ == "__main__": from doctest import testmod testmod() UpperCamelCase = solution_times() print(F"""The time for naive implementation is {times[0]}.""") print(F"""The time for optimized implementation is {times[1]}.""")
125
0
'''simple docstring''' from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class __snake_case ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowerCamelCase__ = 42 class __snake_case ( nn.Module ): '''simple docstring''' def __init__( self , __SCREAMING_SNAKE_CASE=3 , __SCREAMING_SNAKE_CASE=3 , __SCREAMING_SNAKE_CASE=("DownEncoderBlock2D",) , __SCREAMING_SNAKE_CASE=(6_4,) , __SCREAMING_SNAKE_CASE=2 , __SCREAMING_SNAKE_CASE=3_2 , __SCREAMING_SNAKE_CASE="silu" , __SCREAMING_SNAKE_CASE=True , ): super().__init__() snake_case__ : str = layers_per_block snake_case__ : int = torch.nn.Convad( __SCREAMING_SNAKE_CASE , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) snake_case__ : List[Any] = None snake_case__ : List[Any] = nn.ModuleList([] ) # down snake_case__ : Union[str, Any] = block_out_channels[0] for i, down_block_type in enumerate(__SCREAMING_SNAKE_CASE ): snake_case__ : Optional[Any] = output_channel snake_case__ : Union[str, Any] = block_out_channels[i] snake_case__ : int = i == len(__SCREAMING_SNAKE_CASE ) - 1 snake_case__ : str = get_down_block( __SCREAMING_SNAKE_CASE , num_layers=self.layers_per_block , in_channels=__SCREAMING_SNAKE_CASE , out_channels=__SCREAMING_SNAKE_CASE , add_downsample=not is_final_block , resnet_eps=1e-6 , downsample_padding=0 , resnet_act_fn=__SCREAMING_SNAKE_CASE , resnet_groups=__SCREAMING_SNAKE_CASE , attention_head_dim=__SCREAMING_SNAKE_CASE , temb_channels=__SCREAMING_SNAKE_CASE , ) self.down_blocks.append(__SCREAMING_SNAKE_CASE ) # mid snake_case__ : Optional[Any] = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1e-6 , resnet_act_fn=__SCREAMING_SNAKE_CASE , output_scale_factor=1 , resnet_time_scale_shift="""default""" , attention_head_dim=block_out_channels[-1] , resnet_groups=__SCREAMING_SNAKE_CASE , temb_channels=__SCREAMING_SNAKE_CASE , ) # out snake_case__ : Tuple = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=__SCREAMING_SNAKE_CASE , eps=1e-6 ) snake_case__ : Tuple = nn.SiLU() snake_case__ : str = 2 * out_channels if double_z else out_channels snake_case__ : int = nn.Convad(block_out_channels[-1] , __SCREAMING_SNAKE_CASE , 3 , padding=1 ) snake_case__ : Union[str, Any] = False def __UpperCamelCase ( self , __SCREAMING_SNAKE_CASE ): snake_case__ : Optional[Any] = x snake_case__ : int = self.conv_in(__SCREAMING_SNAKE_CASE ) if self.training and self.gradient_checkpointing: def create_custom_forward(__SCREAMING_SNAKE_CASE ): def custom_forward(*__SCREAMING_SNAKE_CASE ): return module(*__SCREAMING_SNAKE_CASE ) return custom_forward # down if is_torch_version(""">=""" , """1.11.0""" ): for down_block in self.down_blocks: snake_case__ : List[Any] = torch.utils.checkpoint.checkpoint( create_custom_forward(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE , use_reentrant=__SCREAMING_SNAKE_CASE ) # middle snake_case__ : List[Any] = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __SCREAMING_SNAKE_CASE , use_reentrant=__SCREAMING_SNAKE_CASE ) else: for down_block in self.down_blocks: snake_case__ : Dict = torch.utils.checkpoint.checkpoint(create_custom_forward(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) # middle snake_case__ : str = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , __SCREAMING_SNAKE_CASE ) else: # down for down_block in self.down_blocks: snake_case__ : List[str] = down_block(__SCREAMING_SNAKE_CASE ) # middle snake_case__ : str = self.mid_block(__SCREAMING_SNAKE_CASE ) # post-process snake_case__ : Any = self.conv_norm_out(__SCREAMING_SNAKE_CASE ) snake_case__ : List[str] = self.conv_act(__SCREAMING_SNAKE_CASE ) snake_case__ : str = self.conv_out(__SCREAMING_SNAKE_CASE ) return sample class __snake_case ( nn.Module ): '''simple docstring''' def __init__( self , __SCREAMING_SNAKE_CASE=3 , __SCREAMING_SNAKE_CASE=3 , __SCREAMING_SNAKE_CASE=("UpDecoderBlock2D",) , __SCREAMING_SNAKE_CASE=(6_4,) , __SCREAMING_SNAKE_CASE=2 , __SCREAMING_SNAKE_CASE=3_2 , __SCREAMING_SNAKE_CASE="silu" , __SCREAMING_SNAKE_CASE="group" , ): super().__init__() snake_case__ : Any = layers_per_block snake_case__ : Optional[Any] = nn.Convad( __SCREAMING_SNAKE_CASE , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) snake_case__ : Union[str, Any] = None snake_case__ : Dict = nn.ModuleList([] ) snake_case__ : Optional[int] = in_channels if norm_type == """spatial""" else None # mid snake_case__ : Tuple = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1e-6 , resnet_act_fn=__SCREAMING_SNAKE_CASE , output_scale_factor=1 , resnet_time_scale_shift="""default""" if norm_type == """group""" else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=__SCREAMING_SNAKE_CASE , temb_channels=__SCREAMING_SNAKE_CASE , ) # up snake_case__ : List[Any] = list(reversed(__SCREAMING_SNAKE_CASE ) ) snake_case__ : Optional[Any] = reversed_block_out_channels[0] for i, up_block_type in enumerate(__SCREAMING_SNAKE_CASE ): snake_case__ : List[Any] = output_channel snake_case__ : Optional[Any] = reversed_block_out_channels[i] snake_case__ : List[str] = i == len(__SCREAMING_SNAKE_CASE ) - 1 snake_case__ : int = get_up_block( __SCREAMING_SNAKE_CASE , num_layers=self.layers_per_block + 1 , in_channels=__SCREAMING_SNAKE_CASE , out_channels=__SCREAMING_SNAKE_CASE , prev_output_channel=__SCREAMING_SNAKE_CASE , add_upsample=not is_final_block , resnet_eps=1e-6 , resnet_act_fn=__SCREAMING_SNAKE_CASE , resnet_groups=__SCREAMING_SNAKE_CASE , attention_head_dim=__SCREAMING_SNAKE_CASE , temb_channels=__SCREAMING_SNAKE_CASE , resnet_time_scale_shift=__SCREAMING_SNAKE_CASE , ) self.up_blocks.append(__SCREAMING_SNAKE_CASE ) snake_case__ : int = output_channel # out if norm_type == "spatial": snake_case__ : List[Any] = SpatialNorm(block_out_channels[0] , __SCREAMING_SNAKE_CASE ) else: snake_case__ : Any = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=__SCREAMING_SNAKE_CASE , eps=1e-6 ) snake_case__ : Tuple = nn.SiLU() snake_case__ : Union[str, Any] = nn.Convad(block_out_channels[0] , __SCREAMING_SNAKE_CASE , 3 , padding=1 ) snake_case__ : int = False def __UpperCamelCase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=None ): snake_case__ : Union[str, Any] = z snake_case__ : Any = self.conv_in(__SCREAMING_SNAKE_CASE ) snake_case__ : Optional[Any] = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(__SCREAMING_SNAKE_CASE ): def custom_forward(*__SCREAMING_SNAKE_CASE ): return module(*__SCREAMING_SNAKE_CASE ) return custom_forward if is_torch_version(""">=""" , """1.11.0""" ): # middle snake_case__ : int = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , use_reentrant=__SCREAMING_SNAKE_CASE ) snake_case__ : int = sample.to(__SCREAMING_SNAKE_CASE ) # up for up_block in self.up_blocks: snake_case__ : List[str] = torch.utils.checkpoint.checkpoint( create_custom_forward(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , use_reentrant=__SCREAMING_SNAKE_CASE ) else: # middle snake_case__ : Dict = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) snake_case__ : Optional[Any] = sample.to(__SCREAMING_SNAKE_CASE ) # up for up_block in self.up_blocks: snake_case__ : str = torch.utils.checkpoint.checkpoint(create_custom_forward(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) else: # middle snake_case__ : List[Any] = self.mid_block(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) snake_case__ : List[Any] = sample.to(__SCREAMING_SNAKE_CASE ) # up for up_block in self.up_blocks: snake_case__ : Dict = up_block(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # post-process if latent_embeds is None: snake_case__ : Optional[Any] = self.conv_norm_out(__SCREAMING_SNAKE_CASE ) else: snake_case__ : str = self.conv_norm_out(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) snake_case__ : Any = self.conv_act(__SCREAMING_SNAKE_CASE ) snake_case__ : Optional[Any] = self.conv_out(__SCREAMING_SNAKE_CASE ) return sample class __snake_case ( nn.Module ): '''simple docstring''' def __init__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE="random" , __SCREAMING_SNAKE_CASE=False , __SCREAMING_SNAKE_CASE=True ): super().__init__() snake_case__ : int = n_e snake_case__ : Optional[int] = vq_embed_dim snake_case__ : int = beta snake_case__ : Optional[int] = legacy snake_case__ : Dict = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) snake_case__ : List[str] = remap if self.remap is not None: self.register_buffer("""used""" , torch.tensor(np.load(self.remap ) ) ) snake_case__ : Optional[Any] = self.used.shape[0] snake_case__ : List[str] = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": snake_case__ : Dict = self.re_embed snake_case__ : List[str] = self.re_embed + 1 print( f"Remapping {self.n_e} indices to {self.re_embed} indices. " f"Using {self.unknown_index} for unknown indices." ) else: snake_case__ : Union[str, Any] = n_e snake_case__ : str = sane_index_shape def __UpperCamelCase ( self , __SCREAMING_SNAKE_CASE ): snake_case__ : Any = inds.shape assert len(__SCREAMING_SNAKE_CASE ) > 1 snake_case__ : Dict = inds.reshape(ishape[0] , -1 ) snake_case__ : Any = self.used.to(__SCREAMING_SNAKE_CASE ) snake_case__ : Dict = (inds[:, :, None] == used[None, None, ...]).long() snake_case__ : List[Any] = match.argmax(-1 ) snake_case__ : List[str] = match.sum(2 ) < 1 if self.unknown_index == "random": snake_case__ : List[str] = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: snake_case__ : Optional[Any] = self.unknown_index return new.reshape(__SCREAMING_SNAKE_CASE ) def __UpperCamelCase ( self , __SCREAMING_SNAKE_CASE ): snake_case__ : List[Any] = inds.shape assert len(__SCREAMING_SNAKE_CASE ) > 1 snake_case__ : int = inds.reshape(ishape[0] , -1 ) snake_case__ : Optional[int] = self.used.to(__SCREAMING_SNAKE_CASE ) if self.re_embed > self.used.shape[0]: # extra token snake_case__ : List[Any] = 0 # simply set to zero snake_case__ : Union[str, Any] = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , __SCREAMING_SNAKE_CASE ) return back.reshape(__SCREAMING_SNAKE_CASE ) def __UpperCamelCase ( self , __SCREAMING_SNAKE_CASE ): # reshape z -> (batch, height, width, channel) and flatten snake_case__ : Any = z.permute(0 , 2 , 3 , 1 ).contiguous() snake_case__ : Optional[Any] = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z snake_case__ : Dict = torch.argmin(torch.cdist(__SCREAMING_SNAKE_CASE , self.embedding.weight ) , dim=1 ) snake_case__ : Union[str, Any] = self.embedding(__SCREAMING_SNAKE_CASE ).view(z.shape ) snake_case__ : List[str] = None snake_case__ : Union[str, Any] = None # compute loss for embedding if not self.legacy: snake_case__ : Tuple = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: snake_case__ : List[Any] = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients snake_case__ : Any = z + (z_q - z).detach() # reshape back to match original input shape snake_case__ : Union[str, Any] = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: snake_case__ : List[Any] = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis snake_case__ : str = self.remap_to_used(__SCREAMING_SNAKE_CASE ) snake_case__ : str = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: snake_case__ : Tuple = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def __UpperCamelCase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): # shape specifying (batch, height, width, channel) if self.remap is not None: snake_case__ : List[Any] = indices.reshape(shape[0] , -1 ) # add batch axis snake_case__ : Optional[int] = self.unmap_to_all(__SCREAMING_SNAKE_CASE ) snake_case__ : Optional[Any] = indices.reshape(-1 ) # flatten again # get quantized latent vectors snake_case__ : int = self.embedding(__SCREAMING_SNAKE_CASE ) if shape is not None: snake_case__ : str = z_q.view(__SCREAMING_SNAKE_CASE ) # reshape back to match original input shape snake_case__ : str = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class __snake_case ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=False ): snake_case__ : Tuple = parameters snake_case__ , snake_case__ : Any = torch.chunk(__SCREAMING_SNAKE_CASE , 2 , dim=1 ) snake_case__ : Union[str, Any] = torch.clamp(self.logvar , -30.0 , 20.0 ) snake_case__ : Optional[int] = deterministic snake_case__ : Optional[int] = torch.exp(0.5 * self.logvar ) snake_case__ : Any = torch.exp(self.logvar ) if self.deterministic: snake_case__ : List[str] = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def __UpperCamelCase ( self , __SCREAMING_SNAKE_CASE = None ): # make sure sample is on the same device as the parameters and has same dtype snake_case__ : Dict = randn_tensor( self.mean.shape , generator=__SCREAMING_SNAKE_CASE , device=self.parameters.device , dtype=self.parameters.dtype ) snake_case__ : Optional[int] = self.mean + self.std * sample return x def __UpperCamelCase ( self , __SCREAMING_SNAKE_CASE=None ): if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def __UpperCamelCase ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=[1, 2, 3] ): if self.deterministic: return torch.Tensor([0.0] ) snake_case__ : Any = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=__SCREAMING_SNAKE_CASE ) def __UpperCamelCase ( self ): return self.mean
38
def __A ( _lowercase = 2_00 ): '''simple docstring''' _A = [1, 2, 5, 10, 20, 50, 1_00, 2_00] _A = [0] * (pence + 1) _A = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(_lowercase , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 73682
484
0
'''simple docstring''' from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, ) @flax.struct.dataclass class A__ ( _snake_case ): lowercase = 42 lowercase = 42 class A__ ( nn.Module ): lowercase = 42 lowercase = (16, 32, 96, 256) lowercase = jnp.floataa def snake_case_ ( self ) -> str: '''simple docstring''' A_ = nn.Conv( self.block_out_channels[0] , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) A_ = [] for i in range(len(self.block_out_channels ) - 1 ): A_ = self.block_out_channels[i] A_ = self.block_out_channels[i + 1] A_ = nn.Conv( UpperCamelCase__ , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(UpperCamelCase__ ) A_ = nn.Conv( UpperCamelCase__ , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(UpperCamelCase__ ) A_ = blocks A_ = nn.Conv( self.conditioning_embedding_channels , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self , UpperCamelCase__ ) -> Optional[int]: '''simple docstring''' A_ = self.conv_in(UpperCamelCase__ ) A_ = nn.silu(UpperCamelCase__ ) for block in self.blocks: A_ = block(UpperCamelCase__ ) A_ = nn.silu(UpperCamelCase__ ) A_ = self.conv_out(UpperCamelCase__ ) return embedding @flax_register_to_config class A__ ( nn.Module , _snake_case , _snake_case ): lowercase = 32 lowercase = 4 lowercase = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) lowercase = False lowercase = (320, 640, 1_280, 1_280) lowercase = 2 lowercase = 8 lowercase = None lowercase = 1_280 lowercase = 0.0 lowercase = False lowercase = jnp.floataa lowercase = True lowercase = 0 lowercase = "rgb" lowercase = (16, 32, 96, 256) def snake_case_ ( self , UpperCamelCase__ ) -> FrozenDict: '''simple docstring''' A_ = (1, self.in_channels, self.sample_size, self.sample_size) A_ = jnp.zeros(UpperCamelCase__ , dtype=jnp.floataa ) A_ = jnp.ones((1,) , dtype=jnp.intaa ) A_ = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) A_ = (1, 3, self.sample_size * 8, self.sample_size * 8) A_ = jnp.zeros(UpperCamelCase__ , dtype=jnp.floataa ) A_ , A_ = jax.random.split(UpperCamelCase__ ) A_ = {"""params""": params_rng, """dropout""": dropout_rng} return self.init(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ )["params"] def snake_case_ ( self ) -> str: '''simple docstring''' A_ = self.block_out_channels A_ = block_out_channels[0] * 4 # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. A_ = self.num_attention_heads or self.attention_head_dim # input A_ = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time A_ = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) A_ = FlaxTimestepEmbedding(UpperCamelCase__ , dtype=self.dtype ) A_ = FlaxControlNetConditioningEmbedding( conditioning_embedding_channels=block_out_channels[0] , block_out_channels=self.conditioning_embedding_out_channels , ) A_ = self.only_cross_attention if isinstance(UpperCamelCase__ , UpperCamelCase__ ): A_ = (only_cross_attention,) * len(self.down_block_types ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ): A_ = (num_attention_heads,) * len(self.down_block_types ) # down A_ = [] A_ = [] A_ = block_out_channels[0] A_ = nn.Conv( UpperCamelCase__ , kernel_size=(1, 1) , padding="""VALID""" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(UpperCamelCase__ ) for i, down_block_type in enumerate(self.down_block_types ): A_ = output_channel A_ = block_out_channels[i] A_ = i == len(UpperCamelCase__ ) - 1 if down_block_type == "CrossAttnDownBlock2D": A_ = FlaxCrossAttnDownBlockaD( in_channels=UpperCamelCase__ , out_channels=UpperCamelCase__ , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , dtype=self.dtype , ) else: A_ = FlaxDownBlockaD( in_channels=UpperCamelCase__ , out_channels=UpperCamelCase__ , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(UpperCamelCase__ ) for _ in range(self.layers_per_block ): A_ = nn.Conv( UpperCamelCase__ , kernel_size=(1, 1) , padding="""VALID""" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(UpperCamelCase__ ) if not is_final_block: A_ = nn.Conv( UpperCamelCase__ , kernel_size=(1, 1) , padding="""VALID""" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(UpperCamelCase__ ) A_ = down_blocks A_ = controlnet_down_blocks # mid A_ = block_out_channels[-1] A_ = FlaxUNetMidBlockaDCrossAttn( in_channels=UpperCamelCase__ , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , dtype=self.dtype , ) A_ = nn.Conv( UpperCamelCase__ , kernel_size=(1, 1) , padding="""VALID""" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = 1.0 , UpperCamelCase__ = True , UpperCamelCase__ = False , ) -> Union[FlaxControlNetOutput, Tuple]: '''simple docstring''' A_ = self.controlnet_conditioning_channel_order if channel_order == "bgr": A_ = jnp.flip(UpperCamelCase__ , axis=1 ) # 1. time if not isinstance(UpperCamelCase__ , jnp.ndarray ): A_ = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(UpperCamelCase__ , jnp.ndarray ) and len(timesteps.shape ) == 0: A_ = timesteps.astype(dtype=jnp.floataa ) A_ = jnp.expand_dims(UpperCamelCase__ , 0 ) A_ = self.time_proj(UpperCamelCase__ ) A_ = self.time_embedding(UpperCamelCase__ ) # 2. pre-process A_ = jnp.transpose(UpperCamelCase__ , (0, 2, 3, 1) ) A_ = self.conv_in(UpperCamelCase__ ) A_ = jnp.transpose(UpperCamelCase__ , (0, 2, 3, 1) ) A_ = self.controlnet_cond_embedding(UpperCamelCase__ ) sample += controlnet_cond # 3. down A_ = (sample,) for down_block in self.down_blocks: if isinstance(UpperCamelCase__ , UpperCamelCase__ ): A_ , A_ = down_block(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , deterministic=not train ) else: A_ , A_ = down_block(UpperCamelCase__ , UpperCamelCase__ , deterministic=not train ) down_block_res_samples += res_samples # 4. mid A_ = self.mid_block(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , deterministic=not train ) # 5. contronet blocks A_ = () for down_block_res_sample, controlnet_block in zip(UpperCamelCase__ , self.controlnet_down_blocks ): A_ = controlnet_block(UpperCamelCase__ ) controlnet_down_block_res_samples += (down_block_res_sample,) A_ = controlnet_down_block_res_samples A_ = self.controlnet_mid_block(UpperCamelCase__ ) # 6. scaling A_ = [sample * conditioning_scale for sample in down_block_res_samples] mid_block_res_sample *= conditioning_scale if not return_dict: return (down_block_res_samples, mid_block_res_sample) return FlaxControlNetOutput( down_block_res_samples=UpperCamelCase__ , mid_block_res_sample=UpperCamelCase__ )
702
'''simple docstring''' import os __lowerCamelCase = {'''I''': 1, '''V''': 5, '''X''': 10, '''L''': 50, '''C''': 100, '''D''': 500, '''M''': 1000} def UpperCAmelCase__ ( UpperCAmelCase__ ) -> int: A_ = 0 A_ = 0 while index < len(UpperCAmelCase__ ) - 1: A_ = SYMBOLS[numerals[index]] A_ = SYMBOLS[numerals[index + 1]] if current_value < next_value: total_value -= current_value else: total_value += current_value index += 1 total_value += SYMBOLS[numerals[index]] return total_value def UpperCAmelCase__ ( UpperCAmelCase__ ) -> str: A_ = """""" A_ = num // 10_00 numerals += m_count * "M" num %= 10_00 A_ = num // 1_00 if c_count == 9: numerals += "CM" c_count -= 9 elif c_count == 4: numerals += "CD" c_count -= 4 if c_count >= 5: numerals += "D" c_count -= 5 numerals += c_count * "C" num %= 1_00 A_ = num // 10 if x_count == 9: numerals += "XC" x_count -= 9 elif x_count == 4: numerals += "XL" x_count -= 4 if x_count >= 5: numerals += "L" x_count -= 5 numerals += x_count * "X" num %= 10 if num == 9: numerals += "IX" num -= 9 elif num == 4: numerals += "IV" num -= 4 if num >= 5: numerals += "V" num -= 5 numerals += num * "I" return numerals def UpperCAmelCase__ ( UpperCAmelCase__ = "/p089_roman.txt" ) -> int: A_ = 0 with open(os.path.dirname(UpperCAmelCase__ ) + roman_numerals_filename ) as filea: A_ = filea.readlines() for line in lines: A_ = line.strip() A_ = parse_roman_numerals(UpperCAmelCase__ ) A_ = generate_roman_numerals(UpperCAmelCase__ ) savings += len(UpperCAmelCase__ ) - len(UpperCAmelCase__ ) return savings if __name__ == "__main__": print(f"""{solution() = }""")
667
0
'''simple docstring''' from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class _snake_case : def __init__( self ,_snake_case ,_snake_case=12 ,_snake_case=7 ,_snake_case=True ,_snake_case=True ,_snake_case=True ,_snake_case=99 ,_snake_case=32 ,_snake_case=32 ,_snake_case=2 ,_snake_case=4 ,_snake_case=37 ,_snake_case=0.1 ,_snake_case=0.1 ,_snake_case=5_12 ,_snake_case=0.02 ,_snake_case=0 ,_snake_case=None ,): UpperCAmelCase_ : int = parent UpperCAmelCase_ : Optional[Any] = batch_size UpperCAmelCase_ : Tuple = seq_length UpperCAmelCase_ : int = is_training UpperCAmelCase_ : Dict = use_input_mask UpperCAmelCase_ : List[str] = use_labels UpperCAmelCase_ : int = vocab_size UpperCAmelCase_ : int = hidden_size UpperCAmelCase_ : Any = projection_dim UpperCAmelCase_ : Dict = num_hidden_layers UpperCAmelCase_ : int = num_attention_heads UpperCAmelCase_ : Optional[Any] = intermediate_size UpperCAmelCase_ : Optional[Any] = dropout UpperCAmelCase_ : Any = attention_dropout UpperCAmelCase_ : List[str] = max_position_embeddings UpperCAmelCase_ : Dict = initializer_range UpperCAmelCase_ : Tuple = scope UpperCAmelCase_ : List[Any] = bos_token_id def UpperCamelCase__ ( self ): UpperCAmelCase_ : List[str] = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) UpperCAmelCase_ : Optional[Any] = None if self.use_input_mask: UpperCAmelCase_ : str = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: UpperCAmelCase_ : Tuple = input_mask.numpy() UpperCAmelCase_ , UpperCAmelCase_ : str = input_mask.shape UpperCAmelCase_ : Union[str, Any] = np.random.randint(1 ,seq_length - 1 ,size=(batch_size,) ) for batch_idx, start_index in enumerate(_snake_case ): UpperCAmelCase_ : Union[str, Any] = 1 UpperCAmelCase_ : Optional[Any] = 0 UpperCAmelCase_ : int = self.get_config() return config, input_ids, tf.convert_to_tensor(_snake_case ) def UpperCamelCase__ ( self ): return BlipTextConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,projection_dim=self.projection_dim ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,dropout=self.dropout ,attention_dropout=self.attention_dropout ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,bos_token_id=self.bos_token_id ,) def UpperCamelCase__ ( self ,_snake_case ,_snake_case ,_snake_case ): UpperCAmelCase_ : Tuple = TFBlipTextModel(config=_snake_case ) UpperCAmelCase_ : Tuple = model(_snake_case ,attention_mask=_snake_case ,training=_snake_case ) UpperCAmelCase_ : int = model(_snake_case ,training=_snake_case ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape ,(self.batch_size, self.hidden_size) ) def UpperCamelCase__ ( self ): UpperCAmelCase_ : Optional[Any] = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Tuple = config_and_inputs UpperCAmelCase_ : List[str] = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class _snake_case (__SCREAMING_SNAKE_CASE , unittest.TestCase): __A : Optional[Any] =(TFBlipTextModel,) if is_tf_available() else () __A : Dict =False __A : Optional[Any] =False __A : Union[str, Any] =False def UpperCamelCase__ ( self ): UpperCAmelCase_ : str = BlipTextModelTester(self ) UpperCAmelCase_ : Dict = ConfigTester(self ,config_class=_snake_case ,hidden_size=37 ) def UpperCamelCase__ ( self ): self.config_tester.run_common_tests() def UpperCamelCase__ ( self ): UpperCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_snake_case ) def UpperCamelCase__ ( self ): pass def UpperCamelCase__ ( self ): pass @unittest.skip(reason="Blip does not use inputs_embeds" ) def UpperCamelCase__ ( self ): pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCamelCase__ ( self ): pass @unittest.skip(reason="BlipTextModel has no base class and is not available in MODEL_MAPPING" ) def UpperCamelCase__ ( self ): pass @slow def UpperCamelCase__ ( self ): for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : Any = TFBlipTextModel.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) def UpperCamelCase__ ( self ,_snake_case=True ): super().test_pt_tf_model_equivalence(allow_missing_keys=_snake_case )
71
"""simple docstring""" import math def _snake_case ( _snake_case : float , _snake_case : float ) -> float: '''simple docstring''' if ( not isinstance(_snake_case , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('power_factor must be a valid float value between -1 and 1.' ) return apparent_power * power_factor def _snake_case ( _snake_case : float , _snake_case : float ) -> float: '''simple docstring''' if ( not isinstance(_snake_case , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('power_factor must be a valid float value between -1 and 1.' ) return apparent_power * math.sqrt(1 - power_factor**2 ) if __name__ == "__main__": import doctest doctest.testmod()
7
0
def __UpperCamelCase ( A ): UpperCamelCase__ = 0 while len(A ) > 1: UpperCamelCase__ = 0 # Consider two files with minimum cost to be merged for _ in range(2 ): UpperCamelCase__ = files.index(min(A ) ) temp += files[min_index] files.pop(A ) files.append(A ) optimal_merge_cost += temp return optimal_merge_cost if __name__ == "__main__": import doctest doctest.testmod()
469
__magic_name__ ={ '''Pillow''': '''Pillow<10.0.0''', '''accelerate''': '''accelerate>=0.20.3''', '''av''': '''av==9.2.0''', '''beautifulsoup4''': '''beautifulsoup4''', '''black''': '''black~=23.1''', '''codecarbon''': '''codecarbon==1.2.0''', '''cookiecutter''': '''cookiecutter==1.7.3''', '''dataclasses''': '''dataclasses''', '''datasets''': '''datasets!=2.5.0''', '''decord''': '''decord==0.6.0''', '''deepspeed''': '''deepspeed>=0.9.3''', '''diffusers''': '''diffusers''', '''dill''': '''dill<0.3.5''', '''evaluate''': '''evaluate>=0.2.0''', '''fairscale''': '''fairscale>0.3''', '''faiss-cpu''': '''faiss-cpu''', '''fastapi''': '''fastapi''', '''filelock''': '''filelock''', '''flax''': '''flax>=0.4.1,<=0.7.0''', '''ftfy''': '''ftfy''', '''fugashi''': '''fugashi>=1.0''', '''GitPython''': '''GitPython<3.1.19''', '''hf-doc-builder''': '''hf-doc-builder>=0.3.0''', '''huggingface-hub''': '''huggingface-hub>=0.14.1,<1.0''', '''importlib_metadata''': '''importlib_metadata''', '''ipadic''': '''ipadic>=1.0.0,<2.0''', '''isort''': '''isort>=5.5.4''', '''jax''': '''jax>=0.2.8,!=0.3.2,<=0.4.13''', '''jaxlib''': '''jaxlib>=0.1.65,<=0.4.13''', '''jieba''': '''jieba''', '''kenlm''': '''kenlm''', '''keras-nlp''': '''keras-nlp>=0.3.1''', '''librosa''': '''librosa''', '''nltk''': '''nltk''', '''natten''': '''natten>=0.14.6''', '''numpy''': '''numpy>=1.17''', '''onnxconverter-common''': '''onnxconverter-common''', '''onnxruntime-tools''': '''onnxruntime-tools>=1.4.2''', '''onnxruntime''': '''onnxruntime>=1.4.0''', '''opencv-python''': '''opencv-python''', '''optuna''': '''optuna''', '''optax''': '''optax>=0.0.8,<=0.1.4''', '''packaging''': '''packaging>=20.0''', '''parameterized''': '''parameterized''', '''phonemizer''': '''phonemizer''', '''protobuf''': '''protobuf''', '''psutil''': '''psutil''', '''pyyaml''': '''pyyaml>=5.1''', '''pydantic''': '''pydantic<2''', '''pytest''': '''pytest>=7.2.0''', '''pytest-timeout''': '''pytest-timeout''', '''pytest-xdist''': '''pytest-xdist''', '''python''': '''python>=3.8.0''', '''ray[tune]''': '''ray[tune]''', '''regex''': '''regex!=2019.12.17''', '''requests''': '''requests''', '''rhoknp''': '''rhoknp>=1.1.0,<1.3.1''', '''rjieba''': '''rjieba''', '''rouge-score''': '''rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1''', '''ruff''': '''ruff>=0.0.241,<=0.0.259''', '''sacrebleu''': '''sacrebleu>=1.4.12,<2.0.0''', '''sacremoses''': '''sacremoses''', '''safetensors''': '''safetensors>=0.3.1''', '''sagemaker''': '''sagemaker>=2.31.0''', '''scikit-learn''': '''scikit-learn''', '''sentencepiece''': '''sentencepiece>=0.1.91,!=0.1.92''', '''sigopt''': '''sigopt''', '''starlette''': '''starlette''', '''sudachipy''': '''sudachipy>=0.6.6''', '''sudachidict_core''': '''sudachidict_core>=20220729''', '''tensorflow-cpu''': '''tensorflow-cpu>=2.6,<2.14''', '''tensorflow''': '''tensorflow>=2.6,<2.14''', '''tensorflow-text''': '''tensorflow-text<2.14''', '''tf2onnx''': '''tf2onnx''', '''timeout-decorator''': '''timeout-decorator''', '''timm''': '''timm''', '''tokenizers''': '''tokenizers>=0.11.1,!=0.11.3,<0.14''', '''torch''': '''torch>=1.9,!=1.12.0''', '''torchaudio''': '''torchaudio''', '''torchvision''': '''torchvision''', '''pyctcdecode''': '''pyctcdecode>=0.4.0''', '''tqdm''': '''tqdm>=4.27''', '''unidic''': '''unidic>=1.0.2''', '''unidic_lite''': '''unidic_lite>=1.0.7''', '''urllib3''': '''urllib3<2.0.0''', '''uvicorn''': '''uvicorn''', }
469
1
"""simple docstring""" from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
510
"""simple docstring""" def snake_case ( _a: list )-> bool: '''simple docstring''' if not isinstance(_a , _a ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(_a ) == 0: raise ValueError('Input list must be a non empty list' ) if len(_a ) == 1: return True lowerCamelCase__ = series[1] - series[0] for index in range(len(_a ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def snake_case ( _a: list )-> float: '''simple docstring''' if not isinstance(_a , _a ): raise ValueError('Input series is not valid, valid series - [2, 4, 6]' ) if len(_a ) == 0: raise ValueError('Input list must be a non empty list' ) lowerCamelCase__ = 0 for val in series: answer += val return answer / len(_a ) if __name__ == "__main__": import doctest doctest.testmod()
510
1
# DISCLAIMER: This code is strongly influenced by https://github.com/pesser/pytorch_diffusion # and https://github.com/hojonathanho/diffusion import math from dataclasses import dataclass from typing import List, Optional, Tuple, Union import numpy as np import torch from diffusers.configuration_utils import ConfigMixin, register_to_config from diffusers.schedulers.scheduling_utils import SchedulerMixin from diffusers.utils import BaseOutput, deprecate @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->DDIM class lowerCAmelCase_ ( _lowercase ): """simple docstring""" UpperCAmelCase__ = 42 UpperCAmelCase__ = None def _a ( __lowercase , __lowercase=0.999 , __lowercase="cosine" , ) -> Union[str, Any]: """simple docstring""" if alpha_transform_type == "cosine": def alpha_bar_fn(__lowercase ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(__lowercase ): return math.exp(t * -12.0 ) else: raise ValueError(F"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) __UpperCamelCase = [] for i in range(__lowercase ): __UpperCamelCase = i / num_diffusion_timesteps __UpperCamelCase = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__lowercase ) / alpha_bar_fn(__lowercase ) , __lowercase ) ) return torch.tensor(__lowercase , dtype=torch.floataa ) class lowerCAmelCase_ ( _lowercase , _lowercase ): """simple docstring""" UpperCAmelCase__ = 1 @register_to_config def __init__( self , _SCREAMING_SNAKE_CASE = 1_000 , _SCREAMING_SNAKE_CASE = 0.0_0_0_1 , _SCREAMING_SNAKE_CASE = 0.0_2 , _SCREAMING_SNAKE_CASE = "linear" , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE = True , _SCREAMING_SNAKE_CASE = 0 , _SCREAMING_SNAKE_CASE = "epsilon" , _SCREAMING_SNAKE_CASE = 1.0 , **_SCREAMING_SNAKE_CASE , ) -> List[Any]: if kwargs.get('set_alpha_to_one' , _SCREAMING_SNAKE_CASE ) is not None: __UpperCamelCase = ( 'The `set_alpha_to_one` argument is deprecated. Please use `set_alpha_to_zero` instead.' ) deprecate('set_alpha_to_one' , '1.0.0' , _SCREAMING_SNAKE_CASE , standard_warn=_SCREAMING_SNAKE_CASE ) __UpperCamelCase = kwargs['set_alpha_to_one'] if trained_betas is not None: __UpperCamelCase = torch.tensor(_SCREAMING_SNAKE_CASE , dtype=torch.floataa ) elif beta_schedule == "linear": __UpperCamelCase = torch.linspace(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. __UpperCamelCase = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , _SCREAMING_SNAKE_CASE , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule __UpperCamelCase = betas_for_alpha_bar(_SCREAMING_SNAKE_CASE ) else: raise NotImplementedError(f"""{beta_schedule} does is not implemented for {self.__class__}""" ) __UpperCamelCase = 1.0 - self.betas __UpperCamelCase = torch.cumprod(self.alphas , dim=0 ) # At every step in inverted ddim, we are looking into the next alphas_cumprod # For the final step, there is no next alphas_cumprod, and the index is out of bounds # `set_alpha_to_zero` decides whether we set this parameter simply to zero # in this case, self.step() just output the predicted noise # or whether we use the final alpha of the "non-previous" one. __UpperCamelCase = torch.tensor(0.0 ) if set_alpha_to_zero else self.alphas_cumprod[-1] # standard deviation of the initial noise distribution __UpperCamelCase = 1.0 # setable values __UpperCamelCase = None __UpperCamelCase = torch.from_numpy(np.arange(0 , _SCREAMING_SNAKE_CASE ).copy().astype(np.intaa ) ) def __lowercase( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> torch.FloatTensor: return sample def __lowercase( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> str: if num_inference_steps > self.config.num_train_timesteps: raise ValueError( f"""`num_inference_steps`: {num_inference_steps} cannot be larger than `self.config.train_timesteps`:""" f""" {self.config.num_train_timesteps} as the unet model trained with this scheduler can only handle""" f""" maximal {self.config.num_train_timesteps} timesteps.""" ) __UpperCamelCase = num_inference_steps __UpperCamelCase = self.config.num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 __UpperCamelCase = (np.arange(0 , _SCREAMING_SNAKE_CASE ) * step_ratio).round().copy().astype(np.intaa ) __UpperCamelCase = torch.from_numpy(_SCREAMING_SNAKE_CASE ).to(_SCREAMING_SNAKE_CASE ) self.timesteps += self.config.steps_offset def __lowercase( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = 0.0 , _SCREAMING_SNAKE_CASE = False , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = True , ) -> Union[DDIMSchedulerOutput, Tuple]: # 1. get previous step value (=t+1) __UpperCamelCase = timestep + self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas # change original implementation to exactly match noise levels for analogous forward process __UpperCamelCase = self.alphas_cumprod[timestep] __UpperCamelCase = ( self.alphas_cumprod[prev_timestep] if prev_timestep < self.config.num_train_timesteps else self.final_alpha_cumprod ) __UpperCamelCase = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf if self.config.prediction_type == "epsilon": __UpperCamelCase = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 __UpperCamelCase = model_output elif self.config.prediction_type == "sample": __UpperCamelCase = model_output __UpperCamelCase = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 elif self.config.prediction_type == "v_prediction": __UpperCamelCase = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output __UpperCamelCase = (alpha_prod_t**0.5) * model_output + (beta_prod_t**0.5) * sample else: raise ValueError( f"""prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample`, or""" ' `v_prediction`' ) # 4. Clip or threshold "predicted x_0" if self.config.clip_sample: __UpperCamelCase = pred_original_sample.clamp( -self.config.clip_sample_range , self.config.clip_sample_range ) # 5. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf __UpperCamelCase = (1 - alpha_prod_t_prev) ** 0.5 * pred_epsilon # 6. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf __UpperCamelCase = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if not return_dict: return (prev_sample, pred_original_sample) return DDIMSchedulerOutput(prev_sample=_SCREAMING_SNAKE_CASE , pred_original_sample=_SCREAMING_SNAKE_CASE ) def __len__( self ) -> Any: return self.config.num_train_timesteps
567
from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { 'bigcode/gpt_bigcode-santacoder': 'https://huggingface.co/bigcode/gpt_bigcode-santacoder/resolve/main/config.json', } class lowerCAmelCase_ ( _lowercase ): """simple docstring""" UpperCAmelCase__ = "gpt_bigcode" UpperCAmelCase__ = ["past_key_values"] UpperCAmelCase__ = { "hidden_size": "n_embd", "max_position_embeddings": "n_positions", "num_attention_heads": "n_head", "num_hidden_layers": "n_layer", } def __init__( self , _SCREAMING_SNAKE_CASE=50_257 , _SCREAMING_SNAKE_CASE=1_024 , _SCREAMING_SNAKE_CASE=768 , _SCREAMING_SNAKE_CASE=12 , _SCREAMING_SNAKE_CASE=12 , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE="gelu_pytorch_tanh" , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=1e-5 , _SCREAMING_SNAKE_CASE=0.0_2 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=50_256 , _SCREAMING_SNAKE_CASE=50_256 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , **_SCREAMING_SNAKE_CASE , ) -> Optional[int]: __UpperCamelCase = vocab_size __UpperCamelCase = n_positions __UpperCamelCase = n_embd __UpperCamelCase = n_layer __UpperCamelCase = n_head __UpperCamelCase = n_inner __UpperCamelCase = activation_function __UpperCamelCase = resid_pdrop __UpperCamelCase = embd_pdrop __UpperCamelCase = attn_pdrop __UpperCamelCase = layer_norm_epsilon __UpperCamelCase = initializer_range __UpperCamelCase = scale_attn_weights __UpperCamelCase = use_cache __UpperCamelCase = attention_softmax_in_fpaa __UpperCamelCase = scale_attention_softmax_in_fpaa __UpperCamelCase = multi_query __UpperCamelCase = bos_token_id __UpperCamelCase = eos_token_id super().__init__(bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE )
567
1
from __future__ import annotations import random import unittest from transformers import TransfoXLConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST, TFTransfoXLForSequenceClassification, TFTransfoXLLMHeadModel, TFTransfoXLModel, ) class a : def __init__( self , A_ , ): '''simple docstring''' _UpperCAmelCase : Optional[Any] = parent _UpperCAmelCase : Dict = 13 _UpperCAmelCase : List[Any] = 7 _UpperCAmelCase : List[Any] = 30 _UpperCAmelCase : int = self.seq_length + self.mem_len _UpperCAmelCase : Union[str, Any] = 15 _UpperCAmelCase : Union[str, Any] = True _UpperCAmelCase : Any = True _UpperCAmelCase : Optional[int] = 99 _UpperCAmelCase : Union[str, Any] = [10, 50, 80] _UpperCAmelCase : List[str] = 32 _UpperCAmelCase : Dict = 32 _UpperCAmelCase : str = 4 _UpperCAmelCase : Optional[Any] = 8 _UpperCAmelCase : Dict = 128 _UpperCAmelCase : List[Any] = 2 _UpperCAmelCase : Any = 2 _UpperCAmelCase : Optional[int] = None _UpperCAmelCase : Union[str, Any] = 1 _UpperCAmelCase : Optional[int] = 0 _UpperCAmelCase : Optional[int] = 3 _UpperCAmelCase : Union[str, Any] = self.vocab_size - 1 _UpperCAmelCase : List[str] = 0.01 def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCAmelCase : str = None if self.use_labels: _UpperCAmelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _UpperCAmelCase : Optional[int] = TransfoXLConfig( vocab_size=self.vocab_size , mem_len=self.mem_len , clamp_len=self.clamp_len , cutoffs=self.cutoffs , d_model=self.hidden_size , d_embed=self.d_embed , n_head=self.num_attention_heads , d_head=self.d_head , d_inner=self.d_inner , div_val=self.div_val , n_layer=self.num_hidden_layers , eos_token_id=self.eos_token_id , pad_token_id=self.vocab_size - 1 , init_range=self.init_range , num_labels=self.num_labels , ) return (config, input_ids_a, input_ids_a, lm_labels) def _UpperCAmelCase ( self ): '''simple docstring''' random.seed(self.seed ) tf.random.set_seed(self.seed ) def _UpperCAmelCase ( self , A_ , A_ , A_ , A_ ): '''simple docstring''' _UpperCAmelCase : Tuple = TFTransfoXLModel(A_ ) _UpperCAmelCase , _UpperCAmelCase : Optional[int] = model(A_ ).to_tuple() _UpperCAmelCase : Dict = {"input_ids": input_ids_a, "mems": mems_a} _UpperCAmelCase , _UpperCAmelCase : Union[str, Any] = model(A_ ).to_tuple() self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(hidden_states_a.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def _UpperCAmelCase ( self , A_ , A_ , A_ , A_ ): '''simple docstring''' _UpperCAmelCase : int = TFTransfoXLLMHeadModel(A_ ) _UpperCAmelCase , _UpperCAmelCase : int = model(A_ ).to_tuple() _UpperCAmelCase : Dict = {"input_ids": input_ids_a, "labels": lm_labels} _UpperCAmelCase , _UpperCAmelCase : Any = model(A_ ).to_tuple() _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = model([input_ids_a, mems_a] ).to_tuple() _UpperCAmelCase : int = {"input_ids": input_ids_a, "mems": mems_a, "labels": lm_labels} _UpperCAmelCase , _UpperCAmelCase : Dict = model(A_ ).to_tuple() self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) self.parent.assertEqual(lm_logits_a.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertListEqual( [mem.shape for mem in mems_a] , [(self.mem_len, self.batch_size, self.hidden_size)] * self.num_hidden_layers , ) def _UpperCAmelCase ( self , A_ , A_ , A_ , A_ ): '''simple docstring''' _UpperCAmelCase : Optional[int] = TFTransfoXLForSequenceClassification(A_ ) _UpperCAmelCase : Dict = model(A_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase : Union[str, Any] = self.prepare_config_and_inputs() ((_UpperCAmelCase) , (_UpperCAmelCase) , (_UpperCAmelCase) , (_UpperCAmelCase)) : Tuple = config_and_inputs _UpperCAmelCase : Optional[Any] = {"input_ids": input_ids_a} return config, inputs_dict @require_tf class a ( UpperCAmelCase , UpperCAmelCase , unittest.TestCase ): _lowercase = ( (TFTransfoXLModel, TFTransfoXLLMHeadModel, TFTransfoXLForSequenceClassification) if is_tf_available() else () ) _lowercase = () if is_tf_available() else () _lowercase = ( { "feature-extraction": TFTransfoXLModel, "text-classification": TFTransfoXLForSequenceClassification, "text-generation": TFTransfoXLLMHeadModel, "zero-shot": TFTransfoXLForSequenceClassification, } if is_tf_available() else {} ) # TODO: add this test when TFTransfoXLLMHead has a linear output layer implemented _lowercase = False _lowercase = False _lowercase = False _lowercase = False def _UpperCAmelCase ( self , A_ , A_ , A_ , A_ , A_ ): '''simple docstring''' if pipeline_test_casse_name == "TextGenerationPipelineTests": # Get `ValueError: AttributeError: 'NoneType' object has no attribute 'new_ones'` or `AssertionError`. # `TransfoXLConfig` was never used in pipeline tests: cannot create a simple # tokenizer. return True return False def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase : int = TFTransfoXLModelTester(self ) _UpperCAmelCase : int = ConfigTester(self , config_class=A_ , d_embed=37 ) def _UpperCAmelCase ( self ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCAmelCase ( self ): '''simple docstring''' self.model_tester.set_seed() _UpperCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_model(*A_ ) def _UpperCAmelCase ( self ): '''simple docstring''' self.model_tester.set_seed() _UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_lm_head(*A_ ) def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_transfo_xl_for_sequence_classification(*A_ ) def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase : Dict = [TFTransfoXLForSequenceClassification] for model_class in self.all_model_classes: _UpperCAmelCase : int = model_class(A_ ) assert isinstance(model.get_input_embeddings() , tf.keras.layers.Layer ) if model_class in list_other_models_with_output_ebd: _UpperCAmelCase : Any = model.get_output_embeddings() assert isinstance(A_ , tf.keras.layers.Layer ) _UpperCAmelCase : Union[str, Any] = model.get_bias() assert name is None else: _UpperCAmelCase : Optional[int] = model.get_output_embeddings() assert x is None _UpperCAmelCase : List[Any] = model.get_bias() assert name is None def _UpperCAmelCase ( self ): '''simple docstring''' pass @slow def _UpperCAmelCase ( self ): '''simple docstring''' for model_name in TF_TRANSFO_XL_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _UpperCAmelCase : List[str] = TFTransfoXLModel.from_pretrained(A_ ) self.assertIsNotNone(A_ ) @unittest.skip(reason="This model doesn't play well with fit() due to not returning a single loss." ) def _UpperCAmelCase ( self ): '''simple docstring''' pass @require_tf class a ( unittest.TestCase ): @unittest.skip("Skip test until #12651 is resolved." ) @slow def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase : Union[str, Any] = TFTransfoXLLMHeadModel.from_pretrained("transfo-xl-wt103" ) # fmt: off _UpperCAmelCase : Union[str, Any] = tf.convert_to_tensor([[33,1297,2,1,1009,4,1109,11739,4762,358,5,25,245,22,1706,17,20098,5,3215,21,37,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,6224,831,16002,2,8,603,78967,29546,23,803,20,25,416,5,8,232,4,277,6,1855,4601,3,29546,54,8,3609,5,57211,49,4,1,277,18,8,1755,15691,3,341,25,416,693,42573,71,17,401,94,31,17919,2,29546,7873,18,1,435,23,11011,755,5,5167,3,7983,98,84,2,29546,3267,8,3609,4,1,4865,1075,2,6087,71,6,346,8,5854,3,29546,824,1400,1868,2,19,160,2,311,8,5496,2,20920,17,25,15097,3,24,24,0]] , dtype=tf.intaa ) # noqa: E231 # fmt: on # In 1991 , the remains of Russian Tsar Nicholas II and his family # ( except for Alexei and Maria ) are discovered . # The voice of Nicholas's young son , Tsarevich Alexei Nikolaevich , narrates the # remainder of the story . 1883 Western Siberia , # a young Grigori Rasputin is asked by his father and a group of men to perform magic . # Rasputin has a vision and denounces one of the men as a horse thief . Although his # father initially slaps him for making such an accusation , Rasputin watches as the # man is chased outside and beaten . Twenty years later , Rasputin sees a vision of # the Virgin Mary , prompting him to become a priest . Rasputin quickly becomes famous , # with people , even a bishop , begging for his blessing . <eod> </s> <eos> # fmt: off _UpperCAmelCase : List[Any] = [33,1297,2,1,1009,4,1109,11739,4762,358,5,25,245,22,1706,17,20098,5,3215,21,37,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,6224,831,16002,2,8,603,78967,29546,23,803,20,25,416,5,8,232,4,277,6,1855,4601,3,29546,54,8,3609,5,57211,49,4,1,277,18,8,1755,15691,3,341,25,416,693,42573,71,17,401,94,31,17919,2,29546,7873,18,1,435,23,11011,755,5,5167,3,7983,98,84,2,29546,3267,8,3609,4,1,4865,1075,2,6087,71,6,346,8,5854,3,29546,824,1400,1868,2,19,160,2,311,8,5496,2,20920,17,25,15097,3,24,24,0,33,1,1857,2,1,1009,4,1109,11739,4762,358,5,25,245,28,1110,3,13,1041,4,24,603,490,2,71477,20098,104447,2,20961,1,2604,4,1,329,3,0] # noqa: E231 # fmt: on # In 1991, the remains of Russian Tsar Nicholas II and his family ( # except for Alexei and Maria ) are discovered. The voice of young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story. # 1883 Western Siberia, a young Grigori Rasputin is asked by his father # and a group of men to perform magic. Rasputin has a vision and # denounces one of the men as a horse thief. Although his father initially # slaps him for making such an accusation, Rasputin watches as the man # is chased outside and beaten. Twenty years later, Rasputin sees a vision # of the Virgin Mary, prompting him to become a priest. # Rasputin quickly becomes famous, with people, even a bishop, begging for # his blessing. <unk> <unk> <eos> In the 1990s, the remains of Russian Tsar # Nicholas II and his family were discovered. The voice of <unk> young son, # Tsarevich Alexei Nikolaevich, narrates the remainder of the story.<eos> _UpperCAmelCase : List[str] = model.generate(A_ , max_length=200 , do_sample=A_ ) self.assertListEqual(output_ids[0].numpy().tolist() , A_ )
300
from __future__ import annotations def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: int , lowerCAmelCase: int ) -> list[list[int]]: _UpperCAmelCase : list[list[int]] = [] create_all_state(1 , lowerCAmelCase , lowerCAmelCase , [] , lowerCAmelCase ) return result def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: int , lowerCAmelCase: int , lowerCAmelCase: int , lowerCAmelCase: list[int] , lowerCAmelCase: list[list[int]] , ) -> None: if level == 0: total_list.append(current_list[:] ) return for i in range(lowerCAmelCase , total_number - level + 2 ): current_list.append(lowerCAmelCase ) create_all_state(i + 1 , lowerCAmelCase , level - 1 , lowerCAmelCase , lowerCAmelCase ) current_list.pop() def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: list[list[int]] ) -> None: for i in total_list: print(*lowerCAmelCase ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = 4 SCREAMING_SNAKE_CASE_ = 2 SCREAMING_SNAKE_CASE_ = generate_all_combinations(n, k) print_all_state(total_list)
300
1
"""simple docstring""" import argparse import copy def _SCREAMING_SNAKE_CASE ( __snake_case : Tuple ): '''simple docstring''' lowercase = {} with open(lowerCamelCase_ ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: lowercase = [] _list.append([line.split()[1], line.split()[2]] ) lowercase = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: lowercase = [] _list.append([line.split()[0], line.split()[2]] ) lowercase = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def _SCREAMING_SNAKE_CASE ( __snake_case : int , __snake_case : int ): '''simple docstring''' with open(lowerCamelCase_ ) as f: lowercase = f.read(1 ) lowercase = start_node lowercase = [] lowercase = start_node lowercase = 0 while visiting not in first_solution: lowercase = 1_00_00 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(lowerCamelCase_ ) and k[0] not in first_solution: lowercase = k[1] lowercase = k[0] first_solution.append(lowerCamelCase_ ) lowercase = distance_of_first_solution + int(lowerCamelCase_ ) lowercase = best_node first_solution.append(lowerCamelCase_ ) lowercase = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 lowercase = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_00_00 ) return first_solution, distance_of_first_solution def _SCREAMING_SNAKE_CASE ( __snake_case : List[Any] , __snake_case : Dict ): '''simple docstring''' lowercase = [] for n in solution[1:-1]: lowercase = solution.index(lowerCamelCase_ ) for kn in solution[1:-1]: lowercase = solution.index(lowerCamelCase_ ) if n == kn: continue lowercase = copy.deepcopy(lowerCamelCase_ ) lowercase = kn lowercase = n lowercase = 0 for k in _tmp[:-1]: lowercase = _tmp[_tmp.index(lowerCamelCase_ ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: lowercase = distance + int(i[1] ) _tmp.append(lowerCamelCase_ ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) lowercase = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda __snake_case : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def _SCREAMING_SNAKE_CASE ( __snake_case : List[str] , __snake_case : Optional[int] , __snake_case : List[Any] , __snake_case : List[str] , __snake_case : str ): '''simple docstring''' lowercase = 1 lowercase = first_solution lowercase = [] lowercase = distance_of_first_solution lowercase = solution while count <= iters: lowercase = find_neighborhood(lowerCamelCase_ , lowerCamelCase_ ) lowercase = 0 lowercase = neighborhood[index_of_best_solution] lowercase = len(lowerCamelCase_ ) - 1 lowercase = False while not found: lowercase = 0 while i < len(lowerCamelCase_ ): if best_solution[i] != solution[i]: lowercase = best_solution[i] lowercase = solution[i] break lowercase = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) lowercase = True lowercase = best_solution[:-1] lowercase = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: lowercase = cost lowercase = solution else: lowercase = index_of_best_solution + 1 lowercase = neighborhood[index_of_best_solution] if len(lowerCamelCase_ ) >= size: tabu_list.pop(0 ) lowercase = count + 1 return best_solution_ever, best_cost def _SCREAMING_SNAKE_CASE ( __snake_case : int=None ): '''simple docstring''' lowercase = generate_neighbours(args.File ) lowercase = generate_first_solution( args.File , lowerCamelCase_ ) lowercase = tabu_search( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , args.Iterations , args.Size , ) print(f'Best solution: {best_sol}, with total distance: {best_cost}.' ) if __name__ == "__main__": _UpperCamelCase : Dict = argparse.ArgumentParser(description='Tabu Search') parser.add_argument( '-f', '--File', type=str, help='Path to the file containing the data', required=True, ) parser.add_argument( '-i', '--Iterations', type=int, help='How many iterations the algorithm should perform', required=True, ) parser.add_argument( '-s', '--Size', type=int, help='Size of the tabu list', required=True ) # Pass the arguments to main method main(parser.parse_args())
706
"""simple docstring""" import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all feature extractors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...feature_extraction_utils import FeatureExtractionMixin from ...utils import CONFIG_NAME, FEATURE_EXTRACTOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) _UpperCamelCase : str = logging.get_logger(__name__) _UpperCamelCase : Dict = OrderedDict( [ ('audio-spectrogram-transformer', 'ASTFeatureExtractor'), ('beit', 'BeitFeatureExtractor'), ('chinese_clip', 'ChineseCLIPFeatureExtractor'), ('clap', 'ClapFeatureExtractor'), ('clip', 'CLIPFeatureExtractor'), ('clipseg', 'ViTFeatureExtractor'), ('conditional_detr', 'ConditionalDetrFeatureExtractor'), ('convnext', 'ConvNextFeatureExtractor'), ('cvt', 'ConvNextFeatureExtractor'), ('data2vec-audio', 'Wav2Vec2FeatureExtractor'), ('data2vec-vision', 'BeitFeatureExtractor'), ('deformable_detr', 'DeformableDetrFeatureExtractor'), ('deit', 'DeiTFeatureExtractor'), ('detr', 'DetrFeatureExtractor'), ('dinat', 'ViTFeatureExtractor'), ('donut-swin', 'DonutFeatureExtractor'), ('dpt', 'DPTFeatureExtractor'), ('encodec', 'EncodecFeatureExtractor'), ('flava', 'FlavaFeatureExtractor'), ('glpn', 'GLPNFeatureExtractor'), ('groupvit', 'CLIPFeatureExtractor'), ('hubert', 'Wav2Vec2FeatureExtractor'), ('imagegpt', 'ImageGPTFeatureExtractor'), ('layoutlmv2', 'LayoutLMv2FeatureExtractor'), ('layoutlmv3', 'LayoutLMv3FeatureExtractor'), ('levit', 'LevitFeatureExtractor'), ('maskformer', 'MaskFormerFeatureExtractor'), ('mctct', 'MCTCTFeatureExtractor'), ('mobilenet_v1', 'MobileNetV1FeatureExtractor'), ('mobilenet_v2', 'MobileNetV2FeatureExtractor'), ('mobilevit', 'MobileViTFeatureExtractor'), ('nat', 'ViTFeatureExtractor'), ('owlvit', 'OwlViTFeatureExtractor'), ('perceiver', 'PerceiverFeatureExtractor'), ('poolformer', 'PoolFormerFeatureExtractor'), ('regnet', 'ConvNextFeatureExtractor'), ('resnet', 'ConvNextFeatureExtractor'), ('segformer', 'SegformerFeatureExtractor'), ('sew', 'Wav2Vec2FeatureExtractor'), ('sew-d', 'Wav2Vec2FeatureExtractor'), ('speech_to_text', 'Speech2TextFeatureExtractor'), ('speecht5', 'SpeechT5FeatureExtractor'), ('swiftformer', 'ViTFeatureExtractor'), ('swin', 'ViTFeatureExtractor'), ('swinv2', 'ViTFeatureExtractor'), ('table-transformer', 'DetrFeatureExtractor'), ('timesformer', 'VideoMAEFeatureExtractor'), ('tvlt', 'TvltFeatureExtractor'), ('unispeech', 'Wav2Vec2FeatureExtractor'), ('unispeech-sat', 'Wav2Vec2FeatureExtractor'), ('van', 'ConvNextFeatureExtractor'), ('videomae', 'VideoMAEFeatureExtractor'), ('vilt', 'ViltFeatureExtractor'), ('vit', 'ViTFeatureExtractor'), ('vit_mae', 'ViTFeatureExtractor'), ('vit_msn', 'ViTFeatureExtractor'), ('wav2vec2', 'Wav2Vec2FeatureExtractor'), ('wav2vec2-conformer', 'Wav2Vec2FeatureExtractor'), ('wavlm', 'Wav2Vec2FeatureExtractor'), ('whisper', 'WhisperFeatureExtractor'), ('xclip', 'CLIPFeatureExtractor'), ('yolos', 'YolosFeatureExtractor'), ] ) _UpperCamelCase : str = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FEATURE_EXTRACTOR_MAPPING_NAMES) def _SCREAMING_SNAKE_CASE ( __snake_case : str ): '''simple docstring''' for module_name, extractors in FEATURE_EXTRACTOR_MAPPING_NAMES.items(): if class_name in extractors: lowercase = model_type_to_module_name(__snake_case ) lowercase = importlib.import_module(f'.{module_name}' , 'transformers.models' ) try: return getattr(__snake_case , __snake_case ) except AttributeError: continue for _, extractor in FEATURE_EXTRACTOR_MAPPING._extra_content.items(): if getattr(__snake_case , '__name__' , __snake_case ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. lowercase = importlib.import_module('transformers' ) if hasattr(__snake_case , __snake_case ): return getattr(__snake_case , __snake_case ) return None def _SCREAMING_SNAKE_CASE ( __snake_case : Union[str, os.PathLike] , __snake_case : Optional[Union[str, os.PathLike]] = None , __snake_case : bool = False , __snake_case : bool = False , __snake_case : Optional[Dict[str, str]] = None , __snake_case : Optional[Union[bool, str]] = None , __snake_case : Optional[str] = None , __snake_case : bool = False , **__snake_case : int , ): '''simple docstring''' lowercase = get_file_from_repo( __snake_case , __snake_case , cache_dir=__snake_case , force_download=__snake_case , resume_download=__snake_case , proxies=__snake_case , use_auth_token=__snake_case , revision=__snake_case , local_files_only=__snake_case , ) if resolved_config_file is None: logger.info( 'Could not locate the feature extractor configuration file, will try to use the model config instead.' ) return {} with open(__snake_case , encoding='utf-8' ) as reader: return json.load(__snake_case ) class a : def __init__( self ): raise EnvironmentError( 'AutoFeatureExtractor is designed to be instantiated ' 'using the `AutoFeatureExtractor.from_pretrained(pretrained_model_name_or_path)` method.' ) @classmethod @replace_list_option_in_docstrings(_lowerCamelCase ) def UpperCamelCase_ ( cls , _lowerCamelCase , **_lowerCamelCase ): lowercase = kwargs.pop('config' , _lowerCamelCase ) lowercase = kwargs.pop('trust_remote_code' , _lowerCamelCase ) lowercase = True lowercase , lowercase = FeatureExtractionMixin.get_feature_extractor_dict(_lowerCamelCase , **_lowerCamelCase ) lowercase = config_dict.get('feature_extractor_type' , _lowerCamelCase ) lowercase = None if "AutoFeatureExtractor" in config_dict.get('auto_map' , {} ): lowercase = config_dict['auto_map']['AutoFeatureExtractor'] # If we don't find the feature extractor class in the feature extractor config, let's try the model config. if feature_extractor_class is None and feature_extractor_auto_map is None: if not isinstance(_lowerCamelCase , _lowerCamelCase ): lowercase = AutoConfig.from_pretrained(_lowerCamelCase , **_lowerCamelCase ) # It could be in `config.feature_extractor_type`` lowercase = getattr(_lowerCamelCase , 'feature_extractor_type' , _lowerCamelCase ) if hasattr(_lowerCamelCase , 'auto_map' ) and "AutoFeatureExtractor" in config.auto_map: lowercase = config.auto_map['AutoFeatureExtractor'] if feature_extractor_class is not None: lowercase = feature_extractor_class_from_name(_lowerCamelCase ) lowercase = feature_extractor_auto_map is not None lowercase = feature_extractor_class is not None or type(_lowerCamelCase ) in FEATURE_EXTRACTOR_MAPPING lowercase = resolve_trust_remote_code( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) if has_remote_code and trust_remote_code: lowercase = get_class_from_dynamic_module( _lowerCamelCase , _lowerCamelCase , **_lowerCamelCase ) lowercase = kwargs.pop('code_revision' , _lowerCamelCase ) if os.path.isdir(_lowerCamelCase ): feature_extractor_class.register_for_auto_class() return feature_extractor_class.from_dict(_lowerCamelCase , **_lowerCamelCase ) elif feature_extractor_class is not None: return feature_extractor_class.from_dict(_lowerCamelCase , **_lowerCamelCase ) # Last try: we use the FEATURE_EXTRACTOR_MAPPING. elif type(_lowerCamelCase ) in FEATURE_EXTRACTOR_MAPPING: lowercase = FEATURE_EXTRACTOR_MAPPING[type(_lowerCamelCase )] return feature_extractor_class.from_dict(_lowerCamelCase , **_lowerCamelCase ) raise ValueError( F'Unrecognized feature extractor in {pretrained_model_name_or_path}. Should have a ' F'`feature_extractor_type` key in its {FEATURE_EXTRACTOR_NAME} of {CONFIG_NAME}, or one of the following ' F'`model_type` keys in its {CONFIG_NAME}: {", ".join(c for c in FEATURE_EXTRACTOR_MAPPING_NAMES.keys() )}' ) @staticmethod def UpperCamelCase_ ( _lowerCamelCase , _lowerCamelCase ): FEATURE_EXTRACTOR_MAPPING.register(_lowerCamelCase , _lowerCamelCase )
134
0
import numpy as np def __lowercase ( snake_case ): """simple docstring""" return 1 / (1 + np.exp(-vector )) def __lowercase ( snake_case ): """simple docstring""" return vector * sigmoid(1.702 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCamelCase : Optional[Any] = { '''configuration_timesformer''': ['''TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TimesformerConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : List[Any] = [ '''TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TimesformerModel''', '''TimesformerForVideoClassification''', '''TimesformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_timesformer import TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimesformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_timesformer import ( TIMESFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimesformerForVideoClassification, TimesformerModel, TimesformerPreTrainedModel, ) else: import sys __lowerCamelCase : Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
216
0
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : '''simple docstring''' def __init__( self : Optional[int] , UpperCamelCase : Dict , UpperCamelCase : Any=13 , UpperCamelCase : List[Any]=32 , UpperCamelCase : int=3 , UpperCamelCase : List[str]=4 , UpperCamelCase : Optional[Any]=[10, 20, 30, 40] , UpperCamelCase : Union[str, Any]=[2, 2, 3, 2] , UpperCamelCase : Tuple=True , UpperCamelCase : int=True , UpperCamelCase : Tuple=37 , UpperCamelCase : Union[str, Any]="gelu" , UpperCamelCase : str=10 , UpperCamelCase : Tuple=0.02 , UpperCamelCase : Tuple=["stage2", "stage3", "stage4"] , UpperCamelCase : Dict=3 , UpperCamelCase : Optional[Any]=None , ): '''simple docstring''' _snake_case : Union[str, Any] = parent _snake_case : Optional[int] = batch_size _snake_case : Optional[int] = image_size _snake_case : Dict = num_channels _snake_case : str = num_stages _snake_case : str = hidden_sizes _snake_case : Optional[Any] = depths _snake_case : List[Any] = is_training _snake_case : Tuple = use_labels _snake_case : Dict = intermediate_size _snake_case : List[Any] = hidden_act _snake_case : List[Any] = type_sequence_label_size _snake_case : Any = initializer_range _snake_case : List[Any] = out_features _snake_case : Optional[Any] = num_labels _snake_case : int = scope _snake_case : Union[str, Any] = num_stages def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _snake_case : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case : str = None if self.use_labels: _snake_case : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case : Union[str, Any] = self.get_config() return config, pixel_values, labels def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def UpperCamelCase_ ( self : Dict ): '''simple docstring''' return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=5_12 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=UpperCamelCase , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=2_56 , auxiliary_num_convs=1 , auxiliary_concat_input=UpperCamelCase , loss_ignore_index=2_55 , num_labels=self.num_labels , ) def UpperCamelCase_ ( self : Union[str, Any] , UpperCamelCase : Optional[Any] , UpperCamelCase : List[str] , UpperCamelCase : Union[str, Any] ): '''simple docstring''' _snake_case : Optional[Any] = UperNetForSemanticSegmentation(config=UpperCamelCase ) model.to(UpperCamelCase ) model.eval() _snake_case : Any = model(UpperCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def UpperCamelCase_ ( self : int ): '''simple docstring''' _snake_case : Tuple = self.prepare_config_and_inputs() ( ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ) : Tuple = config_and_inputs _snake_case : List[str] = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class _lowerCAmelCase ( UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): '''simple docstring''' a_ : Union[str, Any] =(UperNetForSemanticSegmentation,) if is_torch_available() else () a_ : Union[str, Any] ={"""image-segmentation""": UperNetForSemanticSegmentation} if is_torch_available() else {} a_ : int =False a_ : List[str] =False a_ : List[Any] =False a_ : Any =False a_ : Optional[int] =False a_ : Dict =False def UpperCamelCase_ ( self : Any ): '''simple docstring''' _snake_case : Union[str, Any] = UperNetModelTester(self ) _snake_case : Optional[Any] = ConfigTester(self , config_class=UpperCamelCase , has_text_modality=UpperCamelCase , hidden_size=37 ) def UpperCamelCase_ ( self : str ): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' return def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' _snake_case , _snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case : Dict = model_class(UpperCamelCase ) _snake_case : Optional[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case : Tuple = [*signature.parameters.keys()] _snake_case : str = ['pixel_values'] self.assertListEqual(arg_names[:1] , UpperCamelCase ) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*UpperCamelCase ) @unittest.skip(reason='UperNet does not use inputs_embeds' ) def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' pass @unittest.skip(reason='UperNet does not support input and output embeddings' ) def UpperCamelCase_ ( self : int ): '''simple docstring''' pass @unittest.skip(reason='UperNet does not have a base model' ) def UpperCamelCase_ ( self : int ): '''simple docstring''' pass @unittest.skip(reason='UperNet does not have a base model' ) def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip(reason='UperNet has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' pass def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' def check_hidden_states_output(UpperCamelCase : Union[str, Any] , UpperCamelCase : List[str] , UpperCamelCase : int ): _snake_case : List[Any] = model_class(UpperCamelCase ) model.to(UpperCamelCase ) model.eval() with torch.no_grad(): _snake_case : List[Any] = model(**self._prepare_for_class(UpperCamelCase , UpperCamelCase ) ) _snake_case : Any = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _snake_case : Any = self.model_tester.num_stages self.assertEqual(len(UpperCamelCase ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _snake_case , _snake_case : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case : List[Any] = True check_hidden_states_output(UpperCamelCase , UpperCamelCase , UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case : Optional[Any] = True check_hidden_states_output(UpperCamelCase , UpperCamelCase , UpperCamelCase ) def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _snake_case , _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : List[Any] = _config_zero_init(UpperCamelCase ) _snake_case : int = _config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: _snake_case : List[str] = model_class(config=UpperCamelCase ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip(reason='UperNet does not have tied weights' ) def UpperCamelCase_ ( self : Union[str, Any] ): '''simple docstring''' pass @slow def UpperCamelCase_ ( self : int ): '''simple docstring''' for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case : int = UperNetForSemanticSegmentation.from_pretrained(UpperCamelCase ) self.assertIsNotNone(UpperCamelCase ) def lowerCamelCase_ ( )-> Optional[Any]: _snake_case : List[str] = hf_hub_download( repo_id='hf-internal-testing/fixtures_ade20k' , repo_type='dataset' , filename='ADE_val_00000001.jpg' ) _snake_case : int = Image.open(lowerCAmelCase ).convert('RGB' ) return image @require_torch @require_vision @slow class _lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self : Any ): '''simple docstring''' _snake_case : Tuple = AutoImageProcessor.from_pretrained('openmmlab/upernet-swin-tiny' ) _snake_case : Tuple = UperNetForSemanticSegmentation.from_pretrained('openmmlab/upernet-swin-tiny' ).to(UpperCamelCase ) _snake_case : int = prepare_img() _snake_case : List[Any] = processor(images=UpperCamelCase , return_tensors='pt' ).to(UpperCamelCase ) with torch.no_grad(): _snake_case : Dict = model(**UpperCamelCase ) _snake_case : Union[str, Any] = torch.Size((1, model.config.num_labels, 5_12, 5_12) ) self.assertEqual(outputs.logits.shape , UpperCamelCase ) _snake_case : Optional[Any] = torch.tensor( [[-7.59_58, -7.59_58, -7.43_02], [-7.59_58, -7.59_58, -7.43_02], [-7.47_97, -7.47_97, -7.30_68]] ).to(UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCamelCase , atol=1e-4 ) ) def UpperCamelCase_ ( self : Any ): '''simple docstring''' _snake_case : List[str] = AutoImageProcessor.from_pretrained('openmmlab/upernet-convnext-tiny' ) _snake_case : Union[str, Any] = UperNetForSemanticSegmentation.from_pretrained('openmmlab/upernet-convnext-tiny' ).to(UpperCamelCase ) _snake_case : Optional[Any] = prepare_img() _snake_case : Tuple = processor(images=UpperCamelCase , return_tensors='pt' ).to(UpperCamelCase ) with torch.no_grad(): _snake_case : Dict = model(**UpperCamelCase ) _snake_case : int = torch.Size((1, model.config.num_labels, 5_12, 5_12) ) self.assertEqual(outputs.logits.shape , UpperCamelCase ) _snake_case : List[str] = torch.tensor( [[-8.81_10, -8.81_10, -8.65_21], [-8.81_10, -8.81_10, -8.65_21], [-8.77_46, -8.77_46, -8.61_30]] ).to(UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , UpperCamelCase , atol=1e-4 ) )
669
from random import randint, random def lowerCamelCase_ ( lowerCAmelCase: int , lowerCAmelCase: int , lowerCAmelCase: int , lowerCAmelCase: bool = False , lowerCAmelCase: bool = False , lowerCAmelCase: int = 5 , )-> list: _snake_case : Dict = [[-1] * number_of_cells] # Create a highway without any car _snake_case : List[str] = 0 _snake_case : List[str] = max(lowerCAmelCase , 0 ) while i < number_of_cells: _snake_case : Optional[Any] = ( randint(0 , lowerCAmelCase ) if random_speed else initial_speed ) # Place the cars i += ( randint(1 , max_speed * 2 ) if random_frequency else frequency ) # Arbitrary number, may need tuning return highway def lowerCamelCase_ ( lowerCAmelCase: list , lowerCAmelCase: int )-> int: _snake_case : Dict = 0 _snake_case : Optional[Any] = highway_now[car_index + 1 :] for cell in range(len(lowerCAmelCase ) ): # May need a better name for this if cells[cell] != -1: # If the cell is not empty then return distance # we have the distance we wanted distance += 1 # Here if the car is near the end of the highway return distance + get_distance(lowerCAmelCase , -1 ) def lowerCamelCase_ ( lowerCAmelCase: list , lowerCAmelCase: float , lowerCAmelCase: int )-> list: _snake_case : List[Any] = len(lowerCAmelCase ) # Beforce calculations, the highway is empty _snake_case : List[Any] = [-1] * number_of_cells for car_index in range(lowerCAmelCase ): if highway_now[car_index] != -1: # Add 1 to the current speed of the car and cap the speed _snake_case : int = min(highway_now[car_index] + 1 , lowerCAmelCase ) # Number of empty cell before the next car _snake_case : Tuple = get_distance(lowerCAmelCase , lowerCAmelCase ) - 1 # We can't have the car causing an accident _snake_case : Union[str, Any] = min(next_highway[car_index] , lowerCAmelCase ) if random() < probability: # Randomly, a driver will slow down _snake_case : int = max(next_highway[car_index] - 1 , 0 ) return next_highway def lowerCamelCase_ ( lowerCAmelCase: list , lowerCAmelCase: int , lowerCAmelCase: float , lowerCAmelCase: int )-> list: _snake_case : Dict = len(highway[0] ) for i in range(lowerCAmelCase ): _snake_case : Any = update(highway[i] , lowerCAmelCase , lowerCAmelCase ) _snake_case : Tuple = [-1] * number_of_cells for car_index in range(lowerCAmelCase ): _snake_case : Union[str, Any] = next_speeds_calculated[car_index] if speed != -1: # Change the position based on the speed (with % to create the loop) _snake_case : Union[str, Any] = (car_index + speed) % number_of_cells # Commit the change of position _snake_case : Tuple = speed highway.append(lowerCAmelCase ) return highway if __name__ == "__main__": import doctest doctest.testmod()
669
1
import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType __A = logging.get_logger(__name__) class _A ( UpperCamelCase ): """simple docstring""" lowerCamelCase : Union[str, Any] = 'vision-encoder-decoder' lowerCamelCase : int = True def __init__( self : Union[str, Any] , **__SCREAMING_SNAKE_CASE : int ) -> Dict: super().__init__(**__SCREAMING_SNAKE_CASE ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( f'''A configuraton of type {self.model_type} cannot be instantiated because ''' f'''not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}''' ) __UpperCAmelCase =kwargs.pop("""encoder""" ) __UpperCAmelCase =encoder_config.pop("""model_type""" ) __UpperCAmelCase =kwargs.pop("""decoder""" ) __UpperCAmelCase =decoder_config.pop("""model_type""" ) __UpperCAmelCase =AutoConfig.for_model(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =AutoConfig.for_model(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =True @classmethod def _a ( cls : List[str] , __SCREAMING_SNAKE_CASE : PretrainedConfig , __SCREAMING_SNAKE_CASE : PretrainedConfig , **__SCREAMING_SNAKE_CASE : str ) -> PretrainedConfig: logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) __UpperCAmelCase =True __UpperCAmelCase =True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **__SCREAMING_SNAKE_CASE ) def _a ( self : Union[str, Any] ) -> Tuple: __UpperCAmelCase =copy.deepcopy(self.__dict__ ) __UpperCAmelCase =self.encoder.to_dict() __UpperCAmelCase =self.decoder.to_dict() __UpperCAmelCase =self.__class__.model_type return output class _A ( UpperCamelCase ): """simple docstring""" lowerCamelCase : List[Any] = version.parse('1.11' ) @property def _a ( self : Tuple ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def _a ( self : Any ) -> float: return 1e-4 @property def _a ( self : str ) -> Mapping[str, Mapping[int, str]]: return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class _A ( UpperCamelCase ): """simple docstring""" @property def _a ( self : List[Any] ) -> Mapping[str, Mapping[int, str]]: __UpperCAmelCase =OrderedDict() __UpperCAmelCase ={0: """batch""", 1: """past_decoder_sequence + sequence"""} __UpperCAmelCase ={0: """batch""", 1: """past_decoder_sequence + sequence"""} __UpperCAmelCase ={0: """batch""", 1: """encoder_sequence"""} return common_inputs def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : "PreTrainedTokenizerBase" , __SCREAMING_SNAKE_CASE : int = -1 , __SCREAMING_SNAKE_CASE : int = -1 , __SCREAMING_SNAKE_CASE : bool = False , __SCREAMING_SNAKE_CASE : Optional["TensorType"] = None , ) -> Mapping[str, Any]: import torch __UpperCAmelCase =OrderedDict() __UpperCAmelCase =super().generate_dummy_inputs( __SCREAMING_SNAKE_CASE , batch_size=__SCREAMING_SNAKE_CASE , seq_length=__SCREAMING_SNAKE_CASE , is_pair=__SCREAMING_SNAKE_CASE , framework=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase , __UpperCAmelCase =dummy_input["""input_ids"""].shape __UpperCAmelCase =(batch, encoder_sequence, self._config.encoder_hidden_size) __UpperCAmelCase =dummy_input.pop("""input_ids""" ) __UpperCAmelCase =dummy_input.pop("""attention_mask""" ) __UpperCAmelCase =torch.zeros(__SCREAMING_SNAKE_CASE ) return common_inputs class _A ( UpperCamelCase ): """simple docstring""" @property def _a ( self : List[Any] ) -> None: pass def _a ( self : List[Any] , __SCREAMING_SNAKE_CASE : PretrainedConfig ) -> OnnxConfig: return VisionEncoderDecoderEncoderOnnxConfig(__SCREAMING_SNAKE_CASE ) def _a ( self : Dict , __SCREAMING_SNAKE_CASE : PretrainedConfig , __SCREAMING_SNAKE_CASE : PretrainedConfig , __SCREAMING_SNAKE_CASE : str = "default" ) -> OnnxConfig: __UpperCAmelCase =encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )
68
"""simple docstring""" def lowercase_ ( __UpperCAmelCase ) -> str: return " ".join( """""".join(word[::-1] ) if len(__UpperCAmelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words("""Hey wollef sroirraw"""))
299
0
import inspect import unittest from transformers import MobileViTVaConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, MobileViTVaModel from transformers.models.mobilevitva.modeling_mobilevitva import ( MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST, make_divisible, ) if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class lowerCAmelCase_ ( lowerCamelCase__ ): '''simple docstring''' def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : Optional[int] =self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__lowerCamelCase , 'width_multiplier' ) ) class lowerCAmelCase_ : '''simple docstring''' def __init__( self , __UpperCAmelCase , __UpperCAmelCase=13 , __UpperCAmelCase=64 , __UpperCAmelCase=2 , __UpperCAmelCase=3 , __UpperCAmelCase="swish" , __UpperCAmelCase=3 , __UpperCAmelCase=32 , __UpperCAmelCase=0.1 , __UpperCAmelCase=0.02 , __UpperCAmelCase=True , __UpperCAmelCase=True , __UpperCAmelCase=10 , __UpperCAmelCase=None , __UpperCAmelCase=0.25 , __UpperCAmelCase=0.0 , __UpperCAmelCase=0.0 , ): SCREAMING_SNAKE_CASE_ : Dict =parent SCREAMING_SNAKE_CASE_ : List[str] =batch_size SCREAMING_SNAKE_CASE_ : Dict =image_size SCREAMING_SNAKE_CASE_ : Tuple =patch_size SCREAMING_SNAKE_CASE_ : List[Any] =num_channels SCREAMING_SNAKE_CASE_ : Tuple =make_divisible(512 * width_multiplier , divisor=8 ) SCREAMING_SNAKE_CASE_ : Dict =hidden_act SCREAMING_SNAKE_CASE_ : Optional[int] =conv_kernel_size SCREAMING_SNAKE_CASE_ : int =output_stride SCREAMING_SNAKE_CASE_ : List[Any] =classifier_dropout_prob SCREAMING_SNAKE_CASE_ : Dict =use_labels SCREAMING_SNAKE_CASE_ : Tuple =is_training SCREAMING_SNAKE_CASE_ : str =num_labels SCREAMING_SNAKE_CASE_ : Optional[Any] =initializer_range SCREAMING_SNAKE_CASE_ : Tuple =scope SCREAMING_SNAKE_CASE_ : Tuple =width_multiplier SCREAMING_SNAKE_CASE_ : Optional[int] =ffn_dropout SCREAMING_SNAKE_CASE_ : Dict =attn_dropout def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : str =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE_ : Union[str, Any] =None SCREAMING_SNAKE_CASE_ : Dict =None if self.use_labels: SCREAMING_SNAKE_CASE_ : Optional[Any] =ids_tensor([self.batch_size] , self.num_labels ) SCREAMING_SNAKE_CASE_ : Optional[Any] =ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) SCREAMING_SNAKE_CASE_ : Tuple =self.get_config() return config, pixel_values, labels, pixel_labels def __lowerCamelCase ( self ): return MobileViTVaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , width_multiplier=self.width_multiplier , ffn_dropout=self.ffn_dropout_prob , attn_dropout=self.attn_dropout_prob , ) def __lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : int =MobileViTVaModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE_ : Dict =model(__lowerCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : List[str] =self.num_labels SCREAMING_SNAKE_CASE_ : Union[str, Any] =MobileViTVaForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE_ : Dict =model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCamelCase ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : Any =self.num_labels SCREAMING_SNAKE_CASE_ : Optional[int] =MobileViTVaForSemanticSegmentation(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE_ : Optional[Any] =model(__lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) SCREAMING_SNAKE_CASE_ : List[str] =model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : Union[str, Any] =self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE_ : Optional[int] =config_and_inputs SCREAMING_SNAKE_CASE_ : Dict ={'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase_ ( lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): '''simple docstring''' _lowercase = ( (MobileViTVaModel, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation) if is_torch_available() else () ) _lowercase = ( { 'feature-extraction': MobileViTVaModel, 'image-classification': MobileViTVaForImageClassification, 'image-segmentation': MobileViTVaForSemanticSegmentation, } if is_torch_available() else {} ) _lowercase = False _lowercase = False _lowercase = False _lowercase = False def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : Dict =MobileViTVaModelTester(self ) SCREAMING_SNAKE_CASE_ : Optional[int] =MobileViTVaConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase ) def __lowerCamelCase ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='MobileViTV2 does not use inputs_embeds' ) def __lowerCamelCase ( self ): pass @unittest.skip(reason='MobileViTV2 does not support input and output embeddings' ) def __lowerCamelCase ( self ): pass @unittest.skip(reason='MobileViTV2 does not output attentions' ) def __lowerCamelCase ( self ): pass @require_torch_multi_gpu @unittest.skip(reason='Got `CUDA error: misaligned address` for tests after this one being run.' ) def __lowerCamelCase ( self ): pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowerCamelCase ( self ): pass def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : List[Any] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ : Union[str, Any] =model_class(__lowerCamelCase ) SCREAMING_SNAKE_CASE_ : List[Any] =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE_ : str =[*signature.parameters.keys()] SCREAMING_SNAKE_CASE_ : Union[str, Any] =['''pixel_values'''] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : List[Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def __lowerCamelCase ( self ): def check_hidden_states_output(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : Union[str, Any] =model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() with torch.no_grad(): SCREAMING_SNAKE_CASE_ : Optional[Any] =model(**self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) ) SCREAMING_SNAKE_CASE_ : Dict =outputs.hidden_states SCREAMING_SNAKE_CASE_ : Any =5 self.assertEqual(len(__lowerCamelCase ) , __lowerCamelCase ) # MobileViTV2's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. SCREAMING_SNAKE_CASE_ : Optional[Any] =2 for i in range(len(__lowerCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) SCREAMING_SNAKE_CASE_ : Union[str, Any] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ : Any =True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] SCREAMING_SNAKE_CASE_ : Tuple =True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : Optional[Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : str =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*__lowerCamelCase ) @slow def __lowerCamelCase ( self ): for model_name in MOBILEVITV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE_ : Optional[int] =MobileViTVaModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict =Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class lowerCAmelCase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def __lowerCamelCase ( self ): return ( MobileViTImageProcessor.from_pretrained('apple/mobilevitv2-1.0-imagenet1k-256' ) if is_vision_available() else None ) @slow def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : str =MobileViTVaForImageClassification.from_pretrained('apple/mobilevitv2-1.0-imagenet1k-256' ).to( __lowerCamelCase ) SCREAMING_SNAKE_CASE_ : str =self.default_image_processor SCREAMING_SNAKE_CASE_ : Union[str, Any] =prepare_img() SCREAMING_SNAKE_CASE_ : int =image_processor(images=__lowerCamelCase , return_tensors='pt' ).to(__lowerCamelCase ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE_ : Union[str, Any] =model(**__lowerCamelCase ) # verify the logits SCREAMING_SNAKE_CASE_ : str =torch.Size((1, 1_000) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) SCREAMING_SNAKE_CASE_ : Dict =torch.tensor([-1.6_336E00, -7.3_204E-02, -5.1_883E-01] ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1E-4 ) ) @slow def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : Optional[int] =MobileViTVaForSemanticSegmentation.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) SCREAMING_SNAKE_CASE_ : Optional[Any] =model.to(__lowerCamelCase ) SCREAMING_SNAKE_CASE_ : List[Any] =MobileViTImageProcessor.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) SCREAMING_SNAKE_CASE_ : Optional[Any] =prepare_img() SCREAMING_SNAKE_CASE_ : Union[str, Any] =image_processor(images=__lowerCamelCase , return_tensors='pt' ).to(__lowerCamelCase ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE_ : List[Any] =model(**__lowerCamelCase ) SCREAMING_SNAKE_CASE_ : Tuple =outputs.logits # verify the logits SCREAMING_SNAKE_CASE_ : Optional[int] =torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , __lowerCamelCase ) SCREAMING_SNAKE_CASE_ : List[Any] =torch.tensor( [ [[7.0_863, 7.1_525, 6.8_201], [6.6_931, 6.8_770, 6.8_933], [6.2_978, 7.0_366, 6.9_636]], [[-3.7_134, -3.6_712, -3.6_675], [-3.5_825, -3.3_549, -3.4_777], [-3.3_435, -3.3_979, -3.2_857]], [[-2.9_329, -2.8_003, -2.7_369], [-3.0_564, -2.4_780, -2.0_207], [-2.6_889, -1.9_298, -1.7_640]], ] , device=__lowerCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , __lowerCamelCase , atol=1E-4 ) ) @slow def __lowerCamelCase ( self ): SCREAMING_SNAKE_CASE_ : Dict =MobileViTVaForSemanticSegmentation.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) SCREAMING_SNAKE_CASE_ : Dict =model.to(__lowerCamelCase ) SCREAMING_SNAKE_CASE_ : Tuple =MobileViTImageProcessor.from_pretrained('shehan97/mobilevitv2-1.0-voc-deeplabv3' ) SCREAMING_SNAKE_CASE_ : int =prepare_img() SCREAMING_SNAKE_CASE_ : Any =image_processor(images=__lowerCamelCase , return_tensors='pt' ).to(__lowerCamelCase ) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE_ : List[Any] =model(**__lowerCamelCase ) SCREAMING_SNAKE_CASE_ : Optional[Any] =outputs.logits.detach().cpu() SCREAMING_SNAKE_CASE_ : Optional[Any] =image_processor.post_process_semantic_segmentation(outputs=__lowerCamelCase , target_sizes=[(50, 60)] ) SCREAMING_SNAKE_CASE_ : Tuple =torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , __lowerCamelCase ) SCREAMING_SNAKE_CASE_ : Optional[Any] =image_processor.post_process_semantic_segmentation(outputs=__lowerCamelCase ) SCREAMING_SNAKE_CASE_ : Optional[Any] =torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , __lowerCamelCase )
702
from __future__ import annotations __SCREAMING_SNAKE_CASE = '#' class lowerCAmelCase_ : '''simple docstring''' def __init__( self ): SCREAMING_SNAKE_CASE_ : dict ={} def __lowerCamelCase ( self , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : Tuple =self._trie for char in text: if char not in trie: SCREAMING_SNAKE_CASE_ : Optional[int] ={} SCREAMING_SNAKE_CASE_ : Any =trie[char] SCREAMING_SNAKE_CASE_ : Optional[Any] =True def __lowerCamelCase ( self , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : Tuple =self._trie for char in prefix: if char in trie: SCREAMING_SNAKE_CASE_ : Tuple =trie[char] else: return [] return self._elements(__UpperCAmelCase ) def __lowerCamelCase ( self , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : Optional[int] =[] for c, v in d.items(): SCREAMING_SNAKE_CASE_ : List[Any] =[' '] if c == END else [(c + s) for s in self._elements(__UpperCAmelCase )] result.extend(__UpperCAmelCase ) return tuple(__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = Trie() __SCREAMING_SNAKE_CASE = ('depart', 'detergent', 'daring', 'dog', 'deer', 'deal') for word in words: trie.insert_word(word) def SCREAMING_SNAKE_CASE__ ( lowerCAmelCase_ : str ) -> tuple: """simple docstring""" SCREAMING_SNAKE_CASE_ : str =trie.find_word(lowerCAmelCase_ ) return tuple(string + word for word in suffixes ) def SCREAMING_SNAKE_CASE__ ( ) -> None: """simple docstring""" print(autocomplete_using_trie('de' ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
153
0
import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def __init__( self : Optional[Any] , __lowerCamelCase : Tuple , __lowerCamelCase : List[str]=13 , __lowerCamelCase : List[str]=7 , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : Dict=True , __lowerCamelCase : Optional[int]=True , __lowerCamelCase : Optional[int]=99 , __lowerCamelCase : Tuple=32 , __lowerCamelCase : Any=5 , __lowerCamelCase : int=4 , __lowerCamelCase : Any=37 , __lowerCamelCase : Any="gelu" , __lowerCamelCase : int=0.1 , __lowerCamelCase : Optional[int]=0.1 , __lowerCamelCase : Optional[Any]=512 , __lowerCamelCase : Union[str, Any]=16 , __lowerCamelCase : Any=2 , __lowerCamelCase : List[Any]=0.02 , __lowerCamelCase : int=4 , ): SCREAMING_SNAKE_CASE = parent SCREAMING_SNAKE_CASE = batch_size SCREAMING_SNAKE_CASE = seq_length SCREAMING_SNAKE_CASE = is_training SCREAMING_SNAKE_CASE = use_attention_mask SCREAMING_SNAKE_CASE = use_token_type_ids SCREAMING_SNAKE_CASE = use_labels SCREAMING_SNAKE_CASE = vocab_size SCREAMING_SNAKE_CASE = hidden_size SCREAMING_SNAKE_CASE = num_hidden_layers SCREAMING_SNAKE_CASE = num_attention_heads SCREAMING_SNAKE_CASE = intermediate_size SCREAMING_SNAKE_CASE = hidden_act SCREAMING_SNAKE_CASE = hidden_dropout_prob SCREAMING_SNAKE_CASE = attention_probs_dropout_prob SCREAMING_SNAKE_CASE = max_position_embeddings SCREAMING_SNAKE_CASE = type_vocab_size SCREAMING_SNAKE_CASE = type_sequence_label_size SCREAMING_SNAKE_CASE = initializer_range SCREAMING_SNAKE_CASE = num_choices def _snake_case ( self : str ): SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE = None if self.use_attention_mask: SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) SCREAMING_SNAKE_CASE = None if self.use_token_type_ids: SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) SCREAMING_SNAKE_CASE = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _snake_case ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = config_and_inputs SCREAMING_SNAKE_CASE = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class _SCREAMING_SNAKE_CASE ( __snake_case , unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def _snake_case ( self : List[str] ): SCREAMING_SNAKE_CASE = FlaxAlbertModelTester(self ) @slow def _snake_case ( self : Any ): for model_class_name in self.all_model_classes: SCREAMING_SNAKE_CASE = model_class_name.from_pretrained("albert-base-v2" ) SCREAMING_SNAKE_CASE = model(np.ones((1, 1) ) ) self.assertIsNotNone(__lowerCamelCase ) @require_flax class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' @slow def _snake_case ( self : str ): SCREAMING_SNAKE_CASE = FlaxAlbertModel.from_pretrained("albert-base-v2" ) SCREAMING_SNAKE_CASE = np.array([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) SCREAMING_SNAKE_CASE = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) SCREAMING_SNAKE_CASE = model(__lowerCamelCase , attention_mask=__lowerCamelCase )[0] SCREAMING_SNAKE_CASE = (1, 11, 768) self.assertEqual(output.shape , __lowerCamelCase ) SCREAMING_SNAKE_CASE = np.array( [[[-0.6_513, 1.5_035, -0.2_766], [-0.6_515, 1.5_046, -0.2_780], [-0.6_512, 1.5_049, -0.2_784]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , __lowerCamelCase , atol=1e-4 ) )
16
from diffusers.utils.testing_utils import require_onnxruntime @require_onnxruntime class lowercase_ : '''simple docstring''' pass
117
0
"""simple docstring""" import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, require_tf, slow if is_tf_available(): import tensorflow as tf if is_tensorflow_text_available(): from transformers.models.bert import TFBertTokenizer snake_case = ['bert-base-uncased', 'bert-base-cased'] snake_case = 'hf-internal-testing/tiny-bert-tf-only' if is_tf_available(): class UpperCamelCase ( tf.keras.Model ): """simple docstring""" def __init__( self , lowercase__ ) -> List[Any]: """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE = tokenizer SCREAMING_SNAKE_CASE = AutoConfig.from_pretrained(lowercase__ ) SCREAMING_SNAKE_CASE = TFAutoModel.from_config(lowercase__ ) def A ( self , lowercase__ ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE = self.tokenizer(lowercase__ ) SCREAMING_SNAKE_CASE = self.bert(**lowercase__ ) return out["pooler_output"] @require_tf @require_tensorflow_text class UpperCamelCase ( unittest.TestCase ): """simple docstring""" def A ( self ) -> Optional[int]: """simple docstring""" super().setUp() SCREAMING_SNAKE_CASE = [ BertTokenizer.from_pretrained(lowercase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS * 2) ] # repeat for when fast_bert_tokenizer=false SCREAMING_SNAKE_CASE = [TFBertTokenizer.from_pretrained(lowercase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] + [ TFBertTokenizer.from_pretrained(lowercase__ , use_fast_bert_tokenizer=lowercase__ ) for checkpoint in TOKENIZER_CHECKPOINTS ] assert len(self.tokenizers ) == len(self.tf_tokenizers ) SCREAMING_SNAKE_CASE = [ 'This is a straightforward English test sentence.', 'This one has some weird characters\rto\nsee\r\nif those\u00E9break things.', 'Now we\'re going to add some Chinese: 一 二 三 一二三', 'And some much more rare Chinese: 齉 堃 齉堃', 'Je vais aussi écrire en français pour tester les accents', 'Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ', ] SCREAMING_SNAKE_CASE = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def A ( self ) -> Any: """simple docstring""" for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in (self.test_sentences, self.paired_sentences): SCREAMING_SNAKE_CASE = tokenizer(lowercase__ , return_tensors='tf' , padding='longest' ) SCREAMING_SNAKE_CASE = tf_tokenizer(lowercase__ ) for key in python_outputs.keys(): self.assertTrue(tf.reduce_all(python_outputs[key].shape == tf_outputs[key].shape ) ) self.assertTrue(tf.reduce_all(tf.cast(python_outputs[key] , tf.intaa ) == tf_outputs[key] ) ) @slow def A ( self ) -> List[str]: """simple docstring""" for tf_tokenizer in self.tf_tokenizers: SCREAMING_SNAKE_CASE = tf_tokenizer(self.paired_sentences ) SCREAMING_SNAKE_CASE = tf_tokenizer( text=[sentence[0] for sentence in self.paired_sentences] , text_pair=[sentence[1] for sentence in self.paired_sentences] , ) for key in merged_outputs.keys(): self.assertTrue(tf.reduce_all(tf.cast(merged_outputs[key] , tf.intaa ) == separated_outputs[key] ) ) @slow def A ( self ) -> Tuple: """simple docstring""" for tf_tokenizer in self.tf_tokenizers: SCREAMING_SNAKE_CASE = tf.function(lowercase__ ) for test_inputs in (self.test_sentences, self.paired_sentences): SCREAMING_SNAKE_CASE = tf.constant(lowercase__ ) SCREAMING_SNAKE_CASE = compiled_tokenizer(lowercase__ ) SCREAMING_SNAKE_CASE = tf_tokenizer(lowercase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def A ( self ) -> Dict: """simple docstring""" for tf_tokenizer in self.tf_tokenizers: SCREAMING_SNAKE_CASE = ModelToSave(tokenizer=lowercase__ ) SCREAMING_SNAKE_CASE = tf.convert_to_tensor(self.test_sentences ) SCREAMING_SNAKE_CASE = model(lowercase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: SCREAMING_SNAKE_CASE = Path(lowercase__ ) / 'saved.model' model.save(lowercase__ ) SCREAMING_SNAKE_CASE = tf.keras.models.load_model(lowercase__ ) SCREAMING_SNAKE_CASE = loaded_model(lowercase__ ) # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertLessEqual(tf.reduce_max(tf.abs(out - loaded_output ) ) , 1E-5 )
406
"""simple docstring""" import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging snake_case = logging.get_logger(__name__) snake_case = R'\n Args:\n input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`):\n Indices of input sequence tokens in the vocabulary.\n\n Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and\n [`PreTrainedTokenizer.__call__`] for details.\n\n [What are input IDs?](../glossary#input-ids)\n scores (`torch.FloatTensor` of shape `(batch_size, config.vocab_size)`):\n Prediction scores of a language modeling head. These can be scores for each vocabulary token before SoftMax\n or scores for each vocabulary token after SoftMax.\n kwargs (`Dict[str, Any]`, *optional*):\n Additional stopping criteria specific kwargs.\n\n Return:\n `bool`. `False` indicates we should continue, `True` indicates we should stop.\n\n' class UpperCamelCase ( __magic_name__ ): """simple docstring""" @add_start_docstrings(lowercase__ ) def __call__( self , lowercase__ , lowercase__ , **lowercase__ ) -> bool: """simple docstring""" raise NotImplementedError('StoppingCriteria needs to be subclassed' ) class UpperCamelCase ( __magic_name__ ): """simple docstring""" def __init__( self , lowercase__ , lowercase__ = None ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE = max_length SCREAMING_SNAKE_CASE = max_position_embeddings @add_start_docstrings(lowercase__ ) def __call__( self , lowercase__ , lowercase__ , **lowercase__ ) -> bool: """simple docstring""" SCREAMING_SNAKE_CASE = input_ids.shape[-1] SCREAMING_SNAKE_CASE = cur_len >= self.max_length if self.max_position_embeddings is not None and not is_done and cur_len >= self.max_position_embeddings: logger.warning_once( 'This is a friendly reminder - the current text generation call will exceed the model\'s predefined ' f'''maximum length ({self.max_position_embeddings}). Depending on the model, you may observe ''' 'exceptions, performance degradation, or nothing at all.' ) return is_done class UpperCamelCase ( __magic_name__ ): """simple docstring""" def __init__( self , lowercase__ , lowercase__ ) -> Optional[int]: """simple docstring""" warnings.warn( 'The class `MaxNewTokensCriteria` is deprecated. ' f'''Please use `MaxLengthCriteria(max_length={start_length + max_new_tokens})` ''' 'with `max_length = start_length + max_new_tokens` instead.' , lowercase__ , ) SCREAMING_SNAKE_CASE = start_length SCREAMING_SNAKE_CASE = max_new_tokens SCREAMING_SNAKE_CASE = start_length + max_new_tokens @add_start_docstrings(lowercase__ ) def __call__( self , lowercase__ , lowercase__ , **lowercase__ ) -> bool: """simple docstring""" return input_ids.shape[-1] >= self.max_length class UpperCamelCase ( __magic_name__ ): """simple docstring""" def __init__( self , lowercase__ , lowercase__ = None ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE = max_time SCREAMING_SNAKE_CASE = time.time() if initial_timestamp is None else initial_timestamp @add_start_docstrings(lowercase__ ) def __call__( self , lowercase__ , lowercase__ , **lowercase__ ) -> bool: """simple docstring""" return time.time() - self.initial_timestamp > self.max_time class UpperCamelCase ( __magic_name__ ): """simple docstring""" @add_start_docstrings(lowercase__ ) def __call__( self , lowercase__ , lowercase__ , **lowercase__ ) -> bool: """simple docstring""" return any(criteria(lowercase__ , lowercase__ ) for criteria in self ) @property def A ( self ) -> Optional[int]: """simple docstring""" for stopping_criterium in self: if isinstance(lowercase__ , lowercase__ ): return stopping_criterium.max_length elif isinstance(lowercase__ , lowercase__ ): return stopping_criterium.max_length return None def UpperCamelCase_ ( SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ): SCREAMING_SNAKE_CASE = stopping_criteria.max_length SCREAMING_SNAKE_CASE = deepcopy(SCREAMING_SNAKE_CASE_ ) if stopping_max_length is not None and stopping_max_length != max_length: warnings.warn('You set different `max_length` for stopping criteria and `max_length` parameter', SCREAMING_SNAKE_CASE_ ) elif stopping_max_length is None: new_stopping_criteria.append(MaxLengthCriteria(max_length=SCREAMING_SNAKE_CASE_ ) ) return new_stopping_criteria
406
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE__:List[str] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:int = {"""vocab_file""": """spm_char.model"""} SCREAMING_SNAKE_CASE__:Union[str, Any] = { """vocab_file""": { """microsoft/speecht5_asr""": """https://huggingface.co/microsoft/speecht5_asr/resolve/main/spm_char.model""", """microsoft/speecht5_tts""": """https://huggingface.co/microsoft/speecht5_tts/resolve/main/spm_char.model""", """microsoft/speecht5_vc""": """https://huggingface.co/microsoft/speecht5_vc/resolve/main/spm_char.model""", } } SCREAMING_SNAKE_CASE__:Dict = { """microsoft/speecht5_asr""": 1024, """microsoft/speecht5_tts""": 1024, """microsoft/speecht5_vc""": 1024, } class snake_case__ ( snake_case_ ): _snake_case : List[str] = VOCAB_FILES_NAMES _snake_case : List[str] = PRETRAINED_VOCAB_FILES_MAP _snake_case : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _snake_case : Any = ["""input_ids""", """attention_mask"""] def __init__( self , lowerCamelCase , lowerCamelCase="<s>" , lowerCamelCase="</s>" , lowerCamelCase="<unk>" , lowerCamelCase="<pad>" , lowerCamelCase = None , **lowerCamelCase , ): __a = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowerCamelCase , eos_token=lowerCamelCase , unk_token=lowerCamelCase , pad_token=lowerCamelCase , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase , ) __a = vocab_file __a = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowerCamelCase ) @property def a__ ( self ): return self.sp_model.get_piece_size() def a__ ( self ): __a = {self.convert_ids_to_tokens(lowerCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): __a = self.__dict__.copy() __a = None return state def __setstate__( self , lowerCamelCase ): __a = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): __a = {} __a = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def a__ ( self , lowerCamelCase ): return self.sp_model.encode(lowerCamelCase , out_type=lowerCamelCase ) def a__ ( self , lowerCamelCase ): return self.sp_model.piece_to_id(lowerCamelCase ) def a__ ( self , lowerCamelCase ): __a = self.sp_model.IdToPiece(lowerCamelCase ) return token def a__ ( self , lowerCamelCase ): __a = [] __a = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(lowerCamelCase ) + token __a = [] else: current_sub_tokens.append(lowerCamelCase ) out_string += self.sp_model.decode(lowerCamelCase ) return out_string.strip() def a__ ( self , lowerCamelCase , lowerCamelCase=None ): if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def a__ ( self , lowerCamelCase , lowerCamelCase = None , lowerCamelCase = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase , token_ids_a=lowerCamelCase , already_has_special_tokens=lowerCamelCase ) __a = [1] if token_ids_a is None: return ([0] * len(lowerCamelCase )) + suffix_ones return ([0] * len(lowerCamelCase )) + ([0] * len(lowerCamelCase )) + suffix_ones def a__ ( self , lowerCamelCase , lowerCamelCase = None ): if not os.path.isdir(lowerCamelCase ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return __a = os.path.join( lowerCamelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowerCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(lowerCamelCase , "wb" ) as fi: __a = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase ) return (out_vocab_file,)
528
"""simple docstring""" import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def _lowerCamelCase( ): raise RuntimeError("CUDA out of memory." ) class snake_case__ ( nn.Module ): def __init__( self ): super().__init__() __a = nn.Linear(3 , 4 ) __a = nn.BatchNormad(4 ) __a = nn.Linear(4 , 5 ) def a__ ( self , lowerCamelCase ): return self.lineara(self.batchnorm(self.lineara(lowerCamelCase ) ) ) class snake_case__ ( unittest.TestCase ): def a__ ( self ): __a = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(lowerCamelCase ): nonlocal batch_sizes batch_sizes.append(lowerCamelCase ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(lowerCamelCase , [128, 64, 32, 16, 8] ) def a__ ( self ): __a = [] @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(lowerCamelCase , lowerCamelCase ): nonlocal batch_sizes batch_sizes.append(lowerCamelCase ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga __a , __a = mock_training_loop_function("hello" ) self.assertListEqual(lowerCamelCase , [128, 64, 32, 16, 8] ) self.assertListEqual([bs, arga] , [8, "hello"] ) def a__ ( self ): @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(lowerCamelCase ): pass with self.assertRaises(lowerCamelCase ) as cm: mock_training_loop_function() self.assertIn("No executable batch size found, reached zero." , cm.exception.args[0] ) def a__ ( self ): @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(lowerCamelCase ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(lowerCamelCase ) as cm: mock_training_loop_function() self.assertIn("No executable batch size found, reached zero." , cm.exception.args[0] ) def a__ ( self ): @find_executable_batch_size(starting_batch_size=128 ) def mock_training_loop_function(lowerCamelCase , lowerCamelCase , lowerCamelCase ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(lowerCamelCase ) as cm: mock_training_loop_function(128 , "hello" , "world" ) self.assertIn("Batch size was passed into `f`" , cm.exception.args[0] ) self.assertIn("`f(arg1='hello', arg2='world')" , cm.exception.args[0] ) def a__ ( self ): @find_executable_batch_size(starting_batch_size=16 ) def mock_training_loop_function(lowerCamelCase ): raise ValueError("Oops, we had an error!" ) with self.assertRaises(lowerCamelCase ) as cm: mock_training_loop_function() self.assertIn("Oops, we had an error!" , cm.exception.args[0] ) @require_cuda def a__ ( self ): __a = torch.cuda.memory_allocated() __a = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , lowerCamelCase ) __a = release_memory(lowerCamelCase ) self.assertEqual(torch.cuda.memory_allocated() , lowerCamelCase )
528
1
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation _lowercase: Dict = logging.get_logger(__name__) _lowercase: Any = {'''tokenizer_file''': '''tokenizer.json'''} _lowercase: Dict = { '''tokenizer_file''': { '''bigscience/tokenizer''': '''https://huggingface.co/bigscience/tokenizer/blob/main/tokenizer.json''', '''bigscience/bloom-560m''': '''https://huggingface.co/bigscience/bloom-560m/blob/main/tokenizer.json''', '''bigscience/bloom-1b1''': '''https://huggingface.co/bigscience/bloom-1b1/blob/main/tokenizer.json''', '''bigscience/bloom-1b7''': '''https://huggingface.co/bigscience/bloom-1b7/blob/main/tokenizer.json''', '''bigscience/bloom-3b''': '''https://huggingface.co/bigscience/bloom-3b/blob/main/tokenizer.json''', '''bigscience/bloom-7b1''': '''https://huggingface.co/bigscience/bloom-7b1/blob/main/tokenizer.json''', '''bigscience/bloom''': '''https://huggingface.co/bigscience/bloom/blob/main/tokenizer.json''', }, } class lowerCamelCase__ ( UpperCAmelCase ): UpperCamelCase__ =VOCAB_FILES_NAMES UpperCamelCase__ =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ =["input_ids", "attention_mask"] UpperCamelCase__ =None def __init__( self : Any , lowercase__ : List[Any]=None , lowercase__ : Any=None , lowercase__ : int=None , lowercase__ : List[Any]="<unk>" , lowercase__ : Any="<s>" , lowercase__ : List[Any]="</s>" , lowercase__ : Any="<pad>" , lowercase__ : Tuple=False , lowercase__ : str=False , **lowercase__ : Tuple , ): super().__init__( lowercase__ , lowercase__ , tokenizer_file=lowercase__ , unk_token=lowercase__ , bos_token=lowercase__ , eos_token=lowercase__ , pad_token=lowercase__ , add_prefix_space=lowercase__ , clean_up_tokenization_spaces=lowercase__ , **lowercase__ , ) _lowerCAmelCase = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , lowercase__ ) != add_prefix_space: _lowerCAmelCase = getattr(lowercase__ , pre_tok_state.pop('type' ) ) _lowerCAmelCase = add_prefix_space _lowerCAmelCase = pre_tok_class(**lowercase__ ) _lowerCAmelCase = add_prefix_space def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , *lowercase__ : str , **lowercase__ : int ): _lowerCAmelCase = kwargs.get('is_split_into_words' , lowercase__ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with' ' pretokenized inputs.' ) return super()._batch_encode_plus(*lowercase__ , **lowercase__ ) def SCREAMING_SNAKE_CASE__ ( self : Dict , *lowercase__ : Optional[Any] , **lowercase__ : str ): _lowerCAmelCase = kwargs.get('is_split_into_words' , lowercase__ ) if not (self.add_prefix_space or not is_split_into_words): raise Exception( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True to use it with' ' pretokenized inputs.' ) return super()._encode_plus(*lowercase__ , **lowercase__ ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , lowercase__ : str , lowercase__ : Optional[str] = None ): _lowerCAmelCase = self._tokenizer.model.save(lowercase__ , name=lowercase__ ) return tuple(lowercase__ ) def SCREAMING_SNAKE_CASE__ ( self : Tuple , lowercase__ : "Conversation" ): _lowerCAmelCase = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowercase__ , add_special_tokens=lowercase__ ) + [self.eos_token_id] ) if len(lowercase__ ) > self.model_max_length: _lowerCAmelCase = input_ids[-self.model_max_length :] return input_ids
225
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase: List[str] = logging.get_logger(__name__) _lowercase: Optional[Any] = { '''tiiuae/falcon-40b''': '''https://huggingface.co/tiiuae/falcon-40b/resolve/main/config.json''', '''tiiuae/falcon-7b''': '''https://huggingface.co/tiiuae/falcon-7b/resolve/main/config.json''', } class lowerCamelCase__ ( UpperCAmelCase ): UpperCamelCase__ ="falcon" UpperCamelCase__ =["past_key_values"] def __init__( self : Optional[Any] , lowercase__ : List[Any]=6_50_24 , lowercase__ : Optional[Any]=45_44 , lowercase__ : int=32 , lowercase__ : List[Any]=71 , lowercase__ : Any=1e-5 , lowercase__ : Dict=0.0_2 , lowercase__ : Union[str, Any]=True , lowercase__ : Optional[Any]=0.0 , lowercase__ : int=0.0 , lowercase__ : Optional[Any]=None , lowercase__ : List[Any]=False , lowercase__ : Tuple=False , lowercase__ : int=True , lowercase__ : List[Any]=True , lowercase__ : Optional[Any]=False , lowercase__ : Optional[Any]=11 , lowercase__ : Optional[Any]=11 , **lowercase__ : Union[str, Any] , ): _lowerCAmelCase = vocab_size # Backward compatibility with n_embed kwarg _lowerCAmelCase = kwargs.pop('n_embed' , lowercase__ ) _lowerCAmelCase = hidden_size if n_embed is None else n_embed _lowerCAmelCase = num_hidden_layers _lowerCAmelCase = num_attention_heads _lowerCAmelCase = layer_norm_epsilon _lowerCAmelCase = initializer_range _lowerCAmelCase = use_cache _lowerCAmelCase = hidden_dropout _lowerCAmelCase = attention_dropout _lowerCAmelCase = bos_token_id _lowerCAmelCase = eos_token_id _lowerCAmelCase = num_attention_heads if num_kv_heads is None else num_kv_heads _lowerCAmelCase = alibi _lowerCAmelCase = new_decoder_architecture _lowerCAmelCase = multi_query # Ignored when new_decoder_architecture is True _lowerCAmelCase = parallel_attn _lowerCAmelCase = bias super().__init__(bos_token_id=lowercase__ , eos_token_id=lowercase__ , **lowercase__ ) @property def SCREAMING_SNAKE_CASE__ ( self : Tuple ): return self.hidden_size // self.num_attention_heads @property def SCREAMING_SNAKE_CASE__ ( self : List[Any] ): return not self.alibi
225
1
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowercase_ (lowercase__ ): snake_case =['image_processor', 'tokenizer'] snake_case ='LayoutLMv2ImageProcessor' snake_case =('LayoutXLMTokenizer', 'LayoutXLMTokenizerFast') def __init__( self , lowercase_=None , lowercase_=None , **lowercase_) -> List[str]: if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , lowercase_ , ) a__ =kwargs.pop('feature_extractor') a__ =image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.') if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.') super().__init__(lowercase_ , lowercase_) def __call__( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = True , lowercase_ = False , lowercase_ = None , lowercase_ = None , lowercase_ = 0 , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = False , lowercase_ = False , lowercase_ = False , lowercase_ = False , lowercase_ = True , lowercase_ = None , **lowercase_ , ) -> BatchEncoding: # verify input if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( 'You cannot provide bounding boxes ' 'if you initialized the image processor with apply_ocr set to True.') if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( 'You cannot provide word labels if you initialized the image processor with apply_ocr set to True.') if return_overflowing_tokens is True and return_offsets_mapping is False: raise ValueError('You cannot return overflowing tokens without returning the offsets mapping.') # first, apply the image processor a__ =self.image_processor(images=lowercase_ , return_tensors=lowercase_) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(lowercase_ , lowercase_): a__ =[text] # add batch dimension (as the image processor always adds a batch dimension) a__ =features['words'] a__ =self.tokenizer( text=text if text is not None else features['words'] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features['boxes'] , word_labels=lowercase_ , add_special_tokens=lowercase_ , padding=lowercase_ , truncation=lowercase_ , max_length=lowercase_ , stride=lowercase_ , pad_to_multiple_of=lowercase_ , return_token_type_ids=lowercase_ , return_attention_mask=lowercase_ , return_overflowing_tokens=lowercase_ , return_special_tokens_mask=lowercase_ , return_offsets_mapping=lowercase_ , return_length=lowercase_ , verbose=lowercase_ , return_tensors=lowercase_ , **lowercase_ , ) # add pixel values a__ =features.pop('pixel_values') if return_overflowing_tokens is True: a__ =self.get_overflowing_images(lowercase_ , encoded_inputs['overflow_to_sample_mapping']) a__ =images return encoded_inputs def __UpperCamelCase ( self , lowercase_ , lowercase_) -> Dict: # in case there's an overflow, ensure each `input_ids` sample is mapped to its corresponding image a__ =[] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx]) if len(lowercase_) != len(lowercase_): raise ValueError( 'Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got' F""" {len(lowercase_)} and {len(lowercase_)}""") return images_with_overflow def __UpperCamelCase ( self , *lowercase_ , **lowercase_) -> Optional[Any]: return self.tokenizer.batch_decode(*lowercase_ , **lowercase_) def __UpperCamelCase ( self , *lowercase_ , **lowercase_) -> str: return self.tokenizer.decode(*lowercase_ , **lowercase_) @property def __UpperCamelCase ( self) -> Union[str, Any]: return ["input_ids", "bbox", "attention_mask", "image"] @property def __UpperCamelCase ( self) -> List[str]: warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , lowercase_ , ) return self.image_processor_class @property def __UpperCamelCase ( self) -> Optional[int]: warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , lowercase_ , ) return self.image_processor
20
import os import re import sys import traceback import warnings from pathlib import Path from typing import Dict, Optional, Union from uuid import uuida from huggingface_hub import HfFolder, ModelCard, ModelCardData, hf_hub_download, whoami from huggingface_hub.file_download import REGEX_COMMIT_HASH from huggingface_hub.utils import ( EntryNotFoundError, RepositoryNotFoundError, RevisionNotFoundError, is_jinja_available, ) from packaging import version from requests import HTTPError from .. import __version__ from .constants import ( DEPRECATED_REVISION_ARGS, DIFFUSERS_CACHE, HUGGINGFACE_CO_RESOLVE_ENDPOINT, SAFETENSORS_WEIGHTS_NAME, WEIGHTS_NAME, ) from .import_utils import ( ENV_VARS_TRUE_VALUES, _flax_version, _jax_version, _onnxruntime_version, _torch_version, is_flax_available, is_onnx_available, is_torch_available, ) from .logging import get_logger _lowerCAmelCase: Tuple = get_logger(__name__) _lowerCAmelCase: List[str] = Path(__file__).parent / 'model_card_template.md' _lowerCAmelCase: Any = uuida().hex _lowerCAmelCase: List[Any] = os.getenv('HF_HUB_OFFLINE', '').upper() in ENV_VARS_TRUE_VALUES _lowerCAmelCase: int = os.getenv('DISABLE_TELEMETRY', '').upper() in ENV_VARS_TRUE_VALUES _lowerCAmelCase: Tuple = HUGGINGFACE_CO_RESOLVE_ENDPOINT + '/api/telemetry/' def _lowercase( __a : Union[Dict, str, None] = None ): a__ =f"""diffusers/{__version__}; python/{sys.version.split()[0]}; session_id/{SESSION_ID}""" if DISABLE_TELEMETRY or HF_HUB_OFFLINE: return ua + "; telemetry/off" if is_torch_available(): ua += f"""; torch/{_torch_version}""" if is_flax_available(): ua += f"""; jax/{_jax_version}""" ua += f"""; flax/{_flax_version}""" if is_onnx_available(): ua += f"""; onnxruntime/{_onnxruntime_version}""" # CI will set this value to True if os.environ.get('DIFFUSERS_IS_CI' , '' ).upper() in ENV_VARS_TRUE_VALUES: ua += "; is_ci/true" if isinstance(__a , __a ): ua += "; " + "; ".join(f"""{k}/{v}""" for k, v in user_agent.items() ) elif isinstance(__a , __a ): ua += "; " + user_agent return ua def _lowercase( __a : str , __a : Optional[str] = None , __a : Optional[str] = None ): if token is None: a__ =HfFolder.get_token() if organization is None: a__ =whoami(__a )['name'] return f"""{username}/{model_id}""" else: return f"""{organization}/{model_id}""" def _lowercase( __a : Union[str, Any] , __a : Dict ): if not is_jinja_available(): raise ValueError( 'Modelcard rendering is based on Jinja templates.' ' Please make sure to have `jinja` installed before using `create_model_card`.' ' To install it, please run `pip install Jinja2`.' ) if hasattr(__a , 'local_rank' ) and args.local_rank not in [-1, 0]: return a__ =args.hub_token if hasattr(__a , 'hub_token' ) else None a__ =get_full_repo_name(__a , token=__a ) a__ =ModelCard.from_template( card_data=ModelCardData( # Card metadata object that will be converted to YAML block language='en' , license='apache-2.0' , library_name='diffusers' , tags=[] , datasets=args.dataset_name , metrics=[] , ) , template_path=__a , model_name=__a , repo_name=__a , dataset_name=args.dataset_name if hasattr(__a , 'dataset_name' ) else None , learning_rate=args.learning_rate , train_batch_size=args.train_batch_size , eval_batch_size=args.eval_batch_size , gradient_accumulation_steps=( args.gradient_accumulation_steps if hasattr(__a , 'gradient_accumulation_steps' ) else None ) , adam_betaa=args.adam_betaa if hasattr(__a , 'adam_beta1' ) else None , adam_betaa=args.adam_betaa if hasattr(__a , 'adam_beta2' ) else None , adam_weight_decay=args.adam_weight_decay if hasattr(__a , 'adam_weight_decay' ) else None , adam_epsilon=args.adam_epsilon if hasattr(__a , 'adam_epsilon' ) else None , lr_scheduler=args.lr_scheduler if hasattr(__a , 'lr_scheduler' ) else None , lr_warmup_steps=args.lr_warmup_steps if hasattr(__a , 'lr_warmup_steps' ) else None , ema_inv_gamma=args.ema_inv_gamma if hasattr(__a , 'ema_inv_gamma' ) else None , ema_power=args.ema_power if hasattr(__a , 'ema_power' ) else None , ema_max_decay=args.ema_max_decay if hasattr(__a , 'ema_max_decay' ) else None , mixed_precision=args.mixed_precision , ) a__ =os.path.join(args.output_dir , 'README.md' ) model_card.save(__a ) def _lowercase( __a : Optional[str] , __a : Optional[str] = None ): if resolved_file is None or commit_hash is not None: return commit_hash a__ =str(Path(__a ).as_posix() ) a__ =re.search(r'snapshots/([^/]+)/' , __a ) if search is None: return None a__ =search.groups()[0] return commit_hash if REGEX_COMMIT_HASH.match(__a ) else None # Old default cache path, potentially to be migrated. # This logic was more or less taken from `transformers`, with the following differences: # - Diffusers doesn't use custom environment variables to specify the cache path. # - There is no need to migrate the cache format, just move the files to the new location. _lowerCAmelCase: List[str] = os.path.expanduser( os.getenv('HF_HOME', os.path.join(os.getenv('XDG_CACHE_HOME', '~/.cache'), 'huggingface')) ) _lowerCAmelCase: List[str] = os.path.join(hf_cache_home, 'diffusers') def _lowercase( __a : Optional[str] = None , __a : Optional[str] = None ): if new_cache_dir is None: a__ =DIFFUSERS_CACHE if old_cache_dir is None: a__ =old_diffusers_cache a__ =Path(__a ).expanduser() a__ =Path(__a ).expanduser() for old_blob_path in old_cache_dir.glob('**/blobs/*' ): if old_blob_path.is_file() and not old_blob_path.is_symlink(): a__ =new_cache_dir / old_blob_path.relative_to(__a ) new_blob_path.parent.mkdir(parents=__a , exist_ok=__a ) os.replace(__a , __a ) try: os.symlink(__a , __a ) except OSError: logger.warning( 'Could not create symlink between old cache and new cache. If you use an older version of diffusers again, files will be re-downloaded.' ) # At this point, old_cache_dir contains symlinks to the new cache (it can still be used). _lowerCAmelCase: Dict = os.path.join(DIFFUSERS_CACHE, 'version_diffusers_cache.txt') if not os.path.isfile(cache_version_file): _lowerCAmelCase: int = 0 else: with open(cache_version_file) as f: try: _lowerCAmelCase: List[Any] = int(f.read()) except ValueError: _lowerCAmelCase: Any = 0 if cache_version < 1: _lowerCAmelCase: str = os.path.isdir(old_diffusers_cache) and len(os.listdir(old_diffusers_cache)) > 0 if old_cache_is_not_empty: logger.warning( 'The cache for model files in Diffusers v0.14.0 has moved to a new location. Moving your ' 'existing cached models. This is a one-time operation, you can interrupt it or run it ' 'later by calling `diffusers.utils.hub_utils.move_cache()`.' ) try: move_cache() except Exception as e: _lowerCAmelCase: Optional[Any] = '\n'.join(traceback.format_tb(e.__traceback__)) logger.error( F"""There was a problem when trying to move your cache:\n\n{trace}\n{e.__class__.__name__}: {e}\n\nPlease """ 'file an issue at https://github.com/huggingface/diffusers/issues/new/choose, copy paste this whole ' 'message and we will do our best to help.' ) if cache_version < 1: try: os.makedirs(DIFFUSERS_CACHE, exist_ok=True) with open(cache_version_file, 'w') as f: f.write('1') except Exception: logger.warning( F"""There was a problem when trying to write in your cache folder ({DIFFUSERS_CACHE}). Please, ensure """ 'the directory exists and can be written to.' ) def _lowercase( __a : str , __a : Optional[str] = None ): if variant is not None: a__ =weights_name.split('.' ) a__ =splits[:-1] + [variant] + splits[-1:] a__ ='.'.join(__a ) return weights_name def _lowercase( __a : Union[str, Any] , *, __a : Optional[Any] , __a : Optional[Any] , __a : List[Any] , __a : Tuple , __a : Optional[Any] , __a : Dict , __a : str , __a : int , __a : Tuple , __a : Union[str, Any] , __a : int=None , ): a__ =str(__a ) if os.path.isfile(__a ): return pretrained_model_name_or_path elif os.path.isdir(__a ): if os.path.isfile(os.path.join(__a , __a ) ): # Load from a PyTorch checkpoint a__ =os.path.join(__a , __a ) return model_file elif subfolder is not None and os.path.isfile( os.path.join(__a , __a , __a ) ): a__ =os.path.join(__a , __a , __a ) return model_file else: raise EnvironmentError( f"""Error no file named {weights_name} found in directory {pretrained_model_name_or_path}.""" ) else: # 1. First check if deprecated way of loading from branches is used if ( revision in DEPRECATED_REVISION_ARGS and (weights_name == WEIGHTS_NAME or weights_name == SAFETENSORS_WEIGHTS_NAME) and version.parse(version.parse(__a ).base_version ) >= version.parse('0.20.0' ) ): try: a__ =hf_hub_download( __a , filename=_add_variant(__a , __a ) , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , user_agent=__a , subfolder=__a , revision=revision or commit_hash , ) warnings.warn( f"""Loading the variant {revision} from {pretrained_model_name_or_path} via `revision='{revision}'` is deprecated. Loading instead from `revision='main'` with `variant={revision}`. Loading model variants via `revision='{revision}'` will be removed in diffusers v1. Please use `variant='{revision}'` instead.""" , __a , ) return model_file except: # noqa: E722 warnings.warn( f"""You are loading the variant {revision} from {pretrained_model_name_or_path} via `revision='{revision}'`. This behavior is deprecated and will be removed in diffusers v1. One should use `variant='{revision}'` instead. However, it appears that {pretrained_model_name_or_path} currently does not have a {_add_variant(__a , __a )} file in the 'main' branch of {pretrained_model_name_or_path}. \n The Diffusers team and community would be very grateful if you could open an issue: https://github.com/huggingface/diffusers/issues/new with the title '{pretrained_model_name_or_path} is missing {_add_variant(__a , __a )}' so that the correct variant file can be added.""" , __a , ) try: # 2. Load model file as usual a__ =hf_hub_download( __a , filename=__a , cache_dir=__a , force_download=__a , proxies=__a , resume_download=__a , local_files_only=__a , use_auth_token=__a , user_agent=__a , subfolder=__a , revision=revision or commit_hash , ) return model_file except RepositoryNotFoundError: raise EnvironmentError( f"""{pretrained_model_name_or_path} is not a local folder and is not a valid model identifier """ 'listed on \'https://huggingface.co/models\'\nIf this is a private repository, make sure to pass a ' 'token having permission to this repo with `use_auth_token` or log in with `huggingface-cli ' 'login`.' ) except RevisionNotFoundError: raise EnvironmentError( f"""{revision} is not a valid git identifier (branch name, tag name or commit id) that exists for """ 'this model name. Check the model page at ' f"""'https://huggingface.co/{pretrained_model_name_or_path}' for available revisions.""" ) except EntryNotFoundError: raise EnvironmentError( f"""{pretrained_model_name_or_path} does not appear to have a file named {weights_name}.""" ) except HTTPError as err: raise EnvironmentError( f"""There was a specific connection error when trying to load {pretrained_model_name_or_path}:\n{err}""" ) except ValueError: raise EnvironmentError( f"""We couldn't connect to '{HUGGINGFACE_CO_RESOLVE_ENDPOINT}' to load this model, couldn't find it""" f""" in the cached files and it looks like {pretrained_model_name_or_path} is not the path to a""" f""" directory containing a file named {weights_name} or""" ' \nCheckout your internet connection or see how to run the library in' ' offline mode at \'https://huggingface.co/docs/diffusers/installation#offline-mode\'.' ) except EnvironmentError: raise EnvironmentError( f"""Can't load the model for '{pretrained_model_name_or_path}'. If you were trying to load it from """ '\'https://huggingface.co/models\', make sure you don\'t have a local directory with the same name. ' f"""Otherwise, make sure '{pretrained_model_name_or_path}' is the correct path to a directory """ f"""containing a file named {weights_name}""" )
20
1
'''simple docstring''' from __future__ import annotations def _UpperCamelCase ( __UpperCamelCase ,__UpperCamelCase ) -> int: if len(__UpperCamelCase ) < k or k < 0: raise ValueError('Invalid Input' ) lowerCamelCase_ = lowerCamelCase_ = sum(array[:k] ) for i in range(len(__UpperCamelCase ) - k ): lowerCamelCase_ = current_sum - array[i] + array[i + k] lowerCamelCase_ = max(__UpperCamelCase ,__UpperCamelCase ) return max_sum if __name__ == "__main__": from doctest import testmod from random import randint testmod() A_ = [randint(-1_000, 1_000) for i in range(100)] A_ = randint(0, 110) print(f'''The maximum sum of {k} consecutive elements is {max_sum_in_array(array,k)}''')
712
'''simple docstring''' import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class UpperCAmelCase : '''simple docstring''' def __init__( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=13 , SCREAMING_SNAKE_CASE_=2 , SCREAMING_SNAKE_CASE_=24 , SCREAMING_SNAKE_CASE_=16 , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=32 , SCREAMING_SNAKE_CASE_=5 , SCREAMING_SNAKE_CASE_=4 , SCREAMING_SNAKE_CASE_=37 , SCREAMING_SNAKE_CASE_="gelu" , SCREAMING_SNAKE_CASE_=0.1 , SCREAMING_SNAKE_CASE_=0.1 , SCREAMING_SNAKE_CASE_=10 , SCREAMING_SNAKE_CASE_=0.02 , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=2 , SCREAMING_SNAKE_CASE_=2 , ) -> Dict: '''simple docstring''' lowerCamelCase_ = parent lowerCamelCase_ = batch_size lowerCamelCase_ = patch_size lowerCamelCase_ = max_length lowerCamelCase_ = num_mel_bins lowerCamelCase_ = is_training lowerCamelCase_ = use_labels lowerCamelCase_ = hidden_size lowerCamelCase_ = num_hidden_layers lowerCamelCase_ = num_attention_heads lowerCamelCase_ = intermediate_size lowerCamelCase_ = hidden_act lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = type_sequence_label_size lowerCamelCase_ = initializer_range lowerCamelCase_ = scope lowerCamelCase_ = frequency_stride lowerCamelCase_ = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) lowerCamelCase_ = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 lowerCamelCase_ = (self.max_length - self.patch_size) // self.time_stride + 1 lowerCamelCase_ = frequency_out_dimension * time_out_dimension lowerCamelCase_ = num_patches + 2 def UpperCamelCase( self ) -> int: '''simple docstring''' lowerCamelCase_ = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) lowerCamelCase_ = None if self.use_labels: lowerCamelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase_ = self.get_config() return config, input_values, labels def UpperCamelCase( self ) -> str: '''simple docstring''' return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=SCREAMING_SNAKE_CASE_ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def UpperCamelCase( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> List[str]: '''simple docstring''' lowerCamelCase_ = ASTModel(config=SCREAMING_SNAKE_CASE_ ) model.to(SCREAMING_SNAKE_CASE_ ) model.eval() lowerCamelCase_ = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase( self ) -> Optional[Any]: '''simple docstring''' lowerCamelCase_ = self.prepare_config_and_inputs() ( ( lowerCamelCase_ ) ,( lowerCamelCase_ ) ,( lowerCamelCase_ ) , ) = config_and_inputs lowerCamelCase_ = {'input_values': input_values} return config, inputs_dict @require_torch class UpperCAmelCase ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE_ = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE_ = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE_ = False SCREAMING_SNAKE_CASE_ = False SCREAMING_SNAKE_CASE_ = False SCREAMING_SNAKE_CASE_ = False def UpperCamelCase( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) -> List[Any]: '''simple docstring''' if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def UpperCamelCase( self ) -> int: '''simple docstring''' lowerCamelCase_ = ASTModelTester(self ) lowerCamelCase_ = ConfigTester(self , config_class=SCREAMING_SNAKE_CASE_ , has_text_modality=SCREAMING_SNAKE_CASE_ , hidden_size=37 ) def UpperCamelCase( self ) -> Any: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='AST does not use inputs_embeds' ) def UpperCamelCase( self ) -> Any: '''simple docstring''' pass def UpperCamelCase( self ) -> Tuple: '''simple docstring''' lowerCamelCase_ ,lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(SCREAMING_SNAKE_CASE_ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowerCamelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(SCREAMING_SNAKE_CASE_ , nn.Linear ) ) def UpperCamelCase( self ) -> Optional[int]: '''simple docstring''' lowerCamelCase_ ,lowerCamelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase_ = model_class(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase_ = [*signature.parameters.keys()] lowerCamelCase_ = ['input_values'] self.assertListEqual(arg_names[:1] , SCREAMING_SNAKE_CASE_ ) def UpperCamelCase( self ) -> str: '''simple docstring''' lowerCamelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_ ) @slow def UpperCamelCase( self ) -> Optional[int]: '''simple docstring''' for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase_ = ASTModel.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsNotNone(SCREAMING_SNAKE_CASE_ ) def _UpperCamelCase ( ) -> Tuple: lowerCamelCase_ = hf_hub_download( repo_id='nielsr/audio-spectogram-transformer-checkpoint' ,filename='sample_audio.flac' ,repo_type='dataset' ) lowerCamelCase_ ,lowerCamelCase_ = torchaudio.load(__UpperCamelCase ) return audio, sampling_rate @require_torch @require_torchaudio class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase( self ) -> int: '''simple docstring''' return ( ASTFeatureExtractor.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ) if is_torchaudio_available() else None ) @slow def UpperCamelCase( self ) -> Optional[int]: '''simple docstring''' lowerCamelCase_ = self.default_feature_extractor lowerCamelCase_ = ASTForAudioClassification.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ).to(SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = self.default_feature_extractor lowerCamelCase_ ,lowerCamelCase_ = prepare_audio() lowerCamelCase_ = audio.squeeze().numpy() lowerCamelCase_ = feature_extractor(SCREAMING_SNAKE_CASE_ , sampling_rate=SCREAMING_SNAKE_CASE_ , return_tensors='pt' ).to(SCREAMING_SNAKE_CASE_ ) # forward pass with torch.no_grad(): lowerCamelCase_ = model(**SCREAMING_SNAKE_CASE_ ) # verify the logits lowerCamelCase_ = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , SCREAMING_SNAKE_CASE_ ) lowerCamelCase_ = torch.tensor([-0.8_760, -7.0_042, -8.6_602] ).to(SCREAMING_SNAKE_CASE_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , SCREAMING_SNAKE_CASE_ , atol=1E-4 ) )
384
0
import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaProcessor, is_apex_available, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse("1.6"): __A : Optional[int] = True from torch.cuda.amp import autocast __A : Dict = logging.getLogger(__name__) def __lowerCAmelCase( _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None ) -> Tuple: """simple docstring""" return field(default_factory=lambda: default , metadata=_SCREAMING_SNAKE_CASE ) @dataclass class lowerCamelCase: '''simple docstring''' __magic_name__ = field( metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} ) __magic_name__ = field( default=lowerCamelCase_ , metadata={'help': 'Where do you want to store the pretrained models downloaded from huggingface.co'} , ) __magic_name__ = field( default=lowerCamelCase_ , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) __magic_name__ = field( default=0.1 , metadata={'help': 'The dropout ratio for the attention probabilities.'} ) __magic_name__ = field( default=0.1 , metadata={'help': 'The dropout ratio for activations inside the fully connected layer.'} ) __magic_name__ = field( default=0.1 , metadata={ 'help': 'The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler.' } , ) __magic_name__ = field( default=0.1 , metadata={'help': 'The dropout probabilitiy for all 1D convolutional layers in feature extractor.'} , ) __magic_name__ = field( default=0.05 , metadata={ 'help': ( 'Propability of each feature vector along the time axis to be chosen as the start of the vector' 'span to be masked. Approximately ``mask_time_prob * sequence_length // mask_time_length`` feature' 'vectors will be masked along the time axis. This is only relevant if ``apply_spec_augment is True``.' ) } , ) __magic_name__ = field(default=0.0 , metadata={'help': 'The LayerDrop probability.'} ) @dataclass class lowerCamelCase: '''simple docstring''' __magic_name__ = field( default=lowerCamelCase_ , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) __magic_name__ = field( default='train+validation' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) __magic_name__ = field( default=lowerCamelCase_ , metadata={'help': 'Overwrite the cached preprocessed datasets or not.'} ) __magic_name__ = field( default=lowerCamelCase_ , metadata={'help': 'The number of processes to use for the preprocessing.'} , ) __magic_name__ = field( default=lowerCamelCase_ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) __magic_name__ = field( default=lowerCamelCase_ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of validation examples to this ' 'value if set.' ) } , ) __magic_name__ = list_field( default=[',', '?', '.', '!', '-', ';', ':', '""', '%', '\'', '"', '�'] , metadata={'help': 'A list of characters to remove from the transcripts.'} , ) @dataclass class lowerCamelCase: '''simple docstring''' __magic_name__ = 42 __magic_name__ = True __magic_name__ = None __magic_name__ = None __magic_name__ = None __magic_name__ = None def __call__( self , snake_case_ ): _A = [{"input_values": feature["input_values"]} for feature in features] _A = [{"input_ids": feature["labels"]} for feature in features] _A = self.processor.pad( snake_case_ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='pt' , ) _A = self.processor.pad( labels=snake_case_ , padding=self.padding , max_length=self.max_length_labels , pad_to_multiple_of=self.pad_to_multiple_of_labels , return_tensors='pt' , ) # replace padding with -100 to ignore loss correctly _A = labels_batch["input_ids"].masked_fill(labels_batch.attention_mask.ne(1 ) , -100 ) _A = labels return batch class lowerCamelCase( lowerCamelCase_ ): '''simple docstring''' def lowerCAmelCase__ ( self , snake_case_ , snake_case_ ): model.train() _A = self._prepare_inputs(snake_case_ ) if self.use_amp: with autocast(): _A = self.compute_loss(snake_case_ , snake_case_ ) else: _A = self.compute_loss(snake_case_ , snake_case_ ) if self.args.n_gpu > 1: if model.module.config.ctc_loss_reduction == "mean": _A = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": _A = loss.sum() / (inputs["labels"] >= 0).sum() else: raise ValueError(F"{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']" ) if self.args.gradient_accumulation_steps > 1: _A = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(snake_case_ ).backward() elif self.use_apex: with amp.scale_loss(snake_case_ , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(snake_case_ ) else: loss.backward() return loss.detach() def __lowerCAmelCase( ) -> Optional[Any]: """simple docstring""" _A = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _A = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _A = parser.parse_args_into_dataclasses() # Detecting last checkpoint. _A = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _A = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F"Output directory ({training_args.output_dir}) already exists and is not empty. " 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None: logger.info( F"Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change " 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( F"Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}" + F"distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}" ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info('Training/evaluation parameters %s' , _SCREAMING_SNAKE_CASE ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: _A = datasets.load_dataset( 'common_voice' , data_args.dataset_config_name , split=data_args.train_split_name ) _A = datasets.load_dataset('common_voice' , data_args.dataset_config_name , split='test' ) # Create and save tokenizer _A = F"[{''.join(data_args.chars_to_ignore )}]" def remove_special_characters(_SCREAMING_SNAKE_CASE ): _A = re.sub(_SCREAMING_SNAKE_CASE , '' , batch['sentence'] ).lower() + " " return batch _A = train_dataset.map(_SCREAMING_SNAKE_CASE , remove_columns=['sentence'] ) _A = eval_dataset.map(_SCREAMING_SNAKE_CASE , remove_columns=['sentence'] ) def extract_all_chars(_SCREAMING_SNAKE_CASE ): _A = " ".join(batch['text'] ) _A = list(set(_SCREAMING_SNAKE_CASE ) ) return {"vocab": [vocab], "all_text": [all_text]} _A = train_dataset.map( _SCREAMING_SNAKE_CASE , batched=_SCREAMING_SNAKE_CASE , batch_size=-1 , keep_in_memory=_SCREAMING_SNAKE_CASE , remove_columns=train_dataset.column_names , ) _A = train_dataset.map( _SCREAMING_SNAKE_CASE , batched=_SCREAMING_SNAKE_CASE , batch_size=-1 , keep_in_memory=_SCREAMING_SNAKE_CASE , remove_columns=eval_dataset.column_names , ) _A = list(set(vocab_train['vocab'][0] ) | set(vocab_test['vocab'][0] ) ) _A = {v: k for k, v in enumerate(_SCREAMING_SNAKE_CASE )} _A = vocab_dict[" "] del vocab_dict[" "] _A = len(_SCREAMING_SNAKE_CASE ) _A = len(_SCREAMING_SNAKE_CASE ) with open('vocab.json' , 'w' ) as vocab_file: json.dump(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _A = WavaVecaCTCTokenizer( 'vocab.json' , unk_token='[UNK]' , pad_token='[PAD]' , word_delimiter_token='|' , ) _A = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16_000 , padding_value=0.0 , do_normalize=_SCREAMING_SNAKE_CASE , return_attention_mask=_SCREAMING_SNAKE_CASE ) _A = WavaVecaProcessor(feature_extractor=_SCREAMING_SNAKE_CASE , tokenizer=_SCREAMING_SNAKE_CASE ) _A = WavaVecaForCTC.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , activation_dropout=model_args.activation_dropout , attention_dropout=model_args.attention_dropout , hidden_dropout=model_args.hidden_dropout , feat_proj_dropout=model_args.feat_proj_dropout , mask_time_prob=model_args.mask_time_prob , gradient_checkpointing=training_args.gradient_checkpointing , layerdrop=model_args.layerdrop , ctc_loss_reduction='mean' , pad_token_id=processor.tokenizer.pad_token_id , vocab_size=len(processor.tokenizer ) , ) if data_args.max_train_samples is not None: _A = min(len(_SCREAMING_SNAKE_CASE ) , data_args.max_train_samples ) _A = train_dataset.select(range(_SCREAMING_SNAKE_CASE ) ) if data_args.max_val_samples is not None: _A = eval_dataset.select(range(data_args.max_val_samples ) ) _A = torchaudio.transforms.Resample(48_000 , 16_000 ) # Preprocessing the datasets. # We need to read the aduio files as arrays and tokenize the targets. def speech_file_to_array_fn(_SCREAMING_SNAKE_CASE ): _A = torchaudio.load(batch['path'] ) _A = resampler(_SCREAMING_SNAKE_CASE ).squeeze().numpy() _A = 16_000 _A = batch["text"] return batch _A = train_dataset.map( _SCREAMING_SNAKE_CASE , remove_columns=train_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) _A = eval_dataset.map( _SCREAMING_SNAKE_CASE , remove_columns=eval_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) def prepare_dataset(_SCREAMING_SNAKE_CASE ): # check that all files have the correct sampling rate assert ( len(set(batch['sampling_rate'] ) ) == 1 ), F"Make sure all inputs have the same sampling rate of {processor.feature_extractor.sampling_rate}." _A = processor( audio=batch['speech'] , text=batch['target_text'] , sampling_rate=batch['sampling_rate'][0] ) batch.update(_SCREAMING_SNAKE_CASE ) return batch _A = train_dataset.map( _SCREAMING_SNAKE_CASE , remove_columns=train_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=_SCREAMING_SNAKE_CASE , num_proc=data_args.preprocessing_num_workers , ) _A = eval_dataset.map( _SCREAMING_SNAKE_CASE , remove_columns=eval_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=_SCREAMING_SNAKE_CASE , num_proc=data_args.preprocessing_num_workers , ) # Metric _A = datasets.load_metric('wer' ) def compute_metrics(_SCREAMING_SNAKE_CASE ): _A = pred.predictions _A = np.argmax(_SCREAMING_SNAKE_CASE , axis=-1 ) _A = processor.tokenizer.pad_token_id _A = processor.batch_decode(_SCREAMING_SNAKE_CASE ) # we do not want to group tokens when computing the metrics _A = processor.batch_decode(pred.label_ids , group_tokens=_SCREAMING_SNAKE_CASE ) _A = wer_metric.compute(predictions=_SCREAMING_SNAKE_CASE , references=_SCREAMING_SNAKE_CASE ) return {"wer": wer} if model_args.freeze_feature_extractor: model.freeze_feature_extractor() # Data collator _A = DataCollatorCTCWithPadding(processor=_SCREAMING_SNAKE_CASE , padding=_SCREAMING_SNAKE_CASE ) # Initialize our Trainer _A = CTCTrainer( model=_SCREAMING_SNAKE_CASE , data_collator=_SCREAMING_SNAKE_CASE , args=_SCREAMING_SNAKE_CASE , compute_metrics=_SCREAMING_SNAKE_CASE , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=processor.feature_extractor , ) # Training if training_args.do_train: if last_checkpoint is not None: _A = last_checkpoint elif os.path.isdir(model_args.model_name_or_path ): _A = model_args.model_name_or_path else: _A = None # Save the feature_extractor and the tokenizer if is_main_process(training_args.local_rank ): processor.save_pretrained(training_args.output_dir ) _A = trainer.train(resume_from_checkpoint=_SCREAMING_SNAKE_CASE ) trainer.save_model() _A = train_result.metrics _A = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(_SCREAMING_SNAKE_CASE ) ) _A = min(_SCREAMING_SNAKE_CASE , len(_SCREAMING_SNAKE_CASE ) ) trainer.log_metrics('train' , _SCREAMING_SNAKE_CASE ) trainer.save_metrics('train' , _SCREAMING_SNAKE_CASE ) trainer.save_state() # Evaluation _A = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) _A = trainer.evaluate() _A = data_args.max_val_samples if data_args.max_val_samples is not None else len(_SCREAMING_SNAKE_CASE ) _A = min(_SCREAMING_SNAKE_CASE , len(_SCREAMING_SNAKE_CASE ) ) trainer.log_metrics('eval' , _SCREAMING_SNAKE_CASE ) trainer.save_metrics('eval' , _SCREAMING_SNAKE_CASE ) return results if __name__ == "__main__": main()
27
'''simple docstring''' import unittest from transformers import XLMConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMWithLMHeadModel, ) from transformers.models.xlm.modeling_xlm import XLM_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCAmelCase_ : """simple docstring""" def __init__( self , lowerCamelCase , lowerCamelCase=13 , lowerCamelCase=7 , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=True , lowerCamelCase=False , lowerCamelCase=False , lowerCamelCase=False , lowerCamelCase=2 , lowerCamelCase=99 , lowerCamelCase=0 , lowerCamelCase=32 , lowerCamelCase=5 , lowerCamelCase=4 , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=5_12 , lowerCamelCase=2 , lowerCamelCase=0.02 , lowerCamelCase=2 , lowerCamelCase=4 , lowerCamelCase="last" , lowerCamelCase=True , lowerCamelCase=None , lowerCamelCase=0 , ) -> Optional[Any]: '''simple docstring''' UpperCamelCase : Any = parent UpperCamelCase : int = batch_size UpperCamelCase : str = seq_length UpperCamelCase : Dict = is_training UpperCamelCase : int = use_input_lengths UpperCamelCase : int = use_token_type_ids UpperCamelCase : Any = use_labels UpperCamelCase : List[Any] = gelu_activation UpperCamelCase : Optional[int] = sinusoidal_embeddings UpperCamelCase : str = causal UpperCamelCase : Tuple = asm UpperCamelCase : Any = n_langs UpperCamelCase : Any = vocab_size UpperCamelCase : Optional[Any] = n_special UpperCamelCase : Optional[Any] = hidden_size UpperCamelCase : List[str] = num_hidden_layers UpperCamelCase : Optional[int] = num_attention_heads UpperCamelCase : str = hidden_dropout_prob UpperCamelCase : List[Any] = attention_probs_dropout_prob UpperCamelCase : int = max_position_embeddings UpperCamelCase : List[str] = type_sequence_label_size UpperCamelCase : Optional[Any] = initializer_range UpperCamelCase : Union[str, Any] = num_labels UpperCamelCase : int = num_choices UpperCamelCase : Union[str, Any] = summary_type UpperCamelCase : Union[str, Any] = use_proj UpperCamelCase : Optional[int] = scope UpperCamelCase : Any = bos_token_id def SCREAMING_SNAKE_CASE__ ( self ) -> Optional[Any]: '''simple docstring''' UpperCamelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase : Any = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase : List[str] = None if self.use_input_lengths: UpperCamelCase : Optional[Any] = ( ids_tensor([self.batch_size] , vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length UpperCamelCase : List[str] = None if self.use_token_type_ids: UpperCamelCase : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.n_langs ) UpperCamelCase : List[str] = None UpperCamelCase : Union[str, Any] = None UpperCamelCase : Dict = None if self.use_labels: UpperCamelCase : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCamelCase : str = ids_tensor([self.batch_size] , 2 ).float() UpperCamelCase : Optional[int] = ids_tensor([self.batch_size] , self.num_choices ) UpperCamelCase : List[Any] = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def SCREAMING_SNAKE_CASE__ ( self ) -> Union[str, Any]: '''simple docstring''' return XLMConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , num_labels=self.num_labels , bos_token_id=self.bos_token_id , ) def SCREAMING_SNAKE_CASE__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ) -> Dict: '''simple docstring''' UpperCamelCase : Optional[Any] = XLMModel(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() UpperCamelCase : Optional[Any] = model(lowerCamelCase , lengths=lowerCamelCase , langs=lowerCamelCase ) UpperCamelCase : Optional[Any] = model(lowerCamelCase , langs=lowerCamelCase ) UpperCamelCase : List[str] = model(lowerCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ) -> int: '''simple docstring''' UpperCamelCase : Optional[int] = XLMWithLMHeadModel(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() UpperCamelCase : Tuple = model(lowerCamelCase , token_type_ids=lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ) -> Tuple: '''simple docstring''' UpperCamelCase : Union[str, Any] = XLMForQuestionAnsweringSimple(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() UpperCamelCase : List[str] = model(lowerCamelCase ) UpperCamelCase : Optional[int] = model(lowerCamelCase , start_positions=lowerCamelCase , end_positions=lowerCamelCase ) UpperCamelCase : int = outputs self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ) -> int: '''simple docstring''' UpperCamelCase : Optional[int] = XLMForQuestionAnswering(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() UpperCamelCase : List[str] = model(lowerCamelCase ) UpperCamelCase : Any = model( lowerCamelCase , start_positions=lowerCamelCase , end_positions=lowerCamelCase , cls_index=lowerCamelCase , is_impossible=lowerCamelCase , p_mask=lowerCamelCase , ) UpperCamelCase : Optional[Any] = model( lowerCamelCase , start_positions=lowerCamelCase , end_positions=lowerCamelCase , cls_index=lowerCamelCase , is_impossible=lowerCamelCase , ) ((UpperCamelCase) , ) : Any = result_with_labels.to_tuple() UpperCamelCase : Dict = model(lowerCamelCase , start_positions=lowerCamelCase , end_positions=lowerCamelCase ) ((UpperCamelCase) , ) : Tuple = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape , () ) self.parent.assertEqual(result.start_top_log_probs.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual(result.start_top_index.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual( result.end_top_log_probs.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual( result.end_top_index.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual(result.cls_logits.shape , (self.batch_size,) ) def SCREAMING_SNAKE_CASE__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ) -> Optional[Any]: '''simple docstring''' UpperCamelCase : int = XLMForSequenceClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() UpperCamelCase : List[Any] = model(lowerCamelCase ) UpperCamelCase : Dict = model(lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def SCREAMING_SNAKE_CASE__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ) -> List[Any]: '''simple docstring''' UpperCamelCase : Dict = self.num_labels UpperCamelCase : int = XLMForTokenClassification(lowerCamelCase ) model.to(lowerCamelCase ) model.eval() UpperCamelCase : Union[str, Any] = model(lowerCamelCase , attention_mask=lowerCamelCase , labels=lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , ) -> Optional[Any]: '''simple docstring''' UpperCamelCase : Optional[int] = self.num_choices UpperCamelCase : Dict = XLMForMultipleChoice(config=lowerCamelCase ) model.to(lowerCamelCase ) model.eval() UpperCamelCase : Any = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase : Dict = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase : int = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCamelCase : Optional[int] = model( lowerCamelCase , attention_mask=lowerCamelCase , token_type_ids=lowerCamelCase , labels=lowerCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE__ ( self ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase : Union[str, Any] = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) : Optional[int] = config_and_inputs UpperCamelCase : Dict = {"input_ids": input_ids, "token_type_ids": token_type_ids, "lengths": input_lengths} return config, inputs_dict @require_torch class UpperCAmelCase_ ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): """simple docstring""" __SCREAMING_SNAKE_CASE = ( ( XLMModel, XLMWithLMHeadModel, XLMForQuestionAnswering, XLMForSequenceClassification, XLMForQuestionAnsweringSimple, XLMForTokenClassification, XLMForMultipleChoice, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = ( (XLMWithLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable __SCREAMING_SNAKE_CASE = ( { '''feature-extraction''': XLMModel, '''fill-mask''': XLMWithLMHeadModel, '''question-answering''': XLMForQuestionAnsweringSimple, '''text-classification''': XLMForSequenceClassification, '''text-generation''': XLMWithLMHeadModel, '''token-classification''': XLMForTokenClassification, '''zero-shot''': XLMForSequenceClassification, } if is_torch_available() else {} ) def SCREAMING_SNAKE_CASE__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase ) -> List[Any]: '''simple docstring''' if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith("Fast" ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def SCREAMING_SNAKE_CASE__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase=False ) -> Tuple: '''simple docstring''' UpperCamelCase : Tuple = super()._prepare_for_class(lowerCamelCase , lowerCamelCase , return_labels=lowerCamelCase ) if return_labels: if model_class.__name__ == "XLMForQuestionAnswering": UpperCamelCase : List[str] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCamelCase ) UpperCamelCase : Union[str, Any] = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCamelCase ) return inputs_dict def SCREAMING_SNAKE_CASE__ ( self ) -> Optional[int]: '''simple docstring''' UpperCamelCase : Tuple = XLMModelTester(self ) UpperCamelCase : Union[str, Any] = ConfigTester(self , config_class=lowerCamelCase , emb_dim=37 ) def SCREAMING_SNAKE_CASE__ ( self ) -> List[Any]: '''simple docstring''' self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self ) -> Tuple: '''simple docstring''' UpperCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_model(*lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ) -> Tuple: '''simple docstring''' UpperCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_lm_head(*lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ) -> str: '''simple docstring''' UpperCamelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_simple_qa(*lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ) -> List[str]: '''simple docstring''' UpperCamelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_qa(*lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ) -> Optional[Any]: '''simple docstring''' UpperCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_sequence_classif(*lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ) -> Dict: '''simple docstring''' UpperCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_token_classif(*lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_for_multiple_choice(*lowerCamelCase ) def SCREAMING_SNAKE_CASE__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=False , lowerCamelCase=1 ) -> Optional[Any]: '''simple docstring''' self.assertIsInstance(lowerCamelCase , lowerCamelCase ) self.assertListEqual( [isinstance(lowerCamelCase , lowerCamelCase ) for iter_attentions in attentions] , [True] * len(lowerCamelCase ) ) self.assertEqual(len(lowerCamelCase ) , (max_length - min_length) * num_beam_groups ) for idx, iter_attentions in enumerate(lowerCamelCase ): # adds PAD dummy token UpperCamelCase : Dict = min_length + idx + 1 UpperCamelCase : int = min_length + idx + 1 UpperCamelCase : Union[str, Any] = ( batch_size * num_beam_groups, config.num_attention_heads, tgt_len, src_len, ) # check attn size self.assertListEqual( [layer_attention.shape for layer_attention in iter_attentions] , [expected_shape] * len(lowerCamelCase ) ) def SCREAMING_SNAKE_CASE__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=False , lowerCamelCase=1 ) -> Optional[int]: '''simple docstring''' self.assertIsInstance(lowerCamelCase , lowerCamelCase ) self.assertListEqual( [isinstance(lowerCamelCase , lowerCamelCase ) for iter_hidden_states in hidden_states] , [True] * len(lowerCamelCase ) , ) self.assertEqual(len(lowerCamelCase ) , (max_length - min_length) * num_beam_groups ) for idx, iter_hidden_states in enumerate(lowerCamelCase ): # adds PAD dummy token UpperCamelCase : Tuple = min_length + idx + 1 UpperCamelCase : str = (batch_size * num_beam_groups, seq_len, config.hidden_size) # check hidden size self.assertListEqual( [layer_hidden_states.shape for layer_hidden_states in iter_hidden_states] , [expected_shape] * len(lowerCamelCase ) , ) pass @slow def SCREAMING_SNAKE_CASE__ ( self ) -> Optional[int]: '''simple docstring''' for model_name in XLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase : Union[str, Any] = XLMModel.from_pretrained(lowerCamelCase ) self.assertIsNotNone(lowerCamelCase ) @require_torch class UpperCAmelCase_ ( unittest.TestCase ): """simple docstring""" @slow def SCREAMING_SNAKE_CASE__ ( self ) -> List[str]: '''simple docstring''' UpperCamelCase : int = XLMWithLMHeadModel.from_pretrained("xlm-mlm-en-2048" ) model.to(lowerCamelCase ) UpperCamelCase : Optional[int] = torch.tensor([[14, 4_47]] , dtype=torch.long , device=lowerCamelCase ) # the president UpperCamelCase : Any = [ 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, ] # the president the president the president the president the president the president the president the president the president the president # TODO(PVP): this and other input_ids I tried for generation give pretty bad results. Not sure why. Model might just not be made for auto-regressive inference UpperCamelCase : List[Any] = model.generate(lowerCamelCase , do_sample=lowerCamelCase ) self.assertListEqual(output_ids[0].cpu().numpy().tolist() , lowerCamelCase )
173
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase : Optional[int] = logging.get_logger(__name__) lowerCAmelCase : Any = { 'hustvl/yolos-small': 'https://huggingface.co/hustvl/yolos-small/resolve/main/config.json', # See all YOLOS models at https://huggingface.co/models?filter=yolos } class _A ( __magic_name__): SCREAMING_SNAKE_CASE : int = '''yolos''' def __init__( self , _SCREAMING_SNAKE_CASE=768 , _SCREAMING_SNAKE_CASE=12 , _SCREAMING_SNAKE_CASE=12 , _SCREAMING_SNAKE_CASE=3072 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=1e-12 , _SCREAMING_SNAKE_CASE=[512, 864] , _SCREAMING_SNAKE_CASE=16 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=100 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=1 , _SCREAMING_SNAKE_CASE=5 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=5 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=0.1 , **_SCREAMING_SNAKE_CASE , ): """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : List[Any] = hidden_size SCREAMING_SNAKE_CASE_ : List[Any] = num_hidden_layers SCREAMING_SNAKE_CASE_ : Dict = num_attention_heads SCREAMING_SNAKE_CASE_ : List[str] = intermediate_size SCREAMING_SNAKE_CASE_ : Dict = hidden_act SCREAMING_SNAKE_CASE_ : Tuple = hidden_dropout_prob SCREAMING_SNAKE_CASE_ : int = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ : Optional[Any] = initializer_range SCREAMING_SNAKE_CASE_ : Optional[Any] = layer_norm_eps SCREAMING_SNAKE_CASE_ : Union[str, Any] = image_size SCREAMING_SNAKE_CASE_ : Dict = patch_size SCREAMING_SNAKE_CASE_ : Any = num_channels SCREAMING_SNAKE_CASE_ : Optional[int] = qkv_bias SCREAMING_SNAKE_CASE_ : Union[str, Any] = num_detection_tokens SCREAMING_SNAKE_CASE_ : Dict = use_mid_position_embeddings SCREAMING_SNAKE_CASE_ : str = auxiliary_loss # Hungarian matcher SCREAMING_SNAKE_CASE_ : Union[str, Any] = class_cost SCREAMING_SNAKE_CASE_ : List[str] = bbox_cost SCREAMING_SNAKE_CASE_ : Union[str, Any] = giou_cost # Loss coefficients SCREAMING_SNAKE_CASE_ : Tuple = bbox_loss_coefficient SCREAMING_SNAKE_CASE_ : Any = giou_loss_coefficient SCREAMING_SNAKE_CASE_ : Dict = eos_coefficient class _A ( __magic_name__): SCREAMING_SNAKE_CASE : str = version.parse('''1.11''') @property def UpperCAmelCase ( self ): """simple docstring""" return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def UpperCAmelCase ( self ): """simple docstring""" return 1e-4 @property def UpperCAmelCase ( self ): """simple docstring""" return 12
353
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : Optional[int] = logging.get_logger(__name__) lowerCAmelCase : str = { 'microsoft/git-base': 'https://huggingface.co/microsoft/git-base/resolve/main/config.json', } class _A ( __magic_name__): SCREAMING_SNAKE_CASE : Optional[int] = '''git_vision_model''' def __init__( self , _SCREAMING_SNAKE_CASE=768 , _SCREAMING_SNAKE_CASE=3072 , _SCREAMING_SNAKE_CASE=12 , _SCREAMING_SNAKE_CASE=12 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=224 , _SCREAMING_SNAKE_CASE=16 , _SCREAMING_SNAKE_CASE="quick_gelu" , _SCREAMING_SNAKE_CASE=1e-5 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.02 , **_SCREAMING_SNAKE_CASE , ): """simple docstring""" super().__init__(**_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : List[str] = hidden_size SCREAMING_SNAKE_CASE_ : Any = intermediate_size SCREAMING_SNAKE_CASE_ : List[str] = num_hidden_layers SCREAMING_SNAKE_CASE_ : Optional[Any] = num_attention_heads SCREAMING_SNAKE_CASE_ : Optional[Any] = num_channels SCREAMING_SNAKE_CASE_ : Dict = patch_size SCREAMING_SNAKE_CASE_ : Any = image_size SCREAMING_SNAKE_CASE_ : Union[str, Any] = initializer_range SCREAMING_SNAKE_CASE_ : Any = attention_dropout SCREAMING_SNAKE_CASE_ : List[Any] = layer_norm_eps SCREAMING_SNAKE_CASE_ : Tuple = hidden_act @classmethod def UpperCAmelCase ( cls , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ): """simple docstring""" cls._set_token_in_kwargs(_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : str = cls.get_config_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) # get the vision config dict if we are loading from GITConfig if config_dict.get('model_type' ) == "git": SCREAMING_SNAKE_CASE_ : str = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls , 'model_type' ) and config_dict["model_type"] != cls.model_type: logger.warning( f"You are using a model of type {config_dict['model_type']} to instantiate a model of type " f"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) class _A ( __magic_name__): SCREAMING_SNAKE_CASE : Optional[Any] = '''git''' def __init__( self , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=3_0522 , _SCREAMING_SNAKE_CASE=768 , _SCREAMING_SNAKE_CASE=6 , _SCREAMING_SNAKE_CASE=12 , _SCREAMING_SNAKE_CASE=3072 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=1024 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=1e-12 , _SCREAMING_SNAKE_CASE=0 , _SCREAMING_SNAKE_CASE="absolute" , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=101 , _SCREAMING_SNAKE_CASE=102 , _SCREAMING_SNAKE_CASE=None , **_SCREAMING_SNAKE_CASE , ): """simple docstring""" super().__init__(bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , pad_token_id=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) if vision_config is None: SCREAMING_SNAKE_CASE_ : Any = {} logger.info('vision_config is None. initializing the GitVisionConfig with default values.' ) SCREAMING_SNAKE_CASE_ : Optional[Any] = GitVisionConfig(**_SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE_ : Dict = vocab_size SCREAMING_SNAKE_CASE_ : List[Any] = hidden_size SCREAMING_SNAKE_CASE_ : Dict = num_hidden_layers SCREAMING_SNAKE_CASE_ : List[str] = num_attention_heads SCREAMING_SNAKE_CASE_ : List[str] = hidden_act SCREAMING_SNAKE_CASE_ : Union[str, Any] = intermediate_size SCREAMING_SNAKE_CASE_ : Any = hidden_dropout_prob SCREAMING_SNAKE_CASE_ : Tuple = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ : int = max_position_embeddings SCREAMING_SNAKE_CASE_ : Union[str, Any] = initializer_range SCREAMING_SNAKE_CASE_ : int = layer_norm_eps SCREAMING_SNAKE_CASE_ : List[str] = position_embedding_type SCREAMING_SNAKE_CASE_ : List[str] = use_cache SCREAMING_SNAKE_CASE_ : Optional[int] = tie_word_embeddings SCREAMING_SNAKE_CASE_ : List[Any] = num_image_with_embedding SCREAMING_SNAKE_CASE_ : Tuple = bos_token_id SCREAMING_SNAKE_CASE_ : Optional[int] = eos_token_id def UpperCAmelCase ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = copy.deepcopy(self.__dict__ ) SCREAMING_SNAKE_CASE_ : List[str] = self.vision_config.to_dict() SCREAMING_SNAKE_CASE_ : int = self.__class__.model_type return output
353
1
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def lowerCamelCase_ ( _lowerCamelCase : int=None ): if subparsers is not None: lowerCamelCase_ = subparsers.add_parser('''test''' ) else: lowerCamelCase_ = argparse.ArgumentParser('''Accelerate test command''' ) parser.add_argument( '''--config_file''' , default=_lowerCamelCase , help=( '''The path to use to store the config file. Will default to a file named default_config.yaml in the cache ''' '''location, which is the content of the environment `HF_HOME` suffixed with \'accelerate\', or if you don\'t have ''' '''such an environment variable, your cache directory (\'~/.cache\' or the content of `XDG_CACHE_HOME`) suffixed ''' '''with \'huggingface\'.''' ) , ) if subparsers is not None: parser.set_defaults(func=_lowerCamelCase ) return parser def lowerCamelCase_ ( _lowerCamelCase : List[str] ): lowerCamelCase_ = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ['''test_utils''', '''scripts''', '''test_script.py'''] ) if args.config_file is None: lowerCamelCase_ = script_name else: lowerCamelCase_ = F"""--config_file={args.config_file} {script_name}""" lowerCamelCase_ = ['''accelerate-launch'''] + test_args.split() lowerCamelCase_ = execute_subprocess_async(_lowerCamelCase , env=os.environ.copy() ) if result.returncode == 0: print('''Test is a success! You are ready for your distributed training!''' ) def lowerCamelCase_ ( ): lowerCamelCase_ = test_command_parser() lowerCamelCase_ = parser.parse_args() test_command(_lowerCamelCase ) if __name__ == "__main__": main()
142
"""simple docstring""" import importlib.metadata import operator import re import sys from typing import Optional from packaging import version __lowercase : int = { """<""": operator.lt, """<=""": operator.le, """==""": operator.eq, """!=""": operator.ne, """>=""": operator.ge, """>""": operator.gt, } def lowerCamelCase_ ( _lowerCamelCase : Optional[Any] , _lowerCamelCase : List[Any] , _lowerCamelCase : Tuple , _lowerCamelCase : List[str] , _lowerCamelCase : Any , _lowerCamelCase : Dict ): if got_ver is None or want_ver is None: raise ValueError( F"""Unable to compare versions for {requirement}: need={want_ver} found={got_ver}. This is unusual. Consider""" F""" reinstalling {pkg}.""" ) if not ops[op](version.parse(_lowerCamelCase ) , version.parse(_lowerCamelCase ) ): raise ImportError( F"""{requirement} is required for a normal functioning of this module, but found {pkg}=={got_ver}.{hint}""" ) def lowerCamelCase_ ( _lowerCamelCase : str , _lowerCamelCase : Optional[str] = None ): lowerCamelCase_ = F"""\n{hint}""" if hint is not None else '''''' # non-versioned check if re.match(r'''^[\w_\-\d]+$''' , _lowerCamelCase ): lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = requirement, None, None else: lowerCamelCase_ = re.findall(r'''^([^!=<>\s]+)([\s!=<>]{1,2}.+)''' , _lowerCamelCase ) if not match: raise ValueError( '''requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23, but''' F""" got {requirement}""" ) lowerCamelCase_ , lowerCamelCase_ = match[0] lowerCamelCase_ = want_full.split(''',''' ) # there could be multiple requirements lowerCamelCase_ = {} for w in want_range: lowerCamelCase_ = re.findall(r'''^([\s!=<>]{1,2})(.+)''' , _lowerCamelCase ) if not match: raise ValueError( '''requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23,''' F""" but got {requirement}""" ) lowerCamelCase_ , lowerCamelCase_ = match[0] lowerCamelCase_ = want_ver if op not in ops: raise ValueError(F"""{requirement}: need one of {list(ops.keys() )}, but got {op}""" ) # special case if pkg == "python": lowerCamelCase_ = '''.'''.join([str(_lowerCamelCase ) for x in sys.version_info[:3]] ) for op, want_ver in wanted.items(): _compare_versions(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) return # check if any version is installed try: lowerCamelCase_ = importlib.metadata.version(_lowerCamelCase ) except importlib.metadata.PackageNotFoundError: raise importlib.metadata.PackageNotFoundError( F"""The '{requirement}' distribution was not found and is required by this application. {hint}""" ) # check that the right version is installed if version number or a range was provided if want_ver is not None: for op, want_ver in wanted.items(): _compare_versions(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def lowerCamelCase_ ( _lowerCamelCase : List[Any] ): lowerCamelCase_ = '''Try: pip install transformers -U or pip install -e \'.[dev]\' if you\'re working with git main''' return require_version(_lowerCamelCase , _lowerCamelCase )
142
1
'''simple docstring''' from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class A : def __init__( self , snake_case_ , snake_case_=1_3 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=9_9 , snake_case_=3_2 , snake_case_=2 , snake_case_=4 , snake_case_=3_7 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_1_2 , snake_case_=1_6 , snake_case_=2 , snake_case_=0.02 , snake_case_=3 , snake_case_=4 , snake_case_=None , ) -> List[str]: _a = parent _a = 1_3 _a = 7 _a = True _a = True _a = True _a = True _a = 9_9 _a = 3_8_4 _a = 2 _a = 4 _a = 3_7 _a = "gelu" _a = 0.1 _a = 0.1 _a = 5_1_2 _a = 1_6 _a = 2 _a = 0.02 _a = 3 _a = 4 _a = 1_2_8 _a = 2 _a = 9 _a = 1 _a = None def __lowerCAmelCase ( self ) -> Optional[Any]: _a = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _a = None if self.use_input_mask: _a = random_attention_mask([self.batch_size, self.seq_length] ) _a = None if self.use_token_type_ids: _a = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _a = None _a = None _a = None if self.use_labels: _a = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _a = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _a = ids_tensor([self.batch_size] , self.num_choices ) _a = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=snake_case_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowerCAmelCase ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) -> Union[str, Any]: _a = TFConvBertModel(config=snake_case_ ) _a = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _a = [input_ids, input_mask] _a = model(snake_case_ ) _a = model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) -> Tuple: _a = TFConvBertForMaskedLM(config=snake_case_ ) _a = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _a = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCAmelCase ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) -> Dict: _a = self.num_labels _a = TFConvBertForSequenceClassification(config=snake_case_ ) _a = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _a = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) -> Union[str, Any]: _a = self.num_choices _a = TFConvBertForMultipleChoice(config=snake_case_ ) _a = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _a = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _a = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _a = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } _a = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __lowerCAmelCase ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) -> Tuple: _a = self.num_labels _a = TFConvBertForTokenClassification(config=snake_case_ ) _a = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _a = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __lowerCAmelCase ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ) -> Tuple: _a = TFConvBertForQuestionAnswering(config=snake_case_ ) _a = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _a = model(snake_case_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __lowerCAmelCase ( self ) -> List[str]: _a = self.prepare_config_and_inputs() ( ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ( _a ) , ) = config_and_inputs _a = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class A ( UpperCamelCase__ , UpperCamelCase__ , unittest.TestCase ): __UpperCAmelCase : Dict = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) __UpperCAmelCase : str = ( { """feature-extraction""": TFConvBertModel, """fill-mask""": TFConvBertForMaskedLM, """question-answering""": TFConvBertForQuestionAnswering, """text-classification""": TFConvBertForSequenceClassification, """token-classification""": TFConvBertForTokenClassification, """zero-shot""": TFConvBertForSequenceClassification, } if is_tf_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : Optional[int] = False def __lowerCAmelCase ( self ) -> Optional[int]: _a = TFConvBertModelTester(self ) _a = ConfigTester(self , config_class=snake_case_ , hidden_size=3_7 ) def __lowerCAmelCase ( self ) -> List[Any]: self.config_tester.run_common_tests() def __lowerCAmelCase ( self ) -> Optional[int]: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def __lowerCAmelCase ( self ) -> Optional[int]: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case_ ) def __lowerCAmelCase ( self ) -> Union[str, Any]: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*snake_case_ ) def __lowerCAmelCase ( self ) -> Optional[Any]: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case_ ) def __lowerCAmelCase ( self ) -> str: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case_ ) def __lowerCAmelCase ( self ) -> Dict: _a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case_ ) @slow def __lowerCAmelCase ( self ) -> str: _a , _a = self.model_tester.prepare_config_and_inputs_for_common() _a = True _a = True if hasattr(snake_case_ , "use_cache" ): _a = True _a = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) _a = getattr(self.model_tester , "key_length" , snake_case_ ) for model_class in self.all_model_classes: _a = self._prepare_for_class(snake_case_ , snake_case_ ) _a = model_class(snake_case_ ) _a = len(model(snake_case_ ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case_ , saved_model=snake_case_ ) _a = os.path.join(snake_case_ , "saved_model" , "1" ) _a = tf.keras.models.load_model(snake_case_ ) _a = model(snake_case_ ) if self.is_encoder_decoder: _a = outputs["encoder_hidden_states"] _a = outputs["encoder_attentions"] else: _a = outputs["hidden_states"] _a = outputs["attentions"] self.assertEqual(len(snake_case_ ) , snake_case_ ) _a = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(snake_case_ ) , snake_case_ ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def __lowerCAmelCase ( self ) -> str: _a = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) self.assertIsNotNone(snake_case_ ) def __lowerCAmelCase ( self ) -> str: _a , _a = self.model_tester.prepare_config_and_inputs_for_common() _a = True _a = getattr(self.model_tester , "decoder_seq_length" , self.model_tester.seq_length ) _a = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) _a = getattr(self.model_tester , "key_length" , snake_case_ ) _a = getattr(self.model_tester , "key_length" , snake_case_ ) def check_decoder_attentions_output(snake_case_ ): _a = len(snake_case_ ) self.assertEqual(out_len % 2 , 0 ) _a = outputs.decoder_attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(snake_case_ ): _a = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: _a = True _a = False _a = model_class(snake_case_ ) _a = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) _a = len(snake_case_ ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) if self.is_encoder_decoder: _a = model_class(snake_case_ ) _a = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_decoder_attentions_output(snake_case_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] _a = True _a = model_class(snake_case_ ) _a = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) # Check attention is always last and order is fine _a = True _a = True _a = model_class(snake_case_ ) _a = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(snake_case_ ) ) self.assertEqual(model.config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) @require_tf class A ( unittest.TestCase ): @slow def __lowerCAmelCase ( self ) -> int: _a = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) _a = tf.constant([[0, 1, 2, 3, 4, 5]] ) _a = model(snake_case_ )[0] _a = [1, 6, 7_6_8] self.assertEqual(output.shape , snake_case_ ) _a = tf.constant( [ [ [-0.03_475_493, -0.4_686_034, -0.30_638_832], [0.22_637_248, -0.26_988_646, -0.7_423_424], [0.10_324_868, -0.45_013_508, -0.58_280_784], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , snake_case_ , atol=1E-4 )
717
'''simple docstring''' __snake_case : Dict = { "Pillow": "Pillow<10.0.0", "accelerate": "accelerate>=0.20.3", "av": "av==9.2.0", "beautifulsoup4": "beautifulsoup4", "black": "black~=23.1", "codecarbon": "codecarbon==1.2.0", "cookiecutter": "cookiecutter==1.7.3", "dataclasses": "dataclasses", "datasets": "datasets!=2.5.0", "decord": "decord==0.6.0", "deepspeed": "deepspeed>=0.9.3", "diffusers": "diffusers", "dill": "dill<0.3.5", "evaluate": "evaluate>=0.2.0", "fairscale": "fairscale>0.3", "faiss-cpu": "faiss-cpu", "fastapi": "fastapi", "filelock": "filelock", "flax": "flax>=0.4.1,<=0.7.0", "ftfy": "ftfy", "fugashi": "fugashi>=1.0", "GitPython": "GitPython<3.1.19", "hf-doc-builder": "hf-doc-builder>=0.3.0", "huggingface-hub": "huggingface-hub>=0.14.1,<1.0", "importlib_metadata": "importlib_metadata", "ipadic": "ipadic>=1.0.0,<2.0", "isort": "isort>=5.5.4", "jax": "jax>=0.2.8,!=0.3.2,<=0.4.13", "jaxlib": "jaxlib>=0.1.65,<=0.4.13", "jieba": "jieba", "kenlm": "kenlm", "keras-nlp": "keras-nlp>=0.3.1", "librosa": "librosa", "nltk": "nltk", "natten": "natten>=0.14.6", "numpy": "numpy>=1.17", "onnxconverter-common": "onnxconverter-common", "onnxruntime-tools": "onnxruntime-tools>=1.4.2", "onnxruntime": "onnxruntime>=1.4.0", "opencv-python": "opencv-python", "optuna": "optuna", "optax": "optax>=0.0.8,<=0.1.4", "packaging": "packaging>=20.0", "parameterized": "parameterized", "phonemizer": "phonemizer", "protobuf": "protobuf", "psutil": "psutil", "pyyaml": "pyyaml>=5.1", "pydantic": "pydantic<2", "pytest": "pytest>=7.2.0", "pytest-timeout": "pytest-timeout", "pytest-xdist": "pytest-xdist", "python": "python>=3.8.0", "ray[tune]": "ray[tune]", "regex": "regex!=2019.12.17", "requests": "requests", "rhoknp": "rhoknp>=1.1.0,<1.3.1", "rjieba": "rjieba", "rouge-score": "rouge-score!=0.0.7,!=0.0.8,!=0.1,!=0.1.1", "ruff": "ruff>=0.0.241,<=0.0.259", "sacrebleu": "sacrebleu>=1.4.12,<2.0.0", "sacremoses": "sacremoses", "safetensors": "safetensors>=0.3.1", "sagemaker": "sagemaker>=2.31.0", "scikit-learn": "scikit-learn", "sentencepiece": "sentencepiece>=0.1.91,!=0.1.92", "sigopt": "sigopt", "starlette": "starlette", "sudachipy": "sudachipy>=0.6.6", "sudachidict_core": "sudachidict_core>=20220729", "tensorflow-cpu": "tensorflow-cpu>=2.6,<2.14", "tensorflow": "tensorflow>=2.6,<2.14", "tensorflow-text": "tensorflow-text<2.14", "tf2onnx": "tf2onnx", "timeout-decorator": "timeout-decorator", "timm": "timm", "tokenizers": "tokenizers>=0.11.1,!=0.11.3,<0.14", "torch": "torch>=1.9,!=1.12.0", "torchaudio": "torchaudio", "torchvision": "torchvision", "pyctcdecode": "pyctcdecode>=0.4.0", "tqdm": "tqdm>=4.27", "unidic": "unidic>=1.0.2", "unidic_lite": "unidic_lite>=1.0.7", "urllib3": "urllib3<2.0.0", "uvicorn": "uvicorn", }
691
0
'''simple docstring''' import math def lowerCAmelCase_ ( snake_case__ , snake_case__ ): '''simple docstring''' if initial_intensity < 0: raise ValueError('''The value of intensity cannot be negative''' ) # handling of negative values of initial intensity if angle < 0 or angle > 360: raise ValueError('''In Malus Law, the angle is in the range 0-360 degrees''' ) # handling of values out of allowed range return initial_intensity * (math.cos(math.radians(snake_case__ ) ) ** 2) if __name__ == "__main__": import doctest doctest.testmod(name='malus_law')
634
'''simple docstring''' import json import os import unittest from transformers import DebertaTokenizer, DebertaTokenizerFast from transformers.models.deberta.tokenization_deberta import VOCAB_FILES_NAMES from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class A ( __snake_case , unittest.TestCase ): __magic_name__ = DebertaTokenizer __magic_name__ = True __magic_name__ = DebertaTokenizerFast def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt A : Optional[int] = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''[UNK]''', ] A : List[str] = dict(zip(SCREAMING_SNAKE_CASE , range(len(SCREAMING_SNAKE_CASE ) ) ) ) A : List[Any] = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] A : Tuple = {'''unk_token''': '''[UNK]'''} A : Optional[int] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) A : str = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(SCREAMING_SNAKE_CASE ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(SCREAMING_SNAKE_CASE ) ) def __lowerCAmelCase ( self , **SCREAMING_SNAKE_CASE ) -> Optional[Any]: """simple docstring""" kwargs.update(self.special_tokens_map ) return self.tokenizer_class.from_pretrained(self.tmpdirname , **SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE ) -> str: """simple docstring""" A : Union[str, Any] = '''lower newer''' A : List[Any] = '''lower newer''' return input_text, output_text def __lowerCAmelCase ( self ) -> str: """simple docstring""" A : str = self.get_tokenizer() A : List[str] = '''lower newer''' A : int = ['''l''', '''o''', '''w''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] A : Any = tokenizer.tokenize(SCREAMING_SNAKE_CASE ) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) A : str = tokens + [tokenizer.unk_token] A : str = [0, 1, 2, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(SCREAMING_SNAKE_CASE ) , SCREAMING_SNAKE_CASE ) def __lowerCAmelCase ( self ) -> str: """simple docstring""" A : List[Any] = self.get_tokenizer() A : Tuple = tokenizer('''Hello''' , '''World''' ) A : str = [0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1] self.assertListEqual(tokd['''token_type_ids'''] , SCREAMING_SNAKE_CASE ) @slow def __lowerCAmelCase ( self ) -> int: """simple docstring""" A : Union[str, Any] = self.tokenizer_class.from_pretrained('''microsoft/deberta-base''' ) A : Union[str, Any] = tokenizer.encode('''sequence builders''' , add_special_tokens=SCREAMING_SNAKE_CASE ) A : Any = tokenizer.encode('''multi-sequence build''' , add_special_tokens=SCREAMING_SNAKE_CASE ) A : Tuple = tokenizer.encode( '''sequence builders''' , add_special_tokens=SCREAMING_SNAKE_CASE , add_prefix_space=SCREAMING_SNAKE_CASE ) A : Union[str, Any] = tokenizer.encode( '''sequence builders''' , '''multi-sequence build''' , add_special_tokens=SCREAMING_SNAKE_CASE , add_prefix_space=SCREAMING_SNAKE_CASE ) A : List[str] = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE ) A : Union[str, Any] = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) assert encoded_sentence == encoded_text_from_decode assert encoded_pair == encoded_pair_from_decode @slow def __lowerCAmelCase ( self ) -> int: """simple docstring""" A : Any = [self.tokenizer_class] if self.test_rust_tokenizer: tokenizer_classes.append(self.rust_tokenizer_class ) for tokenizer_class in tokenizer_classes: A : List[str] = tokenizer_class.from_pretrained('''microsoft/deberta-base''' ) A : Dict = [ '''ALBERT: A Lite BERT for Self-supervised Learning of Language Representations''', '''ALBERT incorporates two parameter reduction techniques''', '''The first one is a factorized embedding parameterization. By decomposing the large vocabulary''' ''' embedding matrix into two small matrices, we separate the size of the hidden layers from the size of''' ''' vocabulary embedding.''', ] A : List[Any] = tokenizer(SCREAMING_SNAKE_CASE , padding=SCREAMING_SNAKE_CASE ) A : List[str] = [tokenizer.decode(SCREAMING_SNAKE_CASE , skip_special_tokens=SCREAMING_SNAKE_CASE ) for seq in encoding['''input_ids''']] # fmt: off A : List[str] = { '''input_ids''': [ [1, 2118, 11126, 565, 35, 83, 25191, 163, 18854, 13, 12156, 12, 16101, 25376, 13807, 9, 22205, 27893, 1635, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 2118, 11126, 565, 24536, 80, 43797, 4878, 7373, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 133, 78, 65, 16, 10, 3724, 1538, 33183, 11303, 43797, 1938, 4, 870, 24165, 29105, 5, 739, 32644, 33183, 11303, 36173, 88, 80, 650, 7821, 45940, 6, 52, 2559, 5, 1836, 9, 5, 7397, 13171, 31, 5, 1836, 9, 32644, 33183, 11303, 4, 2] ], '''token_type_ids''': [ [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ], '''attention_mask''': [ [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1] ] } # fmt: on A : Any = [ '''ALBERT: A Lite BERT for Self-supervised Learning of Language Representations''', '''ALBERT incorporates two parameter reduction techniques''', '''The first one is a factorized embedding parameterization. By decomposing the large vocabulary''' ''' embedding matrix into two small matrices, we separate the size of the hidden layers from the size of''' ''' vocabulary embedding.''', ] self.assertDictEqual(encoding.data , SCREAMING_SNAKE_CASE ) for expected, decoded in zip(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): self.assertEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE )
634
1
'''simple docstring''' from __future__ import annotations def lowerCAmelCase ( UpperCamelCase__ : int , UpperCamelCase__ : int ): """simple docstring""" if b == 0: return (1, 0) ((__UpperCAmelCase) , (__UpperCAmelCase)) = extended_euclid(UpperCamelCase__ , a % b ) __UpperCAmelCase = a // b return (y, x - k * y) def lowerCAmelCase ( UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int ): """simple docstring""" ((__UpperCAmelCase) , (__UpperCAmelCase)) = extended_euclid(UpperCamelCase__ , UpperCamelCase__ ) __UpperCAmelCase = na * na __UpperCAmelCase = ra * x * na + ra * y * na return (n % m + m) % m def lowerCAmelCase ( UpperCamelCase__ : int , UpperCamelCase__ : int ): """simple docstring""" ((__UpperCAmelCase) , (__UpperCAmelCase)) = extended_euclid(UpperCamelCase__ , UpperCamelCase__ ) if b < 0: __UpperCAmelCase = (b % n + n) % n return b def lowerCAmelCase ( UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : int ): """simple docstring""" __UpperCAmelCase , __UpperCAmelCase = invert_modulo(UpperCamelCase__ , UpperCamelCase__ ), invert_modulo(UpperCamelCase__ , UpperCamelCase__ ) __UpperCAmelCase = na * na __UpperCAmelCase = ra * x * na + ra * y * na return (n % m + m) % m if __name__ == "__main__": from doctest import testmod testmod(name="chinese_remainder_theorem", verbose=True) testmod(name="chinese_remainder_theorem2", verbose=True) testmod(name="invert_modulo", verbose=True) testmod(name="extended_euclid", verbose=True)
654
'''simple docstring''' from __future__ import annotations from statistics import mean def lowerCAmelCase ( UpperCamelCase__ : list[int] , UpperCamelCase__ : list[int] , UpperCamelCase__ : int ): """simple docstring""" __UpperCAmelCase = [0] * no_of_processes __UpperCAmelCase = [0] * no_of_processes # Initialize remaining_time to waiting_time. for i in range(UpperCamelCase__ ): __UpperCAmelCase = burst_time[i] __UpperCAmelCase = [] __UpperCAmelCase = 0 __UpperCAmelCase = 0 # When processes are not completed, # A process whose arrival time has passed \ # and has remaining execution time is put into the ready_process. # The shortest process in the ready_process, target_process is executed. while completed != no_of_processes: __UpperCAmelCase = [] __UpperCAmelCase = -1 for i in range(UpperCamelCase__ ): if (arrival_time[i] <= total_time) and (remaining_time[i] > 0): ready_process.append(UpperCamelCase__ ) if len(UpperCamelCase__ ) > 0: __UpperCAmelCase = ready_process[0] for i in ready_process: if remaining_time[i] < remaining_time[target_process]: __UpperCAmelCase = i total_time += burst_time[target_process] completed += 1 __UpperCAmelCase = 0 __UpperCAmelCase = ( total_time - arrival_time[target_process] - burst_time[target_process] ) else: total_time += 1 return waiting_time def lowerCAmelCase ( UpperCamelCase__ : list[int] , UpperCamelCase__ : int , UpperCamelCase__ : list[int] ): """simple docstring""" __UpperCAmelCase = [0] * no_of_processes for i in range(UpperCamelCase__ ): __UpperCAmelCase = burst_time[i] + waiting_time[i] return turn_around_time if __name__ == "__main__": print("[TEST CASE 01]") __lowerCAmelCase : List[Any] = 4 __lowerCAmelCase : List[Any] = [2, 5, 3, 7] __lowerCAmelCase : Tuple = [0, 0, 0, 0] __lowerCAmelCase : Optional[int] = calculate_waitingtime(arrival_time, burst_time, no_of_processes) __lowerCAmelCase : Dict = calculate_turnaroundtime( burst_time, no_of_processes, waiting_time ) # Printing the Result print("PID\tBurst Time\tArrival Time\tWaiting Time\tTurnaround Time") for i, process_id in enumerate(list(range(1, 5))): print( F"""{process_id}\t{burst_time[i]}\t\t\t{arrival_time[i]}\t\t\t\t""" F"""{waiting_time[i]}\t\t\t\t{turn_around_time[i]}""" ) print(F"""\nAverage waiting time = {mean(waiting_time):.5f}""") print(F"""Average turnaround time = {mean(turn_around_time):.5f}""")
654
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _A : List[str] = { """configuration_llama""": ["""LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LlamaConfig"""], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Optional[int] = ["""LlamaTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : List[Any] = ["""LlamaTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Any = [ """LlamaForCausalLM""", """LlamaModel""", """LlamaPreTrainedModel""", """LlamaForSequenceClassification""", ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys _A : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
100
'''simple docstring''' from typing import Optional, Tuple, Union import torch from einops import rearrange, reduce from diffusers import DDIMScheduler, DDPMScheduler, DiffusionPipeline, ImagePipelineOutput, UNetaDConditionModel from diffusers.schedulers.scheduling_ddim import DDIMSchedulerOutput from diffusers.schedulers.scheduling_ddpm import DDPMSchedulerOutput _snake_case : Optional[Any] = 8 def snake_case_ (UpperCamelCase : List[Any] , UpperCamelCase : Dict=BITS ): '''simple docstring''' _a = x.device _a = (x * 255).int().clamp(0 , 255 ) _a = 2 ** torch.arange(bits - 1 , -1 , -1 , device=UpperCamelCase ) _a = rearrange(UpperCamelCase , '''d -> d 1 1''' ) _a = rearrange(UpperCamelCase , '''b c h w -> b c 1 h w''' ) _a = ((x & mask) != 0).float() _a = rearrange(UpperCamelCase , '''b c d h w -> b (c d) h w''' ) _a = bits * 2 - 1 return bits def snake_case_ (UpperCamelCase : List[Any] , UpperCamelCase : Any=BITS ): '''simple docstring''' _a = x.device _a = (x > 0).int() _a = 2 ** torch.arange(bits - 1 , -1 , -1 , device=UpperCamelCase , dtype=torch.intaa ) _a = rearrange(UpperCamelCase , '''d -> d 1 1''' ) _a = rearrange(UpperCamelCase , '''b (c d) h w -> b c d h w''' , d=8 ) _a = reduce(x * mask , '''b c d h w -> b c h w''' , '''sum''' ) return (dec / 255).clamp(0.0 , 1.0 ) def snake_case_ (self : Union[str, Any] , UpperCamelCase : torch.FloatTensor , UpperCamelCase : int , UpperCamelCase : torch.FloatTensor , UpperCamelCase : float = 0.0 , UpperCamelCase : bool = True , UpperCamelCase : Any=None , UpperCamelCase : bool = True , ): '''simple docstring''' if self.num_inference_steps is None: raise ValueError( '''Number of inference steps is \'None\', you need to run \'set_timesteps\' after creating the scheduler''' ) # See formulas (12) and (16) of DDIM paper https://arxiv.org/pdf/2010.02502.pdf # Ideally, read DDIM paper in-detail understanding # Notation (<variable name> -> <name in paper> # - pred_noise_t -> e_theta(x_t, t) # - pred_original_sample -> f_theta(x_t, t) or x_0 # - std_dev_t -> sigma_t # - eta -> η # - pred_sample_direction -> "direction pointing to x_t" # - pred_prev_sample -> "x_t-1" # 1. get previous step value (=t-1) _a = timestep - self.config.num_train_timesteps // self.num_inference_steps # 2. compute alphas, betas _a = self.alphas_cumprod[timestep] _a = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.final_alpha_cumprod _a = 1 - alpha_prod_t # 3. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _a = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 # 4. Clip "predicted x_0" _a = self.bit_scale if self.config.clip_sample: _a = torch.clamp(UpperCamelCase , -scale , UpperCamelCase ) # 5. compute variance: "sigma_t(η)" -> see formula (16) # σ_t = sqrt((1 − α_t−1)/(1 − α_t)) * sqrt(1 − α_t/α_t−1) _a = self._get_variance(UpperCamelCase , UpperCamelCase ) _a = eta * variance ** 0.5 if use_clipped_model_output: # the model_output is always re-derived from the clipped x_0 in Glide _a = (sample - alpha_prod_t ** 0.5 * pred_original_sample) / beta_prod_t ** 0.5 # 6. compute "direction pointing to x_t" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _a = (1 - alpha_prod_t_prev - std_dev_t**2) ** 0.5 * model_output # 7. compute x_t without "random noise" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf _a = alpha_prod_t_prev ** 0.5 * pred_original_sample + pred_sample_direction if eta > 0: # randn_like does not support generator https://github.com/pytorch/pytorch/issues/27072 _a = model_output.device if torch.is_tensor(UpperCamelCase ) else '''cpu''' _a = torch.randn(model_output.shape , dtype=model_output.dtype , generator=UpperCamelCase ).to(UpperCamelCase ) _a = self._get_variance(UpperCamelCase , UpperCamelCase ) ** 0.5 * eta * noise _a = prev_sample + variance if not return_dict: return (prev_sample,) return DDIMSchedulerOutput(prev_sample=UpperCamelCase , pred_original_sample=UpperCamelCase ) def snake_case_ (self : Any , UpperCamelCase : torch.FloatTensor , UpperCamelCase : int , UpperCamelCase : torch.FloatTensor , UpperCamelCase : str="epsilon" , UpperCamelCase : Dict=None , UpperCamelCase : bool = True , ): '''simple docstring''' _a = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type in ["learned", "learned_range"]: _a , _a = torch.split(UpperCamelCase , sample.shape[1] , dim=1 ) else: _a = None # 1. compute alphas, betas _a = self.alphas_cumprod[t] _a = self.alphas_cumprod[t - 1] if t > 0 else self.one _a = 1 - alpha_prod_t _a = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if prediction_type == "epsilon": _a = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif prediction_type == "sample": _a = model_output else: raise ValueError(f'Unsupported prediction_type {prediction_type}.' ) # 3. Clip "predicted x_0" _a = self.bit_scale if self.config.clip_sample: _a = torch.clamp(UpperCamelCase , -scale , UpperCamelCase ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _a = (alpha_prod_t_prev ** 0.5 * self.betas[t]) / beta_prod_t _a = self.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _a = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise _a = 0 if t > 0: _a = torch.randn( model_output.size() , dtype=model_output.dtype , layout=model_output.layout , generator=UpperCamelCase ).to(model_output.device ) _a = (self._get_variance(UpperCamelCase , predicted_variance=UpperCamelCase ) ** 0.5) * noise _a = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return DDPMSchedulerOutput(prev_sample=UpperCamelCase , pred_original_sample=UpperCamelCase ) class A ( _a ): def __init__( self : Any , lowerCAmelCase_ : UNetaDConditionModel , lowerCAmelCase_ : Union[DDIMScheduler, DDPMScheduler] , lowerCAmelCase_ : Optional[float] = 1.0 , ) -> int: """simple docstring""" super().__init__() _a = bit_scale _a = ( ddim_bit_scheduler_step if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else ddpm_bit_scheduler_step ) self.register_modules(unet=lowerCAmelCase_ , scheduler=lowerCAmelCase_ ) @torch.no_grad() def __call__( self : List[Any] , lowerCAmelCase_ : Optional[int] = 2_56 , lowerCAmelCase_ : Optional[int] = 2_56 , lowerCAmelCase_ : Optional[int] = 50 , lowerCAmelCase_ : Optional[torch.Generator] = None , lowerCAmelCase_ : Optional[int] = 1 , lowerCAmelCase_ : Optional[str] = "pil" , lowerCAmelCase_ : bool = True , **lowerCAmelCase_ : Any , ) -> Union[Tuple, ImagePipelineOutput]: """simple docstring""" _a = torch.randn( (batch_size, self.unet.config.in_channels, height, width) , generator=lowerCAmelCase_ , ) _a = decimal_to_bits(lowerCAmelCase_ ) * self.bit_scale _a = latents.to(self.device ) self.scheduler.set_timesteps(lowerCAmelCase_ ) for t in self.progress_bar(self.scheduler.timesteps ): # predict the noise residual _a = self.unet(lowerCAmelCase_ , lowerCAmelCase_ ).sample # compute the previous noisy sample x_t -> x_t-1 _a = self.scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ).prev_sample _a = bits_to_decimal(lowerCAmelCase_ ) if output_type == "pil": _a = self.numpy_to_pil(lowerCAmelCase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCAmelCase_ )
22
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import BlenderbotSmallConfig, BlenderbotSmallTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel @require_tf class _SCREAMING_SNAKE_CASE : A__ = BlenderbotSmallConfig A__ = {} A__ = 'gelu' def __init__( self : Union[str, Any] , __UpperCamelCase : int , __UpperCamelCase : str=13 , __UpperCamelCase : Dict=7 , __UpperCamelCase : int=True , __UpperCamelCase : Optional[int]=False , __UpperCamelCase : Optional[Any]=99 , __UpperCamelCase : List[str]=32 , __UpperCamelCase : Dict=2 , __UpperCamelCase : Union[str, Any]=4 , __UpperCamelCase : Optional[Any]=37 , __UpperCamelCase : List[str]=0.1 , __UpperCamelCase : Union[str, Any]=0.1 , __UpperCamelCase : str=20 , __UpperCamelCase : Dict=2 , __UpperCamelCase : List[Any]=1 , __UpperCamelCase : Optional[Any]=0 , ) -> List[str]: """simple docstring""" snake_case__ : Dict = parent snake_case__ : List[Any] = batch_size snake_case__ : int = seq_length snake_case__ : Union[str, Any] = is_training snake_case__ : Tuple = use_labels snake_case__ : Tuple = vocab_size snake_case__ : Union[str, Any] = hidden_size snake_case__ : List[str] = num_hidden_layers snake_case__ : Union[str, Any] = num_attention_heads snake_case__ : List[Any] = intermediate_size snake_case__ : int = hidden_dropout_prob snake_case__ : Any = attention_probs_dropout_prob snake_case__ : Any = max_position_embeddings snake_case__ : Dict = eos_token_id snake_case__ : List[Any] = pad_token_id snake_case__ : List[str] = bos_token_id def lowerCAmelCase ( self : Dict ) -> Optional[Any]: """simple docstring""" snake_case__ : Dict = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) snake_case__ : List[Any] = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) snake_case__ : Union[str, Any] = tf.concat([input_ids, eos_tensor] , axis=1 ) snake_case__ : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case__ : Optional[int] = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) snake_case__ : Dict = prepare_blenderbot_small_inputs_dict(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) return config, inputs_dict def lowerCAmelCase ( self : Union[str, Any] , __UpperCamelCase : Tuple , __UpperCamelCase : Tuple ) -> List[str]: """simple docstring""" snake_case__ : Optional[Any] = TFBlenderbotSmallModel(config=__UpperCamelCase ).get_decoder() snake_case__ : Dict = inputs_dict['''input_ids'''] snake_case__ : Optional[int] = input_ids[:1, :] snake_case__ : List[str] = inputs_dict['''attention_mask'''][:1, :] snake_case__ : Union[str, Any] = inputs_dict['''head_mask'''] snake_case__ : List[str] = 1 # first forward pass snake_case__ : Any = model(__UpperCamelCase , attention_mask=__UpperCamelCase , head_mask=__UpperCamelCase , use_cache=__UpperCamelCase ) snake_case__ , snake_case__ : Any = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case__ : Any = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case__ : Optional[Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and snake_case__ : Any = tf.concat([input_ids, next_tokens] , axis=-1 ) snake_case__ : int = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) snake_case__ : Union[str, Any] = model(__UpperCamelCase , attention_mask=__UpperCamelCase )[0] snake_case__ : List[Any] = model(__UpperCamelCase , attention_mask=__UpperCamelCase , past_key_values=__UpperCamelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice snake_case__ : List[Any] = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) snake_case__ : str = output_from_no_past[:, -3:, random_slice_idx] snake_case__ : Optional[int] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__UpperCamelCase , __UpperCamelCase , rtol=1e-3 ) def __UpperCAmelCase ( UpperCamelCase__ :List[str] , UpperCamelCase__ :Union[str, Any] , UpperCamelCase__ :Optional[int] , UpperCamelCase__ :Any=None , UpperCamelCase__ :Tuple=None , UpperCamelCase__ :Optional[int]=None , UpperCamelCase__ :int=None , UpperCamelCase__ :Tuple=None , ) -> Optional[int]: if attention_mask is None: snake_case__ : Tuple = tf.cast(tf.math.not_equal(UpperCamelCase__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: snake_case__ : Optional[Any] = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: snake_case__ : Optional[int] = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: snake_case__ : int = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: snake_case__ : Optional[Any] = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _SCREAMING_SNAKE_CASE (lowercase__, lowercase__, unittest.TestCase ): A__ = ( (TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel) if is_tf_available() else () ) A__ = (TFBlenderbotSmallForConditionalGeneration,) if is_tf_available() else () A__ = ( { 'conversational': TFBlenderbotSmallForConditionalGeneration, 'feature-extraction': TFBlenderbotSmallModel, 'summarization': TFBlenderbotSmallForConditionalGeneration, 'text2text-generation': TFBlenderbotSmallForConditionalGeneration, 'translation': TFBlenderbotSmallForConditionalGeneration, } if is_tf_available() else {} ) A__ = True A__ = False A__ = False def lowerCAmelCase ( self : str ) -> Dict: """simple docstring""" snake_case__ : List[str] = TFBlenderbotSmallModelTester(self ) snake_case__ : List[Any] = ConfigTester(self , config_class=__UpperCamelCase ) def lowerCAmelCase ( self : Tuple ) -> Optional[int]: """simple docstring""" self.config_tester.run_common_tests() def lowerCAmelCase ( self : Union[str, Any] ) -> str: """simple docstring""" snake_case__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__UpperCamelCase ) @require_tokenizers @require_tf class _SCREAMING_SNAKE_CASE (unittest.TestCase ): A__ = [ 'Social anxiety\nWow, I am never shy. Do you have anxiety?\nYes. I end up sweating and blushing and feel like ' ' i\'m going to throw up.\nand why is that?' ] A__ = 'facebook/blenderbot_small-90M' @cached_property def lowerCAmelCase ( self : Dict ) -> Union[str, Any]: """simple docstring""" return BlenderbotSmallTokenizer.from_pretrained('''facebook/blenderbot-90M''' ) @cached_property def lowerCAmelCase ( self : Dict ) -> Dict: """simple docstring""" snake_case__ : Optional[Any] = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def lowerCAmelCase ( self : List[Any] ) -> Optional[int]: """simple docstring""" snake_case__ : Union[str, Any] = self.tokenizer(self.src_text , return_tensors='''tf''' ) snake_case__ : Optional[Any] = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=__UpperCamelCase , ) snake_case__ : str = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=__UpperCamelCase )[0] assert generated_words in ( "i don't know. i just feel like i'm going to throw up. it's not fun.", "i'm not sure. i just feel like i've been feeling like i have to be in a certain place", "i'm not sure. i just feel like i've been in a bad situation.", )
574
'''simple docstring''' import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor _lowercase : List[str] =logging.get_logger(__name__) class _SCREAMING_SNAKE_CASE (lowercase__ ): def __init__( self : Any , *__UpperCamelCase : Optional[Any] , **__UpperCamelCase : Any ) -> None: """simple docstring""" warnings.warn( '''The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use GLPNImageProcessor instead.''' , __UpperCamelCase , ) super().__init__(*__UpperCamelCase , **__UpperCamelCase )
574
1
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' _lowercase : torch.FloatTensor _lowercase : torch.FloatTensor _lowercase : Optional[torch.FloatTensor] = None class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): '''simple docstring''' _lowercase : Optional[int] = 2 @register_to_config def __init__( self , _lowercase = 0.02 , _lowercase = 100 , _lowercase = 1.007 , _lowercase = 80 , _lowercase = 0.05 , _lowercase = 50 , ): """simple docstring""" _lowerCAmelCase = sigma_max # setable values _lowerCAmelCase = None _lowerCAmelCase = None _lowerCAmelCase = None # sigma(t_i) def _lowercase ( self , _lowercase , _lowercase = None ): """simple docstring""" return sample def _lowercase ( self , _lowercase , _lowercase = None ): """simple docstring""" _lowerCAmelCase = num_inference_steps _lowerCAmelCase = np.arange(0 , self.num_inference_steps )[::-1].copy() _lowerCAmelCase = torch.from_numpy(_lowercase ).to(_lowercase ) _lowerCAmelCase = [ ( self.config.sigma_max**2 * (self.config.sigma_min**2 / self.config.sigma_max**2) ** (i / (num_inference_steps - 1)) ) for i in self.timesteps ] _lowerCAmelCase = torch.tensor(_lowercase , dtype=torch.floataa , device=_lowercase ) def _lowercase ( self , _lowercase , _lowercase , _lowercase = None ): """simple docstring""" if self.config.s_min <= sigma <= self.config.s_max: _lowerCAmelCase = min(self.config.s_churn / self.num_inference_steps , 2**0.5 - 1 ) else: _lowerCAmelCase = 0 # sample eps ~ N(0, S_noise^2 * I) _lowerCAmelCase = self.config.s_noise * randn_tensor(sample.shape , generator=_lowercase ).to(sample.device ) _lowerCAmelCase = sigma + gamma * sigma _lowerCAmelCase = sample + ((sigma_hat**2 - sigma**2) ** 0.5 * eps) return sample_hat, sigma_hat def _lowercase ( self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase = True , ): """simple docstring""" _lowerCAmelCase = sample_hat + sigma_hat * model_output _lowerCAmelCase = (sample_hat - pred_original_sample) / sigma_hat _lowerCAmelCase = sample_hat + (sigma_prev - sigma_hat) * derivative if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=_lowercase , derivative=_lowercase , pred_original_sample=_lowercase ) def _lowercase ( self , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase = True , ): """simple docstring""" _lowerCAmelCase = sample_prev + sigma_prev * model_output _lowerCAmelCase = (sample_prev - pred_original_sample) / sigma_prev _lowerCAmelCase = sample_hat + (sigma_prev - sigma_hat) * (0.5 * derivative + 0.5 * derivative_corr) if not return_dict: return (sample_prev, derivative) return KarrasVeOutput( prev_sample=_lowercase , derivative=_lowercase , pred_original_sample=_lowercase ) def _lowercase ( self , _lowercase , _lowercase , _lowercase ): """simple docstring""" raise NotImplementedError()
5
'''simple docstring''' import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('''TEST_SAGEMAKER''' , '''False''' ) ) is not True , reason='''Skipping test because should only be run when releasing minor transformers version''' , ) @pytest.mark.usefixtures('''sm_env''' ) @parameterized_class( [ { '''framework''': '''pytorch''', '''script''': '''run_glue.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 650, '''eval_accuracy''': 0.6, '''eval_loss''': 0.9}, }, { '''framework''': '''tensorflow''', '''script''': '''run_tf.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 600, '''eval_accuracy''': 0.3, '''eval_loss''': 0.9}, }, ] ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase( self ): if self.framework == "pytorch": subprocess.run( F'''cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'''.split() , encoding="utf-8" , check=lowerCamelCase , ) assert hasattr(self , "env" ) def UpperCamelCase( self , lowerCamelCase=1 ): # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F'''{self.env.base_job_name}-single''' , instance_count=lowerCamelCase , instance_type=self.instance_type , debugger_hook_config=lowerCamelCase , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def UpperCamelCase( self , lowerCamelCase ): TrainingJobAnalytics(lowerCamelCase ).export_csv(F'''{self.env.test_path}/{job_name}_metrics.csv''' ) def UpperCamelCase( self ): # create estimator _snake_case = self.create_estimator() # run training estimator.fit() # result dataframe _snake_case = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis _snake_case = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) _snake_case = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping _snake_case = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 999_999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(F'''{estimator.latest_training_job.name}.json''' , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , lowerCamelCase )
672
0
import inspect from typing import Callable, List, Optional, Union import torch from transformers import ( CLIPImageProcessor, CLIPTextModel, CLIPTokenizer, WhisperForConditionalGeneration, WhisperProcessor, ) from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker from diffusers.utils import logging __lowerCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name class __SCREAMING_SNAKE_CASE ( lowercase): def __init__( self : Dict , __UpperCamelCase : WhisperForConditionalGeneration , __UpperCamelCase : WhisperProcessor , __UpperCamelCase : AutoencoderKL , __UpperCamelCase : CLIPTextModel , __UpperCamelCase : CLIPTokenizer , __UpperCamelCase : UNetaDConditionModel , __UpperCamelCase : Union[DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler] , __UpperCamelCase : StableDiffusionSafetyChecker , __UpperCamelCase : CLIPImageProcessor , ): super().__init__() if safety_checker is None: logger.warning( F'''You have disabled the safety checker for {self.__class__} by passing `safety_checker=None`. Ensure''' " that you abide to the conditions of the Stable Diffusion license and do not expose unfiltered" " results in services or applications open to the public. Both the diffusers team and Hugging Face" " strongly recommend to keep the safety filter enabled in all public facing circumstances, disabling" " it only for use-cases that involve analyzing network behavior or auditing its results. For more" " information, please have a look at https://github.com/huggingface/diffusers/pull/254 ." ) self.register_modules( speech_model=__UpperCamelCase , speech_processor=__UpperCamelCase , vae=__UpperCamelCase , text_encoder=__UpperCamelCase , tokenizer=__UpperCamelCase , unet=__UpperCamelCase , scheduler=__UpperCamelCase , feature_extractor=__UpperCamelCase , ) def UpperCAmelCase__ ( self : Optional[Any] , __UpperCamelCase : Optional[Union[str, int]] = "auto" ): if slice_size == "auto": _UpperCAmelCase = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(__UpperCamelCase ) def UpperCAmelCase__ ( self : List[str] ): self.enable_attention_slicing(__UpperCamelCase ) @torch.no_grad() def __call__( self : Optional[Any] , __UpperCamelCase : int , __UpperCamelCase : Tuple=16_000 , __UpperCamelCase : int = 512 , __UpperCamelCase : int = 512 , __UpperCamelCase : int = 50 , __UpperCamelCase : float = 7.5 , __UpperCamelCase : Optional[Union[str, List[str]]] = None , __UpperCamelCase : Optional[int] = 1 , __UpperCamelCase : float = 0.0 , __UpperCamelCase : Optional[torch.Generator] = None , __UpperCamelCase : Optional[torch.FloatTensor] = None , __UpperCamelCase : Optional[str] = "pil" , __UpperCamelCase : bool = True , __UpperCamelCase : Optional[Callable[[int, int, torch.FloatTensor], None]] = None , __UpperCamelCase : int = 1 , **__UpperCamelCase : Optional[Any] , ): _UpperCAmelCase = self.speech_processor.feature_extractor( __UpperCamelCase , return_tensors="pt" , sampling_rate=__UpperCamelCase ).input_features.to(self.device ) _UpperCAmelCase = self.speech_model.generate(__UpperCamelCase , max_length=480_000 ) _UpperCAmelCase = self.speech_processor.tokenizer.batch_decode(__UpperCamelCase , skip_special_tokens=__UpperCamelCase , normalize=__UpperCamelCase )[ 0 ] if isinstance(__UpperCamelCase , __UpperCamelCase ): _UpperCAmelCase = 1 elif isinstance(__UpperCamelCase , __UpperCamelCase ): _UpperCAmelCase = len(__UpperCamelCase ) else: raise ValueError(F'''`prompt` has to be of type `str` or `list` but is {type(__UpperCamelCase )}''' ) if height % 8 != 0 or width % 8 != 0: raise ValueError(F'''`height` and `width` have to be divisible by 8 but are {height} and {width}.''' ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__UpperCamelCase , __UpperCamelCase ) or callback_steps <= 0) ): raise ValueError( F'''`callback_steps` has to be a positive integer but is {callback_steps} of type''' F''' {type(__UpperCamelCase )}.''' ) # get prompt text embeddings _UpperCAmelCase = self.tokenizer( __UpperCamelCase , padding="max_length" , max_length=self.tokenizer.model_max_length , return_tensors="pt" , ) _UpperCAmelCase = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: _UpperCAmelCase = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( "The following part of your input was truncated because CLIP can only handle sequences up to" F''' {self.tokenizer.model_max_length} tokens: {removed_text}''' ) _UpperCAmelCase = text_input_ids[:, : self.tokenizer.model_max_length] _UpperCAmelCase = self.text_encoder(text_input_ids.to(self.device ) )[0] # duplicate text embeddings for each generation per prompt, using mps friendly method _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = text_embeddings.shape _UpperCAmelCase = text_embeddings.repeat(1 , __UpperCamelCase , 1 ) _UpperCAmelCase = text_embeddings.view(bs_embed * num_images_per_prompt , __UpperCamelCase , -1 ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. _UpperCAmelCase = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: _UpperCAmelCase = 42 if negative_prompt is None: _UpperCAmelCase = [""] * batch_size elif type(__UpperCamelCase ) is not type(__UpperCamelCase ): raise TypeError( F'''`negative_prompt` should be the same type to `prompt`, but got {type(__UpperCamelCase )} !=''' F''' {type(__UpperCamelCase )}.''' ) elif isinstance(__UpperCamelCase , __UpperCamelCase ): _UpperCAmelCase = [negative_prompt] elif batch_size != len(__UpperCamelCase ): raise ValueError( F'''`negative_prompt`: {negative_prompt} has batch size {len(__UpperCamelCase )}, but `prompt`:''' F''' {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches''' " the batch size of `prompt`." ) else: _UpperCAmelCase = negative_prompt _UpperCAmelCase = text_input_ids.shape[-1] _UpperCAmelCase = self.tokenizer( __UpperCamelCase , padding="max_length" , max_length=__UpperCamelCase , truncation=__UpperCamelCase , return_tensors="pt" , ) _UpperCAmelCase = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt, using mps friendly method _UpperCAmelCase = uncond_embeddings.shape[1] _UpperCAmelCase = uncond_embeddings.repeat(1 , __UpperCamelCase , 1 ) _UpperCAmelCase = uncond_embeddings.view(batch_size * num_images_per_prompt , __UpperCamelCase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes _UpperCAmelCase = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. _UpperCAmelCase = (batch_size * num_images_per_prompt, self.unet.config.in_channels, height // 8, width // 8) _UpperCAmelCase = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not exist on mps _UpperCAmelCase = torch.randn(__UpperCamelCase , generator=__UpperCamelCase , device="cpu" , dtype=__UpperCamelCase ).to( self.device ) else: _UpperCAmelCase = torch.randn(__UpperCamelCase , generator=__UpperCamelCase , device=self.device , dtype=__UpperCamelCase ) else: if latents.shape != latents_shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) _UpperCAmelCase = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(__UpperCamelCase ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand _UpperCAmelCase = self.scheduler.timesteps.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler _UpperCAmelCase = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] _UpperCAmelCase = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) _UpperCAmelCase = {} if accepts_eta: _UpperCAmelCase = eta for i, t in enumerate(self.progress_bar(__UpperCamelCase ) ): # expand the latents if we are doing classifier free guidance _UpperCAmelCase = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents _UpperCAmelCase = self.scheduler.scale_model_input(__UpperCamelCase , __UpperCamelCase ) # predict the noise residual _UpperCAmelCase = self.unet(__UpperCamelCase , __UpperCamelCase , encoder_hidden_states=__UpperCamelCase ).sample # perform guidance if do_classifier_free_guidance: _UpperCAmelCase , _UpperCAmelCase = noise_pred.chunk(2 ) _UpperCAmelCase = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # compute the previous noisy sample x_t -> x_t-1 _UpperCAmelCase = self.scheduler.step(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , **__UpperCamelCase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) _UpperCAmelCase = 1 / 0.18215 * latents _UpperCAmelCase = self.vae.decode(__UpperCamelCase ).sample _UpperCAmelCase = (image / 2 + 0.5).clamp(0 , 1 ) # we always cast to float32 as this does not cause significant overhead and is compatible with bfloat16 _UpperCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": _UpperCAmelCase = self.numpy_to_pil(__UpperCamelCase ) if not return_dict: return image return StableDiffusionPipelineOutput(images=__UpperCamelCase , nsfw_content_detected=__UpperCamelCase )
700
import json import os import shutil import warnings from argparse import ArgumentParser, Namespace from pathlib import Path from typing import List from ..utils import logging from . import BaseTransformersCLICommand try: from cookiecutter.main import cookiecutter __lowerCAmelCase = True except ImportError: __lowerCAmelCase = False __lowerCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name def __lowerCamelCase ( _lowerCAmelCase ) -> str: return AddNewModelCommand(args.testing , args.testing_file , path=args.path ) class __SCREAMING_SNAKE_CASE ( lowercase): @staticmethod def UpperCAmelCase__ ( __UpperCamelCase : ArgumentParser ): _UpperCAmelCase = parser.add_parser("add-new-model" ) add_new_model_parser.add_argument("--testing" , action="store_true" , help="If in testing mode." ) add_new_model_parser.add_argument("--testing_file" , type=__UpperCamelCase , help="Configuration file on which to run." ) add_new_model_parser.add_argument( "--path" , type=__UpperCamelCase , help="Path to cookiecutter. Should only be used for testing purposes." ) add_new_model_parser.set_defaults(func=__UpperCamelCase ) def __init__( self : List[Any] , __UpperCamelCase : bool , __UpperCamelCase : str , __UpperCamelCase : List[str]=None , *__UpperCamelCase : List[Any] ): _UpperCAmelCase = testing _UpperCAmelCase = testing_file _UpperCAmelCase = path def UpperCAmelCase__ ( self : List[Any] ): warnings.warn( "The command `transformers-cli add-new-model` is deprecated and will be removed in v5 of Transformers. " "It is not actively maintained anymore, so might give a result that won't pass all tests and quality " "checks, you should use `transformers-cli add-new-model-like` instead." ) if not _has_cookiecutter: raise ImportError( "Model creation dependencies are required to use the `add_new_model` command. Install them by running " "the following at the root of your `transformers` clone:\n\n\t$ pip install -e .[modelcreation]\n" ) # Ensure that there is no other `cookiecutter-template-xxx` directory in the current working directory _UpperCAmelCase = [directory for directory in os.listdir() if "cookiecutter-template-" == directory[:22]] if len(__UpperCamelCase ) > 0: raise ValueError( "Several directories starting with `cookiecutter-template-` in current working directory. " "Please clean your directory by removing all folders starting with `cookiecutter-template-` or " "change your working directory." ) _UpperCAmelCase = ( Path(__UpperCamelCase ).parent.parent.parent.parent if self._path is None else Path(self._path ).parent.parent ) _UpperCAmelCase = path_to_transformer_root / "templates" / "adding_a_new_model" # Execute cookiecutter if not self._testing: cookiecutter(str(__UpperCamelCase ) ) else: with open(self._testing_file , "r" ) as configuration_file: _UpperCAmelCase = json.load(__UpperCamelCase ) cookiecutter( str(path_to_cookiecutter if self._path is None else self._path ) , no_input=__UpperCamelCase , extra_context=__UpperCamelCase , ) _UpperCAmelCase = [directory for directory in os.listdir() if "cookiecutter-template-" in directory[:22]][0] # Retrieve configuration with open(directory + "/configuration.json" , "r" ) as configuration_file: _UpperCAmelCase = json.load(__UpperCamelCase ) _UpperCAmelCase = configuration["lowercase_modelname"] _UpperCAmelCase = configuration["generate_tensorflow_pytorch_and_flax"] os.remove(F'''{directory}/configuration.json''' ) _UpperCAmelCase = "PyTorch" in generate_tensorflow_pytorch_and_flax _UpperCAmelCase = "TensorFlow" in generate_tensorflow_pytorch_and_flax _UpperCAmelCase = "Flax" in generate_tensorflow_pytorch_and_flax _UpperCAmelCase = F'''{path_to_transformer_root}/src/transformers/models/{lowercase_model_name}''' os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) os.makedirs(F'''{path_to_transformer_root}/tests/models/{lowercase_model_name}''' , exist_ok=__UpperCamelCase ) # Tests require submodules as they have parent imports with open(F'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/__init__.py''' , "w" ): pass shutil.move( F'''{directory}/__init__.py''' , F'''{model_dir}/__init__.py''' , ) shutil.move( F'''{directory}/configuration_{lowercase_model_name}.py''' , F'''{model_dir}/configuration_{lowercase_model_name}.py''' , ) def remove_copy_lines(__UpperCamelCase : str ): with open(__UpperCamelCase , "r" ) as f: _UpperCAmelCase = f.readlines() with open(__UpperCamelCase , "w" ) as f: for line in lines: if "# Copied from transformers." not in line: f.write(__UpperCamelCase ) if output_pytorch: if not self._testing: remove_copy_lines(F'''{directory}/modeling_{lowercase_model_name}.py''' ) shutil.move( F'''{directory}/modeling_{lowercase_model_name}.py''' , F'''{model_dir}/modeling_{lowercase_model_name}.py''' , ) shutil.move( F'''{directory}/test_modeling_{lowercase_model_name}.py''' , F'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_{lowercase_model_name}.py''' , ) else: os.remove(F'''{directory}/modeling_{lowercase_model_name}.py''' ) os.remove(F'''{directory}/test_modeling_{lowercase_model_name}.py''' ) if output_tensorflow: if not self._testing: remove_copy_lines(F'''{directory}/modeling_tf_{lowercase_model_name}.py''' ) shutil.move( F'''{directory}/modeling_tf_{lowercase_model_name}.py''' , F'''{model_dir}/modeling_tf_{lowercase_model_name}.py''' , ) shutil.move( F'''{directory}/test_modeling_tf_{lowercase_model_name}.py''' , F'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_tf_{lowercase_model_name}.py''' , ) else: os.remove(F'''{directory}/modeling_tf_{lowercase_model_name}.py''' ) os.remove(F'''{directory}/test_modeling_tf_{lowercase_model_name}.py''' ) if output_flax: if not self._testing: remove_copy_lines(F'''{directory}/modeling_flax_{lowercase_model_name}.py''' ) shutil.move( F'''{directory}/modeling_flax_{lowercase_model_name}.py''' , F'''{model_dir}/modeling_flax_{lowercase_model_name}.py''' , ) shutil.move( F'''{directory}/test_modeling_flax_{lowercase_model_name}.py''' , F'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_flax_{lowercase_model_name}.py''' , ) else: os.remove(F'''{directory}/modeling_flax_{lowercase_model_name}.py''' ) os.remove(F'''{directory}/test_modeling_flax_{lowercase_model_name}.py''' ) shutil.move( F'''{directory}/{lowercase_model_name}.md''' , F'''{path_to_transformer_root}/docs/source/en/model_doc/{lowercase_model_name}.md''' , ) shutil.move( F'''{directory}/tokenization_{lowercase_model_name}.py''' , F'''{model_dir}/tokenization_{lowercase_model_name}.py''' , ) shutil.move( F'''{directory}/tokenization_fast_{lowercase_model_name}.py''' , F'''{model_dir}/tokenization_{lowercase_model_name}_fast.py''' , ) from os import fdopen, remove from shutil import copymode, move from tempfile import mkstemp def replace(__UpperCamelCase : str , __UpperCamelCase : str , __UpperCamelCase : List[str] ): # Create temp file _UpperCAmelCase , _UpperCAmelCase = mkstemp() _UpperCAmelCase = False with fdopen(__UpperCamelCase , "w" ) as new_file: with open(__UpperCamelCase ) as old_file: for line in old_file: new_file.write(__UpperCamelCase ) if line_to_copy_below in line: _UpperCAmelCase = True for line_to_copy in lines_to_copy: new_file.write(__UpperCamelCase ) if not line_found: raise ValueError(F'''Line {line_to_copy_below} was not found in file.''' ) # Copy the file permissions from the old file to the new file copymode(__UpperCamelCase , __UpperCamelCase ) # Remove original file remove(__UpperCamelCase ) # Move new file move(__UpperCamelCase , __UpperCamelCase ) def skip_units(__UpperCamelCase : List[Any] ): return ( ("generating PyTorch" in line and not output_pytorch) or ("generating TensorFlow" in line and not output_tensorflow) or ("generating Flax" in line and not output_flax) ) def replace_in_files(__UpperCamelCase : Union[str, Any] ): with open(__UpperCamelCase ) as datafile: _UpperCAmelCase = [] _UpperCAmelCase = False _UpperCAmelCase = False for line in datafile: if "# To replace in: " in line and "##" not in line: _UpperCAmelCase = line.split("\"" )[1] _UpperCAmelCase = skip_units(__UpperCamelCase ) elif "# Below: " in line and "##" not in line: _UpperCAmelCase = line.split("\"" )[1] _UpperCAmelCase = skip_units(__UpperCamelCase ) elif "# End." in line and "##" not in line: if not skip_file and not skip_snippet: replace(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) _UpperCAmelCase = [] elif "# Replace with" in line and "##" not in line: _UpperCAmelCase = [] elif "##" not in line: lines_to_copy.append(__UpperCamelCase ) remove(__UpperCamelCase ) replace_in_files(F'''{directory}/to_replace_{lowercase_model_name}.py''' ) os.rmdir(__UpperCamelCase )
129
0
"""simple docstring""" from __future__ import annotations def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase ) -> Optional[int]: print(F"""Vertex\tShortest Distance from vertex {src}""" ) for i, d in enumerate(snake_case_ ): print(F"""{i}\t\t{d}""" ) def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> str: for j in range(snake_case_ ): lowercase__, lowercase__, lowercase__: int = (graph[j][k] for k in ['''src''', '''dst''', '''weight''']) if distance[u] != float('''inf''' ) and distance[u] + w < distance[v]: return True return False def SCREAMING_SNAKE_CASE__ ( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) -> list[float]: lowercase__: Dict = [float('''inf''' )] * vertex_count lowercase__: List[str] = 0.0 for _ in range(vertex_count - 1 ): for j in range(snake_case_ ): lowercase__, lowercase__, lowercase__: Union[str, Any] = (graph[j][k] for k in ['''src''', '''dst''', '''weight''']) if distance[u] != float('''inf''' ) and distance[u] + w < distance[v]: lowercase__: List[Any] = distance[u] + w lowercase__: List[str] = check_negative_cycle(snake_case_ , snake_case_ , snake_case_ ) if negative_cycle_exists: raise Exception('''Negative cycle found''' ) return distance if __name__ == "__main__": import doctest doctest.testmod() __A = int(input("Enter number of vertices: ").strip()) __A = int(input("Enter number of edges: ").strip()) __A = [{} for _ in range(E)] for i in range(E): print("Edge ", i + 1) __A = ( int(x) for x in input("Enter source, destination, weight: ").strip().split(" ") ) __A = {'''src''': src, '''dst''': dest, '''weight''': weight} __A = int(input("\nEnter shortest path source:").strip()) __A = bellman_ford(graph, V, E, source) print_distance(shortest_distance, 0)
586
'''simple docstring''' def UpperCamelCase_ ( snake_case_ : list[int] ) -> list[list[int]]: '''simple docstring''' __lowerCAmelCase = [] if len(snake_case_ ) == 1: return [nums.copy()] for _ in range(len(snake_case_ ) ): __lowerCAmelCase = nums.pop(0 ) __lowerCAmelCase = permute(snake_case_ ) for perm in permutations: perm.append(snake_case_ ) result.extend(snake_case_ ) nums.append(snake_case_ ) return result def UpperCamelCase_ ( snake_case_ : Any ) -> Dict: '''simple docstring''' def backtrack(snake_case_ : List[Any] ): if start == len(snake_case_ ) - 1: output.append(nums[:] ) else: for i in range(snake_case_ , len(snake_case_ ) ): __lowerCAmelCase , __lowerCAmelCase = nums[i], nums[start] backtrack(start + 1 ) __lowerCAmelCase , __lowerCAmelCase = nums[i], nums[start] # backtrack __lowerCAmelCase = [] backtrack(0 ) return output if __name__ == "__main__": import doctest # use res to print the data in permute2 function _A : Union[str, Any] = permutea([1, 2, 3]) print(res) doctest.testmod()
427
0
'''simple docstring''' import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class a_ ( unittest.TestCase ): lowerCamelCase__ : Tuple = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def lowerCAmelCase__ ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): a_ = hf_hub_download( repo_id="""nateraw/video-demo""" , filename="""archery.mp4""" , repo_type="""dataset""" ) a_ = VideoClassificationPipeline(model=UpperCAmelCase , image_processor=UpperCAmelCase , top_k=2 ) a_ = [ example_video_filepath, """https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4""", ] return video_classifier, examples def lowerCAmelCase__ ( self , UpperCAmelCase , UpperCAmelCase ): for example in examples: a_ = video_classifier(UpperCAmelCase ) self.assertEqual( UpperCAmelCase , [ {"""score""": ANY(UpperCAmelCase ), """label""": ANY(UpperCAmelCase )}, {"""score""": ANY(UpperCAmelCase ), """label""": ANY(UpperCAmelCase )}, ] , ) @require_torch def lowerCAmelCase__ ( self ): a_ = """hf-internal-testing/tiny-random-VideoMAEForVideoClassification""" a_ = VideoMAEFeatureExtractor( size={"""shortest_edge""": 10} , crop_size={"""height""": 10, """width""": 10} ) a_ = pipeline( """video-classification""" , model=UpperCAmelCase , feature_extractor=UpperCAmelCase , frame_sampling_rate=4 ) a_ = hf_hub_download(repo_id="""nateraw/video-demo""" , filename="""archery.mp4""" , repo_type="""dataset""" ) a_ = video_classifier(UpperCAmelCase , top_k=2 ) self.assertEqual( nested_simplify(UpperCAmelCase , decimals=4 ) , [{"""score""": 0.51_99, """label""": """LABEL_0"""}, {"""score""": 0.48_01, """label""": """LABEL_1"""}] , ) a_ = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(UpperCAmelCase , decimals=4 ) , [ [{"""score""": 0.51_99, """label""": """LABEL_0"""}, {"""score""": 0.48_01, """label""": """LABEL_1"""}], [{"""score""": 0.51_99, """label""": """LABEL_0"""}, {"""score""": 0.48_01, """label""": """LABEL_1"""}], ] , ) @require_tf def lowerCAmelCase__ ( self ): pass
705
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, Encoder, VectorQuantizer @dataclass class a_ ( UpperCamelCase__ ): lowerCamelCase__ : torch.FloatTensor class a_ ( UpperCamelCase__ , UpperCamelCase__ ): @register_to_config def __init__( self , UpperCAmelCase = 3 , UpperCAmelCase = 3 , UpperCAmelCase = ("DownEncoderBlock2D",) , UpperCAmelCase = ("UpDecoderBlock2D",) , UpperCAmelCase = (64,) , UpperCAmelCase = 1 , UpperCAmelCase = "silu" , UpperCAmelCase = 3 , UpperCAmelCase = 32 , UpperCAmelCase = 2_56 , UpperCAmelCase = 32 , UpperCAmelCase = None , UpperCAmelCase = 0.1_82_15 , UpperCAmelCase = "group" , ): super().__init__() # pass init params to Encoder a_ = Encoder( in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , down_block_types=UpperCAmelCase , block_out_channels=UpperCAmelCase , layers_per_block=UpperCAmelCase , act_fn=UpperCAmelCase , norm_num_groups=UpperCAmelCase , double_z=UpperCAmelCase , ) a_ = vq_embed_dim if vq_embed_dim is not None else latent_channels a_ = nn.Convad(UpperCAmelCase , UpperCAmelCase , 1 ) a_ = VectorQuantizer(UpperCAmelCase , UpperCAmelCase , beta=0.25 , remap=UpperCAmelCase , sane_index_shape=UpperCAmelCase ) a_ = nn.Convad(UpperCAmelCase , UpperCAmelCase , 1 ) # pass init params to Decoder a_ = Decoder( in_channels=UpperCAmelCase , out_channels=UpperCAmelCase , up_block_types=UpperCAmelCase , block_out_channels=UpperCAmelCase , layers_per_block=UpperCAmelCase , act_fn=UpperCAmelCase , norm_num_groups=UpperCAmelCase , norm_type=UpperCAmelCase , ) @apply_forward_hook def lowerCAmelCase__ ( self , UpperCAmelCase , UpperCAmelCase = True ): a_ = self.encoder(UpperCAmelCase ) a_ = self.quant_conv(UpperCAmelCase ) if not return_dict: return (h,) return VQEncoderOutput(latents=UpperCAmelCase ) @apply_forward_hook def lowerCAmelCase__ ( self , UpperCAmelCase , UpperCAmelCase = False , UpperCAmelCase = True ): # also go through quantization layer if not force_not_quantize: a_ , a_ , a_ = self.quantize(UpperCAmelCase ) else: a_ = h a_ = self.post_quant_conv(UpperCAmelCase ) a_ = self.decoder(UpperCAmelCase , quant if self.config.norm_type == """spatial""" else None ) if not return_dict: return (dec,) return DecoderOutput(sample=UpperCAmelCase ) def lowerCAmelCase__ ( self , UpperCAmelCase , UpperCAmelCase = True ): a_ = sample a_ = self.encode(UpperCAmelCase ).latents a_ = self.decode(UpperCAmelCase ).sample if not return_dict: return (dec,) return DecoderOutput(sample=UpperCAmelCase )
511
0
'''simple docstring''' from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _UpperCAmelCase : Any = { '''configuration_autoformer''': [ '''AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''AutoformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCAmelCase : int = [ '''AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''AutoformerForPrediction''', '''AutoformerModel''', '''AutoformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_autoformer import ( AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_autoformer import ( AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, AutoformerForPrediction, AutoformerModel, AutoformerPreTrainedModel, ) else: import sys _UpperCAmelCase : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
107
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowerCamelCase = { '''configuration_whisper''': ['''WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''WhisperConfig''', '''WhisperOnnxConfig'''], '''feature_extraction_whisper''': ['''WhisperFeatureExtractor'''], '''processing_whisper''': ['''WhisperProcessor'''], '''tokenization_whisper''': ['''WhisperTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = ['''WhisperTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = [ '''WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''WhisperForConditionalGeneration''', '''WhisperModel''', '''WhisperPreTrainedModel''', '''WhisperForAudioClassification''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = [ '''TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWhisperForConditionalGeneration''', '''TFWhisperModel''', '''TFWhisperPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase = [ '''FlaxWhisperForConditionalGeneration''', '''FlaxWhisperModel''', '''FlaxWhisperPreTrainedModel''', '''FlaxWhisperForAudioClassification''', ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys __lowerCamelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
288
0
import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def __a ( __UpperCAmelCase = 3 ): if isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise TypeError('''number of qubits must be a integer.''' ) if number_of_qubits <= 0: raise ValueError('''number of qubits must be > 0.''' ) if math.floor(__UpperCAmelCase ) != number_of_qubits: raise ValueError('''number of qubits must be exact integer.''' ) if number_of_qubits > 10: raise ValueError('''number of qubits too large to simulate(>10).''' ) a__ = QuantumRegister(__UpperCAmelCase , '''qr''' ) a__ = ClassicalRegister(__UpperCAmelCase , '''cr''' ) a__ = QuantumCircuit(__UpperCAmelCase , __UpperCAmelCase ) a__ = number_of_qubits for i in range(__UpperCAmelCase ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(__UpperCAmelCase ): quantum_circuit.cp(np.pi / 2 ** (counter - j) , __UpperCAmelCase , __UpperCAmelCase ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(__UpperCAmelCase , number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(__UpperCAmelCase , __UpperCAmelCase ) # simulate with 10000 shots a__ = Aer.get_backend('''qasm_simulator''' ) a__ = execute(__UpperCAmelCase , __UpperCAmelCase , shots=1_0000 ) return job.result().get_counts(__UpperCAmelCase ) if __name__ == "__main__": print( f'Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}' )
148
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available a_ : Optional[Any] = { 'configuration_mask2former': [ 'MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Mask2FormerConfig', ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : int = ['Mask2FormerImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : List[Any] = [ 'MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'Mask2FormerForUniversalSegmentation', 'Mask2FormerModel', 'Mask2FormerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_maskaformer import MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskaFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_maskaformer import MaskaFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskaformer import ( MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskaFormerForUniversalSegmentation, MaskaFormerModel, MaskaFormerPreTrainedModel, ) else: import sys a_ : Any = _LazyModule(__name__, globals()['__file__'], _import_structure)
148
1
from json import JSONDecodeError # Workaround for requests.exceptions.JSONDecodeError import requests def A ( lowercase__ : str = "isbn/0140328726" ) -> dict: UpperCamelCase__ :Optional[Any] = olid.strip().strip("""/""" ) # Remove leading/trailing whitespace & slashes if new_olid.count("""/""" ) != 1: UpperCamelCase__ :str = f"""{olid} is not a valid Open Library olid""" raise ValueError(a_ ) return requests.get(f"""https://openlibrary.org/{new_olid}.json""" ).json() def A ( lowercase__ : dict ) -> dict: UpperCamelCase__ :List[str] = { '''title''': '''Title''', '''publish_date''': '''Publish date''', '''authors''': '''Authors''', '''number_of_pages''': '''Number of pages:''', '''first_sentence''': '''First sentence''', '''isbn_10''': '''ISBN (10)''', '''isbn_13''': '''ISBN (13)''', } UpperCamelCase__ :Optional[int] = {better_key: ol_book_data[key] for key, better_key in desired_keys.items()} UpperCamelCase__ :str = [ get_openlibrary_data(author["""key"""] )['''name'''] for author in data['''Authors'''] ] UpperCamelCase__ :Optional[int] = data['''First sentence''']['''value'''] for key, value in data.items(): if isinstance(a_ , a_ ): UpperCamelCase__ :Union[str, Any] = ''', '''.join(a_ ) return data if __name__ == "__main__": import doctest doctest.testmod() while True: UpperCamelCase = input("\nEnter the ISBN code to search (or \'quit\' to stop): ").strip() if isbn.lower() in ("", "q", "quit", "exit", "stop"): break if len(isbn) not in (10, 13) or not isbn.isdigit(): print(f'''Sorry, {isbn} is not a valid ISBN. Please, input a valid ISBN.''') continue print(f'''\nSearching Open Library for ISBN: {isbn}...\n''') try: UpperCamelCase = summarize_book(get_openlibrary_data(f'''isbn/{isbn}''')) print("\n".join(f'''{key}: {value}''' for key, value in book_summary.items())) except JSONDecodeError: # Workaround for requests.exceptions.RequestException: print(f'''Sorry, there are no results for ISBN: {isbn}.''')
45
"""simple docstring""" from __future__ import annotations import unittest from transformers import MobileBertConfig, is_tf_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_MODEL_FOR_PRETRAINING_MAPPING, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertModel, ) @require_tf class __magic_name__ ( __UpperCAmelCase , __UpperCAmelCase , unittest.TestCase ): __A : Any = ( ( TFMobileBertModel, TFMobileBertForMaskedLM, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertForMultipleChoice, ) if is_tf_available() else () ) __A : List[Any] = ( { "feature-extraction": TFMobileBertModel, "fill-mask": TFMobileBertForMaskedLM, "question-answering": TFMobileBertForQuestionAnswering, "text-classification": TFMobileBertForSequenceClassification, "token-classification": TFMobileBertForTokenClassification, "zero-shot": TFMobileBertForSequenceClassification, } if is_tf_available() else {} ) __A : List[str] = False __A : int = False def __snake_case ( self : Optional[Any] , snake_case__ : Optional[Any] , snake_case__ : Any , snake_case__ : int=False ): '''simple docstring''' lowercase :Union[str, Any] = super()._prepare_for_class(snake_case__ , snake_case__ , return_labels=snake_case__ ) if return_labels: if model_class in get_values(snake_case__ ): lowercase :Tuple = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) return inputs_dict class __magic_name__ ( __UpperCAmelCase ): def __init__( self : Any , snake_case__ : Dict , snake_case__ : Dict=1_3 , snake_case__ : Tuple=7 , snake_case__ : Optional[Any]=True , snake_case__ : Union[str, Any]=True , snake_case__ : str=True , snake_case__ : Optional[Any]=True , snake_case__ : Any=9_9 , snake_case__ : Optional[Any]=3_2 , snake_case__ : Optional[Any]=3_2 , snake_case__ : Any=2 , snake_case__ : Optional[int]=4 , snake_case__ : List[Any]=3_7 , snake_case__ : Optional[int]="gelu" , snake_case__ : List[Any]=0.1 , snake_case__ : str=0.1 , snake_case__ : List[Any]=5_1_2 , snake_case__ : List[str]=1_6 , snake_case__ : Union[str, Any]=2 , snake_case__ : Optional[Any]=0.02 , snake_case__ : Optional[Any]=3 , snake_case__ : Dict=4 , snake_case__ : int=None , ): '''simple docstring''' lowercase :Tuple = parent lowercase :Tuple = batch_size lowercase :Optional[Any] = seq_length lowercase :Optional[Any] = is_training lowercase :Optional[Any] = use_input_mask lowercase :List[Any] = use_token_type_ids lowercase :str = use_labels lowercase :List[str] = vocab_size lowercase :str = hidden_size lowercase :Optional[int] = num_hidden_layers lowercase :Dict = num_attention_heads lowercase :Any = intermediate_size lowercase :List[str] = hidden_act lowercase :Optional[Any] = hidden_dropout_prob lowercase :List[Any] = attention_probs_dropout_prob lowercase :List[Any] = max_position_embeddings lowercase :List[Any] = type_vocab_size lowercase :Union[str, Any] = type_sequence_label_size lowercase :Union[str, Any] = initializer_range lowercase :Any = num_labels lowercase :int = num_choices lowercase :Dict = scope lowercase :Dict = embedding_size def __snake_case ( self : Tuple ): '''simple docstring''' lowercase :Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase :int = None if self.use_input_mask: lowercase :int = random_attention_mask([self.batch_size, self.seq_length] ) lowercase :Tuple = None if self.use_token_type_ids: lowercase :int = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase :Union[str, Any] = None lowercase :int = None lowercase :str = None if self.use_labels: lowercase :int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase :str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase :Dict = ids_tensor([self.batch_size] , self.num_choices ) lowercase :Optional[int] = MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , embedding_size=self.embedding_size , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __snake_case ( self : Union[str, Any] , snake_case__ : Tuple , snake_case__ : str , snake_case__ : Tuple , snake_case__ : str , snake_case__ : Any , snake_case__ : Tuple , snake_case__ : Tuple ): '''simple docstring''' lowercase :Dict = TFMobileBertModel(config=snake_case__ ) lowercase :Union[str, Any] = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase :List[Any] = model(snake_case__ ) lowercase :Optional[int] = [input_ids, input_mask] lowercase :Optional[int] = model(snake_case__ ) lowercase :Union[str, Any] = model(snake_case__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def __snake_case ( self : List[Any] , snake_case__ : Dict , snake_case__ : Optional[int] , snake_case__ : int , snake_case__ : str , snake_case__ : Dict , snake_case__ : Tuple , snake_case__ : Optional[int] ): '''simple docstring''' lowercase :Any = TFMobileBertForMaskedLM(config=snake_case__ ) lowercase :Any = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase :int = model(snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __snake_case ( self : Tuple , snake_case__ : Any , snake_case__ : Optional[Any] , snake_case__ : List[Any] , snake_case__ : Dict , snake_case__ : str , snake_case__ : List[Any] , snake_case__ : Tuple ): '''simple docstring''' lowercase :Optional[Any] = TFMobileBertForNextSentencePrediction(config=snake_case__ ) lowercase :Tuple = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase :Optional[Any] = model(snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def __snake_case ( self : Dict , snake_case__ : Optional[Any] , snake_case__ : Optional[int] , snake_case__ : str , snake_case__ : Optional[int] , snake_case__ : int , snake_case__ : Union[str, Any] , snake_case__ : Dict ): '''simple docstring''' lowercase :int = TFMobileBertForPreTraining(config=snake_case__ ) lowercase :Any = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase :List[Any] = model(snake_case__ ) self.parent.assertEqual( result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def __snake_case ( self : Optional[Any] , snake_case__ : Tuple , snake_case__ : Tuple , snake_case__ : str , snake_case__ : Optional[Any] , snake_case__ : List[str] , snake_case__ : Dict , snake_case__ : Optional[Any] ): '''simple docstring''' lowercase :List[Any] = self.num_labels lowercase :List[Any] = TFMobileBertForSequenceClassification(config=snake_case__ ) lowercase :Dict = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase :List[Any] = model(snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __snake_case ( self : Any , snake_case__ : Optional[int] , snake_case__ : Tuple , snake_case__ : Union[str, Any] , snake_case__ : List[Any] , snake_case__ : List[Any] , snake_case__ : List[Any] , snake_case__ : Optional[Any] ): '''simple docstring''' lowercase :Tuple = self.num_choices lowercase :Any = TFMobileBertForMultipleChoice(config=snake_case__ ) lowercase :Any = tf.tile(tf.expand_dims(snake_case__ , 1 ) , (1, self.num_choices, 1) ) lowercase :Union[str, Any] = tf.tile(tf.expand_dims(snake_case__ , 1 ) , (1, self.num_choices, 1) ) lowercase :List[Any] = tf.tile(tf.expand_dims(snake_case__ , 1 ) , (1, self.num_choices, 1) ) lowercase :Dict = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } lowercase :Optional[Any] = model(snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __snake_case ( self : Any , snake_case__ : str , snake_case__ : Tuple , snake_case__ : int , snake_case__ : Tuple , snake_case__ : Dict , snake_case__ : Optional[Any] , snake_case__ : Dict ): '''simple docstring''' lowercase :List[Any] = self.num_labels lowercase :List[str] = TFMobileBertForTokenClassification(config=snake_case__ ) lowercase :int = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase :int = model(snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __snake_case ( self : List[str] , snake_case__ : List[Any] , snake_case__ : Optional[int] , snake_case__ : Optional[int] , snake_case__ : List[Any] , snake_case__ : Dict , snake_case__ : Dict , snake_case__ : str ): '''simple docstring''' lowercase :Union[str, Any] = TFMobileBertForQuestionAnswering(config=snake_case__ ) lowercase :List[Any] = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} lowercase :str = model(snake_case__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __snake_case ( self : List[Any] ): '''simple docstring''' lowercase :Dict = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) :Dict = config_and_inputs lowercase :Optional[Any] = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict def __snake_case ( self : Optional[Any] ): '''simple docstring''' lowercase :List[Any] = TFMobileBertModelTest.TFMobileBertModelTester(self ) lowercase :List[str] = ConfigTester(self , config_class=snake_case__ , hidden_size=3_7 ) def __snake_case ( self : List[Any] ): '''simple docstring''' self.config_tester.run_common_tests() def __snake_case ( self : Union[str, Any] ): '''simple docstring''' lowercase :List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*snake_case__ ) def __snake_case ( self : Any ): '''simple docstring''' lowercase :List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*snake_case__ ) def __snake_case ( self : Optional[Any] ): '''simple docstring''' lowercase :str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*snake_case__ ) def __snake_case ( self : Union[str, Any] ): '''simple docstring''' lowercase :Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*snake_case__ ) def __snake_case ( self : Optional[int] ): '''simple docstring''' lowercase :int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*snake_case__ ) def __snake_case ( self : List[str] ): '''simple docstring''' lowercase :Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*snake_case__ ) def __snake_case ( self : List[str] ): '''simple docstring''' lowercase :Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*snake_case__ ) def __snake_case ( self : Dict ): '''simple docstring''' lowercase :Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*snake_case__ ) @slow def __snake_case ( self : int ): '''simple docstring''' for model_name in ["google/mobilebert-uncased"]: lowercase :List[str] = TFMobileBertModel.from_pretrained(snake_case__ ) self.assertIsNotNone(snake_case__ ) @require_tf class __magic_name__ ( unittest.TestCase ): @slow def __snake_case ( self : Tuple ): '''simple docstring''' lowercase :int = TFMobileBertForPreTraining.from_pretrained('''google/mobilebert-uncased''' ) lowercase :Optional[Any] = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowercase :List[Any] = model(snake_case__ )[0] lowercase :Union[str, Any] = [1, 6, 3_0_5_2_2] self.assertEqual(output.shape , snake_case__ ) lowercase :Optional[int] = tf.constant( [ [ [-4.5_91_95_47, -9.24_82_95, -9.64_52_56], [-6.7_30_61_75, -6.44_02_84, -6.6_05_28_37], [-7.2_74_35_06, -6.7_84_79_15, -6.02_46_73], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , snake_case__ , atol=1e-4 )
677
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpec from ...tokenization_utils_base import PreTrainedTokenizerBase __SCREAMING_SNAKE_CASE =logging.get_logger(__name__) __SCREAMING_SNAKE_CASE ={ """allenai/longformer-base-4096""": """https://huggingface.co/allenai/longformer-base-4096/resolve/main/config.json""", """allenai/longformer-large-4096""": """https://huggingface.co/allenai/longformer-large-4096/resolve/main/config.json""", """allenai/longformer-large-4096-finetuned-triviaqa""": ( """https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/config.json""" ), """allenai/longformer-base-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/config.json""" ), """allenai/longformer-large-4096-extra.pos.embd.only""": ( """https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/config.json""" ), } class __magic_name__ ( __UpperCAmelCase): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = "longformer" def __init__( self: Union[str, Any] , _lowerCamelCase: Union[List[int], int] = 5_12 , _lowerCamelCase: int = 2 , _lowerCamelCase: int = 1 , _lowerCamelCase: int = 0 , _lowerCamelCase: int = 2 , _lowerCamelCase: int = 3_05_22 , _lowerCamelCase: int = 7_68 , _lowerCamelCase: int = 12 , _lowerCamelCase: int = 12 , _lowerCamelCase: int = 30_72 , _lowerCamelCase: str = "gelu" , _lowerCamelCase: float = 0.1 , _lowerCamelCase: float = 0.1 , _lowerCamelCase: int = 5_12 , _lowerCamelCase: int = 2 , _lowerCamelCase: float = 0.02 , _lowerCamelCase: float = 1E-12 , _lowerCamelCase: bool = False , **_lowerCamelCase: Optional[int] , ): super().__init__(pad_token_id=_lowerCamelCase , **_lowerCamelCase ) SCREAMING_SNAKE_CASE_ = attention_window SCREAMING_SNAKE_CASE_ = sep_token_id SCREAMING_SNAKE_CASE_ = bos_token_id SCREAMING_SNAKE_CASE_ = eos_token_id SCREAMING_SNAKE_CASE_ = vocab_size SCREAMING_SNAKE_CASE_ = hidden_size SCREAMING_SNAKE_CASE_ = num_hidden_layers SCREAMING_SNAKE_CASE_ = num_attention_heads SCREAMING_SNAKE_CASE_ = hidden_act SCREAMING_SNAKE_CASE_ = intermediate_size SCREAMING_SNAKE_CASE_ = hidden_dropout_prob SCREAMING_SNAKE_CASE_ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ = max_position_embeddings SCREAMING_SNAKE_CASE_ = type_vocab_size SCREAMING_SNAKE_CASE_ = initializer_range SCREAMING_SNAKE_CASE_ = layer_norm_eps SCREAMING_SNAKE_CASE_ = onnx_export class __magic_name__ ( __UpperCAmelCase): '''simple docstring''' def __init__( self: Optional[int] , _lowerCamelCase: "PretrainedConfig" , _lowerCamelCase: str = "default" , _lowerCamelCase: "List[PatchingSpec]" = None ): super().__init__(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) SCREAMING_SNAKE_CASE_ = True @property def _A ( self: Tuple ): if self.task == "multiple-choice": SCREAMING_SNAKE_CASE_ = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: SCREAMING_SNAKE_CASE_ = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''global_attention_mask''', dynamic_axis), ] ) @property def _A ( self: Any ): SCREAMING_SNAKE_CASE_ = super().outputs if self.task == "default": SCREAMING_SNAKE_CASE_ = {0: '''batch'''} return outputs @property def _A ( self: Optional[int] ): return 1E-4 @property def _A ( self: List[str] ): # needs to be >= 14 to support tril operator return max(super().default_onnx_opset , 14 ) def _A ( self: List[Any] , _lowerCamelCase: "PreTrainedTokenizerBase" , _lowerCamelCase: int = -1 , _lowerCamelCase: int = -1 , _lowerCamelCase: bool = False , _lowerCamelCase: Optional[TensorType] = None , ): SCREAMING_SNAKE_CASE_ = super().generate_dummy_inputs( preprocessor=_lowerCamelCase , batch_size=_lowerCamelCase , seq_length=_lowerCamelCase , is_pair=_lowerCamelCase , framework=_lowerCamelCase ) import torch # for some reason, replacing this code by inputs["global_attention_mask"] = torch.randint(2, inputs["input_ids"].shape, dtype=torch.int64) # makes the export fail randomly SCREAMING_SNAKE_CASE_ = torch.zeros_like(inputs['''input_ids'''] ) # make every second token global SCREAMING_SNAKE_CASE_ = 1 return inputs
89
import gc import random import unittest import numpy as np import torch from transformers import CLIPImageProcessor, CLIPVisionConfig, CLIPVisionModel from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEImgaImgPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import floats_tensor, load_image, load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class __magic_name__ ( __UpperCAmelCase , unittest.TestCase): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = ShapEImgaImgPipeline SCREAMING_SNAKE_CASE__ : Dict = ["image"] SCREAMING_SNAKE_CASE__ : List[Any] = ["image"] SCREAMING_SNAKE_CASE__ : List[Any] = [ "num_images_per_prompt", "num_inference_steps", "generator", "latents", "guidance_scale", "frame_size", "output_type", "return_dict", ] SCREAMING_SNAKE_CASE__ : Optional[int] = False @property def _A ( self: Optional[Any] ): return 32 @property def _A ( self: Optional[int] ): return 32 @property def _A ( self: List[Any] ): return self.time_input_dim * 4 @property def _A ( self: Any ): return 8 @property def _A ( self: int ): torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=64 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=1 , ) SCREAMING_SNAKE_CASE_ = CLIPVisionModel(_lowerCamelCase ) return model @property def _A ( self: List[Any] ): SCREAMING_SNAKE_CASE_ = CLIPImageProcessor( crop_size=2_24 , do_center_crop=_lowerCamelCase , do_normalize=_lowerCamelCase , do_resize=_lowerCamelCase , image_mean=[0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73] , image_std=[0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11] , resample=3 , size=2_24 , ) return image_processor @property def _A ( self: Dict ): torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ = { '''num_attention_heads''': 2, '''attention_head_dim''': 16, '''embedding_dim''': self.time_input_dim, '''num_embeddings''': 32, '''embedding_proj_dim''': self.text_embedder_hidden_size, '''time_embed_dim''': self.time_embed_dim, '''num_layers''': 1, '''clip_embed_dim''': self.time_input_dim * 2, '''additional_embeddings''': 0, '''time_embed_act_fn''': '''gelu''', '''norm_in_type''': '''layer''', '''embedding_proj_norm_type''': '''layer''', '''encoder_hid_proj_type''': None, '''added_emb_type''': None, } SCREAMING_SNAKE_CASE_ = PriorTransformer(**_lowerCamelCase ) return model @property def _A ( self: List[Any] ): torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ = { '''param_shapes''': ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), '''d_latent''': self.time_input_dim, '''d_hidden''': self.renderer_dim, '''n_output''': 12, '''background''': ( 0.1, 0.1, 0.1, ), } SCREAMING_SNAKE_CASE_ = ShapERenderer(**_lowerCamelCase ) return model def _A ( self: Optional[int] ): SCREAMING_SNAKE_CASE_ = self.dummy_prior SCREAMING_SNAKE_CASE_ = self.dummy_image_encoder SCREAMING_SNAKE_CASE_ = self.dummy_image_processor SCREAMING_SNAKE_CASE_ = self.dummy_renderer SCREAMING_SNAKE_CASE_ = HeunDiscreteScheduler( beta_schedule='''exp''' , num_train_timesteps=10_24 , prediction_type='''sample''' , use_karras_sigmas=_lowerCamelCase , clip_sample=_lowerCamelCase , clip_sample_range=1.0 , ) SCREAMING_SNAKE_CASE_ = { '''prior''': prior, '''image_encoder''': image_encoder, '''image_processor''': image_processor, '''renderer''': renderer, '''scheduler''': scheduler, } return components def _A ( self: Optional[Any] , _lowerCamelCase: List[Any] , _lowerCamelCase: Optional[Any]=0 ): SCREAMING_SNAKE_CASE_ = floats_tensor((1, 3, 64, 64) , rng=random.Random(_lowerCamelCase ) ).to(_lowerCamelCase ) if str(_lowerCamelCase ).startswith('''mps''' ): SCREAMING_SNAKE_CASE_ = torch.manual_seed(_lowerCamelCase ) else: SCREAMING_SNAKE_CASE_ = torch.Generator(device=_lowerCamelCase ).manual_seed(_lowerCamelCase ) SCREAMING_SNAKE_CASE_ = { '''image''': input_image, '''generator''': generator, '''num_inference_steps''': 1, '''frame_size''': 32, '''output_type''': '''np''', } return inputs def _A ( self: Any ): SCREAMING_SNAKE_CASE_ = '''cpu''' SCREAMING_SNAKE_CASE_ = self.get_dummy_components() SCREAMING_SNAKE_CASE_ = self.pipeline_class(**_lowerCamelCase ) SCREAMING_SNAKE_CASE_ = pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) SCREAMING_SNAKE_CASE_ = pipe(**self.get_dummy_inputs(_lowerCamelCase ) ) SCREAMING_SNAKE_CASE_ = output.images[0] SCREAMING_SNAKE_CASE_ = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) SCREAMING_SNAKE_CASE_ = np.array( [ 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def _A ( self: str ): # NOTE: Larger batch sizes cause this test to timeout, only test on smaller batches self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def _A ( self: Tuple ): SCREAMING_SNAKE_CASE_ = torch_device == '''cpu''' SCREAMING_SNAKE_CASE_ = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=_lowerCamelCase , relax_max_difference=_lowerCamelCase , ) def _A ( self: Tuple ): SCREAMING_SNAKE_CASE_ = self.get_dummy_components() SCREAMING_SNAKE_CASE_ = self.pipeline_class(**_lowerCamelCase ) SCREAMING_SNAKE_CASE_ = pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) SCREAMING_SNAKE_CASE_ = 1 SCREAMING_SNAKE_CASE_ = 2 SCREAMING_SNAKE_CASE_ = self.get_dummy_inputs(_lowerCamelCase ) for key in inputs.keys(): if key in self.batch_params: SCREAMING_SNAKE_CASE_ = batch_size * [inputs[key]] SCREAMING_SNAKE_CASE_ = pipe(**_lowerCamelCase , num_images_per_prompt=_lowerCamelCase )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class __magic_name__ ( unittest.TestCase): '''simple docstring''' def _A ( self: Any ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _A ( self: Optional[int] ): SCREAMING_SNAKE_CASE_ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/shap_e/corgi.png''' ) SCREAMING_SNAKE_CASE_ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/shap_e/test_shap_e_img2img_out.npy''' ) SCREAMING_SNAKE_CASE_ = ShapEImgaImgPipeline.from_pretrained('''openai/shap-e-img2img''' ) SCREAMING_SNAKE_CASE_ = pipe.to(_lowerCamelCase ) pipe.set_progress_bar_config(disable=_lowerCamelCase ) SCREAMING_SNAKE_CASE_ = torch.Generator(device=_lowerCamelCase ).manual_seed(0 ) SCREAMING_SNAKE_CASE_ = pipe( _lowerCamelCase , generator=_lowerCamelCase , guidance_scale=3.0 , num_inference_steps=64 , frame_size=64 , output_type='''np''' , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(_lowerCamelCase , _lowerCamelCase )
89
1
import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate.utils import patch_environment @require_huggingface_suite class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' def UpperCamelCase_ ( self : List[str] ): __A = inspect.getfile(accelerate.test_utils ) __A = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["scripts", "external_deps", "test_metrics.py"] ) from accelerate.test_utils.scripts.external_deps import test_metrics # noqa: F401 __A = test_metrics @require_cpu def UpperCamelCase_ ( self : str ): debug_launcher(self.test_metrics.main ,num_processes=1 ) @require_cpu def UpperCamelCase_ ( self : str ): debug_launcher(self.test_metrics.main ) @require_single_gpu def UpperCamelCase_ ( self : List[str] ): self.test_metrics.main() @require_multi_gpu def UpperCamelCase_ ( self : int ): print(f'''Found {torch.cuda.device_count()} devices.''' ) __A = ["torchrun", f'''--nproc_per_node={torch.cuda.device_count()}''', self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(A ,env=os.environ.copy() )
55
from numpy import exp, pi, sqrt def UpperCAmelCase ( a_ , a_ = 0.0 , a_ = 1.0 ) -> int: """simple docstring""" return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
55
1
'''simple docstring''' import inspect import unittest from datasets import load_dataset from packaging import version from transformers import BeitConfig from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( MODEL_MAPPING, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation, BeitModel, ) from transformers.models.beit.modeling_beit import BEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): import PIL from PIL import Image from transformers import BeitImageProcessor class lowerCAmelCase__ : """simple docstring""" def __init__( self : str , A__ : str , A__ : List[str]=1_0_0 , A__ : Optional[Any]=1_3 , A__ : List[Any]=3_0 , A__ : Dict=2 , A__ : Optional[int]=3 , A__ : Tuple=True , A__ : List[str]=True , A__ : Any=3_2 , A__ : int=4 , A__ : Tuple=4 , A__ : Optional[int]=3_7 , A__ : Tuple="gelu" , A__ : Dict=0.1 , A__ : Dict=0.1 , A__ : Union[str, Any]=1_0 , A__ : str=0.02 , A__ : Optional[int]=3 , A__ : Dict=None , A__ : Dict=[0, 1, 2, 3] , ) -> Union[str, Any]: '''simple docstring''' a__ : Union[str, Any] = parent a__ : Tuple = 1_0_0 a__ : List[str] = batch_size a__ : Optional[int] = image_size a__ : Any = patch_size a__ : List[Any] = num_channels a__ : List[Any] = is_training a__ : str = use_labels a__ : Dict = hidden_size a__ : Optional[Any] = num_hidden_layers a__ : Optional[Any] = num_attention_heads a__ : Any = intermediate_size a__ : Optional[Any] = hidden_act a__ : Dict = hidden_dropout_prob a__ : Tuple = attention_probs_dropout_prob a__ : Optional[int] = type_sequence_label_size a__ : int = initializer_range a__ : Dict = scope a__ : int = out_indices a__ : Optional[Any] = num_labels # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) a__ : Tuple = (image_size // patch_size) ** 2 a__ : str = num_patches + 1 def __lowerCAmelCase ( self : Tuple ) -> List[str]: '''simple docstring''' a__ : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) a__ : Dict = None a__ : Optional[Any] = None if self.use_labels: a__ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a__ : List[Any] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) a__ : Optional[int] = self.get_config() return config, pixel_values, labels, pixel_labels def __lowerCAmelCase ( self : str ) -> Tuple: '''simple docstring''' return BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A__ , initializer_range=self.initializer_range , out_indices=self.out_indices , ) def __lowerCAmelCase ( self : Any , A__ : int , A__ : str , A__ : List[Any] , A__ : Optional[Any] ) -> List[Any]: '''simple docstring''' a__ : int = BeitModel(config=A__ ) model.to(A__ ) model.eval() a__ : List[Any] = model(A__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self : Any , A__ : List[Any] , A__ : Optional[int] , A__ : int , A__ : Any ) -> Dict: '''simple docstring''' a__ : Dict = BeitForMaskedImageModeling(config=A__ ) model.to(A__ ) model.eval() a__ : Any = model(A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def __lowerCAmelCase ( self : Union[str, Any] , A__ : List[str] , A__ : Union[str, Any] , A__ : Any , A__ : int ) -> Tuple: '''simple docstring''' a__ : Union[str, Any] = self.type_sequence_label_size a__ : int = BeitForImageClassification(A__ ) model.to(A__ ) model.eval() a__ : List[str] = model(A__ , labels=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images a__ : List[str] = 1 a__ : Union[str, Any] = BeitForImageClassification(A__ ) model.to(A__ ) model.eval() a__ : Dict = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) a__ : Any = model(A__ , labels=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __lowerCAmelCase ( self : Union[str, Any] , A__ : List[Any] , A__ : Union[str, Any] , A__ : str , A__ : Dict ) -> List[Any]: '''simple docstring''' a__ : int = self.num_labels a__ : Dict = BeitForSemanticSegmentation(A__ ) model.to(A__ ) model.eval() a__ : str = model(A__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) a__ : List[str] = model(A__ , labels=A__ ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size * 2, self.image_size * 2) ) def __lowerCAmelCase ( self : int ) -> Tuple: '''simple docstring''' a__ : Any = self.prepare_config_and_inputs() a__ : List[Any] = config_and_inputs a__ : Dict = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase__ ( lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): """simple docstring""" __UpperCamelCase = ( (BeitModel, BeitForImageClassification, BeitForMaskedImageModeling, BeitForSemanticSegmentation) if is_torch_available() else () ) __UpperCamelCase = ( { "feature-extraction": BeitModel, "image-classification": BeitForImageClassification, "image-segmentation": BeitForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCamelCase = False __UpperCamelCase = False __UpperCamelCase = False def __lowerCAmelCase ( self : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' a__ : Union[str, Any] = BeitModelTester(self ) a__ : List[Any] = ConfigTester(self , config_class=A__ , has_text_modality=A__ , hidden_size=3_7 ) def __lowerCAmelCase ( self : Any ) -> Optional[int]: '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='''BEiT does not use inputs_embeds''' ) def __lowerCAmelCase ( self : List[str] ) -> Dict: '''simple docstring''' pass @require_torch_multi_gpu @unittest.skip(reason='''BEiT has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`''' ) def __lowerCAmelCase ( self : Any ) -> List[str]: '''simple docstring''' pass def __lowerCAmelCase ( self : Any ) -> Optional[Any]: '''simple docstring''' a__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a__ : int = model_class(A__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) a__ : Union[str, Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A__ , nn.Linear ) ) def __lowerCAmelCase ( self : Any ) -> Dict: '''simple docstring''' a__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: a__ : List[str] = model_class(A__ ) a__ : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic a__ : Union[str, Any] = [*signature.parameters.keys()] a__ : Optional[int] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A__ ) def __lowerCAmelCase ( self : str ) -> int: '''simple docstring''' a__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A__ ) def __lowerCAmelCase ( self : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' a__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A__ ) def __lowerCAmelCase ( self : List[str] ) -> Any: '''simple docstring''' a__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A__ ) def __lowerCAmelCase ( self : Union[str, Any] ) -> Dict: '''simple docstring''' a__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*A__ ) def __lowerCAmelCase ( self : List[str] ) -> Optional[int]: '''simple docstring''' if not self.model_tester.is_training: return a__ : str = self.model_tester.prepare_config_and_inputs_for_common() a__ : str = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if model_class in [*get_values(A__ ), BeitForMaskedImageModeling]: continue a__ : str = model_class(A__ ) model.to(A__ ) model.train() a__ : List[str] = self._prepare_for_class(A__ , A__ , return_labels=A__ ) a__ : Dict = model(**A__ ).loss loss.backward() def __lowerCAmelCase ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' a__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return a__ : List[Any] = False a__ : Dict = True for model_class in self.all_model_classes: # we don't test BeitForMaskedImageModeling if ( model_class in [*get_values(A__ ), BeitForMaskedImageModeling] or not model_class.supports_gradient_checkpointing ): continue a__ : Tuple = model_class(A__ ) model.gradient_checkpointing_enable() model.to(A__ ) model.train() a__ : List[str] = self._prepare_for_class(A__ , A__ , return_labels=A__ ) a__ : int = model(**A__ ).loss loss.backward() def __lowerCAmelCase ( self : Optional[int] ) -> Union[str, Any]: '''simple docstring''' a__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() a__ : int = _config_zero_init(A__ ) for model_class in self.all_model_classes: a__ : Optional[int] = model_class(config=A__ ) for name, param in model.named_parameters(): # we skip lambda parameters as these require special initial values # determined by config.layer_scale_init_value if "lambda" in name: continue if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) @slow def __lowerCAmelCase ( self : List[Any] ) -> int: '''simple docstring''' for model_name in BEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: a__ : Union[str, Any] = BeitModel.from_pretrained(A__ ) self.assertIsNotNone(A__ ) def __a ( ): a__ : Union[str, Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowerCAmelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self : List[Any] ) -> Dict: '''simple docstring''' return BeitImageProcessor.from_pretrained('''microsoft/beit-base-patch16-224''' ) if is_vision_available() else None @slow def __lowerCAmelCase ( self : int ) -> Optional[Any]: '''simple docstring''' a__ : Tuple = BeitForMaskedImageModeling.from_pretrained('''microsoft/beit-base-patch16-224-pt22k''' ).to(A__ ) a__ : Dict = self.default_image_processor a__ : List[str] = prepare_img() a__ : Union[str, Any] = image_processor(images=A__ , return_tensors='''pt''' ).pixel_values.to(A__ ) # prepare bool_masked_pos a__ : str = torch.ones((1, 1_9_6) , dtype=torch.bool ).to(A__ ) # forward pass with torch.no_grad(): a__ : Union[str, Any] = model(pixel_values=A__ , bool_masked_pos=A__ ) a__ : List[str] = outputs.logits # verify the logits a__ : Dict = torch.Size((1, 1_9_6, 8_1_9_2) ) self.assertEqual(logits.shape , A__ ) a__ : Dict = torch.tensor( [[-3.2_437, 0.5_072, -13.9_174], [-3.2_456, 0.4_948, -13.9_401], [-3.2_033, 0.5_121, -13.8_550]] ).to(A__ ) self.assertTrue(torch.allclose(logits[bool_masked_pos][:3, :3] , A__ , atol=1E-2 ) ) @slow def __lowerCAmelCase ( self : Optional[Any] ) -> Dict: '''simple docstring''' a__ : Union[str, Any] = BeitForImageClassification.from_pretrained('''microsoft/beit-base-patch16-224''' ).to(A__ ) a__ : Dict = self.default_image_processor a__ : Optional[int] = prepare_img() a__ : Union[str, Any] = image_processor(images=A__ , return_tensors='''pt''' ).to(A__ ) # forward pass with torch.no_grad(): a__ : Tuple = model(**A__ ) a__ : int = outputs.logits # verify the logits a__ : Tuple = torch.Size((1, 1_0_0_0) ) self.assertEqual(logits.shape , A__ ) a__ : str = torch.tensor([-1.2_385, -1.0_987, -1.0_108] ).to(A__ ) self.assertTrue(torch.allclose(logits[0, :3] , A__ , atol=1E-4 ) ) a__ : int = 2_8_1 self.assertEqual(logits.argmax(-1 ).item() , A__ ) @slow def __lowerCAmelCase ( self : int ) -> str: '''simple docstring''' a__ : Dict = BeitForImageClassification.from_pretrained('''microsoft/beit-large-patch16-224-pt22k-ft22k''' ).to( A__ ) a__ : Any = self.default_image_processor a__ : Any = prepare_img() a__ : int = image_processor(images=A__ , return_tensors='''pt''' ).to(A__ ) # forward pass with torch.no_grad(): a__ : Optional[int] = model(**A__ ) a__ : Optional[int] = outputs.logits # verify the logits a__ : Optional[Any] = torch.Size((1, 2_1_8_4_1) ) self.assertEqual(logits.shape , A__ ) a__ : Tuple = torch.tensor([1.6_881, -0.2_787, 0.5_901] ).to(A__ ) self.assertTrue(torch.allclose(logits[0, :3] , A__ , atol=1E-4 ) ) a__ : int = 2_3_9_6 self.assertEqual(logits.argmax(-1 ).item() , A__ ) @slow def __lowerCAmelCase ( self : Any ) -> Dict: '''simple docstring''' a__ : int = BeitForSemanticSegmentation.from_pretrained('''microsoft/beit-base-finetuned-ade-640-640''' ) a__ : Optional[int] = model.to(A__ ) a__ : Tuple = BeitImageProcessor(do_resize=A__ , size=6_4_0 , do_center_crop=A__ ) a__ : Optional[Any] = load_dataset('''hf-internal-testing/fixtures_ade20k''' , split='''test''' ) a__ : Optional[Any] = Image.open(ds[0]['''file'''] ) a__ : List[str] = image_processor(images=A__ , return_tensors='''pt''' ).to(A__ ) # forward pass with torch.no_grad(): a__ : str = model(**A__ ) a__ : Any = outputs.logits # verify the logits a__ : Any = torch.Size((1, 1_5_0, 1_6_0, 1_6_0) ) self.assertEqual(logits.shape , A__ ) a__ : Tuple = version.parse(PIL.__version__ ) < version.parse('''9.0.0''' ) if is_pillow_less_than_a: a__ : int = torch.tensor( [ [[-4.9_225, -2.3_954, -3.0_522], [-2.8_822, -1.0_046, -1.7_561], [-2.9_549, -1.3_228, -2.1_347]], [[-5.8_168, -3.4_129, -4.0_778], [-3.8_651, -2.2_214, -3.0_277], [-3.8_356, -2.4_643, -3.3_535]], [[-0.0_078, 3.9_952, 4.0_754], [2.9_856, 4.6_944, 5.0_035], [3.2_413, 4.7_813, 4.9_969]], ] , device=A__ , ) else: a__ : List[Any] = torch.tensor( [ [[-4.8_960, -2.3_688, -3.0_355], [-2.8_478, -0.9_836, -1.7_418], [-2.9_449, -1.3_332, -2.1_456]], [[-5.8_081, -3.4_124, -4.1_006], [-3.8_561, -2.2_081, -3.0_323], [-3.8_365, -2.4_601, -3.3_669]], [[-0.0_309, 3.9_868, 4.0_540], [2.9_640, 4.6_877, 4.9_976], [3.2_081, 4.7_690, 4.9_942]], ] , device=A__ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , A__ , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self : Optional[int] ) -> Tuple: '''simple docstring''' a__ : Tuple = BeitForSemanticSegmentation.from_pretrained('''microsoft/beit-base-finetuned-ade-640-640''' ) a__ : int = model.to(A__ ) a__ : List[Any] = BeitImageProcessor(do_resize=A__ , size=6_4_0 , do_center_crop=A__ ) a__ : Optional[int] = load_dataset('''hf-internal-testing/fixtures_ade20k''' , split='''test''' ) a__ : Dict = Image.open(ds[0]['''file'''] ) a__ : Optional[Any] = image_processor(images=A__ , return_tensors='''pt''' ).to(A__ ) # forward pass with torch.no_grad(): a__ : Union[str, Any] = model(**A__ ) a__ : Union[str, Any] = outputs.logits.detach().cpu() a__ : Dict = image_processor.post_process_semantic_segmentation(outputs=A__ , target_sizes=[(5_0_0, 3_0_0)] ) a__ : Union[str, Any] = torch.Size((5_0_0, 3_0_0) ) self.assertEqual(segmentation[0].shape , A__ ) a__ : List[str] = image_processor.post_process_semantic_segmentation(outputs=A__ ) a__ : Any = torch.Size((1_6_0, 1_6_0) ) self.assertEqual(segmentation[0].shape , A__ )
704
'''simple docstring''' import os def __a ( ): with open(os.path.dirname(lowerCAmelCase__ ) + '''/grid.txt''' ) as f: a__ : Optional[int] = [] # noqa: E741 for _ in range(20 ): l.append([int(lowerCAmelCase__ ) for x in f.readline().split()] ) a__ : List[str] = 0 # right for i in range(20 ): for j in range(17 ): a__ : int = l[i][j] * l[i][j + 1] * l[i][j + 2] * l[i][j + 3] if temp > maximum: a__ : str = temp # down for i in range(17 ): for j in range(20 ): a__ : List[str] = l[i][j] * l[i + 1][j] * l[i + 2][j] * l[i + 3][j] if temp > maximum: a__ : Dict = temp # diagonal 1 for i in range(17 ): for j in range(17 ): a__ : str = l[i][j] * l[i + 1][j + 1] * l[i + 2][j + 2] * l[i + 3][j + 3] if temp > maximum: a__ : Optional[Any] = temp # diagonal 2 for i in range(17 ): for j in range(3 , 20 ): a__ : str = l[i][j] * l[i + 1][j - 1] * l[i + 2][j - 2] * l[i + 3][j - 3] if temp > maximum: a__ : Tuple = temp return maximum if __name__ == "__main__": print(solution())
340
0
'''simple docstring''' def lowerCAmelCase_ ( snake_case_ : int ) -> int: '''simple docstring''' assert isinstance(snake_case_ , snake_case_ ), f"""The input value of [n={number}] is not an integer""" if number == 1: return 2 elif number < 1: UpperCAmelCase_ = f"""The input value of [n={number}] has to be > 0""" raise ValueError(snake_case_ ) else: UpperCAmelCase_ = sylvester(number - 1 ) UpperCAmelCase_ = num - 1 UpperCAmelCase_ = num return lower * upper + 1 if __name__ == "__main__": print(f"The 8th number in Sylvester's sequence: {sylvester(8)}")
78
'''simple docstring''' import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin a : List[Any] = get_tests_dir('''fixtures/test_sentencepiece_with_bytefallback.model''') @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = GPTSwaTokenizer __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = False def A ( self : int ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __snake_case = GPTSwaTokenizer(a_ , eos_token="<unk>" , bos_token="<unk>" , pad_token="<unk>" ) tokenizer.save_pretrained(self.tmpdirname ) def A ( self : str , a_ : List[Any] ): """simple docstring""" __snake_case = "This is a test" __snake_case = "This is a test" return input_text, output_text def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = "<s>" __snake_case = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(a_ ) , 2_000 ) def A ( self : Optional[int] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 2_000 ) def A ( self : Dict ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = tokenizer.tokenize("This is a test" ) self.assertListEqual(a_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , [465, 287, 265, 631, 842] ) __snake_case = tokenizer.tokenize("I was born in 92000, and this is falsé." ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] , ) # fmt: on __snake_case = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual( a_ , [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260] , ) __snake_case = tokenizer.convert_ids_to_tokens(a_ ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] ) # fmt: on def A ( self : List[str] ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = ["This is a test", "I was born in 92000, and this is falsé."] __snake_case = [ [465, 287, 265, 631, 842], [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(a_ , a_ ): self.assertListEqual(tokenizer.encode_fast(a_ ) , a_ ) # Test that decode_fast returns the input text for text, token_ids in zip(a_ , a_ ): self.assertEqual(tokenizer.decode_fast(a_ ) , a_ ) @slow def A ( self : Any ): """simple docstring""" __snake_case = [ "<|python|>def fibonacci(n)\n if n < 0:\n print('Incorrect input')", "Hey there, how are you doing this fine day?", "This is a text with a trailing spaces followed by a dot .", "Häj sväjs lillebrör! =)", "Det är inget fel på Mr. Cool", ] # fmt: off __snake_case = {"input_ids": [[63_423, 5, 6_811, 14_954, 282, 816, 3_821, 63_466, 63_425, 63_462, 18, 63_978, 678, 301, 1_320, 63_423, 63_455, 63_458, 18, 63_982, 4_246, 3_940, 1_901, 47_789, 5_547, 18_994], [19_630, 1_100, 63_446, 1_342, 633, 544, 4_488, 593, 5_102, 2_416, 63_495, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_652, 428, 268, 1_936, 515, 268, 58_593, 22_413, 9_106, 546, 268, 33_213, 63_979, 698, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [55_130, 63_450, 924, 63_449, 2_249, 4_062, 1_558, 318, 63_504, 21_498, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [509, 377, 2_827, 2_559, 332, 6_575, 63_443, 26_801, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_ , model_name="AI-Sweden/gpt-sw3-126m" , sequences=a_ , )
69
0
import gc import unittest from diffusers import FlaxDPMSolverMultistepScheduler, FlaxStableDiffusionPipeline from diffusers.utils import is_flax_available, slow from diffusers.utils.testing_utils import require_flax if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard @slow @require_flax class __snake_case ( unittest.TestCase ): def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ = FlaxStableDiffusionPipeline.from_pretrained( 'stabilityai/stable-diffusion-2' ,revision='bf16' ,dtype=jnp.bfloataa ,) lowerCAmelCase__ = 'A painting of a squirrel eating a burger' lowerCAmelCase__ = jax.device_count() lowerCAmelCase__ = num_samples * [prompt] lowerCAmelCase__ = sd_pipe.prepare_inputs(a_ ) lowerCAmelCase__ = replicate(a_ ) lowerCAmelCase__ = shard(a_ ) lowerCAmelCase__ = jax.random.PRNGKey(0 ) lowerCAmelCase__ = jax.random.split(a_ ,jax.device_count() ) lowerCAmelCase__ = sd_pipe(a_ ,a_ ,a_ ,num_inference_steps=25 ,jit=a_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) lowerCAmelCase__ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) lowerCAmelCase__ = images[0, 253:256, 253:256, -1] lowerCAmelCase__ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) lowerCAmelCase__ = jnp.array([0.4238, 0.4414, 0.4395, 0.4453, 0.4629, 0.4590, 0.4531, 0.45508, 0.4512] ) print(f'output_slice: {output_slice}' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2 def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ = 'stabilityai/stable-diffusion-2' lowerCAmelCase__ , lowerCAmelCase__ = FlaxDPMSolverMultistepScheduler.from_pretrained(a_ ,subfolder='scheduler' ) lowerCAmelCase__ , lowerCAmelCase__ = FlaxStableDiffusionPipeline.from_pretrained( a_ ,scheduler=a_ ,revision='bf16' ,dtype=jnp.bfloataa ,) lowerCAmelCase__ = scheduler_params lowerCAmelCase__ = 'A painting of a squirrel eating a burger' lowerCAmelCase__ = jax.device_count() lowerCAmelCase__ = num_samples * [prompt] lowerCAmelCase__ = sd_pipe.prepare_inputs(a_ ) lowerCAmelCase__ = replicate(a_ ) lowerCAmelCase__ = shard(a_ ) lowerCAmelCase__ = jax.random.PRNGKey(0 ) lowerCAmelCase__ = jax.random.split(a_ ,jax.device_count() ) lowerCAmelCase__ = sd_pipe(a_ ,a_ ,a_ ,num_inference_steps=25 ,jit=a_ )[0] assert images.shape == (jax.device_count(), 1, 768, 768, 3) lowerCAmelCase__ = images.reshape((images.shape[0] * images.shape[1],) + images.shape[-3:] ) lowerCAmelCase__ = images[0, 253:256, 253:256, -1] lowerCAmelCase__ = jnp.asarray(jax.device_get(image_slice.flatten() ) ) lowerCAmelCase__ = jnp.array([0.4336, 0.42969, 0.4453, 0.4199, 0.4297, 0.4531, 0.4434, 0.4434, 0.4297] ) print(f'output_slice: {output_slice}' ) assert jnp.abs(output_slice - expected_slice ).max() < 1e-2
604
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __snake_case : def __init__( self ,a_ ,a_=13 ,a_=32 ,a_=2 ,a_=3 ,a_=16 ,a_=[1, 2, 1] ,a_=[2, 2, 4] ,a_=2 ,a_=2.0 ,a_=True ,a_=0.0 ,a_=0.0 ,a_=0.1 ,a_="gelu" ,a_=False ,a_=True ,a_=0.02 ,a_=1e-5 ,a_=True ,a_=None ,a_=True ,a_=10 ,a_=8 ,): """simple docstring""" lowerCAmelCase__ = parent lowerCAmelCase__ = batch_size lowerCAmelCase__ = image_size lowerCAmelCase__ = patch_size lowerCAmelCase__ = num_channels lowerCAmelCase__ = embed_dim lowerCAmelCase__ = depths lowerCAmelCase__ = num_heads lowerCAmelCase__ = window_size lowerCAmelCase__ = mlp_ratio lowerCAmelCase__ = qkv_bias lowerCAmelCase__ = hidden_dropout_prob lowerCAmelCase__ = attention_probs_dropout_prob lowerCAmelCase__ = drop_path_rate lowerCAmelCase__ = hidden_act lowerCAmelCase__ = use_absolute_embeddings lowerCAmelCase__ = patch_norm lowerCAmelCase__ = layer_norm_eps lowerCAmelCase__ = initializer_range lowerCAmelCase__ = is_training lowerCAmelCase__ = scope lowerCAmelCase__ = use_labels lowerCAmelCase__ = type_sequence_label_size lowerCAmelCase__ = encoder_stride def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase__ = None if self.use_labels: lowerCAmelCase__ = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) lowerCAmelCase__ = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" return SwinvaConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,depths=self.depths ,num_heads=self.num_heads ,window_size=self.window_size ,mlp_ratio=self.mlp_ratio ,qkv_bias=self.qkv_bias ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,drop_path_rate=self.drop_path_rate ,hidden_act=self.hidden_act ,use_absolute_embeddings=self.use_absolute_embeddings ,path_norm=self.patch_norm ,layer_norm_eps=self.layer_norm_eps ,initializer_range=self.initializer_range ,encoder_stride=self.encoder_stride ,) def SCREAMING_SNAKE_CASE_ ( self ,a_ ,a_ ,a_ ): """simple docstring""" lowerCAmelCase__ = SwinvaModel(config=a_ ) model.to(a_ ) model.eval() lowerCAmelCase__ = model(a_ ) lowerCAmelCase__ = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) lowerCAmelCase__ = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, expected_seq_len, expected_dim) ) def SCREAMING_SNAKE_CASE_ ( self ,a_ ,a_ ,a_ ): """simple docstring""" lowerCAmelCase__ = SwinvaForMaskedImageModeling(config=a_ ) model.to(a_ ) model.eval() lowerCAmelCase__ = model(a_ ) self.parent.assertEqual( result.logits.shape ,(self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images lowerCAmelCase__ = 1 lowerCAmelCase__ = SwinvaForMaskedImageModeling(a_ ) model.to(a_ ) model.eval() lowerCAmelCase__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCAmelCase__ = model(a_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, 1, self.image_size, self.image_size) ) def SCREAMING_SNAKE_CASE_ ( self ,a_ ,a_ ,a_ ): """simple docstring""" lowerCAmelCase__ = self.type_sequence_label_size lowerCAmelCase__ = SwinvaForImageClassification(a_ ) model.to(a_ ) model.eval() lowerCAmelCase__ = model(a_ ,labels=a_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ = self.prepare_config_and_inputs() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = config_and_inputs lowerCAmelCase__ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class __snake_case ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , unittest.TestCase ): SCREAMING_SNAKE_CASE__ = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) SCREAMING_SNAKE_CASE__ = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ = SwinvaModelTester(self ) lowerCAmelCase__ = ConfigTester(self ,config_class=a_ ,embed_dim=37 ) def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" pass def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ = model_class(a_ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) lowerCAmelCase__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a_ ,nn.Linear ) ) def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase__ = model_class(a_ ) lowerCAmelCase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase__ = [*signature.parameters.keys()] lowerCAmelCase__ = ['pixel_values'] self.assertListEqual(arg_names[:1] ,a_ ) def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ = True for model_class in self.all_model_classes: lowerCAmelCase__ = True lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): lowerCAmelCase__ = model(**self._prepare_for_class(a_ ,a_ ) ) lowerCAmelCase__ = outputs.attentions lowerCAmelCase__ = len(self.model_tester.depths ) self.assertEqual(len(a_ ) ,a_ ) # check that output_attentions also work using config del inputs_dict["output_attentions"] lowerCAmelCase__ = True lowerCAmelCase__ = config.window_size**2 lowerCAmelCase__ = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): lowerCAmelCase__ = model(**self._prepare_for_class(a_ ,a_ ) ) lowerCAmelCase__ = outputs.attentions self.assertEqual(len(a_ ) ,a_ ) self.assertListEqual( list(attentions[0].shape[-3:] ) ,[self.model_tester.num_heads[0], window_size_squared, window_size_squared] ,) lowerCAmelCase__ = len(a_ ) # Check attention is always last and order is fine lowerCAmelCase__ = True lowerCAmelCase__ = True lowerCAmelCase__ = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): lowerCAmelCase__ = model(**self._prepare_for_class(a_ ,a_ ) ) if hasattr(self.model_tester ,'num_hidden_states_types' ): lowerCAmelCase__ = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states lowerCAmelCase__ = 2 self.assertEqual(out_len + added_hidden_states ,len(a_ ) ) lowerCAmelCase__ = outputs.attentions self.assertEqual(len(a_ ) ,a_ ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) ,[self.model_tester.num_heads[0], window_size_squared, window_size_squared] ,) def SCREAMING_SNAKE_CASE_ ( self ,a_ ,a_ ,a_ ,a_ ): """simple docstring""" lowerCAmelCase__ = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): lowerCAmelCase__ = model(**self._prepare_for_class(a_ ,a_ ) ) lowerCAmelCase__ = outputs.hidden_states lowerCAmelCase__ = getattr( self.model_tester ,'expected_num_hidden_layers' ,len(self.model_tester.depths ) + 1 ) self.assertEqual(len(a_ ) ,a_ ) # Swinv2 has a different seq_length lowerCAmelCase__ = ( config.patch_size if isinstance(config.patch_size ,collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) ,[num_patches, self.model_tester.embed_dim] ,) lowerCAmelCase__ = outputs.reshaped_hidden_states self.assertEqual(len(a_ ) ,a_ ) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = reshaped_hidden_states[0].shape lowerCAmelCase__ = ( reshaped_hidden_states[0].view(a_ ,a_ ,height * width ).permute(0 ,2 ,1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) ,[num_patches, self.model_tester.embed_dim] ,) def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size ,collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: lowerCAmelCase__ = True self.check_hidden_states_output(a_ ,a_ ,a_ ,a_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ = True self.check_hidden_states_output(a_ ,a_ ,a_ ,a_ ) def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ = 3 lowerCAmelCase__ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size ,collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) lowerCAmelCase__ = ( config.patch_size if isinstance(config.patch_size ,collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) lowerCAmelCase__ = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) lowerCAmelCase__ = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: lowerCAmelCase__ = True self.check_hidden_states_output(a_ ,a_ ,a_ ,(padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase__ = True self.check_hidden_states_output(a_ ,a_ ,a_ ,(padded_height, padded_width) ) def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*a_ ) def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a_ ) @slow def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase__ = SwinvaModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase__ = _config_zero_init(a_ ) for model_class in self.all_model_classes: lowerCAmelCase__ = model_class(config=a_ ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() ,[0.0, 1.0] ,msg=f'Parameter {name} of model {model_class} seems not properly initialized' ,) @require_vision @require_torch class __snake_case ( unittest.TestCase ): @cached_property def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def SCREAMING_SNAKE_CASE_ ( self ): """simple docstring""" lowerCAmelCase__ = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( a_ ) lowerCAmelCase__ = self.default_image_processor lowerCAmelCase__ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) lowerCAmelCase__ = image_processor(images=a_ ,return_tensors='pt' ).to(a_ ) # forward pass with torch.no_grad(): lowerCAmelCase__ = model(**a_ ) # verify the logits lowerCAmelCase__ = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape ,a_ ) lowerCAmelCase__ = torch.tensor([-0.3947, -0.4306, 0.0026] ).to(a_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,a_ ,atol=1e-4 ) )
604
1
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __A = logging.get_logger(__name__) __A = { "microsoft/wavlm-base": "https://huggingface.co/microsoft/wavlm-base/resolve/main/config.json", # See all WavLM models at https://huggingface.co/models?filter=wavlm } class _A ( _a ): """simple docstring""" lowerCamelCase : Union[str, Any] = "wavlm" def __init__( self : Tuple , __SCREAMING_SNAKE_CASE : List[Any]=32 , __SCREAMING_SNAKE_CASE : Optional[Any]=768 , __SCREAMING_SNAKE_CASE : Optional[int]=12 , __SCREAMING_SNAKE_CASE : List[Any]=12 , __SCREAMING_SNAKE_CASE : int=3072 , __SCREAMING_SNAKE_CASE : Optional[int]="gelu" , __SCREAMING_SNAKE_CASE : int=0.1 , __SCREAMING_SNAKE_CASE : List[str]=0.1 , __SCREAMING_SNAKE_CASE : str=0.1 , __SCREAMING_SNAKE_CASE : int=0.0 , __SCREAMING_SNAKE_CASE : List[Any]=0.1 , __SCREAMING_SNAKE_CASE : List[str]=0.1 , __SCREAMING_SNAKE_CASE : Dict=0.02 , __SCREAMING_SNAKE_CASE : Any=1e-5 , __SCREAMING_SNAKE_CASE : List[str]="group" , __SCREAMING_SNAKE_CASE : List[Any]="gelu" , __SCREAMING_SNAKE_CASE : int=(512, 512, 512, 512, 512, 512, 512) , __SCREAMING_SNAKE_CASE : Any=(5, 2, 2, 2, 2, 2, 2) , __SCREAMING_SNAKE_CASE : int=(10, 3, 3, 3, 3, 2, 2) , __SCREAMING_SNAKE_CASE : Optional[int]=False , __SCREAMING_SNAKE_CASE : Any=128 , __SCREAMING_SNAKE_CASE : int=16 , __SCREAMING_SNAKE_CASE : Any=320 , __SCREAMING_SNAKE_CASE : Optional[Any]=800 , __SCREAMING_SNAKE_CASE : str=False , __SCREAMING_SNAKE_CASE : str=True , __SCREAMING_SNAKE_CASE : List[Any]=0.05 , __SCREAMING_SNAKE_CASE : int=10 , __SCREAMING_SNAKE_CASE : Tuple=2 , __SCREAMING_SNAKE_CASE : Tuple=0.0 , __SCREAMING_SNAKE_CASE : str=10 , __SCREAMING_SNAKE_CASE : Tuple=320 , __SCREAMING_SNAKE_CASE : Optional[Any]=2 , __SCREAMING_SNAKE_CASE : List[Any]=0.1 , __SCREAMING_SNAKE_CASE : Any=100 , __SCREAMING_SNAKE_CASE : Tuple=256 , __SCREAMING_SNAKE_CASE : Dict=256 , __SCREAMING_SNAKE_CASE : Optional[Any]=0.1 , __SCREAMING_SNAKE_CASE : int="mean" , __SCREAMING_SNAKE_CASE : List[str]=False , __SCREAMING_SNAKE_CASE : Tuple=False , __SCREAMING_SNAKE_CASE : str=256 , __SCREAMING_SNAKE_CASE : Union[str, Any]=(512, 512, 512, 512, 1500) , __SCREAMING_SNAKE_CASE : Any=(5, 3, 3, 1, 1) , __SCREAMING_SNAKE_CASE : Optional[int]=(1, 2, 3, 1, 1) , __SCREAMING_SNAKE_CASE : int=512 , __SCREAMING_SNAKE_CASE : int=80 , __SCREAMING_SNAKE_CASE : List[Any]=0 , __SCREAMING_SNAKE_CASE : Any=1 , __SCREAMING_SNAKE_CASE : List[Any]=2 , __SCREAMING_SNAKE_CASE : str=False , __SCREAMING_SNAKE_CASE : Any=3 , __SCREAMING_SNAKE_CASE : Optional[Any]=2 , __SCREAMING_SNAKE_CASE : List[Any]=3 , __SCREAMING_SNAKE_CASE : Dict=None , **__SCREAMING_SNAKE_CASE : int , ) -> Optional[int]: super().__init__(**__SCREAMING_SNAKE_CASE , pad_token_id=__SCREAMING_SNAKE_CASE , bos_token_id=__SCREAMING_SNAKE_CASE , eos_token_id=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =hidden_size __UpperCAmelCase =feat_extract_norm __UpperCAmelCase =feat_extract_activation __UpperCAmelCase =list(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =list(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =list(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =conv_bias __UpperCAmelCase =num_buckets __UpperCAmelCase =max_bucket_distance __UpperCAmelCase =num_conv_pos_embeddings __UpperCAmelCase =num_conv_pos_embedding_groups __UpperCAmelCase =len(self.conv_dim ) __UpperCAmelCase =num_hidden_layers __UpperCAmelCase =intermediate_size __UpperCAmelCase =hidden_act __UpperCAmelCase =num_attention_heads __UpperCAmelCase =hidden_dropout __UpperCAmelCase =attention_dropout __UpperCAmelCase =activation_dropout __UpperCAmelCase =feat_proj_dropout __UpperCAmelCase =final_dropout __UpperCAmelCase =layerdrop __UpperCAmelCase =layer_norm_eps __UpperCAmelCase =initializer_range __UpperCAmelCase =num_ctc_classes __UpperCAmelCase =vocab_size __UpperCAmelCase =do_stable_layer_norm __UpperCAmelCase =use_weighted_layer_sum __UpperCAmelCase =classifier_proj_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==""" """ `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =""" f''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' f''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __UpperCAmelCase =apply_spec_augment __UpperCAmelCase =mask_time_prob __UpperCAmelCase =mask_time_length __UpperCAmelCase =mask_time_min_masks __UpperCAmelCase =mask_feature_prob __UpperCAmelCase =mask_feature_length # parameters for pretraining with codevector quantized representations __UpperCAmelCase =num_codevectors_per_group __UpperCAmelCase =num_codevector_groups __UpperCAmelCase =contrastive_logits_temperature __UpperCAmelCase =num_negatives __UpperCAmelCase =codevector_dim __UpperCAmelCase =proj_codevector_dim __UpperCAmelCase =diversity_loss_weight # ctc loss __UpperCAmelCase =ctc_loss_reduction __UpperCAmelCase =ctc_zero_infinity # adapter __UpperCAmelCase =add_adapter __UpperCAmelCase =adapter_kernel_size __UpperCAmelCase =adapter_stride __UpperCAmelCase =num_adapter_layers __UpperCAmelCase =output_hidden_size or hidden_size # SequenceClassification-specific parameter. Feel free to ignore for other classes. __UpperCAmelCase =classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. __UpperCAmelCase =list(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =list(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =list(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =xvector_output_dim @property def _a ( self : List[str] ) -> List[Any]: return functools.reduce(operator.mul , self.conv_stride , 1 )
68
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) _UpperCamelCase : str = { "configuration_perceiver": ["PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PerceiverConfig", "PerceiverOnnxConfig"], "tokenization_perceiver": ["PerceiverTokenizer"], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : List[Any] = ["PerceiverFeatureExtractor"] _UpperCamelCase : Optional[int] = ["PerceiverImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _UpperCamelCase : List[str] = [ "PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST", "PerceiverForImageClassificationConvProcessing", "PerceiverForImageClassificationFourier", "PerceiverForImageClassificationLearned", "PerceiverForMaskedLM", "PerceiverForMultimodalAutoencoding", "PerceiverForOpticalFlow", "PerceiverForSequenceClassification", "PerceiverLayer", "PerceiverModel", "PerceiverPreTrainedModel", ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys _UpperCamelCase : int = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
599
0
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
78
"""simple docstring""" import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) _a = [ ["""attention""", """attn"""], ["""encoder_attention""", """encoder_attn"""], ["""q_lin""", """q_proj"""], ["""k_lin""", """k_proj"""], ["""v_lin""", """v_proj"""], ["""out_lin""", """out_proj"""], ["""norm_embeddings""", """layernorm_embedding"""], ["""position_embeddings""", """embed_positions"""], ["""embeddings""", """embed_tokens"""], ["""ffn.lin""", """fc"""], ] def lowerCamelCase__ ( __snake_case ) -> Optional[Any]: """simple docstring""" if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: _UpperCamelCase = k.replace(__snake_case, __snake_case ) if k.startswith('''encoder''' ): _UpperCamelCase = k.replace('''.attn''', '''.self_attn''' ) _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''final_layer_norm''' ) elif k.startswith('''decoder''' ): _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''encoder_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm3''', '''final_layer_norm''' ) return k def lowerCamelCase__ ( __snake_case ) -> Optional[int]: """simple docstring""" _UpperCamelCase = [ '''model.encoder.layernorm_embedding.weight''', '''model.encoder.layernorm_embedding.bias''', '''model.decoder.layernorm_embedding.weight''', '''model.decoder.layernorm_embedding.bias''', ] for k in keys: _UpperCamelCase = sd.pop(__snake_case ) _UpperCamelCase = k.replace('''layernorm_embedding''', '''layer_norm''' ) assert new_k not in sd _UpperCamelCase = v _a = ["""START"""] @torch.no_grad() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> int: """simple docstring""" _UpperCamelCase = torch.load(__snake_case, map_location='''cpu''' ) _UpperCamelCase = model['''model'''] _UpperCamelCase = BlenderbotConfig.from_json_file(__snake_case ) _UpperCamelCase = BlenderbotForConditionalGeneration(__snake_case ) _UpperCamelCase = m.model.state_dict().keys() _UpperCamelCase = [] _UpperCamelCase = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue _UpperCamelCase = rename_state_dict_key(__snake_case ) if new_k not in valid_keys: failures.append([k, new_k] ) else: _UpperCamelCase = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(__snake_case ) m.model.load_state_dict(__snake_case, strict=__snake_case ) m.half() m.save_pretrained(__snake_case ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument("""--src_path""", type=str, help="""like blenderbot-model.bin""") parser.add_argument("""--save_dir""", default="""hf_blenderbot""", type=str, help="""Where to save converted model.""") parser.add_argument( """--hf_config_json""", default="""blenderbot-3b-config.json""", type=str, help="""Path to config to use""" ) _a = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
78
1
import numpy as np import torch from torch.utils.data import DataLoader from accelerate.utils.dataclasses import DistributedType class UpperCamelCase__ : def __init__( self : Optional[Any], __lowerCamelCase : Any=2, __lowerCamelCase : List[str]=3, __lowerCamelCase : Dict=64, __lowerCamelCase : List[Any]=None ) -> List[str]: UpperCamelCase__ : Dict = np.random.default_rng(__lowerCamelCase ) UpperCamelCase__ : Union[str, Any] = length UpperCamelCase__ : Any = rng.normal(size=(length,) ).astype(np.floataa ) UpperCamelCase__ : Optional[int] = a * self.x + b + rng.normal(scale=0.1, size=(length,) ).astype(np.floataa ) def __len__( self : Optional[int] ) -> Optional[Any]: return self.length def __getitem__( self : Dict, __lowerCamelCase : Optional[int] ) -> str: return {"x": self.x[i], "y": self.y[i]} class UpperCamelCase__ ( torch.nn.Module ): def __init__( self : Dict, __lowerCamelCase : str=0, __lowerCamelCase : Any=0, __lowerCamelCase : List[str]=False ) -> Any: super().__init__() UpperCamelCase__ : Dict = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) UpperCamelCase__ : Union[str, Any] = torch.nn.Parameter(torch.tensor([2, 3] ).float() ) UpperCamelCase__ : List[str] = True def __lowercase( self : Union[str, Any], __lowerCamelCase : Dict=None ) -> Any: if self.first_batch: print(f'Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}' ) UpperCamelCase__ : List[str] = False return x * self.a[0] + self.b[0] class UpperCamelCase__ ( torch.nn.Module ): def __init__( self : Optional[Any], __lowerCamelCase : int=0, __lowerCamelCase : List[Any]=0, __lowerCamelCase : Any=False ) -> int: super().__init__() UpperCamelCase__ : int = torch.nn.Parameter(torch.tensor(__lowerCamelCase ).float() ) UpperCamelCase__ : Optional[int] = torch.nn.Parameter(torch.tensor(__lowerCamelCase ).float() ) UpperCamelCase__ : List[Any] = True def __lowercase( self : Union[str, Any], __lowerCamelCase : str=None ) -> int: if self.first_batch: print(f'Model dtype: {self.a.dtype}, {self.b.dtype}. Input dtype: {x.dtype}' ) UpperCamelCase__ : Any = False return x * self.a + self.b def _lowercase ( __lowerCamelCase : List[str] ,__lowerCamelCase : List[Any] = 16 ) -> List[str]: '''simple docstring''' from datasets import load_dataset from transformers import AutoTokenizer UpperCamelCase__ : List[Any] = AutoTokenizer.from_pretrained('''bert-base-cased''' ) UpperCamelCase__ : int = {'train': 'tests/test_samples/MRPC/train.csv', 'validation': 'tests/test_samples/MRPC/dev.csv'} UpperCamelCase__ : Optional[Any] = load_dataset('''csv''' ,data_files=__SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Union[str, Any] = datasets['train'].unique('''label''' ) UpperCamelCase__ : str = {v: i for i, v in enumerate(__SCREAMING_SNAKE_CASE )} def tokenize_function(__lowerCamelCase : Dict ): # max_length=None => use the model max length (it's actually the default) UpperCamelCase__ : Union[str, Any] = tokenizer( examples['''sentence1'''] ,examples['''sentence2'''] ,truncation=__SCREAMING_SNAKE_CASE ,max_length=__SCREAMING_SNAKE_CASE ,padding='''max_length''' ) if "label" in examples: UpperCamelCase__ : int = [label_to_id[l] for l in examples['label']] return outputs # Apply the method we just defined to all the examples in all the splits of the dataset UpperCamelCase__ : List[Any] = datasets.map( __SCREAMING_SNAKE_CASE ,batched=__SCREAMING_SNAKE_CASE ,remove_columns=['''sentence1''', '''sentence2''', '''label'''] ,) def collate_fn(__lowerCamelCase : str ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(__SCREAMING_SNAKE_CASE ,padding='''max_length''' ,max_length=128 ,return_tensors='''pt''' ) return tokenizer.pad(__SCREAMING_SNAKE_CASE ,padding='''longest''' ,return_tensors='''pt''' ) # Instantiate dataloaders. UpperCamelCase__ : Tuple = DataLoader(tokenized_datasets['''train'''] ,shuffle=__SCREAMING_SNAKE_CASE ,collate_fn=__SCREAMING_SNAKE_CASE ,batch_size=2 ) UpperCamelCase__ : Tuple = DataLoader(tokenized_datasets['''validation'''] ,shuffle=__SCREAMING_SNAKE_CASE ,collate_fn=__SCREAMING_SNAKE_CASE ,batch_size=1 ) return train_dataloader, eval_dataloader
344
# Usage: # ./gen-card-allenai-wmt16.py import os from pathlib import Path def _lowercase ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) -> int: UpperCamelCase__ : Dict = { 'en': 'Machine learning is great, isn\'t it?', 'ru': 'Машинное обучение - это здорово, не так ли?', 'de': 'Maschinelles Lernen ist großartig, nicht wahr?', } # BLUE scores as follows: # "pair": [fairseq, transformers] UpperCamelCase__ : List[str] = { 'wmt16-en-de-dist-12-1': [28.3, 27.52], 'wmt16-en-de-dist-6-1': [27.4, 27.11], 'wmt16-en-de-12-1': [26.9, 25.75], } UpperCamelCase__ : Optional[int] = F"""{src_lang}-{tgt_lang}""" UpperCamelCase__ : Tuple = F""" --- language: - {src_lang} - {tgt_lang} thumbnail: tags: - translation - wmt16 - allenai license: apache-2.0 datasets: - wmt16 metrics: - bleu --- # FSMT ## Model description This is a ported version of fairseq-based [wmt16 transformer](https://github.com/jungokasai/deep-shallow/) for {src_lang}-{tgt_lang}. For more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation](https://arxiv.org/abs/2006.10369). All 3 models are available: * [wmt16-en-de-dist-12-1](https://huggingface.co/allenai/wmt16-en-de-dist-12-1) * [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1) * [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1) ## Intended uses & limitations #### How to use ```python from transformers import FSMTForConditionalGeneration, FSMTTokenizer mname = \"allenai/{model_name}\" tokenizer = FSMTTokenizer.from_pretrained(mname) model = FSMTForConditionalGeneration.from_pretrained(mname) input = \"{texts[src_lang]}\" input_ids = tokenizer.encode(input, return_tensors=\"pt\") outputs = model.generate(input_ids) decoded = tokenizer.decode(outputs[0], skip_special_tokens=True) print(decoded) # {texts[tgt_lang]} ``` #### Limitations and bias ## Training data Pretrained weights were left identical to the original model released by allenai. For more details, please, see the [paper](https://arxiv.org/abs/2006.10369). ## Eval results Here are the BLEU scores: model | fairseq | transformers -------|---------|---------- {model_name} | {scores[model_name][0]} | {scores[model_name][1]} The score is slightly below the score reported in the paper, as the researchers don't use `sacrebleu` and measure the score on tokenized outputs. `transformers` score was measured using `sacrebleu` on detokenized outputs. The score was calculated using this code: ```bash git clone https://github.com/huggingface/transformers cd transformers export PAIR={pair} export DATA_DIR=data/$PAIR export SAVE_DIR=data/$PAIR export BS=8 export NUM_BEAMS=5 mkdir -p $DATA_DIR sacrebleu -t wmt16 -l $PAIR --echo src > $DATA_DIR/val.source sacrebleu -t wmt16 -l $PAIR --echo ref > $DATA_DIR/val.target echo $PAIR PYTHONPATH=\"src:examples/seq2seq\" python examples/seq2seq/run_eval.py allenai/{model_name} $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS ``` ## Data Sources - [training, etc.](http://www.statmt.org/wmt16/) - [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372) ### BibTeX entry and citation info ``` @misc{{kasai2020deep, title={{Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation}}, author={{Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith}}, year={{2020}}, eprint={{2006.10369}}, archivePrefix={{arXiv}}, primaryClass={{cs.CL}} }} ``` """ model_card_dir.mkdir(parents=__SCREAMING_SNAKE_CASE , exist_ok=__SCREAMING_SNAKE_CASE ) UpperCamelCase__ : Tuple = os.path.join(__SCREAMING_SNAKE_CASE , 'README.md' ) print(F"""Generating {path}""" ) with open(__SCREAMING_SNAKE_CASE , 'w' , encoding='utf-8' ) as f: f.write(__SCREAMING_SNAKE_CASE ) # make sure we are under the root of the project UpperCAmelCase__ : Any = Path(__file__).resolve().parent.parent.parent UpperCAmelCase__ : Optional[Any] = repo_dir / '''model_cards''' for model_name in ["wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1"]: UpperCAmelCase__ : List[str] = model_cards_dir / '''allenai''' / model_name write_model_card(model_card_dir, src_lang='''en''', tgt_lang='''de''', model_name=model_name)
410
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from diffusers import DDIMScheduler, KandinskyVaaPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class A_ ( _UpperCAmelCase , unittest.TestCase ): """simple docstring""" lowercase : Union[str, Any] = KandinskyVaaPipeline lowercase : Union[str, Any] = [ "image_embeds", "negative_image_embeds", ] lowercase : List[str] = ["image_embeds", "negative_image_embeds"] lowercase : int = [ "generator", "height", "width", "latents", "guidance_scale", "num_inference_steps", "return_dict", "guidance_scale", "num_images_per_prompt", "output_type", "return_dict", ] lowercase : Optional[int] = False @property def lowercase_ ( self ) -> Optional[int]: return 32 @property def lowercase_ ( self ) -> int: return 32 @property def lowercase_ ( self ) -> List[Any]: return self.time_input_dim @property def lowercase_ ( self ) -> int: return self.time_input_dim * 4 @property def lowercase_ ( self ) -> Union[str, Any]: return 1_00 @property def lowercase_ ( self ) -> Optional[Any]: torch.manual_seed(0 ) a : Tuple = { 'in_channels': 4, # Out channels is double in channels because predicts mean and variance 'out_channels': 8, 'addition_embed_type': 'image', 'down_block_types': ('ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D'), 'up_block_types': ('SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'), 'mid_block_type': 'UNetMidBlock2DSimpleCrossAttn', 'block_out_channels': (self.block_out_channels_a, self.block_out_channels_a * 2), 'layers_per_block': 1, 'encoder_hid_dim': self.text_embedder_hidden_size, 'encoder_hid_dim_type': 'image_proj', 'cross_attention_dim': self.cross_attention_dim, 'attention_head_dim': 4, 'resnet_time_scale_shift': 'scale_shift', 'class_embed_type': None, } a : Optional[Any] = UNetaDConditionModel(**__UpperCAmelCase ) return model @property def lowercase_ ( self ) -> Dict: return { "block_out_channels": [32, 64], "down_block_types": ["DownEncoderBlock2D", "AttnDownEncoderBlock2D"], "in_channels": 3, "latent_channels": 4, "layers_per_block": 1, "norm_num_groups": 8, "norm_type": "spatial", "num_vq_embeddings": 12, "out_channels": 3, "up_block_types": [ "AttnUpDecoderBlock2D", "UpDecoderBlock2D", ], "vq_embed_dim": 4, } @property def lowercase_ ( self ) -> Dict: torch.manual_seed(0 ) a : Dict = VQModel(**self.dummy_movq_kwargs ) return model def lowercase_ ( self ) -> List[Any]: a : Tuple = self.dummy_unet a : int = self.dummy_movq a : int = DDIMScheduler( num_train_timesteps=10_00 , beta_schedule='linear' , beta_start=0.0_0085 , beta_end=0.012 , clip_sample=__UpperCAmelCase , set_alpha_to_one=__UpperCAmelCase , steps_offset=1 , prediction_type='epsilon' , thresholding=__UpperCAmelCase , ) a : int = { 'unet': unet, 'scheduler': scheduler, 'movq': movq, } return components def lowercase_ ( self , __UpperCAmelCase , __UpperCAmelCase=0 ) -> Union[str, Any]: a : List[str] = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(__UpperCAmelCase ) ).to(__UpperCAmelCase ) a : int = floats_tensor((1, self.text_embedder_hidden_size) , rng=random.Random(seed + 1 ) ).to( __UpperCAmelCase ) if str(__UpperCAmelCase ).startswith('mps' ): a : Optional[Any] = torch.manual_seed(__UpperCAmelCase ) else: a : int = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) a : Any = { 'image_embeds': image_embeds, 'negative_image_embeds': negative_image_embeds, 'generator': generator, 'height': 64, 'width': 64, 'guidance_scale': 4.0, 'num_inference_steps': 2, 'output_type': 'np', } return inputs def lowercase_ ( self ) -> List[str]: a : str = 'cpu' a : Any = self.get_dummy_components() a : Optional[Any] = self.pipeline_class(**__UpperCAmelCase ) a : int = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) a : str = pipe(**self.get_dummy_inputs(__UpperCAmelCase ) ) a : Dict = output.images a : Any = pipe( **self.get_dummy_inputs(__UpperCAmelCase ) , return_dict=__UpperCAmelCase , )[0] a : Union[str, Any] = image[0, -3:, -3:, -1] a : Dict = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) a : Optional[int] = np.array( [0.623_7976, 1.0, 0.3644_1332, 1.0, 0.7063_9634, 0.2987_7186, 0.8565_2125, 0.521_6843, 0.5445_4046] ) assert ( np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 ), f' expected_slice {expected_slice}, but got {image_slice.flatten()}' assert ( np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 ), f' expected_slice {expected_slice}, but got {image_from_tuple_slice.flatten()}' @slow @require_torch_gpu class A_ ( unittest.TestCase ): """simple docstring""" def lowercase_ ( self ) -> Optional[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ ( self ) -> Optional[Any]: a : Dict = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/kandinskyv22/kandinskyv22_text2img_cat_fp16.npy' ) a : Optional[Any] = KandinskyVaaPriorPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-prior' , torch_dtype=torch.floataa ) pipe_prior.to(__UpperCAmelCase ) a : str = KandinskyVaaPipeline.from_pretrained( 'kandinsky-community/kandinsky-2-2-decoder' , torch_dtype=torch.floataa ) a : str = pipeline.to(__UpperCAmelCase ) pipeline.set_progress_bar_config(disable=__UpperCAmelCase ) a : List[Any] = 'red cat, 4k photo' a : Optional[int] = torch.Generator(device='cuda' ).manual_seed(0 ) a : List[str] = pipe_prior( __UpperCAmelCase , generator=__UpperCAmelCase , num_inference_steps=5 , negative_prompt='' , ).to_tuple() a : int = torch.Generator(device='cuda' ).manual_seed(0 ) a : Any = pipeline( image_embeds=__UpperCAmelCase , negative_image_embeds=__UpperCAmelCase , generator=__UpperCAmelCase , num_inference_steps=1_00 , output_type='np' , ) a : Optional[int] = output.images[0] assert image.shape == (5_12, 5_12, 3) assert_mean_pixel_difference(__UpperCAmelCase , __UpperCAmelCase )
709
"""simple docstring""" import warnings from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class A_ ( _UpperCAmelCase ): """simple docstring""" lowercase : str = ["image_processor", "tokenizer"] lowercase : Union[str, Any] = "FlavaImageProcessor" lowercase : Dict = ("BertTokenizer", "BertTokenizerFast") def __init__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , **__UpperCAmelCase ) -> Tuple: a : List[str] = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , __UpperCAmelCase , ) a : Any = kwargs.pop('feature_extractor' ) a : str = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(__UpperCAmelCase , __UpperCAmelCase ) a : Optional[Any] = self.image_processor def __call__( self , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = True , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = None , __UpperCAmelCase = 0 , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = False , __UpperCAmelCase = True , __UpperCAmelCase = None , **__UpperCAmelCase , ) -> List[str]: if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: a : Tuple = self.tokenizer( text=__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , padding=__UpperCAmelCase , truncation=__UpperCAmelCase , max_length=__UpperCAmelCase , stride=__UpperCAmelCase , pad_to_multiple_of=__UpperCAmelCase , return_token_type_ids=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , return_overflowing_tokens=__UpperCAmelCase , return_special_tokens_mask=__UpperCAmelCase , return_offsets_mapping=__UpperCAmelCase , return_length=__UpperCAmelCase , verbose=__UpperCAmelCase , return_tensors=__UpperCAmelCase , **__UpperCAmelCase , ) if images is not None: a : Tuple = self.image_processor( __UpperCAmelCase , return_image_mask=__UpperCAmelCase , return_codebook_pixels=__UpperCAmelCase , return_tensors=__UpperCAmelCase , **__UpperCAmelCase , ) if text is not None and images is not None: encoding.update(__UpperCAmelCase ) return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__UpperCAmelCase ) , tensor_type=__UpperCAmelCase ) def lowercase_ ( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[Any]: return self.tokenizer.batch_decode(*__UpperCAmelCase , **__UpperCAmelCase ) def lowercase_ ( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> Optional[int]: return self.tokenizer.decode(*__UpperCAmelCase , **__UpperCAmelCase ) @property def lowercase_ ( self ) -> str: a : str = self.tokenizer.model_input_names a : List[str] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def lowercase_ ( self ) -> List[Any]: warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , __UpperCAmelCase , ) return self.image_processor_class @property def lowercase_ ( self ) -> Any: warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , __UpperCAmelCase , ) return self.image_processor
509
0
a__ = { "joule": 1.0, "kilojoule": 1000, "megajoule": 1000000, "gigajoule": 1000000000, "wattsecond": 1.0, "watthour": 3600, "kilowatthour": 3600000, "newtonmeter": 1.0, "calorie_nutr": 4186.8, "kilocalorie_nutr": 4186800.00, "electronvolt": 1.6_0217_6634E-19, "britishthermalunit_it": 1055.05585, "footpound": 1.355818, } def __UpperCAmelCase ( __a : str ,__a : str ,__a : float ) -> float: """simple docstring""" if to_type not in ENERGY_CONVERSION or from_type not in ENERGY_CONVERSION: _a : Optional[Any] = ( F"""Incorrect 'from_type' or 'to_type' value: {from_type!r}, {to_type!r}\n""" F"""Valid values are: {', '.join(__a )}""" ) raise ValueError(__a ) return value * ENERGY_CONVERSION[from_type] / ENERGY_CONVERSION[to_type] if __name__ == "__main__": import doctest doctest.testmod()
14
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def UpperCAmelCase ( A__: int , A__: str , A__: List[Any]=None , A__: Dict=None ) -> List[str]: if attention_mask is None: __lowerCamelCase : int = tf.cast(tf.math.not_equal(A__ , config.pad_token_id ) , tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class __lowercase: '''simple docstring''' __a : Any = OPTConfig __a : Union[str, Any] = {} __a : Any = 'gelu' def __init__( self , __a , __a=13 , __a=7 , __a=True , __a=False , __a=99 , __a=16 , __a=2 , __a=4 , __a=4 , __a="gelu" , __a=0.1 , __a=0.1 , __a=20 , __a=2 , __a=1 , __a=0 , __a=16 , __a=16 , ): __lowerCamelCase : Dict = parent __lowerCamelCase : List[str] = batch_size __lowerCamelCase : Tuple = seq_length __lowerCamelCase : int = is_training __lowerCamelCase : Optional[int] = use_labels __lowerCamelCase : Optional[int] = vocab_size __lowerCamelCase : Any = hidden_size __lowerCamelCase : Tuple = num_hidden_layers __lowerCamelCase : Optional[int] = num_attention_heads __lowerCamelCase : Any = intermediate_size __lowerCamelCase : Union[str, Any] = hidden_act __lowerCamelCase : int = hidden_dropout_prob __lowerCamelCase : List[Any] = attention_probs_dropout_prob __lowerCamelCase : Dict = max_position_embeddings __lowerCamelCase : str = eos_token_id __lowerCamelCase : int = pad_token_id __lowerCamelCase : Union[str, Any] = bos_token_id __lowerCamelCase : Dict = embed_dim __lowerCamelCase : Tuple = word_embed_proj_dim __lowerCamelCase : Any = False def snake_case_ ( self ): __lowerCamelCase : str = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowerCamelCase : int = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowerCamelCase : Tuple = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowerCamelCase : Any = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=__a , **self.config_updates , ) __lowerCamelCase : int = prepare_opt_inputs_dict(__a , __a ) return config, inputs_dict def snake_case_ ( self , __a , __a ): __lowerCamelCase : Optional[int] = TFOPTModel(config=__a ) __lowerCamelCase : Dict = inputs_dict['input_ids'] __lowerCamelCase : List[Any] = input_ids[:1, :] __lowerCamelCase : Optional[int] = inputs_dict['attention_mask'][:1, :] __lowerCamelCase : Any = 1 # first forward pass __lowerCamelCase : int = model(__a , attention_mask=__a , use_cache=__a ) __lowerCamelCase , __lowerCamelCase : Tuple = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowerCamelCase : Dict = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowerCamelCase : Union[str, Any] = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowerCamelCase : Union[str, Any] = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowerCamelCase : Dict = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowerCamelCase : Dict = model(__a , attention_mask=__a )[0] __lowerCamelCase : str = model(__a , attention_mask=__a , past_key_values=__a )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowerCamelCase : Dict = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowerCamelCase : Dict = output_from_no_past[:, -3:, random_slice_idx] __lowerCamelCase : Any = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__a , __a , rtol=1E-3 ) @require_tf class __lowercase( lowercase__ , lowercase__ , unittest.TestCase ): '''simple docstring''' __a : List[Any] = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () __a : List[Any] = (TFOPTForCausalLM,) if is_tf_available() else () __a : List[str] = ( {'feature-extraction': TFOPTModel, 'text-generation': TFOPTForCausalLM} if is_tf_available() else {} ) __a : List[Any] = False __a : Dict = False __a : Dict = False __a : int = 10 def snake_case_ ( self ): __lowerCamelCase : Optional[int] = TFOPTModelTester(self ) __lowerCamelCase : List[Any] = ConfigTester(self , config_class=__a ) def snake_case_ ( self ): self.config_tester.run_common_tests() def snake_case_ ( self ): __lowerCamelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__a ) def snake_case_ ( self ): __lowerCamelCase , __lowerCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(__a , __a ): if hasattr(__a , 'weight' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(__a , 'weight' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 10, config.vocab_size + 10]: # build the embeddings __lowerCamelCase : Any = model_class(config=__a ) __lowerCamelCase : Union[str, Any] = _get_word_embedding_weight(__a , model.get_input_embeddings() ) __lowerCamelCase : Any = _get_word_embedding_weight(__a , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(__a ) __lowerCamelCase : int = _get_word_embedding_weight(__a , model.get_input_embeddings() ) __lowerCamelCase : Tuple = _get_word_embedding_weight(__a , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. __lowerCamelCase : Tuple = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , __a ) # check that weights remain the same after resizing __lowerCamelCase : List[Any] = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCamelCase : str = False self.assertTrue(__a ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , __a ) __lowerCamelCase : Dict = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowerCamelCase : Tuple = False self.assertTrue(__a ) def UpperCAmelCase ( A__: Tuple ) -> Dict: return tf.constant(A__ , dtype=tf.intaa ) @require_tf class __lowercase( unittest.TestCase ): '''simple docstring''' __a : str = 99 def snake_case_ ( self ): __lowerCamelCase : str = tf.ones((4, 1) , dtype=tf.intaa ) * 2 __lowerCamelCase : int = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) __lowerCamelCase : int = input_ids.shape[0] __lowerCamelCase : Optional[int] = OPTConfig( vocab_size=self.vocab_size , hidden_size=24 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class __lowercase( unittest.TestCase ): '''simple docstring''' @slow def snake_case_ ( self ): __lowerCamelCase : str = TFOPTModel.from_pretrained('facebook/opt-350m' ) __lowerCamelCase : Dict = _long_tensor([[0, 31414, 232, 328, 740, 1140, 12695, 69, 46078, 1588, 2]] ) __lowerCamelCase : List[Any] = tf.not_equal(__a , model.config.pad_token_id ) with tf.GradientTape(): __lowerCamelCase : Dict = model(input_ids=__a , attention_mask=__a ).last_hidden_state __lowerCamelCase : List[Any] = (1, 11, 512) self.assertEqual(output.shape , __a ) __lowerCamelCase : str = tf.constant( [[-0.2_873, -1.9_218, -0.3_033], [-1.2_710, -0.1_338, -0.1_902], [0.4_095, 0.1_214, -1.3_121]] ) self.assertTrue(np.allclose(output[:, :3, :3] , __a , atol=4E-3 ) ) __lowerCamelCase : int = tf.function(__a , jit_compile=__a ) __lowerCamelCase : Optional[int] = xla_generate(__a , __a )[0] self.assertTrue(np.allclose(output[:, :3, :3] , __a , atol=4E-2 ) ) @require_tf @slow class __lowercase( unittest.TestCase ): '''simple docstring''' def snake_case_ ( self ): super().setUp() __lowerCamelCase : str = 'facebook/opt-350m' def snake_case_ ( self ): __lowerCamelCase : int = TFOPTForCausalLM.from_pretrained(self.path_model ) __lowerCamelCase : List[Any] = GPTaTokenizer.from_pretrained(self.path_model ) __lowerCamelCase : Optional[int] = [ 'Today is a beautiful day and I want to', 'In the city of', 'Paris is the capital of France and', 'Computers and mobile phones have taken', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False __lowerCamelCase : str = tokenizer(__a , return_tensors='tf' , padding=__a , add_special_tokens=__a ) __lowerCamelCase : Tuple = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) __lowerCamelCase : List[str] = tf.constant( [ [1.3_851, -13.8_923, -10.5_229, -10.7_533, -0.2_309, -10.2_384, -0.5_365, -9.0_947, -5.1_670], [-4.7_073, -10.6_276, -3.9_415, -21.5_242, -0.2_822, -0.2_822, -0.2_822, -0.2_822, -0.2_822], [0.6_247, -3.4_229, -8.9_179, -1.4_297, -14.1_650, 1.4_146, -9.0_218, -0.2_703, -0.2_703], [6.4_783, -1.9_913, -10.7_926, -2.3_336, 1.5_092, -0.9_974, -6.8_213, 1.3_477, 1.3_477], ] ) self.assertTrue(np.allclose(__a , __a , atol=1E-4 ) ) __lowerCamelCase : Union[str, Any] = tf.function(__a , jit_compile=__a ) __lowerCamelCase : List[Any] = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(__a , __a , atol=1E-4 ) ) @require_tf @slow class __lowercase( unittest.TestCase ): '''simple docstring''' @property def snake_case_ ( self ): return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def snake_case_ ( self ): __lowerCamelCase : List[str] = 'facebook/opt-125m' __lowerCamelCase : Dict = [ 'Today is a beautiful day and I want to', 'In the city of New York, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] __lowerCamelCase : Tuple = [] __lowerCamelCase : str = GPTaTokenizer.from_pretrained(__a ) __lowerCamelCase : Dict = TFOPTForCausalLM.from_pretrained(__a ) for prompt in self.prompts: __lowerCamelCase : Dict = tokenizer(__a , return_tensors='tf' ).input_ids __lowerCamelCase : int = model.generate(__a , max_length=10 ) __lowerCamelCase : List[Any] = tokenizer.batch_decode(__a , skip_special_tokens=__a ) predicted_outputs += generated_string self.assertListEqual(__a , __a ) def snake_case_ ( self ): __lowerCamelCase : int = 'facebook/opt-350m' __lowerCamelCase : List[Any] = GPTaTokenizer.from_pretrained(__a ) __lowerCamelCase : str = TFOPTForCausalLM.from_pretrained(__a ) __lowerCamelCase : Optional[int] = 'left' # use different length sentences to test batching __lowerCamelCase : List[Any] = [ 'Hello, my dog is a little', 'Today, I', ] __lowerCamelCase : Optional[int] = tokenizer(__a , return_tensors='tf' , padding=__a ) __lowerCamelCase : Tuple = inputs['input_ids'] __lowerCamelCase : Optional[Any] = model.generate(input_ids=__a , attention_mask=inputs['attention_mask'] ) __lowerCamelCase : Any = tokenizer(sentences[0] , return_tensors='tf' ).input_ids __lowerCamelCase : Optional[Any] = model.generate(input_ids=__a ) __lowerCamelCase : List[Any] = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['attention_mask'][-1] , tf.intaa ) ) __lowerCamelCase : Union[str, Any] = tokenizer(sentences[1] , return_tensors='tf' ).input_ids __lowerCamelCase : Any = model.generate(input_ids=__a , max_length=model.config.max_length - num_paddings ) __lowerCamelCase : Optional[int] = tokenizer.batch_decode(__a , skip_special_tokens=__a ) __lowerCamelCase : str = tokenizer.decode(output_non_padded[0] , skip_special_tokens=__a ) __lowerCamelCase : Union[str, Any] = tokenizer.decode(output_padded[0] , skip_special_tokens=__a ) __lowerCamelCase : Any = [ 'Hello, my dog is a little bit of a dork.\nI\'m a little bit', 'Today, I was in the middle of a conversation with a friend about the', ] self.assertListEqual(__a , __a ) self.assertListEqual(__a , [non_padded_sentence, padded_sentence] ) def snake_case_ ( self ): __lowerCamelCase : Any = 'facebook/opt-350m' __lowerCamelCase : str = [ 'Today is a beautiful day and I want to', 'In the city of San Francisco, the city', 'Paris is the capital of France and the capital', 'Computers and mobile phones have taken over the', ] __lowerCamelCase : int = [] __lowerCamelCase : Tuple = GPTaTokenizer.from_pretrained(__a ) __lowerCamelCase : List[Any] = TFOPTForCausalLM.from_pretrained(__a ) for prompt in self.prompts: __lowerCamelCase : Optional[Any] = tokenizer(__a , return_tensors='tf' ).input_ids __lowerCamelCase : List[Any] = model.generate(__a , max_length=10 ) __lowerCamelCase : Union[str, Any] = tokenizer.batch_decode(__a , skip_special_tokens=__a ) predicted_outputs += generated_string self.assertListEqual(__a , __a )
594
0
"""simple docstring""" from __future__ import annotations import copy import inspect import json import math import os import tempfile import unittest from importlib import import_module import numpy as np from transformers import ViTMAEConfig from transformers.file_utils import cached_property, is_tf_available, is_vision_available from transformers.testing_utils import require_tf, require_vision, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTMAEForPreTraining, TFViTMAEModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class A__ : """simple docstring""" def __init__( self: Dict , __a: Union[str, Any] , __a: List[Any]=13 , __a: Optional[Any]=30 , __a: List[Any]=2 , __a: Dict=3 , __a: str=True , __a: Any=True , __a: List[str]=32 , __a: Optional[Any]=2 , __a: int=4 , __a: Optional[Any]=37 , __a: Union[str, Any]="gelu" , __a: str=0.1 , __a: str=0.1 , __a: Tuple=10 , __a: Tuple=0.02 , __a: Optional[int]=3 , __a: Optional[Any]=0.6 , __a: Dict=None , )-> List[str]: lowerCamelCase : Any = parent lowerCamelCase : Optional[int] = batch_size lowerCamelCase : Dict = image_size lowerCamelCase : List[str] = patch_size lowerCamelCase : int = num_channels lowerCamelCase : str = is_training lowerCamelCase : Union[str, Any] = use_labels lowerCamelCase : Union[str, Any] = hidden_size lowerCamelCase : int = num_hidden_layers lowerCamelCase : List[str] = num_attention_heads lowerCamelCase : Optional[Any] = intermediate_size lowerCamelCase : Any = hidden_act lowerCamelCase : Tuple = hidden_dropout_prob lowerCamelCase : Optional[Any] = attention_probs_dropout_prob lowerCamelCase : List[Any] = type_sequence_label_size lowerCamelCase : Tuple = initializer_range lowerCamelCase : Dict = mask_ratio lowerCamelCase : int = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) lowerCamelCase : Union[str, Any] = (image_size // patch_size) ** 2 lowerCamelCase : int = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def a__ ( self: int )-> str: lowerCamelCase : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase : Union[str, Any] = None if self.use_labels: lowerCamelCase : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCamelCase : Union[str, Any] = self.get_config() return config, pixel_values, labels def a__ ( self: List[str] )-> List[str]: return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , decoder_hidden_size=self.hidden_size , decoder_num_hidden_layers=self.num_hidden_layers , decoder_num_attention_heads=self.num_attention_heads , decoder_intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A_ , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def a__ ( self: List[str] , __a: int , __a: str , __a: Union[str, Any] )-> Tuple: lowerCamelCase : List[str] = TFViTMAEModel(config=A_ ) lowerCamelCase : Any = model(A_ , training=A_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def a__ ( self: Optional[Any] , __a: Dict , __a: Union[str, Any] , __a: List[str] )-> Optional[int]: lowerCamelCase : List[str] = TFViTMAEForPreTraining(A_ ) lowerCamelCase : int = model(A_ , training=A_ ) # expected sequence length = num_patches lowerCamelCase : str = (self.image_size // self.patch_size) ** 2 lowerCamelCase : Tuple = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images lowerCamelCase : List[str] = 1 lowerCamelCase : List[str] = TFViTMAEForPreTraining(A_ ) lowerCamelCase : Any = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowerCamelCase : Union[str, Any] = model(A_ , training=A_ ) lowerCamelCase : Union[str, Any] = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def a__ ( self: Dict )-> int: lowerCamelCase : List[str] = self.prepare_config_and_inputs() ((lowerCamelCase) , (lowerCamelCase) , (lowerCamelCase)) : List[Any] = config_and_inputs lowerCamelCase : Union[str, Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class A__ ( _lowercase , _lowercase , unittest.TestCase): """simple docstring""" snake_case__ : int =(TFViTMAEModel, TFViTMAEForPreTraining) if is_tf_available() else () snake_case__ : str ={'''feature-extraction''': TFViTMAEModel} if is_tf_available() else {} snake_case__ : Any =False snake_case__ : Any =False snake_case__ : str =False snake_case__ : str =False def a__ ( self: List[Any] )-> Optional[int]: lowerCamelCase : List[str] = TFViTMAEModelTester(self ) lowerCamelCase : Optional[int] = ConfigTester(self , config_class=A_ , has_text_modality=A_ , hidden_size=37 ) def a__ ( self: int )-> Optional[int]: self.config_tester.run_common_tests() @unittest.skip(reason="""ViTMAE does not use inputs_embeds""" ) def a__ ( self: Any )-> Any: pass def a__ ( self: List[str] )-> List[str]: lowerCamelCase , lowerCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase : str = model_class(A_ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) lowerCamelCase : int = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A_ , tf.keras.layers.Layer ) ) def a__ ( self: Dict )-> Optional[int]: lowerCamelCase , lowerCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase : str = model_class(A_ ) lowerCamelCase : Optional[int] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase : Optional[int] = [*signature.parameters.keys()] lowerCamelCase : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , A_ ) def a__ ( self: List[Any] )-> List[str]: lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A_ ) def a__ ( self: Optional[Any] )-> int: lowerCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*A_ ) def a__ ( self: List[str] )-> Optional[Any]: # make the mask reproducible np.random.seed(2 ) lowerCamelCase , lowerCamelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase : str = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase : List[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase : str = model_class(A_ ) lowerCamelCase : Dict = self._prepare_for_class(A_ , A_ ) lowerCamelCase : Optional[int] = model(A_ , noise=A_ ) lowerCamelCase : int = copy.deepcopy(self._prepare_for_class(A_ , A_ ) ) lowerCamelCase : str = model(**A_ , noise=A_ ) lowerCamelCase : Optional[int] = outputs_dict[0].numpy() lowerCamelCase : Any = outputs_keywords[0].numpy() self.assertLess(np.sum(np.abs(output_dict - output_keywords ) ) , 1e-6 ) def a__ ( self: List[Any] )-> Dict: # make the mask reproducible np.random.seed(2 ) lowerCamelCase , lowerCamelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase : Optional[Any] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase : str = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) def prepare_numpy_arrays(__a: Union[str, Any] ): lowerCamelCase : Union[str, Any] = {} for k, v in inputs_dict.items(): if tf.is_tensor(A_ ): lowerCamelCase : Tuple = v.numpy() else: lowerCamelCase : int = np.array(A_ ) return inputs_np_dict for model_class in self.all_model_classes: lowerCamelCase : List[str] = model_class(A_ ) lowerCamelCase : Dict = self._prepare_for_class(A_ , A_ ) lowerCamelCase : Dict = prepare_numpy_arrays(A_ ) lowerCamelCase : Any = model(A_ , noise=A_ ) lowerCamelCase : int = model(**A_ , noise=A_ ) self.assert_outputs_same(A_ , A_ ) def a__ ( self: Tuple , __a: Dict , __a: Union[str, Any] , __a: Optional[Any] )-> Dict: # make masks reproducible np.random.seed(2 ) lowerCamelCase : Optional[Any] = int((tf_model.config.image_size // tf_model.config.patch_size) ** 2 ) lowerCamelCase : Any = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowerCamelCase : Dict = tf.constant(A_ ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument lowerCamelCase : List[str] = tf_noise super().check_pt_tf_models(A_ , A_ , A_ ) def a__ ( self: Optional[int] )-> Any: # make mask reproducible np.random.seed(2 ) lowerCamelCase , lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase : Tuple = { module_member for model_class in self.all_model_classes for module in (import_module(model_class.__module__ ),) for module_member_name in dir(A_ ) if module_member_name.endswith("""MainLayer""" ) # This condition is required, since `modeling_tf_clip.py` has 3 classes whose names end with `MainLayer`. and module_member_name[: -len("""MainLayer""" )] == model_class.__name__[: -len("""Model""" )] for module_member in (getattr(A_ , A_ ),) if isinstance(A_ , A_ ) and tf.keras.layers.Layer in module_member.__bases__ and getattr(A_ , """_keras_serializable""" , A_ ) } lowerCamelCase : Optional[Any] = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase : str = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowerCamelCase : Optional[Any] = tf.convert_to_tensor(A_ ) inputs_dict.update({"""noise""": noise} ) for main_layer_class in tf_main_layer_classes: lowerCamelCase : List[str] = main_layer_class(A_ ) lowerCamelCase : Union[str, Any] = { name: tf.keras.Input(tensor.shape[1:] , dtype=tensor.dtype ) for name, tensor in inputs_dict.items() } lowerCamelCase : List[str] = tf.keras.Model(A_ , outputs=main_layer(A_ ) ) lowerCamelCase : Optional[int] = model(A_ ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase : Any = os.path.join(A_ , """keras_model.h5""" ) model.save(A_ ) lowerCamelCase : str = tf.keras.models.load_model( A_ , custom_objects={main_layer_class.__name__: main_layer_class} ) assert isinstance(A_ , tf.keras.Model ) lowerCamelCase : Any = model(A_ ) self.assert_outputs_same(A_ , A_ ) @slow def a__ ( self: Any )-> str: # make mask reproducible np.random.seed(2 ) lowerCamelCase , lowerCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase : Tuple = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase : Optional[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase : Optional[int] = model_class(A_ ) lowerCamelCase : List[str] = self._prepare_for_class(A_ , A_ ) lowerCamelCase : str = model(A_ , noise=A_ ) if model_class.__name__ == "TFViTMAEModel": lowerCamelCase : Any = outputs.last_hidden_state.numpy() lowerCamelCase : List[Any] = 0 else: lowerCamelCase : Any = outputs.logits.numpy() lowerCamelCase : List[Any] = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(A_ , saved_model=A_ ) lowerCamelCase : Optional[int] = model_class.from_pretrained(A_ ) lowerCamelCase : Tuple = model(A_ , noise=A_ ) if model_class.__name__ == "TFViTMAEModel": lowerCamelCase : List[Any] = after_outputs["""last_hidden_state"""].numpy() lowerCamelCase : Dict = 0 else: lowerCamelCase : Tuple = after_outputs["""logits"""].numpy() lowerCamelCase : int = 0 lowerCamelCase : Any = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(A_ , 1e-5 ) def a__ ( self: str )-> Any: # make mask reproducible np.random.seed(2 ) lowerCamelCase , lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() lowerCamelCase : Tuple = int((config.image_size // config.patch_size) ** 2 ) lowerCamelCase : List[Any] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) for model_class in self.all_model_classes: lowerCamelCase : List[Any] = model_class(A_ ) lowerCamelCase : Any = self._prepare_for_class(A_ , A_ ) lowerCamelCase : Dict = model(A_ , noise=A_ ) lowerCamelCase : Any = model.get_config() # make sure that returned config is jsonifiable, which is required by keras json.dumps(A_ ) lowerCamelCase : Optional[int] = model_class.from_config(model.get_config() ) # make sure it also accepts a normal config lowerCamelCase : List[Any] = model_class.from_config(model.config ) lowerCamelCase : Optional[int] = new_model(A_ ) # Build model new_model.set_weights(model.get_weights() ) lowerCamelCase : Any = new_model(A_ , noise=A_ ) self.assert_outputs_same(A_ , A_ ) @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.""" ) def a__ ( self: Optional[int] )-> Optional[Any]: pass @unittest.skip(reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load""" ) def a__ ( self: List[Any] )-> Dict: pass @slow def a__ ( self: Union[str, Any] )-> List[str]: lowerCamelCase : Dict = TFViTMAEModel.from_pretrained("""google/vit-base-patch16-224""" ) self.assertIsNotNone(A_ ) def snake_case ( ) -> str: lowerCamelCase : Tuple = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class A__ ( unittest.TestCase): """simple docstring""" @cached_property def a__ ( self: Dict )-> Union[str, Any]: return ViTImageProcessor.from_pretrained("""facebook/vit-mae-base""" ) if is_vision_available() else None @slow def a__ ( self: Union[str, Any] )-> Optional[int]: # make random mask reproducible across the PT and TF model np.random.seed(2 ) lowerCamelCase : int = TFViTMAEForPreTraining.from_pretrained("""facebook/vit-mae-base""" ) lowerCamelCase : int = self.default_image_processor lowerCamelCase : Optional[Any] = prepare_img() lowerCamelCase : Any = image_processor(images=A_ , return_tensors="""tf""" ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) lowerCamelCase : Optional[Any] = ViTMAEConfig() lowerCamelCase : int = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) lowerCamelCase : Union[str, Any] = np.random.uniform(size=(1, num_patches) ) # forward pass lowerCamelCase : int = model(**A_ , noise=A_ ) # verify the logits lowerCamelCase : Optional[Any] = tf.convert_to_tensor([1, 196, 768] ) self.assertEqual(outputs.logits.shape , A_ ) lowerCamelCase : Dict = tf.convert_to_tensor( [[-0.05_48, -1.70_23, -0.93_25], [0.37_21, -0.56_70, -0.22_33], [0.82_35, -1.38_78, -0.35_24]] ) tf.debugging.assert_near(outputs.logits[0, :3, :3] , A_ , atol=1e-4 )
700
"""simple docstring""" import unittest from knapsack import greedy_knapsack as kp class A__ ( unittest.TestCase): """simple docstring""" def a__ ( self: Optional[int] )-> Union[str, Any]: lowerCamelCase : Tuple = [10, 20, 30, 40, 50, 60] lowerCamelCase : Union[str, Any] = [2, 4, 6, 8, 10, 12] lowerCamelCase : Union[str, Any] = 100 self.assertEqual(kp.calc_profit(__a , __a , __a ) , 210 ) def a__ ( self: str )-> str: self.assertRaisesRegex(__a , """max_weight must greater than zero.""" ) def a__ ( self: str )-> List[Any]: self.assertRaisesRegex(__a , """Weight can not be negative.""" ) def a__ ( self: Any )-> Dict: self.assertRaisesRegex(__a , """Profit can not be negative.""" ) def a__ ( self: Optional[Any] )-> List[Any]: self.assertRaisesRegex(__a , """max_weight must greater than zero.""" ) def a__ ( self: Optional[Any] )-> Tuple: self.assertRaisesRegex( __a , """The length of profit and weight must be same.""" ) if __name__ == "__main__": unittest.main()
42
0
__magic_name__: int = "0.18.2" from .configuration_utils import ConfigMixin from .utils import ( OptionalDependencyNotAvailable, is_flax_available, is_inflect_available, is_invisible_watermark_available, is_k_diffusion_available, is_k_diffusion_version, is_librosa_available, is_note_seq_available, is_onnx_available, is_scipy_available, is_torch_available, is_torchsde_available, is_transformers_available, is_transformers_version, is_unidecode_available, logging, ) try: if not is_onnx_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_onnx_objects import * # noqa F403 else: from .pipelines import OnnxRuntimeModel try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_pt_objects import * # noqa F403 else: from .models import ( AutoencoderKL, ControlNetModel, ModelMixin, PriorTransformer, TaFilmDecoder, TransformeraDModel, UNetaDModel, UNetaDConditionModel, UNetaDModel, UNetaDConditionModel, VQModel, ) from .optimization import ( get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, get_scheduler, ) from .pipelines import ( AudioPipelineOutput, ConsistencyModelPipeline, DanceDiffusionPipeline, DDIMPipeline, DDPMPipeline, DiffusionPipeline, DiTPipeline, ImagePipelineOutput, KarrasVePipeline, LDMPipeline, LDMSuperResolutionPipeline, PNDMPipeline, RePaintPipeline, ScoreSdeVePipeline, ) from .schedulers import ( CMStochasticIterativeScheduler, DDIMInverseScheduler, DDIMParallelScheduler, DDIMScheduler, DDPMParallelScheduler, DDPMScheduler, DEISMultistepScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, HeunDiscreteScheduler, IPNDMScheduler, KarrasVeScheduler, KDPMaAncestralDiscreteScheduler, KDPMaDiscreteScheduler, PNDMScheduler, RePaintScheduler, SchedulerMixin, ScoreSdeVeScheduler, UnCLIPScheduler, UniPCMultistepScheduler, VQDiffusionScheduler, ) from .training_utils import EMAModel try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .schedulers import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .schedulers import DPMSolverSDEScheduler try: if not (is_torch_available() and is_transformers_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipelines import ( AltDiffusionImgaImgPipeline, AltDiffusionPipeline, AudioLDMPipeline, CycleDiffusionPipeline, IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ImageTextPipelineOutput, KandinskyImgaImgPipeline, KandinskyInpaintPipeline, KandinskyPipeline, KandinskyPriorPipeline, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaControlnetPipeline, KandinskyVaaImgaImgPipeline, KandinskyVaaInpaintPipeline, KandinskyVaaPipeline, KandinskyVaaPriorEmbaEmbPipeline, KandinskyVaaPriorPipeline, LDMTextToImagePipeline, PaintByExamplePipeline, SemanticStableDiffusionPipeline, ShapEImgaImgPipeline, ShapEPipeline, StableDiffusionAttendAndExcitePipeline, StableDiffusionControlNetImgaImgPipeline, StableDiffusionControlNetInpaintPipeline, StableDiffusionControlNetPipeline, StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionImageVariationPipeline, StableDiffusionImgaImgPipeline, StableDiffusionInpaintPipeline, StableDiffusionInpaintPipelineLegacy, StableDiffusionInstructPixaPixPipeline, StableDiffusionLatentUpscalePipeline, StableDiffusionLDMaDPipeline, StableDiffusionModelEditingPipeline, StableDiffusionPanoramaPipeline, StableDiffusionParadigmsPipeline, StableDiffusionPipeline, StableDiffusionPipelineSafe, StableDiffusionPixaPixZeroPipeline, StableDiffusionSAGPipeline, StableDiffusionUpscalePipeline, StableUnCLIPImgaImgPipeline, StableUnCLIPPipeline, TextToVideoSDPipeline, TextToVideoZeroPipeline, UnCLIPImageVariationPipeline, UnCLIPPipeline, UniDiffuserModel, UniDiffuserPipeline, UniDiffuserTextDecoder, VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, VideoToVideoSDPipeline, VQDiffusionPipeline, ) try: if not (is_torch_available() and is_transformers_available() and is_invisible_watermark_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_invisible_watermark_objects import * # noqa F403 else: from .pipelines import StableDiffusionXLImgaImgPipeline, StableDiffusionXLPipeline try: if not (is_torch_available() and is_transformers_available() and is_k_diffusion_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipelines import StableDiffusionKDiffusionPipeline try: if not (is_torch_available() and is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_onnx_objects import * # noqa F403 else: from .pipelines import ( OnnxStableDiffusionImgaImgPipeline, OnnxStableDiffusionInpaintPipeline, OnnxStableDiffusionInpaintPipelineLegacy, OnnxStableDiffusionPipeline, OnnxStableDiffusionUpscalePipeline, StableDiffusionOnnxPipeline, ) try: if not (is_torch_available() and is_librosa_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_librosa_objects import * # noqa F403 else: from .pipelines import AudioDiffusionPipeline, Mel try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .pipelines import SpectrogramDiffusionPipeline try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_flax_objects import * # noqa F403 else: from .models.controlnet_flax import FlaxControlNetModel from .models.modeling_flax_utils import FlaxModelMixin from .models.unet_ad_condition_flax import FlaxUNetaDConditionModel from .models.vae_flax import FlaxAutoencoderKL from .pipelines import FlaxDiffusionPipeline from .schedulers import ( FlaxDDIMScheduler, FlaxDDPMScheduler, FlaxDPMSolverMultistepScheduler, FlaxKarrasVeScheduler, FlaxLMSDiscreteScheduler, FlaxPNDMScheduler, FlaxSchedulerMixin, FlaxScoreSdeVeScheduler, ) try: if not (is_flax_available() and is_transformers_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_flax_and_transformers_objects import * # noqa F403 else: from .pipelines import ( FlaxStableDiffusionControlNetPipeline, FlaxStableDiffusionImgaImgPipeline, FlaxStableDiffusionInpaintPipeline, FlaxStableDiffusionPipeline, ) try: if not (is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_note_seq_objects import * # noqa F403 else: from .pipelines import MidiProcessor
324
from __future__ import annotations from functools import lru_cache from math import ceil __magic_name__: Tuple = 100 __magic_name__: Any = set(range(3, NUM_PRIMES, 2)) primes.add(2) __magic_name__: int for prime in range(3, ceil(NUM_PRIMES**0.5), 2): if prime not in primes: continue primes.difference_update(set(range(prime * prime, NUM_PRIMES, prime))) @lru_cache(maxsize=100 ) def UpperCamelCase ( _A ): """simple docstring""" if number_to_partition < 0: return set() elif number_to_partition == 0: return {1} __magic_name__ : set[int] = set() __magic_name__ : int __magic_name__ : int for prime in primes: if prime > number_to_partition: continue for sub in partition(number_to_partition - prime ): ret.add(sub * prime ) return ret def UpperCamelCase ( _A = 5000 ): """simple docstring""" for number_to_partition in range(1, _A ): if len(partition(_A ) ) > number_unique_partitions: return number_to_partition return None if __name__ == "__main__": print(F"""{solution() = }""")
324
1
'''simple docstring''' def _snake_case ( A , A ) -> str: _enforce_args(A , A ) if n == 0: return 0 lowerCAmelCase__ = float('''-inf''' ) for i in range(1 , n + 1 ): lowerCAmelCase__ = max( A , prices[i - 1] + naive_cut_rod_recursive(n - i , A ) ) return max_revue def _snake_case ( A , A ) -> str: _enforce_args(A , A ) lowerCAmelCase__ = [float('''-inf''' ) for _ in range(n + 1 )] return _top_down_cut_rod_recursive(A , A , A ) def _snake_case ( A , A , A ) -> str: if max_rev[n] >= 0: return max_rev[n] elif n == 0: return 0 else: lowerCAmelCase__ = float('''-inf''' ) for i in range(1 , n + 1 ): lowerCAmelCase__ = max( A , prices[i - 1] + _top_down_cut_rod_recursive(n - i , A , A ) , ) lowerCAmelCase__ = max_revenue return max_rev[n] def _snake_case ( A , A ) -> Union[str, Any]: _enforce_args(A , A ) # length(max_rev) = n + 1, to accommodate for the revenue obtainable from a rod of # length 0. lowerCAmelCase__ = [float('''-inf''' ) for _ in range(n + 1 )] lowerCAmelCase__ = 0 for i in range(1 , n + 1 ): lowerCAmelCase__ = max_rev[i] for j in range(1 , i + 1 ): lowerCAmelCase__ = max(A , prices[j - 1] + max_rev[i - j] ) lowerCAmelCase__ = max_revenue_i return max_rev[n] def _snake_case ( A , A ) -> Optional[int]: if n < 0: lowerCAmelCase__ = F"""n must be greater than or equal to 0. Got n = {n}""" raise ValueError(A ) if n > len(A ): lowerCAmelCase__ = ( '''Each integral piece of rod must have a corresponding price. ''' F"""Got n = {n} but length of prices = {len(A )}""" ) raise ValueError(A ) def _snake_case ( ) -> Union[str, Any]: lowerCAmelCase__ = [6, 10, 12, 15, 20, 23] lowerCAmelCase__ = len(A ) # the best revenue comes from cutting the rod into 6 pieces, each # of length 1 resulting in a revenue of 6 * 6 = 36. lowerCAmelCase__ = 36 lowerCAmelCase__ = top_down_cut_rod(A , A ) lowerCAmelCase__ = bottom_up_cut_rod(A , A ) lowerCAmelCase__ = naive_cut_rod_recursive(A , A ) assert expected_max_revenue == max_rev_top_down assert max_rev_top_down == max_rev_bottom_up assert max_rev_bottom_up == max_rev_naive if __name__ == "__main__": main()
705
'''simple docstring''' from collections import Counter from timeit import timeit def _snake_case ( A = "" , ) -> bool: return sum(c % 2 for c in Counter(input_str.replace(''' ''' , '''''' ).lower() ).values() ) < 2 def _snake_case ( A = "" ) -> bool: if len(A ) == 0: return True lowerCAmelCase__ = input_str.replace(''' ''' , '''''' ).lower() # character_freq_dict: Stores the frequency of every character in the input string lowerCAmelCase__ = {} for character in lower_case_input_str: lowerCAmelCase__ = character_freq_dict.get(A , 0 ) + 1 lowerCAmelCase__ = 0 for character_count in character_freq_dict.values(): if character_count % 2: odd_char += 1 if odd_char > 1: return False return True def _snake_case ( A = "" ) -> None: print('''\nFor string = ''' , A , ''':''' ) print( '''> can_string_be_rearranged_as_palindrome_counter()''' , '''\tans =''' , can_string_be_rearranged_as_palindrome_counter(A ) , '''\ttime =''' , timeit( '''z.can_string_be_rearranged_as_palindrome_counter(z.check_str)''' , setup='''import __main__ as z''' , ) , '''seconds''' , ) print( '''> can_string_be_rearranged_as_palindrome()''' , '''\tans =''' , can_string_be_rearranged_as_palindrome(A ) , '''\ttime =''' , timeit( '''z.can_string_be_rearranged_as_palindrome(z.check_str)''' , setup='''import __main__ as z''' , ) , '''seconds''' , ) if __name__ == "__main__": __UpperCAmelCase = input( '''Enter string to determine if it can be rearranged as a palindrome or not: ''' ).strip() benchmark(check_str) __UpperCAmelCase = can_string_be_rearranged_as_palindrome_counter(check_str) print(f"""{check_str} can {'' if status else 'not '}be rearranged as a palindrome""")
98
0
'''simple docstring''' import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class __snake_case ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' lowerCamelCase__ = (PNDMScheduler,) lowerCamelCase__ = (('''num_inference_steps''', 50),) def __UpperCamelCase ( self , **__SCREAMING_SNAKE_CASE ): snake_case__ : List[str] = { """num_train_timesteps""": 1_0_0_0, """beta_start""": 0.0001, """beta_end""": 0.02, """beta_schedule""": """linear""", } config.update(**__SCREAMING_SNAKE_CASE ) return config def __UpperCamelCase ( self , __SCREAMING_SNAKE_CASE=0 , **__SCREAMING_SNAKE_CASE ): snake_case__ : List[str] = dict(self.forward_default_kwargs ) snake_case__ : Union[str, Any] = kwargs.pop("""num_inference_steps""" , __SCREAMING_SNAKE_CASE ) snake_case__ : List[str] = self.dummy_sample snake_case__ : Any = 0.1 * sample snake_case__ : Tuple = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: snake_case__ : Dict = self.get_scheduler_config(**__SCREAMING_SNAKE_CASE ) snake_case__ : Optional[int] = scheduler_class(**__SCREAMING_SNAKE_CASE ) scheduler.set_timesteps(__SCREAMING_SNAKE_CASE ) # copy over dummy past residuals snake_case__ : str = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(__SCREAMING_SNAKE_CASE ) snake_case__ : Dict = scheduler_class.from_pretrained(__SCREAMING_SNAKE_CASE ) new_scheduler.set_timesteps(__SCREAMING_SNAKE_CASE ) # copy over dummy past residuals snake_case__ : Any = dummy_past_residuals[:] snake_case__ : int = scheduler.step_prk(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ).prev_sample snake_case__ : str = new_scheduler.step_prk(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" snake_case__ : Tuple = scheduler.step_plms(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ).prev_sample snake_case__ : Optional[Any] = new_scheduler.step_plms(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __UpperCamelCase ( self ): pass def __UpperCamelCase ( self , __SCREAMING_SNAKE_CASE=0 , **__SCREAMING_SNAKE_CASE ): snake_case__ : List[Any] = dict(self.forward_default_kwargs ) snake_case__ : str = kwargs.pop("""num_inference_steps""" , __SCREAMING_SNAKE_CASE ) snake_case__ : str = self.dummy_sample snake_case__ : Optional[Any] = 0.1 * sample snake_case__ : int = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: snake_case__ : List[Any] = self.get_scheduler_config() snake_case__ : str = scheduler_class(**__SCREAMING_SNAKE_CASE ) scheduler.set_timesteps(__SCREAMING_SNAKE_CASE ) # copy over dummy past residuals (must be after setting timesteps) snake_case__ : Dict = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(__SCREAMING_SNAKE_CASE ) snake_case__ : Union[str, Any] = scheduler_class.from_pretrained(__SCREAMING_SNAKE_CASE ) # copy over dummy past residuals new_scheduler.set_timesteps(__SCREAMING_SNAKE_CASE ) # copy over dummy past residual (must be after setting timesteps) snake_case__ : List[str] = dummy_past_residuals[:] snake_case__ : int = scheduler.step_prk(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ).prev_sample snake_case__ : Dict = new_scheduler.step_prk(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" snake_case__ : int = scheduler.step_plms(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ).prev_sample snake_case__ : int = new_scheduler.step_plms(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def __UpperCamelCase ( self , **__SCREAMING_SNAKE_CASE ): snake_case__ : Any = self.scheduler_classes[0] snake_case__ : int = self.get_scheduler_config(**__SCREAMING_SNAKE_CASE ) snake_case__ : Dict = scheduler_class(**__SCREAMING_SNAKE_CASE ) snake_case__ : Optional[int] = 1_0 snake_case__ : int = self.dummy_model() snake_case__ : Any = self.dummy_sample_deter scheduler.set_timesteps(__SCREAMING_SNAKE_CASE ) for i, t in enumerate(scheduler.prk_timesteps ): snake_case__ : Dict = model(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) snake_case__ : List[Any] = scheduler.step_prk(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): snake_case__ : str = model(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) snake_case__ : List[Any] = scheduler.step_plms(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ).prev_sample return sample def __UpperCamelCase ( self ): snake_case__ : str = dict(self.forward_default_kwargs ) snake_case__ : str = kwargs.pop("""num_inference_steps""" , __SCREAMING_SNAKE_CASE ) for scheduler_class in self.scheduler_classes: snake_case__ : Optional[Any] = self.get_scheduler_config() snake_case__ : str = scheduler_class(**__SCREAMING_SNAKE_CASE ) snake_case__ : Any = self.dummy_sample snake_case__ : Tuple = 0.1 * sample if num_inference_steps is not None and hasattr(__SCREAMING_SNAKE_CASE , """set_timesteps""" ): scheduler.set_timesteps(__SCREAMING_SNAKE_CASE ) elif num_inference_steps is not None and not hasattr(__SCREAMING_SNAKE_CASE , """set_timesteps""" ): snake_case__ : int = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) snake_case__ : Any = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] snake_case__ : Any = dummy_past_residuals[:] snake_case__ : Optional[Any] = scheduler.step_prk(__SCREAMING_SNAKE_CASE , 0 , __SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ).prev_sample snake_case__ : Optional[Any] = scheduler.step_prk(__SCREAMING_SNAKE_CASE , 1 , __SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) snake_case__ : Tuple = scheduler.step_plms(__SCREAMING_SNAKE_CASE , 0 , __SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ).prev_sample snake_case__ : Dict = scheduler.step_plms(__SCREAMING_SNAKE_CASE , 1 , __SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def __UpperCamelCase ( self ): for timesteps in [1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=__SCREAMING_SNAKE_CASE ) def __UpperCamelCase ( self ): for steps_offset in [0, 1]: self.check_over_configs(steps_offset=__SCREAMING_SNAKE_CASE ) snake_case__ : List[str] = self.scheduler_classes[0] snake_case__ : Optional[Any] = self.get_scheduler_config(steps_offset=1 ) snake_case__ : Tuple = scheduler_class(**__SCREAMING_SNAKE_CASE ) scheduler.set_timesteps(1_0 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [9_0_1, 8_5_1, 8_5_1, 8_0_1, 8_0_1, 7_5_1, 7_5_1, 7_0_1, 7_0_1, 6_5_1, 6_5_1, 6_0_1, 6_0_1, 5_0_1, 4_0_1, 3_0_1, 2_0_1, 1_0_1, 1] ) , ) def __UpperCamelCase ( self ): for beta_start, beta_end in zip([0.0001, 0.001] , [0.002, 0.02] ): self.check_over_configs(beta_start=__SCREAMING_SNAKE_CASE , beta_end=__SCREAMING_SNAKE_CASE ) def __UpperCamelCase ( self ): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=__SCREAMING_SNAKE_CASE ) def __UpperCamelCase ( self ): for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=__SCREAMING_SNAKE_CASE ) def __UpperCamelCase ( self ): for t in [1, 5, 1_0]: self.check_over_forward(time_step=__SCREAMING_SNAKE_CASE ) def __UpperCamelCase ( self ): for t, num_inference_steps in zip([1, 5, 1_0] , [1_0, 5_0, 1_0_0] ): self.check_over_forward(num_inference_steps=__SCREAMING_SNAKE_CASE ) def __UpperCamelCase ( self ): # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 snake_case__ : List[str] = 2_7 for scheduler_class in self.scheduler_classes: snake_case__ : str = self.dummy_sample snake_case__ : str = 0.1 * sample snake_case__ : int = self.get_scheduler_config() snake_case__ : List[Any] = scheduler_class(**__SCREAMING_SNAKE_CASE ) scheduler.set_timesteps(__SCREAMING_SNAKE_CASE ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): snake_case__ : Union[str, Any] = scheduler.step_prk(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ).prev_sample def __UpperCamelCase ( self ): with self.assertRaises(__SCREAMING_SNAKE_CASE ): snake_case__ : Tuple = self.scheduler_classes[0] snake_case__ : Optional[int] = self.get_scheduler_config() snake_case__ : str = scheduler_class(**__SCREAMING_SNAKE_CASE ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def __UpperCamelCase ( self ): snake_case__ : List[Any] = self.full_loop() snake_case__ : Optional[int] = torch.sum(torch.abs(__SCREAMING_SNAKE_CASE ) ) snake_case__ : int = torch.mean(torch.abs(__SCREAMING_SNAKE_CASE ) ) assert abs(result_sum.item() - 198.1318 ) < 1e-2 assert abs(result_mean.item() - 0.2580 ) < 1e-3 def __UpperCamelCase ( self ): snake_case__ : Tuple = self.full_loop(prediction_type="""v_prediction""" ) snake_case__ : List[Any] = torch.sum(torch.abs(__SCREAMING_SNAKE_CASE ) ) snake_case__ : List[str] = torch.mean(torch.abs(__SCREAMING_SNAKE_CASE ) ) assert abs(result_sum.item() - 67.3986 ) < 1e-2 assert abs(result_mean.item() - 0.0878 ) < 1e-3 def __UpperCamelCase ( self ): # We specify different beta, so that the first alpha is 0.99 snake_case__ : Optional[int] = self.full_loop(set_alpha_to_one=__SCREAMING_SNAKE_CASE , beta_start=0.01 ) snake_case__ : str = torch.sum(torch.abs(__SCREAMING_SNAKE_CASE ) ) snake_case__ : Dict = torch.mean(torch.abs(__SCREAMING_SNAKE_CASE ) ) assert abs(result_sum.item() - 230.0399 ) < 1e-2 assert abs(result_mean.item() - 0.2995 ) < 1e-3 def __UpperCamelCase ( self ): # We specify different beta, so that the first alpha is 0.99 snake_case__ : Dict = self.full_loop(set_alpha_to_one=__SCREAMING_SNAKE_CASE , beta_start=0.01 ) snake_case__ : List[Any] = torch.sum(torch.abs(__SCREAMING_SNAKE_CASE ) ) snake_case__ : List[Any] = torch.mean(torch.abs(__SCREAMING_SNAKE_CASE ) ) assert abs(result_sum.item() - 186.9482 ) < 1e-2 assert abs(result_mean.item() - 0.2434 ) < 1e-3
38
import logging import os import sys from dataclasses import dataclass, field from importlib import import_module from typing import Dict, List, Optional, Tuple import numpy as np from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch import nn from utils_ner import Split, TokenClassificationDataset, TokenClassificationTask import transformers from transformers import ( AutoConfig, AutoModelForTokenClassification, AutoTokenizer, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _SCREAMING_SNAKE_CASE = logging.getLogger(__name__) @dataclass class SCREAMING_SNAKE_CASE_ : __lowerCAmelCase = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) __lowerCAmelCase = field( default=__lowerCAmelCase , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) __lowerCAmelCase = field( default="""NER""" , metadata={"""help""": """Task type to fine tune in training (e.g. NER, POS, etc)"""} ) __lowerCAmelCase = field( default=__lowerCAmelCase , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) __lowerCAmelCase = field(default=__lowerCAmelCase , metadata={"""help""": """Set this flag to use fast tokenization."""} ) # If you want to tweak more attributes on your tokenizer, you should do it in a distinct script, # or just modify its tokenizer_config.json. __lowerCAmelCase = field( default=__lowerCAmelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) @dataclass class SCREAMING_SNAKE_CASE_ : __lowerCAmelCase = field( metadata={"""help""": """The input data dir. Should contain the .txt files for a CoNLL-2003-formatted task."""} ) __lowerCAmelCase = field( default=__lowerCAmelCase , metadata={"""help""": """Path to a file containing all labels. If not specified, CoNLL-2003 labels are used."""} , ) __lowerCAmelCase = field( default=128 , metadata={ """help""": ( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) __lowerCAmelCase = field( default=__lowerCAmelCase , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) def lowercase( ) -> Optional[Any]: '''simple docstring''' # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. UpperCamelCase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(""".json""" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. UpperCamelCase , UpperCamelCase , UpperCamelCase = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: UpperCamelCase , UpperCamelCase , UpperCamelCase = parser.parse_args_into_dataclasses() if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" """ --overwrite_output_dir to overcome.""" ) UpperCamelCase = import_module("""tasks""" ) try: UpperCamelCase = getattr(UpperCamelCase_ , model_args.task_type ) UpperCamelCase = token_classification_task_clazz() except AttributeError: raise ValueError( f"""Task {model_args.task_type} needs to be defined as a TokenClassificationTask subclass in {module}. """ f"""Available tasks classes are: {TokenClassificationTask.__subclasses__()}""" ) # Setup logging logging.basicConfig( format="""%(asctime)s - %(levelname)s - %(name)s - %(message)s""" , datefmt="""%m/%d/%Y %H:%M:%S""" , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( """Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s""" , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info("""Training/evaluation parameters %s""" , UpperCamelCase_ ) # Set seed set_seed(training_args.seed ) # Prepare CONLL-2003 task UpperCamelCase = token_classification_task.get_labels(data_args.labels ) UpperCamelCase = dict(enumerate(UpperCamelCase_ ) ) UpperCamelCase = len(UpperCamelCase_ ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. UpperCamelCase = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=UpperCamelCase_ , idalabel=UpperCamelCase_ , labelaid={label: i for i, label in enumerate(UpperCamelCase_ )} , cache_dir=model_args.cache_dir , ) UpperCamelCase = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast , ) UpperCamelCase = AutoModelForTokenClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(""".ckpt""" in model_args.model_name_or_path ) , config=UpperCamelCase_ , cache_dir=model_args.cache_dir , ) # Get datasets UpperCamelCase = ( TokenClassificationDataset( token_classification_task=UpperCamelCase_ , data_dir=data_args.data_dir , tokenizer=UpperCamelCase_ , labels=UpperCamelCase_ , model_type=config.model_type , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.train , ) if training_args.do_train else None ) UpperCamelCase = ( TokenClassificationDataset( token_classification_task=UpperCamelCase_ , data_dir=data_args.data_dir , tokenizer=UpperCamelCase_ , labels=UpperCamelCase_ , model_type=config.model_type , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.dev , ) if training_args.do_eval else None ) def align_predictions(UpperCamelCase_ , UpperCamelCase_ ) -> Tuple[List[int], List[int]]: UpperCamelCase = np.argmax(UpperCamelCase_ , axis=2 ) UpperCamelCase , UpperCamelCase = preds.shape UpperCamelCase = [[] for _ in range(UpperCamelCase_ )] UpperCamelCase = [[] for _ in range(UpperCamelCase_ )] for i in range(UpperCamelCase_ ): for j in range(UpperCamelCase_ ): if label_ids[i, j] != nn.CrossEntropyLoss().ignore_index: out_label_list[i].append(label_map[label_ids[i][j]] ) preds_list[i].append(label_map[preds[i][j]] ) return preds_list, out_label_list def compute_metrics(UpperCamelCase_ ) -> Dict: UpperCamelCase , UpperCamelCase = align_predictions(p.predictions , p.label_ids ) return { "accuracy_score": accuracy_score(UpperCamelCase_ , UpperCamelCase_ ), "precision": precision_score(UpperCamelCase_ , UpperCamelCase_ ), "recall": recall_score(UpperCamelCase_ , UpperCamelCase_ ), "f1": fa_score(UpperCamelCase_ , UpperCamelCase_ ), } # Data collator UpperCamelCase = DataCollatorWithPadding(UpperCamelCase_ , pad_to_multiple_of=8 ) if training_args.fpaa else None # Initialize our Trainer UpperCamelCase = Trainer( model=UpperCamelCase_ , args=UpperCamelCase_ , train_dataset=UpperCamelCase_ , eval_dataset=UpperCamelCase_ , compute_metrics=UpperCamelCase_ , data_collator=UpperCamelCase_ , ) # Training if training_args.do_train: trainer.train( model_path=model_args.model_name_or_path if os.path.isdir(model_args.model_name_or_path ) else None ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_process_zero(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation UpperCamelCase = {} if training_args.do_eval: logger.info("""*** Evaluate ***""" ) UpperCamelCase = trainer.evaluate() UpperCamelCase = os.path.join(training_args.output_dir , """eval_results.txt""" ) if trainer.is_world_process_zero(): with open(UpperCamelCase_ , """w""" ) as writer: logger.info("""***** Eval results *****""" ) for key, value in result.items(): logger.info(""" %s = %s""" , UpperCamelCase_ , UpperCamelCase_ ) writer.write("""%s = %s\n""" % (key, value) ) results.update(UpperCamelCase_ ) # Predict if training_args.do_predict: UpperCamelCase = TokenClassificationDataset( token_classification_task=UpperCamelCase_ , data_dir=data_args.data_dir , tokenizer=UpperCamelCase_ , labels=UpperCamelCase_ , model_type=config.model_type , max_seq_length=data_args.max_seq_length , overwrite_cache=data_args.overwrite_cache , mode=Split.test , ) UpperCamelCase , UpperCamelCase , UpperCamelCase = trainer.predict(UpperCamelCase_ ) UpperCamelCase , UpperCamelCase = align_predictions(UpperCamelCase_ , UpperCamelCase_ ) UpperCamelCase = os.path.join(training_args.output_dir , """test_results.txt""" ) if trainer.is_world_process_zero(): with open(UpperCamelCase_ , """w""" ) as writer: for key, value in metrics.items(): logger.info(""" %s = %s""" , UpperCamelCase_ , UpperCamelCase_ ) writer.write("""%s = %s\n""" % (key, value) ) # Save predictions UpperCamelCase = os.path.join(training_args.output_dir , """test_predictions.txt""" ) if trainer.is_world_process_zero(): with open(UpperCamelCase_ , """w""" ) as writer: with open(os.path.join(data_args.data_dir , """test.txt""" ) , """r""" ) as f: token_classification_task.write_predictions_to_file(UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) return results def lowercase( UpperCamelCase_ ) -> Dict: '''simple docstring''' # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
537
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL UpperCamelCase_ = logging.get_logger(__name__) class snake_case ( __lowercase ): a_ : int = ['''pixel_values'''] def __init__( self , __UpperCAmelCase = True , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = PILImageResampling.BILINEAR , __UpperCAmelCase = True , __UpperCAmelCase = 1 / 2_55 , __UpperCAmelCase = True , __UpperCAmelCase = None , __UpperCAmelCase = None , **__UpperCAmelCase , ) ->None: super().__init__(**__a) a_ = size if size is not None else {"""shortest_edge""": 3_84} a_ = get_size_dict(__a , default_to_square=__a) a_ = do_resize a_ = size # Default value set here for backwards compatibility where the value in config is None a_ = crop_pct if crop_pct is not None else 2_24 / 2_56 a_ = resample a_ = do_rescale a_ = rescale_factor a_ = do_normalize a_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN a_ = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = PILImageResampling.BICUBIC , __UpperCAmelCase = None , **__UpperCAmelCase , ) ->np.ndarray: a_ = get_size_dict(__a , default_to_square=__a) if "shortest_edge" not in size: raise ValueError(F'''Size dictionary must contain \'shortest_edge\' key. Got {size.keys()}''') a_ = size["""shortest_edge"""] if shortest_edge < 3_84: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct a_ = int(shortest_edge / crop_pct) a_ = get_resize_output_image_size(__a , size=__a , default_to_square=__a) a_ = resize(image=__a , size=__a , resample=__a , data_format=__a , **__a) # then crop to (shortest_edge, shortest_edge) return center_crop(image=__a , size=(shortest_edge, shortest_edge) , data_format=__a , **__a) else: # warping (no cropping) when evaluated at 384 or larger return resize( __a , size=(shortest_edge, shortest_edge) , resample=__a , data_format=__a , **__a) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , **__UpperCAmelCase , ) ->Union[str, Any]: return rescale(__a , scale=__a , data_format=__a , **__a) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , **__UpperCAmelCase , ) ->np.ndarray: return normalize(__a , mean=__a , std=__a , data_format=__a , **__a) def UpperCAmelCase__ ( self , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = None , __UpperCAmelCase = ChannelDimension.FIRST , **__UpperCAmelCase , ) ->PIL.Image.Image: a_ = do_resize if do_resize is not None else self.do_resize a_ = crop_pct if crop_pct is not None else self.crop_pct a_ = resample if resample is not None else self.resample a_ = do_rescale if do_rescale is not None else self.do_rescale a_ = rescale_factor if rescale_factor is not None else self.rescale_factor a_ = do_normalize if do_normalize is not None else self.do_normalize a_ = image_mean if image_mean is not None else self.image_mean a_ = image_std if image_std is not None else self.image_std a_ = size if size is not None else self.size a_ = get_size_dict(__a , default_to_square=__a) a_ = make_list_of_images(__a) if not valid_images(__a): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray.") if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True.") if do_resize and size["shortest_edge"] < 3_84 and crop_pct is None: raise ValueError("crop_pct must be specified if size < 384.") if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True.") if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True.") # All transformations expect numpy arrays. a_ = [to_numpy_array(__a) for image in images] if do_resize: a_ = [self.resize(image=__a , size=__a , crop_pct=__a , resample=__a) for image in images] if do_rescale: a_ = [self.rescale(image=__a , scale=__a) for image in images] if do_normalize: a_ = [self.normalize(image=__a , mean=__a , std=__a) for image in images] a_ = [to_channel_dimension_format(__a , __a) for image in images] a_ = {"""pixel_values""": images} return BatchFeature(data=__a , tensor_type=__a)
720
"""simple docstring""" from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class snake_case : a_ : List[str] a_ : Optional[str] = None # Automatically constructed a_ : ClassVar[str] = "dict" a_ : ClassVar[Any] = None a_ : str = field(default="""Translation""" , init=SCREAMING_SNAKE_CASE_ , repr=SCREAMING_SNAKE_CASE_ ) def __call__( self) ->Tuple: return pa.struct({lang: pa.string() for lang in sorted(self.languages)}) def UpperCAmelCase__ ( self) ->Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value return {k: Value("string") for k in sorted(self.languages)} @dataclass class snake_case : a_ : Optional[List] = None a_ : Optional[int] = None a_ : Optional[str] = None # Automatically constructed a_ : ClassVar[str] = "dict" a_ : ClassVar[Any] = None a_ : str = field(default="""TranslationVariableLanguages""" , init=SCREAMING_SNAKE_CASE_ , repr=SCREAMING_SNAKE_CASE_ ) def UpperCAmelCase__ ( self) ->Optional[int]: a_ = sorted(set(self.languages)) if self.languages else None a_ = len(self.languages) if self.languages else None def __call__( self) ->Any: return pa.struct({"language": pa.list_(pa.string()), "translation": pa.list_(pa.string())}) def UpperCAmelCase__ ( self , __UpperCAmelCase) ->int: a_ = set(self.languages) if self.languages and set(__UpperCAmelCase) - lang_set: raise ValueError( F'''Some languages in example ({", ".join(sorted(set(__UpperCAmelCase) - lang_set))}) are not in valid set ({", ".join(__UpperCAmelCase)}).''') # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. a_ = [] for lang, text in translation_dict.items(): if isinstance(__UpperCAmelCase , __UpperCAmelCase): translation_tuples.append((lang, text)) else: translation_tuples.extend([(lang, el) for el in text]) # Ensure translations are in ascending order by language code. a_ , a_ = zip(*sorted(__UpperCAmelCase)) return {"language": languages, "translation": translations} def UpperCAmelCase__ ( self) ->Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Sequence, Value return { "language": Sequence(Value("string")), "translation": Sequence(Value("string")), }
210
0
'''simple docstring''' def UpperCamelCase__ ( ) -> list[list[int]]: '''simple docstring''' return [list(range(10_00 - i , -10_00 - i , -1 ) ) for i in range(10_00 )] A_ : Any = generate_large_matrix() A_ : str = ( [[4, 3, 2, -1], [3, 2, 1, -1], [1, 1, -1, -2], [-1, -1, -2, -3]], [[3, 2], [1, 0]], [[7, 7, 6]], [[7, 7, 6], [-1, -2, -3]], grid, ) def UpperCamelCase__ ( __magic_name__ : list[list[int]] ) -> None: '''simple docstring''' assert all(row == sorted(__magic_name__ , reverse=__magic_name__ ) for row in grid ) assert all(list(__magic_name__ ) == sorted(__magic_name__ , reverse=__magic_name__ ) for col in zip(*__magic_name__ ) ) def UpperCamelCase__ ( __magic_name__ : list[int] ) -> int: '''simple docstring''' snake_case__ : Dict = 0 snake_case__ : Tuple = len(__magic_name__ ) - 1 # Edge cases such as no values or all numbers are negative. if not array or array[0] < 0: return 0 while right + 1 > left: snake_case__ : Optional[Any] = (left + right) // 2 snake_case__ : List[Any] = array[mid] # Num must be negative and the index must be greater than or equal to 0. if num < 0 and array[mid - 1] >= 0: return mid if num >= 0: snake_case__ : str = mid + 1 else: snake_case__ : str = mid - 1 # No negative numbers so return the last index of the array + 1 which is the length. return len(__magic_name__ ) def UpperCamelCase__ ( __magic_name__ : list[list[int]] ) -> int: '''simple docstring''' snake_case__ : Dict = 0 snake_case__ : Any = len(grid[0] ) for i in range(len(__magic_name__ ) ): snake_case__ : int = find_negative_index(grid[i][:bound] ) total += bound return (len(__magic_name__ ) * len(grid[0] )) - total def UpperCamelCase__ ( __magic_name__ : list[list[int]] ) -> int: '''simple docstring''' return len([number for row in grid for number in row if number < 0] ) def UpperCamelCase__ ( __magic_name__ : list[list[int]] ) -> int: '''simple docstring''' snake_case__ : List[str] = 0 for row in grid: for i, number in enumerate(__magic_name__ ): if number < 0: total += len(__magic_name__ ) - i break return total def UpperCamelCase__ ( ) -> None: '''simple docstring''' from timeit import timeit print("""Running benchmarks""" ) snake_case__ : Tuple = ( """from __main__ import count_negatives_binary_search, """ """count_negatives_brute_force, count_negatives_brute_force_with_break, grid""" ) for func in ( "count_negatives_binary_search", # took 0.7727 seconds "count_negatives_brute_force_with_break", # took 4.6505 seconds "count_negatives_brute_force", # took 12.8160 seconds ): snake_case__ : Union[str, Any] = timeit(f"{func}(grid=grid)" , setup=__magic_name__ , number=5_00 ) print(f"{func}() took {time:0.4f} seconds" ) if __name__ == "__main__": import doctest doctest.testmod() benchmark()
38
class A : '''simple docstring''' def __init__( self : Optional[int] ) -> Dict: """simple docstring""" A__ = {} def a_ ( self : Any ) -> None: """simple docstring""" print(self.vertex ) for i in self.vertex: print(__lowerCAmelCase , """ -> """ , """ -> """.join([str(__lowerCAmelCase ) for j in self.vertex[i]] ) ) def a_ ( self : Optional[int] , __lowerCAmelCase : int , __lowerCAmelCase : int ) -> None: """simple docstring""" if from_vertex in self.vertex: self.vertex[from_vertex].append(__lowerCAmelCase ) else: # else make a new vertex A__ = [to_vertex] def a_ ( self : List[str] ) -> None: """simple docstring""" A__ = [False] * len(self.vertex ) # call the recursive helper function for i in range(len(self.vertex ) ): if not visited[i]: self.dfs_recursive(__lowerCAmelCase , __lowerCAmelCase ) def a_ ( self : List[str] , __lowerCAmelCase : int , __lowerCAmelCase : list ) -> None: """simple docstring""" A__ = True print(__lowerCAmelCase , end=""" """ ) # Recur for all the vertices that are adjacent to this node for i in self.vertex: if not visited[i]: self.dfs_recursive(__lowerCAmelCase , __lowerCAmelCase ) if __name__ == "__main__": A : Tuple = Graph() g.add_edge(0, 1) g.add_edge(0, 2) g.add_edge(1, 2) g.add_edge(2, 0) g.add_edge(2, 3) g.add_edge(3, 3) g.print_graph() print('''DFS:''') g.dfs() # OUTPUT: # 0 -> 1 -> 2 # 1 -> 2 # 2 -> 0 -> 3 # 3 -> 3 # DFS: # 0 1 2 3
176
0
'''simple docstring''' from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging __SCREAMING_SNAKE_CASE : str = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( __lowerCAmelCase ): lowercase__ = ['''audio_values''', '''audio_mask'''] def __init__( self , __UpperCamelCase=2048 , __UpperCamelCase=1 , __UpperCamelCase=[16, 16] , __UpperCamelCase=128 , __UpperCamelCase=4_4100 , __UpperCamelCase=86 , __UpperCamelCase=2048 , __UpperCamelCase=0.0 , **__UpperCamelCase , ): '''simple docstring''' super().__init__( feature_size=lowerCAmelCase_ , sampling_rate=lowerCAmelCase_ , padding_value=lowerCAmelCase_ , **lowerCAmelCase_ , ) __a : Tuple = spectrogram_length __a : List[Any] = num_channels __a : Optional[Any] = patch_size __a : Union[str, Any] = feature_size // self.patch_size[1] __a : Tuple = n_fft __a : Optional[int] = sampling_rate // hop_length_to_sampling_rate __a : str = sampling_rate __a : str = padding_value __a : List[str] = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=lowerCAmelCase_ , min_frequency=0.0 , max_frequency=2_2050.0 , sampling_rate=lowerCAmelCase_ , norm="""slaney""" , mel_scale="""slaney""" , ).T def __lowerCamelCase ( self , __UpperCamelCase ): '''simple docstring''' __a : Optional[Any] = spectrogram( lowerCAmelCase_ , window_function(self.n_fft , """hann""" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel="""dB""" , db_range=8_0.0 , ) __a : List[Any] = log_spec[:, :-1] __a : Any = log_spec - 2_0.0 __a : str = np.clip(log_spec / 4_0.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = False , __UpperCamelCase = False , **__UpperCamelCase , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( """This feature extractor is set to support sampling rate""" f""" of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled""" f""" with {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) __a : Union[str, Any] = isinstance(lowerCAmelCase_ , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"""Only mono-channel audio is supported for input to {self}""" ) __a : List[str] = is_batched_numpy or ( isinstance(lowerCAmelCase_ , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __a : Optional[Any] = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(lowerCAmelCase_ , np.ndarray ): __a : int = np.asarray(lowerCAmelCase_ , dtype=np.floataa ) elif isinstance(lowerCAmelCase_ , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __a : Union[str, Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __a : Dict = [np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis __a : Dict = [ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , lowerCAmelCase_ ): __a : str = [np.asarray(lowerCAmelCase_ , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask __a : List[Any] = max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: __a : Tuple = [ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] __a : str = np.array(lowerCAmelCase_ ).astype(np.floataa ) # convert into correct format for padding __a : Any = max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch __a : List[Any] = np.ones([len(lowerCAmelCase_ ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) __a : Optional[int] = padded_audio_features * self.padding_value for i in range(len(lowerCAmelCase_ ) ): __a : Tuple = audio_features[i] __a : Dict = feature # return as BatchFeature if return_attention_mask: __a : Tuple = {"""audio_values""": padded_audio_features, """audio_mask""": audio_mask} else: __a : Optional[Any] = {"""audio_values""": padded_audio_features} __a : Optional[Any] = BatchFeature(data=lowerCAmelCase_ , tensor_type=lowerCAmelCase_ ) return encoded_inputs
717
'''simple docstring''' import warnings from functools import wraps from typing import Callable def _snake_case ( lowercase ) -> Callable: @wraps(lowercase ) def _inner_fn(*lowercase , **lowercase ): warnings.warn( (F"""'{fn.__name__}' is experimental and might be subject to breaking changes in the future.""") , lowercase , ) return fn(*lowercase , **lowercase ) return _inner_fn
697
0
"""simple docstring""" class a__ ( A__ ): pass class a__ ( A__ ): pass class a__ : def __init__( self :Dict ): '''simple docstring''' UpperCamelCase_ : Union[str, Any] =[ [], [], [], ] def lowerCamelCase_ ( self :str , _lowerCamelCase :int , _lowerCamelCase :int ): '''simple docstring''' try: if len(self.queues[priority] ) >= 100: raise OverflowError('Maximum queue size is 100' ) self.queues[priority].append(snake_case__ ) except IndexError: raise ValueError('Valid priorities are 0, 1, and 2' ) def lowerCamelCase_ ( self :int ): '''simple docstring''' for queue in self.queues: if queue: return queue.pop(0 ) raise UnderFlowError('All queues are empty' ) def __str__( self :List[Any] ): '''simple docstring''' return "\n".join(f'''Priority {i}: {q}''' for i, q in enumerate(self.queues ) ) class a__ : def __init__( self :Tuple ): '''simple docstring''' UpperCamelCase_ : Union[str, Any] =[] def lowerCamelCase_ ( self :Dict , _lowerCamelCase :int ): '''simple docstring''' if len(self.queue ) == 100: raise OverFlowError('Maximum queue size is 100' ) self.queue.append(snake_case__ ) def lowerCamelCase_ ( self :Dict ): '''simple docstring''' if not self.queue: raise UnderFlowError('The queue is empty' ) else: UpperCamelCase_ : Dict =min(self.queue ) self.queue.remove(snake_case__ ) return data def __str__( self :Tuple ): '''simple docstring''' return str(self.queue ) def A_ ( ): UpperCamelCase_ : List[Any] =FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 1_00 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 1_28 ) print(__lowercase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(__lowercase ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def A_ ( ): UpperCamelCase_ : Optional[Any] =ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(1_00 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(1_28 ) print(__lowercase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(__lowercase ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
357
"""simple docstring""" import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class lowerCAmelCase__ : def __init__( self : Any , snake_case__ : List[str] , snake_case__ : Union[str, Any]=1_3 , snake_case__ : Dict=7 , snake_case__ : Optional[int]=True , snake_case__ : Optional[int]=True , snake_case__ : int=False , snake_case__ : Optional[Any]=True , snake_case__ : Optional[int]=9_9 , snake_case__ : List[Any]=3_2 , snake_case__ : Optional[Any]=5 , snake_case__ : Union[str, Any]=4 , snake_case__ : List[Any]=3_7 , snake_case__ : List[str]="gelu" , snake_case__ : List[Any]=0.1 , snake_case__ : Tuple=0.1 , snake_case__ : Optional[Any]=5_1_2 , snake_case__ : Tuple=1_6 , snake_case__ : str=2 , snake_case__ : Optional[int]=0.02 , snake_case__ : int=3 , snake_case__ : Union[str, Any]=4 , snake_case__ : Optional[int]=None , ): '''simple docstring''' UpperCAmelCase__ : List[str] = parent UpperCAmelCase__ : Optional[Any] = batch_size UpperCAmelCase__ : Any = seq_length UpperCAmelCase__ : int = is_training UpperCAmelCase__ : Dict = use_input_mask UpperCAmelCase__ : List[Any] = use_token_type_ids UpperCAmelCase__ : Union[str, Any] = use_labels UpperCAmelCase__ : Union[str, Any] = vocab_size UpperCAmelCase__ : Union[str, Any] = hidden_size UpperCAmelCase__ : Union[str, Any] = num_hidden_layers UpperCAmelCase__ : str = num_attention_heads UpperCAmelCase__ : int = intermediate_size UpperCAmelCase__ : int = hidden_act UpperCAmelCase__ : str = hidden_dropout_prob UpperCAmelCase__ : Optional[int] = attention_probs_dropout_prob UpperCAmelCase__ : List[Any] = max_position_embeddings UpperCAmelCase__ : str = type_vocab_size UpperCAmelCase__ : Union[str, Any] = type_sequence_label_size UpperCAmelCase__ : str = initializer_range UpperCAmelCase__ : Tuple = num_labels UpperCAmelCase__ : Union[str, Any] = num_choices UpperCAmelCase__ : Tuple = scope def __a ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase__ : Optional[int] = None if self.use_input_mask: UpperCAmelCase__ : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ : Dict = None if self.use_token_type_ids: UpperCAmelCase__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase__ : str = None UpperCAmelCase__ : Optional[Any] = None UpperCAmelCase__ : int = None if self.use_labels: UpperCAmelCase__ : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase__ : Tuple = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase__ : Dict = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __a ( self : Tuple ): '''simple docstring''' return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case__ , initializer_range=self.initializer_range , ) def __a ( self : Optional[int] , snake_case__ : Optional[int] , snake_case__ : Optional[int] , snake_case__ : Tuple , snake_case__ : Dict , snake_case__ : Tuple , snake_case__ : List[str] , snake_case__ : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ : Optional[Any] = LlamaModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase__ : List[str] = model(snake_case__ , attention_mask=snake_case__ ) UpperCAmelCase__ : Any = model(snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __a ( self : Any , snake_case__ : Union[str, Any] , snake_case__ : Optional[Any] , snake_case__ : List[Any] , snake_case__ : Dict , snake_case__ : Optional[int] , snake_case__ : Tuple , snake_case__ : Tuple , snake_case__ : str , snake_case__ : int , ): '''simple docstring''' UpperCAmelCase__ : List[Any] = True UpperCAmelCase__ : str = LlamaModel(snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase__ : Optional[Any] = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , encoder_attention_mask=snake_case__ , ) UpperCAmelCase__ : str = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , ) UpperCAmelCase__ : List[str] = model(snake_case__ , attention_mask=snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __a ( self : List[str] , snake_case__ : Dict , snake_case__ : Optional[int] , snake_case__ : int , snake_case__ : Optional[int] , snake_case__ : List[str] , snake_case__ : Optional[int] , snake_case__ : List[Any] , snake_case__ : List[str] , snake_case__ : Union[str, Any] , ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = LlamaForCausalLM(config=snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase__ : Optional[Any] = model(snake_case__ , attention_mask=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __a ( self : List[str] , snake_case__ : Union[str, Any] , snake_case__ : List[Any] , snake_case__ : int , snake_case__ : Optional[int] , snake_case__ : List[Any] , snake_case__ : int , snake_case__ : Optional[Any] , snake_case__ : Dict , snake_case__ : Optional[int] , ): '''simple docstring''' UpperCAmelCase__ : List[str] = True UpperCAmelCase__ : str = True UpperCAmelCase__ : Union[str, Any] = LlamaForCausalLM(config=snake_case__ ) model.to(snake_case__ ) model.eval() # first forward pass UpperCAmelCase__ : List[str] = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , encoder_attention_mask=snake_case__ , use_cache=snake_case__ , ) UpperCAmelCase__ : Optional[Any] = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids UpperCAmelCase__ : Any = ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCAmelCase__ : Optional[int] = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and UpperCAmelCase__ : str = torch.cat([input_ids, next_tokens] , dim=-1 ) UpperCAmelCase__ : List[Any] = torch.cat([input_mask, next_mask] , dim=-1 ) UpperCAmelCase__ : str = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , encoder_attention_mask=snake_case__ , output_hidden_states=snake_case__ , )["hidden_states"][0] UpperCAmelCase__ : str = model( snake_case__ , attention_mask=snake_case__ , encoder_hidden_states=snake_case__ , encoder_attention_mask=snake_case__ , past_key_values=snake_case__ , output_hidden_states=snake_case__ , )["hidden_states"][0] # select random slice UpperCAmelCase__ : Dict = ids_tensor((1,) , output_from_past.shape[-1] ).item() UpperCAmelCase__ : str = output_from_no_past[:, -3:, random_slice_idx].detach() UpperCAmelCase__ : Any = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1e-3 ) ) def __a ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : int = self.prepare_config_and_inputs() ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) : Optional[Any] = config_and_inputs UpperCAmelCase__ : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class lowerCAmelCase__ ( __magic_name__ , __magic_name__ , __magic_name__ , unittest.TestCase ): SCREAMING_SNAKE_CASE_ =(LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE_ =(LlamaForCausalLM,) if is_torch_available() else () SCREAMING_SNAKE_CASE_ =( { '''feature-extraction''': LlamaModel, '''text-classification''': LlamaForSequenceClassification, '''text-generation''': LlamaForCausalLM, '''zero-shot''': LlamaForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE_ =False SCREAMING_SNAKE_CASE_ =False def __a ( self : int ): '''simple docstring''' UpperCAmelCase__ : Optional[int] = LlamaModelTester(self ) UpperCAmelCase__ : str = ConfigTester(self , config_class=snake_case__ , hidden_size=3_7 ) def __a ( self : Union[str, Any] ): '''simple docstring''' self.config_tester.run_common_tests() def __a ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case__ ) def __a ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase__ : List[Any] = type self.model_tester.create_and_check_model(*snake_case__ ) def __a ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Tuple = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : str = 3 UpperCAmelCase__ : Any = input_dict["input_ids"] UpperCAmelCase__ : Any = input_ids.ne(1 ).to(snake_case__ ) UpperCAmelCase__ : str = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) UpperCAmelCase__ : Optional[Any] = LlamaForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase__ : Optional[int] = model(snake_case__ , attention_mask=snake_case__ , labels=snake_case__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def __a ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : List[Any] = 3 UpperCAmelCase__ : Any = "single_label_classification" UpperCAmelCase__ : Tuple = input_dict["input_ids"] UpperCAmelCase__ : str = input_ids.ne(1 ).to(snake_case__ ) UpperCAmelCase__ : List[Any] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) UpperCAmelCase__ : Any = LlamaForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase__ : List[str] = model(snake_case__ , attention_mask=snake_case__ , labels=snake_case__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def __a ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : Dict = 3 UpperCAmelCase__ : Tuple = "multi_label_classification" UpperCAmelCase__ : Any = input_dict["input_ids"] UpperCAmelCase__ : str = input_ids.ne(1 ).to(snake_case__ ) UpperCAmelCase__ : Any = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) UpperCAmelCase__ : Dict = LlamaForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() UpperCAmelCase__ : int = model(snake_case__ , attention_mask=snake_case__ , labels=snake_case__ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip("LLaMA buffers include complex numbers, which breaks this test" ) def __a ( self : str ): '''simple docstring''' pass @parameterized.expand([("linear",), ("dynamic",)] ) def __a ( self : List[str] , snake_case__ : int ): '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ : str = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ : List[str] = ids_tensor([1, 1_0] , config.vocab_size ) UpperCAmelCase__ : int = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(4_2 ) # Fixed seed at init time so the two models get the same random weights UpperCAmelCase__ : str = LlamaModel(snake_case__ ) original_model.to(snake_case__ ) original_model.eval() UpperCAmelCase__ : List[Any] = original_model(snake_case__ ).last_hidden_state UpperCAmelCase__ : Tuple = original_model(snake_case__ ).last_hidden_state set_seed(4_2 ) # Fixed seed at init time so the two models get the same random weights UpperCAmelCase__ : str = {"type": scaling_type, "factor": 10.0} UpperCAmelCase__ : Optional[int] = LlamaModel(snake_case__ ) scaled_model.to(snake_case__ ) scaled_model.eval() UpperCAmelCase__ : Tuple = scaled_model(snake_case__ ).last_hidden_state UpperCAmelCase__ : Union[str, Any] = scaled_model(snake_case__ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(snake_case__ , snake_case__ , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(snake_case__ , snake_case__ , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(snake_case__ , snake_case__ , atol=1e-5 ) ) @require_torch class lowerCAmelCase__ ( unittest.TestCase ): @unittest.skip("Logits are not exactly the same, once we fix the instabalities somehow, will update!" ) @slow def __a ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : Union[str, Any] = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] UpperCAmelCase__ : Any = LlamaForCausalLM.from_pretrained("meta-llama/Llama-2-7b-hf" , device_map="auto" ) UpperCAmelCase__ : Any = model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 UpperCAmelCase__ : Optional[Any] = torch.tensor([[-6.6550, -4.1227, -4.9859, -3.2406, 0.8262, -3.0033, 1.2964, -3.3699]] ) torch.testing.assert_close(out.mean(-1 ) , snake_case__ , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off UpperCAmelCase__ : Optional[int] = torch.tensor([-12.8281, -7.4453, -0.4639, -8.0625, -7.2500, -8.0000, -6.4883, -7.7695, -7.8438, -7.0312, -6.2188, -7.1328, -1.8496, 1.9961, -8.6250, -6.7227, -12.8281, -6.9492, -7.0742, -7.7852, -7.5820, -7.9062, -6.9375, -7.9805, -8.3438, -8.1562, -8.0469, -7.6250, -7.7422, -7.3398,] ) # fmt: on torch.testing.assert_close(out[0, 0, :3_0] , snake_case__ , atol=1e-5 , rtol=1e-5 ) @unittest.skip("Logits are not exactly the same, once we fix the instabalities somehow, will update!" ) @slow def __a ( self : Any ): '''simple docstring''' UpperCAmelCase__ : List[str] = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] UpperCAmelCase__ : Dict = LlamaForCausalLM.from_pretrained("meta-llama/Llama-2-13b-hf" , device_map="auto" ) UpperCAmelCase__ : List[str] = model(torch.tensor(snake_case__ ) ) # Expected mean on dim = -1 UpperCAmelCase__ : str = torch.tensor([[-2.0622, -1.2794, -1.1638, -0.9788, -1.4603, -1.0238, -1.7893, -1.4411]] ) torch.testing.assert_close(out.mean(-1 ) , snake_case__ , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off UpperCAmelCase__ : str = torch.tensor([-8.1406, -8.0547, 2.7461, -1.2344, -0.1448, -1.8262, -1.0020, -1.8154, -1.6895, -1.8516, -2.3574, -0.9277, 3.7598, 6.5742, -1.2998, -0.1177, -8.1406, -2.9688, -2.9199, -3.1699, -3.5254, -2.3555, -2.7988, -3.4141, -2.8262, -4.5195, -3.3379, -3.3164, -2.7832, -3.0273] ) # fmt: on torch.testing.assert_close(out[0, 0, :3_0] , snake_case__ , atol=1e-5 , rtol=1e-5 ) @unittest.skip("Logits are not exactly the same, once we fix the instabalities somehow, will update!" ) @slow def __a ( self : int ): '''simple docstring''' UpperCAmelCase__ : List[Any] = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] UpperCAmelCase__ : Optional[Any] = LlamaForCausalLM.from_pretrained("meta-llama/Llama-2-13b-chat-hf" , device_map="auto" ) UpperCAmelCase__ : Dict = model(torch.tensor(snake_case__ ) ) # Expected mean on dim = -1 UpperCAmelCase__ : List[str] = torch.tensor([[-0.8562, -1.8520, -0.7551, -0.4162, -1.5161, -1.2038, -2.4823, -2.3254]] ) torch.testing.assert_close(out.mean(-1 ) , snake_case__ , atol=1e-2 , rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off UpperCAmelCase__ : str = torch.tensor([-2.2227, 4.8828, 0.9023, -0.4578, -0.7871, -0.1033, -0.6221, -0.5786, -0.7803, -1.0674, -1.2920, -0.1570, 0.8008, 2.0723, -0.9497, 0.2771, -2.2227, -0.7612, -1.4346, -1.2061, -1.6426, -0.3000, -0.7139, -1.1934, -1.8691, -1.6973, -1.5947, -1.2705, -0.3523, -0.5513] ) # fmt: on torch.testing.assert_close(out.mean(-1 ) , snake_case__ , atol=1e-2 , rtol=1e-2 ) @unittest.skip( "Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test" ) @slow def __a ( self : Any ): '''simple docstring''' UpperCAmelCase__ : Tuple = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] UpperCAmelCase__ : Optional[int] = LlamaForCausalLM.from_pretrained("meta-llama/Llama-2-70b-hf" , device_map="auto" ) UpperCAmelCase__ : List[Any] = model(torch.tensor(snake_case__ ) ) UpperCAmelCase__ : str = torch.tensor( [[-4.2327, -3.3360, -4.6665, -4.7631, -1.8180, -3.4170, -1.4211, -3.1810]] , dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ) , snake_case__ , atol=1e-2 , rtol=1e-2 ) # fmt: off UpperCAmelCase__ : Any = torch.tensor([-9.4922, -3.9551, 1.7998, -5.6758, -5.1055, -5.8984, -4.8320, -6.8086, -6.5391, -5.6172, -5.5820, -5.5352, 1.7881, 3.6289, -6.5117, -3.4785, -9.5000, -6.0352, -6.8125, -6.0195, -6.6836, -5.4727, -6.2812, -6.0391, -7.3398, -7.4297, -7.4844, -6.5820, -5.8789, -5.5312] ) # fmt: on torch.testing.assert_close(out[0, 0, :3_0] , snake_case__ , atol=1e-5 , rtol=1e-5 ) @unittest.skip("Model is curently gated" ) @slow def __a ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : List[str] = "Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the \"princi" UpperCAmelCase__ : Any = "Simply put, the theory of relativity states that " UpperCAmelCase__ : int = LlamaTokenizer.from_pretrained("meta-llama/Llama-2-13b-chat-hf" ) UpperCAmelCase__ : List[Any] = tokenizer.encode(snake_case__ , return_tensors="pt" ) UpperCAmelCase__ : Union[str, Any] = LlamaForCausalLM.from_pretrained( "meta-llama/Llama-2-13b-chat-hf" , device_map="sequential" , use_safetensors=snake_case__ ) # greedy generation outputs UpperCAmelCase__ : Optional[int] = model.generate(snake_case__ , max_new_tokens=6_4 , top_p=snake_case__ , temperature=1 , do_sample=snake_case__ ) UpperCAmelCase__ : Tuple = tokenizer.decode(generated_ids[0] , skip_special_tokens=snake_case__ ) self.assertEqual(snake_case__ , snake_case__ )
438
0
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase : int = logging.get_logger(__name__) lowercase : Any = { '''BAAI/AltCLIP''': '''https://huggingface.co/BAAI/AltCLIP/resolve/main/config.json''', # See all AltCLIP models at https://huggingface.co/models?filter=altclip } class UpperCAmelCase_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' A : int = 'altclip_text_model' def __init__( self , _SCREAMING_SNAKE_CASE=25_0002 , _SCREAMING_SNAKE_CASE=1024 , _SCREAMING_SNAKE_CASE=24 , _SCREAMING_SNAKE_CASE=16 , _SCREAMING_SNAKE_CASE=4096 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=514 , _SCREAMING_SNAKE_CASE=1 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=1e-05 , _SCREAMING_SNAKE_CASE=1 , _SCREAMING_SNAKE_CASE=0 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE="absolute" , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=768 , **_SCREAMING_SNAKE_CASE , ) -> List[str]: super().__init__(pad_token_id=_SCREAMING_SNAKE_CASE , bos_token_id=_SCREAMING_SNAKE_CASE , eos_token_id=_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) snake_case_ : Optional[Any] = vocab_size snake_case_ : List[str] = hidden_size snake_case_ : Dict = num_hidden_layers snake_case_ : List[str] = num_attention_heads snake_case_ : Optional[int] = hidden_act snake_case_ : Optional[Any] = intermediate_size snake_case_ : List[str] = hidden_dropout_prob snake_case_ : Union[str, Any] = attention_probs_dropout_prob snake_case_ : Optional[Any] = max_position_embeddings snake_case_ : Any = type_vocab_size snake_case_ : int = initializer_range snake_case_ : str = initializer_factor snake_case_ : List[str] = layer_norm_eps snake_case_ : Any = position_embedding_type snake_case_ : Optional[int] = use_cache snake_case_ : int = project_dim class UpperCAmelCase_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' A : List[Any] = 'altclip_vision_model' def __init__( self , _SCREAMING_SNAKE_CASE=768 , _SCREAMING_SNAKE_CASE=3072 , _SCREAMING_SNAKE_CASE=512 , _SCREAMING_SNAKE_CASE=12 , _SCREAMING_SNAKE_CASE=12 , _SCREAMING_SNAKE_CASE=3 , _SCREAMING_SNAKE_CASE=224 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE="quick_gelu" , _SCREAMING_SNAKE_CASE=1e-5 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=1.0 , **_SCREAMING_SNAKE_CASE , ) -> Optional[Any]: super().__init__(**_SCREAMING_SNAKE_CASE ) snake_case_ : Tuple = hidden_size snake_case_ : Dict = intermediate_size snake_case_ : Optional[int] = projection_dim snake_case_ : int = num_hidden_layers snake_case_ : Union[str, Any] = num_attention_heads snake_case_ : int = num_channels snake_case_ : Optional[int] = patch_size snake_case_ : Dict = image_size snake_case_ : Tuple = initializer_range snake_case_ : Optional[Any] = initializer_factor snake_case_ : Any = attention_dropout snake_case_ : Tuple = layer_norm_eps snake_case_ : int = hidden_act @classmethod def _lowerCAmelCase ( cls , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> "PretrainedConfig": cls._set_token_in_kwargs(_SCREAMING_SNAKE_CASE ) snake_case_ : str = cls.get_config_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) # get the vision config dict if we are loading from AltCLIPConfig if config_dict.get("model_type" ) == "altclip": snake_case_ : List[Any] = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict['model_type']} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) class UpperCAmelCase_ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' A : str = 'altclip' A : Dict = True def __init__( self , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=768 , _SCREAMING_SNAKE_CASE=2.6592 , **_SCREAMING_SNAKE_CASE ) -> Optional[int]: # If `_config_dict` exist, we use them for the backward compatibility. # We pop out these 2 attributes before calling `super().__init__` to avoid them being saved (which causes a lot # of confusion!). snake_case_ : Optional[Any] = kwargs.pop("text_config_dict" , _SCREAMING_SNAKE_CASE ) snake_case_ : Dict = kwargs.pop("vision_config_dict" , _SCREAMING_SNAKE_CASE ) super().__init__(**_SCREAMING_SNAKE_CASE ) # Instead of simply assigning `[text|vision]_config_dict` to `[text|vision]_config`, we use the values in # `[text|vision]_config_dict` to update the values in `[text|vision]_config`. The values should be same in most # cases, but we don't want to break anything regarding `_config_dict` that existed before commit `8827e1b2`. if text_config_dict is not None: if text_config is None: snake_case_ : Any = {} # This is the complete result when using `text_config_dict`. snake_case_ : str = AltCLIPTextConfig(**_SCREAMING_SNAKE_CASE ).to_dict() # Give a warning if the values exist in both `_text_config_dict` and `text_config` but being different. for key, value in _text_config_dict.items(): if key in text_config and value != text_config[key] and key not in ["transformers_version"]: # If specified in `text_config_dict` if key in text_config_dict: snake_case_ : Optional[Any] = ( f'''`{key}` is found in both `text_config_dict` and `text_config` but with different values. ''' f'''The value `text_config_dict["{key}"]` will be used instead.''' ) # If inferred from default argument values (just to be super careful) else: snake_case_ : Dict = ( f'''`text_config_dict` is provided which will be used to initialize `AltCLIPTextConfig`. The ''' f'''value `text_config["{key}"]` will be overriden.''' ) logger.warning(_SCREAMING_SNAKE_CASE ) # Update all values in `text_config` with the ones in `_text_config_dict`. text_config.update(_text_config_dict ) if vision_config_dict is not None: if vision_config is None: snake_case_ : str = {} # This is the complete result when using `vision_config_dict`. snake_case_ : Optional[Any] = AltCLIPVisionConfig(**_SCREAMING_SNAKE_CASE ).to_dict() # convert keys to string instead of integer if "id2label" in _vision_config_dict: snake_case_ : Any = { str(_SCREAMING_SNAKE_CASE ): value for key, value in _vision_config_dict["id2label"].items() } # Give a warning if the values exist in both `_vision_config_dict` and `vision_config` but being different. for key, value in _vision_config_dict.items(): if key in vision_config and value != vision_config[key] and key not in ["transformers_version"]: # If specified in `vision_config_dict` if key in vision_config_dict: snake_case_ : Dict = ( f'''`{key}` is found in both `vision_config_dict` and `vision_config` but with different ''' f'''values. The value `vision_config_dict["{key}"]` will be used instead.''' ) # If inferred from default argument values (just to be super careful) else: snake_case_ : Optional[Any] = ( f'''`vision_config_dict` is provided which will be used to initialize `AltCLIPVisionConfig`. ''' f'''The value `vision_config["{key}"]` will be overriden.''' ) logger.warning(_SCREAMING_SNAKE_CASE ) # Update all values in `vision_config` with the ones in `_vision_config_dict`. vision_config.update(_vision_config_dict ) if text_config is None: snake_case_ : Tuple = {} logger.info("`text_config` is `None`. Initializing the `AltCLIPTextConfig` with default values." ) if vision_config is None: snake_case_ : Optional[Any] = {} logger.info("`vision_config` is `None`. initializing the `AltCLIPVisionConfig` with default values." ) snake_case_ : List[Any] = AltCLIPTextConfig(**_SCREAMING_SNAKE_CASE ) snake_case_ : Union[str, Any] = AltCLIPVisionConfig(**_SCREAMING_SNAKE_CASE ) snake_case_ : Optional[int] = projection_dim snake_case_ : Tuple = logit_scale_init_value snake_case_ : List[Any] = 1.0 @classmethod def _lowerCAmelCase ( cls , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Optional[Any]: return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **_SCREAMING_SNAKE_CASE ) def _lowerCAmelCase ( self ) -> List[str]: snake_case_ : Any = copy.deepcopy(self.__dict__ ) snake_case_ : Union[str, Any] = self.text_config.to_dict() snake_case_ : List[Any] = self.vision_config.to_dict() snake_case_ : Tuple = self.__class__.model_type return output
719
import heapq import sys import numpy as np lowercase : str = tuple[int, int] class UpperCAmelCase_ : '''simple docstring''' def __init__( self ) -> Optional[int]: snake_case_ : int = [] snake_case_ : int = set() def _lowerCAmelCase ( self ) -> List[Any]: if not self.empty(): return self.elements[0][0] else: return float("inf" ) def _lowerCAmelCase ( self ) -> Optional[int]: return len(self.elements ) == 0 def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> str: if item not in self.set: heapq.heappush(self.elements , (priority, item) ) self.set.add(_SCREAMING_SNAKE_CASE ) else: # update # print("update", item) snake_case_ : Any = [] ((snake_case_) , (snake_case_)) : Optional[Any] = heapq.heappop(self.elements ) while x != item: temp.append((pri, x) ) ((snake_case_) , (snake_case_)) : Optional[int] = heapq.heappop(self.elements ) temp.append((priority, item) ) for pro, xxx in temp: heapq.heappush(self.elements , (pro, xxx) ) def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE ) -> Any: if item in self.set: self.set.remove(_SCREAMING_SNAKE_CASE ) snake_case_ : Optional[Any] = [] ((snake_case_) , (snake_case_)) : Tuple = heapq.heappop(self.elements ) while x != item: temp.append((pro, x) ) ((snake_case_) , (snake_case_)) : Dict = heapq.heappop(self.elements ) for prito, yyy in temp: heapq.heappush(self.elements , (prito, yyy) ) def _lowerCAmelCase ( self ) -> Optional[Any]: return self.elements[0][1] def _lowerCAmelCase ( self ) -> Optional[int]: ((snake_case_) , (snake_case_)) : Any = heapq.heappop(self.elements ) self.set.remove(_SCREAMING_SNAKE_CASE ) return (priority, item) def lowerCAmelCase__ ( _a : TPos , _a : TPos ): # euclidean distance snake_case_ : Optional[Any] = np.array(_a ) snake_case_ : Dict = np.array(_a ) return np.linalg.norm(a - b ) def lowerCAmelCase__ ( _a : TPos , _a : TPos ): # integer division by time variable return consistent_heuristic(_a , _a ) // t def lowerCAmelCase__ ( _a : TPos , _a : TPos ): # manhattan distance return abs(p[0] - goal[0] ) + abs(p[1] - goal[1] ) def lowerCAmelCase__ ( _a : TPos , _a : int , _a : TPos , _a : dict[TPos, float] ): snake_case_ : List[Any] = g_function[start] + Wa * heuristics[i](_a , _a ) return ans def lowerCAmelCase__ ( _a : Tuple , _a : Union[str, Any] , _a : List[Any] ): snake_case_ : Dict = np.chararray((n, n) ) for i in range(_a ): for j in range(_a ): snake_case_ : List[str] = "*" for i in range(_a ): for j in range(_a ): if (j, (n - 1) - i) in blocks: snake_case_ : Optional[Any] = "#" snake_case_ : Optional[int] = "-" snake_case_ : Tuple = back_pointer[goal] while x != start: ((snake_case_) , (snake_case_)) : Any = x # print(x) snake_case_ : List[str] = "-" snake_case_ : Dict = back_pointer[x] snake_case_ : str = "-" for i in range(_a ): for j in range(_a ): if (i, j) == (0, n - 1): print(grid[i][j] , end=" " ) print("<-- End position" , end=" " ) else: print(grid[i][j] , end=" " ) print() print("^" ) print("Start position" ) print() print("# is an obstacle" ) print("- is the path taken by algorithm" ) print("PATH TAKEN BY THE ALGORITHM IS:-" ) snake_case_ : List[Any] = back_pointer[goal] while x != start: print(_a , end=" " ) snake_case_ : int = back_pointer[x] print(_a ) sys.exit() def lowerCAmelCase__ ( _a : TPos ): if p[0] < 0 or p[0] > n - 1: return False if p[1] < 0 or p[1] > n - 1: return False return True def lowerCAmelCase__ ( _a : int , _a : str , _a : Optional[int] , _a : List[str] , _a : Optional[Any] , _a : Dict , _a : Tuple , _a : Any , ): for itera in range(_a ): open_list[itera].remove_element(_a ) # print("s", s) # print("j", j) ((snake_case_) , (snake_case_)) : Tuple = s snake_case_ : Dict = (x - 1, y) snake_case_ : Union[str, Any] = (x + 1, y) snake_case_ : List[str] = (x, y + 1) snake_case_ : int = (x, y - 1) for neighbours in [left, right, up, down]: if neighbours not in blocks: if valid(_a ) and neighbours not in visited: # print("neighbour", neighbours) visited.add(_a ) snake_case_ : Optional[int] = -1 snake_case_ : int = float("inf" ) if valid(_a ) and g_function[neighbours] > g_function[s] + 1: snake_case_ : int = g_function[s] + 1 snake_case_ : Any = s if neighbours not in close_list_anchor: open_list[0].put(_a , key(_a , 0 , _a , _a ) ) if neighbours not in close_list_inad: for var in range(1 , _a ): if key(_a , _a , _a , _a ) <= Wa * key( _a , 0 , _a , _a ): open_list[j].put( _a , key(_a , _a , _a , _a ) ) def lowerCAmelCase__ ( ): snake_case_ : Union[str, Any] = [] for x in range(1 , 5 ): for y in range(1 , 6 ): some_list.append((x, y) ) for x in range(15 , 20 ): some_list.append((x, 17) ) for x in range(10 , 19 ): for y in range(1 , 15 ): some_list.append((x, y) ) # L block for x in range(1 , 4 ): for y in range(12 , 19 ): some_list.append((x, y) ) for x in range(3 , 13 ): for y in range(16 , 19 ): some_list.append((x, y) ) return some_list lowercase : Union[str, Any] = {0: consistent_heuristic, 1: heuristic_a, 2: heuristic_a} lowercase : Union[str, Any] = [ (0, 1), (1, 1), (2, 1), (3, 1), (4, 1), (5, 1), (6, 1), (7, 1), (8, 1), (9, 1), (10, 1), (11, 1), (12, 1), (13, 1), (14, 1), (15, 1), (16, 1), (17, 1), (18, 1), (19, 1), ] lowercase : int = make_common_ground() lowercase : Optional[Any] = blocks_blk # hyper parameters lowercase : Optional[Any] = 1 lowercase : str = 1 lowercase : Any = 20 lowercase : str = 3 # one consistent and two other inconsistent # start and end destination lowercase : Any = (0, 0) lowercase : int = (n - 1, n - 1) lowercase : str = 1 def lowerCAmelCase__ ( _a : TPos , _a : TPos , _a : int ): snake_case_ : List[str] = {start: 0, goal: float("inf" )} snake_case_ : List[Any] = {start: -1, goal: -1} snake_case_ : Optional[Any] = [] snake_case_ : Dict = set() for i in range(_a ): open_list.append(PriorityQueue() ) open_list[i].put(_a , key(_a , _a , _a , _a ) ) snake_case_ : list[int] = [] snake_case_ : list[int] = [] while open_list[0].minkey() < float("inf" ): for i in range(1 , _a ): # print(open_list[0].minkey(), open_list[i].minkey()) if open_list[i].minkey() <= Wa * open_list[0].minkey(): global t t += 1 if g_function[goal] <= open_list[i].minkey(): if g_function[goal] < float("inf" ): do_something(_a , _a , _a ) else: snake_case_ , snake_case_ : Optional[Any] = open_list[i].top_show() visited.add(_a ) expand_state( _a , _a , _a , _a , _a , _a , _a , _a , ) close_list_inad.append(_a ) else: if g_function[goal] <= open_list[0].minkey(): if g_function[goal] < float("inf" ): do_something(_a , _a , _a ) else: snake_case_ : Tuple = open_list[0].top_show() visited.add(_a ) expand_state( _a , 0 , _a , _a , _a , _a , _a , _a , ) close_list_anchor.append(_a ) print("No path found to goal" ) print() for i in range(n - 1 , -1 , -1 ): for j in range(_a ): if (j, i) in blocks: print("#" , end=" " ) elif (j, i) in back_pointer: if (j, i) == (n - 1, n - 1): print("*" , end=" " ) else: print("-" , end=" " ) else: print("*" , end=" " ) if (j, i) == (n - 1, n - 1): print("<-- End position" , end=" " ) print() print("^" ) print("Start position" ) print() print("# is an obstacle" ) print("- is the path taken by algorithm" ) if __name__ == "__main__": multi_a_star(start, goal, n_heuristic)
114
0
"""simple docstring""" import argparse import shlex import runhouse as rh if __name__ == "__main__": # Refer to https://runhouse-docs.readthedocs-hosted.com/en/latest/api/python/cluster.html#hardware-setup for cloud access # setup instructions, if using on-demand hardware # If user passes --user <user> --host <host> --key_path <key_path> <example> <args>, fill them in as BYO cluster # If user passes --instance <instance> --provider <provider> <example> <args>, fill them in as on-demand cluster # Throw an error if user passes both BYO and on-demand cluster args # Otherwise, use default values UpperCAmelCase_ : List[str] = argparse.ArgumentParser() parser.add_argument("""--user""", type=str, default="""ubuntu""") parser.add_argument("""--host""", type=str, default="""localhost""") parser.add_argument("""--key_path""", type=str, default=None) parser.add_argument("""--instance""", type=str, default="""V100:1""") parser.add_argument("""--provider""", type=str, default="""cheapest""") parser.add_argument("""--use_spot""", type=bool, default=False) parser.add_argument("""--example""", type=str, default="""pytorch/text-generation/run_generation.py""") UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = parser.parse_known_args() if args.host != "localhost": if args.instance != "V100:1" or args.provider != "cheapest": raise ValueError("""Cannot specify both BYO and on-demand cluster args""") UpperCAmelCase_ : List[Any] = rh.cluster( name="""rh-cluster""", ips=[args.host], ssh_creds={"""ssh_user""": args.user, """ssh_private_key""": args.key_path} ) else: UpperCAmelCase_ : Optional[Any] = rh.cluster( name="""rh-cluster""", instance_type=args.instance, provider=args.provider, use_spot=args.use_spot ) UpperCAmelCase_ : Tuple = args.example.rsplit("""/""", 1)[0] # Set up remote environment cluster.install_packages(["""pip:./"""]) # Installs transformers from local source # Note transformers is copied into the home directory on the remote machine, so we can install from there cluster.run([f'''pip install -r transformers/examples/{example_dir}/requirements.txt''']) cluster.run(["""pip install torch --upgrade --extra-index-url https://download.pytorch.org/whl/cu117"""]) # Run example. You can bypass the CLI wrapper and paste your own code here. cluster.run([f'''python transformers/examples/{args.example} {' '.join(shlex.quote(arg) for arg in unknown)}''']) # Alternatively, we can just import and run a training function (especially if there's no wrapper CLI): # from my_script... import train # reqs = ['pip:./', 'torch', 'datasets', 'accelerate', 'evaluate', 'tqdm', 'scipy', 'scikit-learn', 'tensorboard'] # launch_train_gpu = rh.function(fn=train, # system=gpu, # reqs=reqs, # name='train_bert_glue') # # We can pass in arguments just like we would to a function: # launch_train_gpu(num_epochs = 3, lr = 2e-5, seed = 42, batch_size = 16 # stream_logs=True)
512
"""simple docstring""" class lowerCAmelCase__ : '''simple docstring''' def __init__( self : int , lowercase_ : List[str] , lowercase_ : str , lowercase_ : Tuple): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[int] = name SCREAMING_SNAKE_CASE_ : List[Any] = value SCREAMING_SNAKE_CASE_ : Optional[Any] = weight def __repr__( self : str): '''simple docstring''' return F'{self.__class__.__name__}({self.name}, {self.value}, {self.weight})' def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): '''simple docstring''' return self.value def _SCREAMING_SNAKE_CASE ( self : Tuple): '''simple docstring''' return self.name def _SCREAMING_SNAKE_CASE ( self : List[str]): '''simple docstring''' return self.weight def _SCREAMING_SNAKE_CASE ( self : str): '''simple docstring''' return self.value / self.weight def _A (__a , __a , __a ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE_ : Dict = [] for i in range(len(__a ) ): menu.append(Things(name[i] , value[i] , weight[i] ) ) return menu def _A (__a , __a , __a ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = sorted(__a , key=__a , reverse=__a ) SCREAMING_SNAKE_CASE_ : Tuple = [] SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : int = 0.0, 0.0 for i in range(len(__a ) ): if (total_cost + items_copy[i].get_weight()) <= max_cost: result.append(items_copy[i] ) total_cost += items_copy[i].get_weight() total_value += items_copy[i].get_value() return (result, total_value) def _A () -> Optional[Any]: """simple docstring""" if __name__ == "__main__": import doctest doctest.testmod()
512
1
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def _lowerCamelCase ( _a ): """simple docstring""" _lowerCamelCase = 3_8_4 if "tiny" in model_name: _lowerCamelCase = [3, 3, 9, 3] _lowerCamelCase = [9_6, 1_9_2, 3_8_4, 7_6_8] if "small" in model_name: _lowerCamelCase = [3, 3, 2_7, 3] _lowerCamelCase = [9_6, 1_9_2, 3_8_4, 7_6_8] if "base" in model_name: _lowerCamelCase = [3, 3, 2_7, 3] _lowerCamelCase = [1_2_8, 2_5_6, 5_1_2, 1_0_2_4] _lowerCamelCase = 5_1_2 if "large" in model_name: _lowerCamelCase = [3, 3, 2_7, 3] _lowerCamelCase = [1_9_2, 3_8_4, 7_6_8, 1_5_3_6] _lowerCamelCase = 7_6_8 if "xlarge" in model_name: _lowerCamelCase = [3, 3, 2_7, 3] _lowerCamelCase = [2_5_6, 5_1_2, 1_0_2_4, 2_0_4_8] _lowerCamelCase = 1_0_2_4 # set label information _lowerCamelCase = 1_5_0 _lowerCamelCase = '''huggingface/label-files''' _lowerCamelCase = '''ade20k-id2label.json''' _lowerCamelCase = json.load(open(hf_hub_download(snake_case__ , snake_case__ , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(snake_case__ ): v for k, v in idalabel.items()} _lowerCamelCase = {v: k for k, v in idalabel.items()} _lowerCamelCase = ConvNextConfig( depths=snake_case__ , hidden_sizes=snake_case__ , out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] ) _lowerCamelCase = UperNetConfig( backbone_config=snake_case__ , auxiliary_in_channels=snake_case__ , num_labels=snake_case__ , idalabel=snake_case__ , labelaid=snake_case__ , ) return config def _lowerCamelCase ( _a ): """simple docstring""" _lowerCamelCase = [] # fmt: off # stem rename_keys.append(('''backbone.downsample_layers.0.0.weight''', '''backbone.embeddings.patch_embeddings.weight''') ) rename_keys.append(('''backbone.downsample_layers.0.0.bias''', '''backbone.embeddings.patch_embeddings.bias''') ) rename_keys.append(('''backbone.downsample_layers.0.1.weight''', '''backbone.embeddings.layernorm.weight''') ) rename_keys.append(('''backbone.downsample_layers.0.1.bias''', '''backbone.embeddings.layernorm.bias''') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F'''backbone.stages.{i}.{j}.gamma''', F'''backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter''') ) rename_keys.append((F'''backbone.stages.{i}.{j}.depthwise_conv.weight''', F'''backbone.encoder.stages.{i}.layers.{j}.dwconv.weight''') ) rename_keys.append((F'''backbone.stages.{i}.{j}.depthwise_conv.bias''', F'''backbone.encoder.stages.{i}.layers.{j}.dwconv.bias''') ) rename_keys.append((F'''backbone.stages.{i}.{j}.norm.weight''', F'''backbone.encoder.stages.{i}.layers.{j}.layernorm.weight''') ) rename_keys.append((F'''backbone.stages.{i}.{j}.norm.bias''', F'''backbone.encoder.stages.{i}.layers.{j}.layernorm.bias''') ) rename_keys.append((F'''backbone.stages.{i}.{j}.pointwise_conv1.weight''', F'''backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight''') ) rename_keys.append((F'''backbone.stages.{i}.{j}.pointwise_conv1.bias''', F'''backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias''') ) rename_keys.append((F'''backbone.stages.{i}.{j}.pointwise_conv2.weight''', F'''backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight''') ) rename_keys.append((F'''backbone.stages.{i}.{j}.pointwise_conv2.bias''', F'''backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias''') ) if i > 0: rename_keys.append((F'''backbone.downsample_layers.{i}.0.weight''', F'''backbone.encoder.stages.{i}.downsampling_layer.0.weight''') ) rename_keys.append((F'''backbone.downsample_layers.{i}.0.bias''', F'''backbone.encoder.stages.{i}.downsampling_layer.0.bias''') ) rename_keys.append((F'''backbone.downsample_layers.{i}.1.weight''', F'''backbone.encoder.stages.{i}.downsampling_layer.1.weight''') ) rename_keys.append((F'''backbone.downsample_layers.{i}.1.bias''', F'''backbone.encoder.stages.{i}.downsampling_layer.1.bias''') ) rename_keys.append((F'''backbone.norm{i}.weight''', F'''backbone.hidden_states_norms.stage{i+1}.weight''') ) rename_keys.append((F'''backbone.norm{i}.bias''', F'''backbone.hidden_states_norms.stage{i+1}.bias''') ) # decode head rename_keys.extend( [ ('''decode_head.conv_seg.weight''', '''decode_head.classifier.weight'''), ('''decode_head.conv_seg.bias''', '''decode_head.classifier.bias'''), ('''auxiliary_head.conv_seg.weight''', '''auxiliary_head.classifier.weight'''), ('''auxiliary_head.conv_seg.bias''', '''auxiliary_head.classifier.bias'''), ] ) # fmt: on return rename_keys def _lowerCamelCase ( _a , _a , _a ): """simple docstring""" _lowerCamelCase = dct.pop(snake_case__ ) _lowerCamelCase = val def _lowerCamelCase ( _a , _a , _a ): """simple docstring""" _lowerCamelCase = { '''upernet-convnext-tiny''': '''https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth''', '''upernet-convnext-small''': '''https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth''', '''upernet-convnext-base''': '''https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth''', '''upernet-convnext-large''': '''https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth''', '''upernet-convnext-xlarge''': '''https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth''', } _lowerCamelCase = model_name_to_url[model_name] _lowerCamelCase = torch.hub.load_state_dict_from_url(snake_case__ , map_location='''cpu''' )['''state_dict'''] _lowerCamelCase = get_upernet_config(snake_case__ ) _lowerCamelCase = UperNetForSemanticSegmentation(snake_case__ ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): _lowerCamelCase = state_dict.pop(snake_case__ ) if "bn" in key: _lowerCamelCase = key.replace('''bn''' , '''batch_norm''' ) _lowerCamelCase = val # rename keys _lowerCamelCase = create_rename_keys(snake_case__ ) for src, dest in rename_keys: rename_key(snake_case__ , snake_case__ , snake_case__ ) model.load_state_dict(snake_case__ ) # verify on image _lowerCamelCase = '''https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg''' _lowerCamelCase = Image.open(requests.get(snake_case__ , stream=snake_case__ ).raw ).convert('''RGB''' ) _lowerCamelCase = SegformerImageProcessor() _lowerCamelCase = processor(snake_case__ , return_tensors='''pt''' ).pixel_values with torch.no_grad(): _lowerCamelCase = model(snake_case__ ) if model_name == "upernet-convnext-tiny": _lowerCamelCase = torch.tensor( [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] ) elif model_name == "upernet-convnext-small": _lowerCamelCase = torch.tensor( [[-8.8236, -8.8236, -8.6771], [-8.8236, -8.8236, -8.6771], [-8.7638, -8.7638, -8.6240]] ) elif model_name == "upernet-convnext-base": _lowerCamelCase = torch.tensor( [[-8.8558, -8.8558, -8.6905], [-8.8558, -8.8558, -8.6905], [-8.7669, -8.7669, -8.6021]] ) elif model_name == "upernet-convnext-large": _lowerCamelCase = torch.tensor( [[-8.6660, -8.6660, -8.6210], [-8.6660, -8.6660, -8.6210], [-8.6310, -8.6310, -8.5964]] ) elif model_name == "upernet-convnext-xlarge": _lowerCamelCase = torch.tensor( [[-8.4980, -8.4980, -8.3977], [-8.4980, -8.4980, -8.3977], [-8.4379, -8.4379, -8.3412]] ) print('''Logits:''' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , snake_case__ , atol=1E-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(snake_case__ ) print(F'''Saving processor to {pytorch_dump_folder_path}''' ) processor.save_pretrained(snake_case__ ) if push_to_hub: print(F'''Pushing model and processor for {model_name} to hub''' ) model.push_to_hub(F'''openmmlab/{model_name}''' ) processor.push_to_hub(F'''openmmlab/{model_name}''' ) if __name__ == "__main__": _UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="upernet-convnext-tiny", type=str, choices=[F'upernet-convnext-{size}' for size in ["tiny", "small", "base", "large", "xlarge"]], help="Name of the ConvNext UperNet model you\'d like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) _UpperCAmelCase = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
700
def _lowerCamelCase ( _a , _a ): """simple docstring""" return base * power(_a , (exponent - 1) ) if exponent else 1 if __name__ == "__main__": print("Raise base to the power of exponent using recursion...") _UpperCAmelCase = int(input("Enter the base: ").strip()) _UpperCAmelCase = int(input("Enter the exponent: ").strip()) _UpperCAmelCase = power(base, abs(exponent)) if exponent < 0: # power() does not properly deal w/ negative exponents _UpperCAmelCase = 1 / result print(F'{base} to the power of {exponent} is {result}')
297
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE__:Optional[Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE__:Tuple = { """SCUT-DLVCLab/lilt-roberta-en-base""": ( """https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json""" ), } class snake_case__ ( snake_case_ ): _snake_case : Union[str, Any] = """lilt""" def __init__( self , lowerCamelCase=30522 , lowerCamelCase=768 , lowerCamelCase=12 , lowerCamelCase=12 , lowerCamelCase=3072 , lowerCamelCase="gelu" , lowerCamelCase=0.1 , lowerCamelCase=0.1 , lowerCamelCase=512 , lowerCamelCase=2 , lowerCamelCase=0.02 , lowerCamelCase=1E-12 , lowerCamelCase=0 , lowerCamelCase="absolute" , lowerCamelCase=None , lowerCamelCase=4 , lowerCamelCase=1024 , **lowerCamelCase , ): super().__init__(pad_token_id=lowerCamelCase , **lowerCamelCase ) __a = vocab_size __a = hidden_size __a = num_hidden_layers __a = num_attention_heads __a = hidden_act __a = intermediate_size __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = max_position_embeddings __a = type_vocab_size __a = initializer_range __a = layer_norm_eps __a = position_embedding_type __a = classifier_dropout __a = channel_shrink_ratio __a = max_ad_position_embeddings
528
"""simple docstring""" def _lowerCamelCase( a , a ): __a = 0 __a = len(a ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __a = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(a ): return None __a = sorted_collection[point] if current_item == item: return point else: if point < left: __a = left __a = point elif point > right: __a = right __a = point else: if item < current_item: __a = point - 1 else: __a = point + 1 return None def _lowerCamelCase( a , a , a , a ): # avoid divided by 0 during interpolation if sorted_collection[left] == sorted_collection[right]: if sorted_collection[left] == item: return left else: return None __a = left + ((item - sorted_collection[left]) * (right - left)) // ( sorted_collection[right] - sorted_collection[left] ) # out of range check if point < 0 or point >= len(a ): return None if sorted_collection[point] == item: return point elif point < left: return interpolation_search_by_recursion(a , a , a , a ) elif point > right: return interpolation_search_by_recursion(a , a , a , a ) else: if sorted_collection[point] > item: return interpolation_search_by_recursion( a , a , a , point - 1 ) else: return interpolation_search_by_recursion( a , a , point + 1 , a ) def _lowerCamelCase( a ): if collection != sorted(a ): raise ValueError("Collection must be ascending sorted" ) return True if __name__ == "__main__": import sys SCREAMING_SNAKE_CASE__:List[str] = 0 if debug == 1: SCREAMING_SNAKE_CASE__:Tuple = [10, 30, 40, 45, 50, 66, 77, 93] try: __assert_sorted(collection) except ValueError: sys.exit("""Sequence must be ascending sorted to apply interpolation search""") SCREAMING_SNAKE_CASE__:int = 67 SCREAMING_SNAKE_CASE__:List[str] = interpolation_search(collection, target) if result is not None: print(F'''{target} found at positions: {result}''') else: print("""Not found""")
528
1
import warnings from ...utils import logging from .image_processing_imagegpt import ImageGPTImageProcessor lowerCAmelCase__ = logging.get_logger(__name__) class a__ ( snake_case ): """simple docstring""" def __init__( self , *lowercase , **lowercase ) -> None: '''simple docstring''' warnings.warn( "The class ImageGPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers." " Please use ImageGPTImageProcessor instead." , lowercase , ) super().__init__(*lowercase , **lowercase )
626
from math import factorial def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: int = 1_0_0 ) -> int: '''simple docstring''' return sum(map(SCREAMING_SNAKE_CASE_ , str(factorial(SCREAMING_SNAKE_CASE_ ) ) ) ) if __name__ == "__main__": print(solution(int(input("""Enter the Number: """).strip())))
626
1
from __future__ import annotations from typing import Any class __UpperCAmelCase : """simple docstring""" def __init__( self , __A , __A , __A = 0 ): __a , __a = row, column __a = [[default_value for c in range(__A )] for r in range(__A )] def __str__( self ): __a = f'''Matrix consist of {self.row} rows and {self.column} columns\n''' # Make string identifier __a = 0 for row_vector in self.array: for obj in row_vector: __a = max(__A , len(str(__A ) ) ) __a = f'''%{max_element_length}s''' # Make string and return def single_line(__A ) -> str: nonlocal string_format_identifier __a = """[""" line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(__A ) for row_vector in self.array ) return s def __repr__( self ): return str(self ) def snake_case_ ( self , __A ): if not (isinstance(__A , (list, tuple) ) and len(__A ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self , __A ): assert self.validate_indicies(__A ) return self.array[loc[0]][loc[1]] def __setitem__( self , __A , __A ): assert self.validate_indicies(__A ) __a = value def __add__( self , __A ): assert isinstance(__A , __A ) assert self.row == another.row and self.column == another.column # Add __a = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __a = self[r, c] + another[r, c] return result def __neg__( self ): __a = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __a = -self[r, c] return result def __sub__( self , __A ): return self + (-another) def __mul__( self , __A ): if isinstance(__A , (int, float) ): # Scalar multiplication __a = Matrix(self.row , self.column ) for r in range(self.row ): for c in range(self.column ): __a = self[r, c] * another return result elif isinstance(__A , __A ): # Matrix multiplication assert self.column == another.row __a = Matrix(self.row , another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __a = f'''Unsupported type given for another ({type(__A )})''' raise TypeError(__A ) def snake_case_ ( self ): __a = Matrix(self.column , self.row ) for r in range(self.row ): for c in range(self.column ): __a = self[r, c] return result def snake_case_ ( self , __A , __A ): assert isinstance(__A , __A ) and isinstance(__A , __A ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __a = v.transpose() __a = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def a (): # a^(-1) __a = Matrix(3 , 3 , 0 ) for i in range(3 ): __a = 1 print(f'''a^(-1) is {ainv}''' ) # u, v __a = Matrix(3 , 1 , 0 ) __a , __a , __a = 1, 2, -3 __a = Matrix(3 , 1 , 0 ) __a , __a , __a = 4, -2, 5 print(f'''u is {u}''' ) print(f'''v is {v}''' ) print(f'''uv^T is {u * v.transpose()}''' ) # Sherman Morrison print(f'''(a + uv^T)^(-1) is {ainv.sherman_morrison(lowerCAmelCase__ , lowerCAmelCase__ )}''' ) def a (): import doctest doctest.testmod() testa()
99
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class __UpperCAmelCase ( unittest.TestCase ): """simple docstring""" def __init__( self , __A , __A=13 , __A=3 , __A=224 , __A=30 , __A=400 , __A=True , __A=None , __A=True , __A=[0.5, 0.5, 0.5] , __A=[0.5, 0.5, 0.5] , ): __a = size if size is not None else {"""height""": 18, """width""": 18} __a = parent __a = batch_size __a = num_channels __a = image_size __a = min_resolution __a = max_resolution __a = do_resize __a = size __a = do_normalize __a = image_mean __a = image_std def snake_case_ ( self ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class __UpperCAmelCase ( __A , unittest.TestCase ): """simple docstring""" _lowerCamelCase = ViTImageProcessor if is_vision_available() else None def snake_case_ ( self ): __a = EfficientFormerImageProcessorTester(self ) @property def snake_case_ ( self ): return self.image_proc_tester.prepare_image_processor_dict() def snake_case_ ( self ): __a = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__A , """image_mean""" ) ) self.assertTrue(hasattr(__A , """image_std""" ) ) self.assertTrue(hasattr(__A , """do_normalize""" ) ) self.assertTrue(hasattr(__A , """do_resize""" ) ) self.assertTrue(hasattr(__A , """size""" ) ) def snake_case_ ( self ): pass def snake_case_ ( self ): # Initialize image_processor __a = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __a = prepare_image_inputs(self.image_proc_tester , equal_resolution=__A ) for image in image_inputs: self.assertIsInstance(__A , Image.Image ) # Test not batched input __a = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched __a = image_processor(__A , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def snake_case_ ( self ): # Initialize image_processor __a = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __a = prepare_image_inputs(self.image_proc_tester , equal_resolution=__A , numpify=__A ) for image in image_inputs: self.assertIsInstance(__A , np.ndarray ) # Test not batched input __a = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched __a = image_processor(__A , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) def snake_case_ ( self ): # Initialize image_processor __a = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __a = prepare_image_inputs(self.image_proc_tester , equal_resolution=__A , torchify=__A ) for image in image_inputs: self.assertIsInstance(__A , torch.Tensor ) # Test not batched input __a = image_processor(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , ) # Test batched __a = image_processor(__A , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["""height"""], self.image_proc_tester.size["""width"""], ) , )
99
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, is_vision_available, ) lowercase_ = {"configuration_vit": ["VIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "ViTConfig", "ViTOnnxConfig"]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = ["ViTFeatureExtractor"] lowercase_ = ["ViTImageProcessor"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ "VIT_PRETRAINED_MODEL_ARCHIVE_LIST", "ViTForImageClassification", "ViTForMaskedImageModeling", "ViTModel", "ViTPreTrainedModel", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ "TFViTForImageClassification", "TFViTModel", "TFViTPreTrainedModel", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ "FlaxViTForImageClassification", "FlaxViTModel", "FlaxViTPreTrainedModel", ] if TYPE_CHECKING: from .configuration_vit import VIT_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTConfig, ViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_vit import ViTFeatureExtractor from .image_processing_vit import ViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit import ( VIT_PRETRAINED_MODEL_ARCHIVE_LIST, ViTForImageClassification, ViTForMaskedImageModeling, ViTModel, ViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit import TFViTForImageClassification, TFViTModel, TFViTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel, FlaxViTPreTrainedModel else: import sys lowercase_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
717
import argparse import json from dataclasses import dataclass, field from functools import partial from pathlib import Path from typing import Callable, Dict, List, Tuple import timm import torch import torch.nn as nn from classy_vision.models.regnet import RegNet, RegNetParams, RegNetYaagf, RegNetYaagf, RegNetYaaagf from huggingface_hub import cached_download, hf_hub_url from torch import Tensor from vissl.models.model_helpers import get_trunk_forward_outputs from transformers import AutoImageProcessor, RegNetConfig, RegNetForImageClassification, RegNetModel from transformers.utils import logging logging.set_verbosity_info() lowercase_ = logging.get_logger() @dataclass class A : """simple docstring""" lowerCamelCase = 42 lowerCamelCase = field(default_factory=_UpperCAmelCase ) lowerCamelCase = field(default_factory=_UpperCAmelCase ) def snake_case__ ( self : Union[str, Any],lowercase_ : Dict,lowercase_ : Tensor,lowercase_ : Tensor )-> Tuple: '''simple docstring''' A__ = len(list(m.modules() ) ) == 1 or isinstance(lowercase_,nn.Convad ) or isinstance(lowercase_,nn.BatchNormad ) if has_not_submodules: self.traced.append(lowercase_ ) def __call__( self : Tuple,lowercase_ : Tensor )-> Any: '''simple docstring''' for m in self.module.modules(): self.handles.append(m.register_forward_hook(self._forward_hook ) ) self.module(lowercase_ ) [x.remove() for x in self.handles] return self @property def snake_case__ ( self : Optional[int] )-> Optional[int]: '''simple docstring''' return list(filter(lambda lowercase_ : len(list(x.state_dict().keys() ) ) > 0,self.traced ) ) @dataclass class A : """simple docstring""" lowerCamelCase = 42 lowerCamelCase = 42 lowerCamelCase = 1 lowerCamelCase = field(default_factory=_UpperCAmelCase ) lowerCamelCase = field(default_factory=_UpperCAmelCase ) lowerCamelCase = True def __call__( self : str,lowercase_ : Tensor )-> Dict: '''simple docstring''' A__ = Tracker(self.dest )(lowercase_ ).parametrized A__ = Tracker(self.src )(lowercase_ ).parametrized A__ = list(filter(lambda lowercase_ : type(lowercase_ ) not in self.src_skip,lowercase_ ) ) A__ = list(filter(lambda lowercase_ : type(lowercase_ ) not in self.dest_skip,lowercase_ ) ) if len(lowercase_ ) != len(lowercase_ ) and self.raise_if_mismatch: raise Exception( F'Numbers of operations are different. Source module has {len(lowercase_ )} operations while' F' destination module has {len(lowercase_ )}.' ) for dest_m, src_m in zip(lowercase_,lowercase_ ): dest_m.load_state_dict(src_m.state_dict() ) if self.verbose == 1: print(F'Transfered from={src_m} to={dest_m}' ) class A ( nn.Module ): """simple docstring""" def __init__( self : Any,lowercase_ : nn.Module )-> int: '''simple docstring''' super().__init__() A__ = [] # - get the stem feature_blocks.append(('conv1', model.stem) ) # - get all the feature blocks for k, v in model.trunk_output.named_children(): assert k.startswith('block' ), F'Unexpected layer name {k}' A__ = len(lowercase_ ) + 1 feature_blocks.append((F'res{block_index}', v) ) A__ = nn.ModuleDict(lowercase_ ) def snake_case__ ( self : List[Any],lowercase_ : Tensor )-> Any: '''simple docstring''' return get_trunk_forward_outputs( lowercase_,out_feat_keys=lowercase_,feature_blocks=self._feature_blocks,) class A ( _UpperCAmelCase ): """simple docstring""" def snake_case__ ( self : List[Any],lowercase_ : str )-> str: '''simple docstring''' A__ = x.split('-' ) return x_split[0] + x_split[1] + "_" + "".join(x_split[2:] ) def __getitem__( self : Optional[Any],lowercase_ : str )-> Callable[[], Tuple[nn.Module, Dict]]: '''simple docstring''' if x not in self: A__ = self.convert_name_to_timm(lowercase_ ) A__ = partial(lambda: (timm.create_model(lowercase_,pretrained=lowercase_ ).eval(), None) ) else: A__ = super().__getitem__(lowercase_ ) return val class A ( _UpperCAmelCase ): """simple docstring""" def __getitem__( self : Tuple,lowercase_ : str )-> Callable[[], nn.Module]: '''simple docstring''' if "seer" in x and "in1k" not in x: A__ = RegNetModel else: A__ = RegNetForImageClassification return val def _snake_case( SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : List[Tuple[str, str]] ) -> Dict: '''simple docstring''' for from_key, to_key in keys: A__ = from_state_dict[from_key].clone() print(f'Copied key={from_key} to={to_key}' ) return to_state_dict def _snake_case( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Callable[[], nn.Module] , SCREAMING_SNAKE_CASE__ : Callable[[], nn.Module] , SCREAMING_SNAKE_CASE__ : RegNetConfig , SCREAMING_SNAKE_CASE__ : Path , SCREAMING_SNAKE_CASE__ : bool = True , ) -> Any: '''simple docstring''' print(f'Converting {name}...' ) with torch.no_grad(): A__ , A__ = from_model_func() A__ = our_model_func(SCREAMING_SNAKE_CASE__ ).eval() A__ = ModuleTransfer(src=SCREAMING_SNAKE_CASE__ , dest=SCREAMING_SNAKE_CASE__ , raise_if_mismatch=SCREAMING_SNAKE_CASE__ ) A__ = torch.randn((1, 3, 224, 224) ) module_transfer(SCREAMING_SNAKE_CASE__ ) if from_state_dict is not None: A__ = [] # for seer - in1k finetuned we have to manually copy the head if "seer" in name and "in1k" in name: A__ = [('0.clf.0.weight', 'classifier.1.weight'), ('0.clf.0.bias', 'classifier.1.bias')] A__ = manually_copy_vissl_head(SCREAMING_SNAKE_CASE__ , our_model.state_dict() , SCREAMING_SNAKE_CASE__ ) our_model.load_state_dict(SCREAMING_SNAKE_CASE__ ) A__ = our_model(SCREAMING_SNAKE_CASE__ , output_hidden_states=SCREAMING_SNAKE_CASE__ ) A__ = ( our_outputs.logits if isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) else our_outputs.last_hidden_state ) A__ = from_model(SCREAMING_SNAKE_CASE__ ) A__ = from_output[-1] if type(SCREAMING_SNAKE_CASE__ ) is list else from_output # now since I don't want to use any config files, vissl seer model doesn't actually have an head, so let's just check the last hidden state if "seer" in name and "in1k" in name: A__ = our_outputs.hidden_states[-1] assert torch.allclose(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ), "The model logits don't match the original one." if push_to_hub: our_model.push_to_hub( repo_path_or_name=save_directory / name , commit_message='Add model' , use_temp_dir=SCREAMING_SNAKE_CASE__ , ) A__ = 224 if 'seer' not in name else 384 # we can use the convnext one A__ = AutoImageProcessor.from_pretrained('facebook/convnext-base-224-22k-1k' , size=SCREAMING_SNAKE_CASE__ ) image_processor.push_to_hub( repo_path_or_name=save_directory / name , commit_message='Add image processor' , use_temp_dir=SCREAMING_SNAKE_CASE__ , ) print(f'Pushed {name}' ) def _snake_case( SCREAMING_SNAKE_CASE__ : Path , SCREAMING_SNAKE_CASE__ : str = None , SCREAMING_SNAKE_CASE__ : bool = True ) -> List[Any]: '''simple docstring''' A__ = 'imagenet-1k-id2label.json' A__ = 1000 A__ = (1, num_labels) A__ = 'huggingface/label-files' A__ = num_labels A__ = json.load(open(cached_download(hf_hub_url(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , repo_type='dataset' ) ) , 'r' ) ) A__ = {int(SCREAMING_SNAKE_CASE__ ): v for k, v in idalabel.items()} A__ = idalabel A__ = {v: k for k, v in idalabel.items()} A__ = partial(SCREAMING_SNAKE_CASE__ , num_labels=SCREAMING_SNAKE_CASE__ , idalabel=SCREAMING_SNAKE_CASE__ , labelaid=SCREAMING_SNAKE_CASE__ ) A__ = { 'regnet-x-002': ImageNetPreTrainedConfig( depths=[1, 1, 4, 7] , hidden_sizes=[24, 56, 152, 368] , groups_width=8 , layer_type='x' ), 'regnet-x-004': ImageNetPreTrainedConfig( depths=[1, 2, 7, 12] , hidden_sizes=[32, 64, 160, 384] , groups_width=16 , layer_type='x' ), 'regnet-x-006': ImageNetPreTrainedConfig( depths=[1, 3, 5, 7] , hidden_sizes=[48, 96, 240, 528] , groups_width=24 , layer_type='x' ), 'regnet-x-008': ImageNetPreTrainedConfig( depths=[1, 3, 7, 5] , hidden_sizes=[64, 128, 288, 672] , groups_width=16 , layer_type='x' ), 'regnet-x-016': ImageNetPreTrainedConfig( depths=[2, 4, 10, 2] , hidden_sizes=[72, 168, 408, 912] , groups_width=24 , layer_type='x' ), 'regnet-x-032': ImageNetPreTrainedConfig( depths=[2, 6, 15, 2] , hidden_sizes=[96, 192, 432, 1008] , groups_width=48 , layer_type='x' ), 'regnet-x-040': ImageNetPreTrainedConfig( depths=[2, 5, 14, 2] , hidden_sizes=[80, 240, 560, 1360] , groups_width=40 , layer_type='x' ), 'regnet-x-064': ImageNetPreTrainedConfig( depths=[2, 4, 10, 1] , hidden_sizes=[168, 392, 784, 1624] , groups_width=56 , layer_type='x' ), 'regnet-x-080': ImageNetPreTrainedConfig( depths=[2, 5, 15, 1] , hidden_sizes=[80, 240, 720, 1920] , groups_width=120 , layer_type='x' ), 'regnet-x-120': ImageNetPreTrainedConfig( depths=[2, 5, 11, 1] , hidden_sizes=[224, 448, 896, 2240] , groups_width=112 , layer_type='x' ), 'regnet-x-160': ImageNetPreTrainedConfig( depths=[2, 6, 13, 1] , hidden_sizes=[256, 512, 896, 2048] , groups_width=128 , layer_type='x' ), 'regnet-x-320': ImageNetPreTrainedConfig( depths=[2, 7, 13, 1] , hidden_sizes=[336, 672, 1344, 2520] , groups_width=168 , layer_type='x' ), # y variant 'regnet-y-002': ImageNetPreTrainedConfig(depths=[1, 1, 4, 7] , hidden_sizes=[24, 56, 152, 368] , groups_width=8 ), 'regnet-y-004': ImageNetPreTrainedConfig( depths=[1, 3, 6, 6] , hidden_sizes=[48, 104, 208, 440] , groups_width=8 ), 'regnet-y-006': ImageNetPreTrainedConfig( depths=[1, 3, 7, 4] , hidden_sizes=[48, 112, 256, 608] , groups_width=16 ), 'regnet-y-008': ImageNetPreTrainedConfig( depths=[1, 3, 8, 2] , hidden_sizes=[64, 128, 320, 768] , groups_width=16 ), 'regnet-y-016': ImageNetPreTrainedConfig( depths=[2, 6, 17, 2] , hidden_sizes=[48, 120, 336, 888] , groups_width=24 ), 'regnet-y-032': ImageNetPreTrainedConfig( depths=[2, 5, 13, 1] , hidden_sizes=[72, 216, 576, 1512] , groups_width=24 ), 'regnet-y-040': ImageNetPreTrainedConfig( depths=[2, 6, 12, 2] , hidden_sizes=[128, 192, 512, 1088] , groups_width=64 ), 'regnet-y-064': ImageNetPreTrainedConfig( depths=[2, 7, 14, 2] , hidden_sizes=[144, 288, 576, 1296] , groups_width=72 ), 'regnet-y-080': ImageNetPreTrainedConfig( depths=[2, 4, 10, 1] , hidden_sizes=[168, 448, 896, 2016] , groups_width=56 ), 'regnet-y-120': ImageNetPreTrainedConfig( depths=[2, 5, 11, 1] , hidden_sizes=[224, 448, 896, 2240] , groups_width=112 ), 'regnet-y-160': ImageNetPreTrainedConfig( depths=[2, 4, 11, 1] , hidden_sizes=[224, 448, 1232, 3024] , groups_width=112 ), 'regnet-y-320': ImageNetPreTrainedConfig( depths=[2, 5, 12, 1] , hidden_sizes=[232, 696, 1392, 3712] , groups_width=232 ), # models created by SEER -> https://arxiv.org/abs/2202.08360 'regnet-y-320-seer': RegNetConfig(depths=[2, 5, 12, 1] , hidden_sizes=[232, 696, 1392, 3712] , groups_width=232 ), 'regnet-y-640-seer': RegNetConfig(depths=[2, 5, 12, 1] , hidden_sizes=[328, 984, 1968, 4920] , groups_width=328 ), 'regnet-y-1280-seer': RegNetConfig( depths=[2, 7, 17, 1] , hidden_sizes=[528, 1056, 2904, 7392] , groups_width=264 ), 'regnet-y-2560-seer': RegNetConfig( depths=[3, 7, 16, 1] , hidden_sizes=[640, 1696, 2544, 5088] , groups_width=640 ), 'regnet-y-10b-seer': ImageNetPreTrainedConfig( depths=[2, 7, 17, 1] , hidden_sizes=[2020, 4040, 11110, 28280] , groups_width=1010 ), # finetuned on imagenet 'regnet-y-320-seer-in1k': ImageNetPreTrainedConfig( depths=[2, 5, 12, 1] , hidden_sizes=[232, 696, 1392, 3712] , groups_width=232 ), 'regnet-y-640-seer-in1k': ImageNetPreTrainedConfig( depths=[2, 5, 12, 1] , hidden_sizes=[328, 984, 1968, 4920] , groups_width=328 ), 'regnet-y-1280-seer-in1k': ImageNetPreTrainedConfig( depths=[2, 7, 17, 1] , hidden_sizes=[528, 1056, 2904, 7392] , groups_width=264 ), 'regnet-y-2560-seer-in1k': ImageNetPreTrainedConfig( depths=[3, 7, 16, 1] , hidden_sizes=[640, 1696, 2544, 5088] , groups_width=640 ), 'regnet-y-10b-seer-in1k': ImageNetPreTrainedConfig( depths=[2, 7, 17, 1] , hidden_sizes=[2020, 4040, 11110, 28280] , groups_width=1010 ), } A__ = NameToOurModelFuncMap() A__ = NameToFromModelFuncMap() # add seer weights logic def load_using_classy_vision(SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Callable[[], nn.Module] ) -> Tuple[nn.Module, Dict]: A__ = torch.hub.load_state_dict_from_url(SCREAMING_SNAKE_CASE__ , model_dir=str(SCREAMING_SNAKE_CASE__ ) , map_location='cpu' ) A__ = model_func() # check if we have a head, if yes add it A__ = files['classy_state_dict']['base_model']['model'] A__ = model_state_dict['trunk'] model.load_state_dict(SCREAMING_SNAKE_CASE__ ) return model.eval(), model_state_dict["heads"] # pretrained A__ = partial( SCREAMING_SNAKE_CASE__ , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet32d/seer_regnet32gf_model_iteration244000.torch' , lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) , ) A__ = partial( SCREAMING_SNAKE_CASE__ , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet64/seer_regnet64gf_model_final_checkpoint_phase0.torch' , lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) , ) A__ = partial( SCREAMING_SNAKE_CASE__ , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/swav_ig1b_regnet128Gf_cnstant_bs32_node16_sinkhorn10_proto16k_syncBN64_warmup8k/model_final_checkpoint_phase0.torch' , lambda: FakeRegNetVisslWrapper(RegNetYaaagf() ) , ) A__ = partial( SCREAMING_SNAKE_CASE__ , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_regnet10B/model_iteration124500_conso.torch' , lambda: FakeRegNetVisslWrapper( RegNet(RegNetParams(depth=27 , group_width=1010 , w_a=1744 , w_a=620.83 , w_m=2.52 ) ) ) , ) # IN1K finetuned A__ = partial( SCREAMING_SNAKE_CASE__ , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet32_finetuned_in1k_model_final_checkpoint_phase78.torch' , lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) , ) A__ = partial( SCREAMING_SNAKE_CASE__ , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet64_finetuned_in1k_model_final_checkpoint_phase78.torch' , lambda: FakeRegNetVisslWrapper(RegNetYaagf() ) , ) A__ = partial( SCREAMING_SNAKE_CASE__ , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_regnet128_finetuned_in1k_model_final_checkpoint_phase78.torch' , lambda: FakeRegNetVisslWrapper(RegNetYaaagf() ) , ) A__ = partial( SCREAMING_SNAKE_CASE__ , 'https://dl.fbaipublicfiles.com/vissl/model_zoo/seer_finetuned/seer_10b_finetuned_in1k_model_phase28_conso.torch' , lambda: FakeRegNetVisslWrapper( RegNet(RegNetParams(depth=27 , group_width=1010 , w_a=1744 , w_a=620.83 , w_m=2.52 ) ) ) , ) if model_name: convert_weight_and_push( SCREAMING_SNAKE_CASE__ , names_to_from_model_map[model_name] , names_to_ours_model_map[model_name] , names_to_config[model_name] , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ) else: for model_name, config in names_to_config.items(): convert_weight_and_push( SCREAMING_SNAKE_CASE__ , names_to_from_model_map[model_name] , names_to_ours_model_map[model_name] , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , ) return config, expected_shape if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default=None, type=str, help=( "The name of the model you wish to convert, it must be one of the supported regnet* architecture," " currently: regnetx-*, regnety-*. If `None`, all of them will the converted." ), ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=Path, required=True, help="Path to the output PyTorch model directory.", ) parser.add_argument( "--push_to_hub", default=True, type=bool, required=False, help="If True, push model and image processor to the hub.", ) lowercase_ = parser.parse_args() lowercase_ = args.pytorch_dump_folder_path pytorch_dump_folder_path.mkdir(exist_ok=True, parents=True) convert_weights_and_push(pytorch_dump_folder_path, args.model_name, args.push_to_hub)
586
0
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets _lowerCAmelCase: Union[str, Any] = '\\n@inproceedings{lin-2004-rouge,\n title = "{ROUGE}: A Package for Automatic Evaluation of Summaries",\n author = "Lin, Chin-Yew",\n booktitle = "Text Summarization Branches Out",\n month = jul,\n year = "2004",\n address = "Barcelona, Spain",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/W04-1013",\n pages = "74--81",\n}\n' _lowerCAmelCase: Dict = '\\nROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for\nevaluating automatic summarization and machine translation software in natural language processing.\nThe metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation.\n\nNote that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters.\n\nThis metrics is a wrapper around Google Research reimplementation of ROUGE:\nhttps://github.com/google-research/google-research/tree/master/rouge\n' _lowerCAmelCase: List[Any] = '\nCalculates average rouge scores for a list of hypotheses and references\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n rouge_types: A list of rouge types to calculate.\n Valid names:\n `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring,\n `"rougeL"`: Longest common subsequence based scoring.\n `"rougeLSum"`: rougeLsum splits text using `"\n"`.\n See details in https://github.com/huggingface/datasets/issues/617\n use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes.\n use_aggregator: Return aggregates if this is set to True\nReturns:\n rouge1: rouge_1 (precision, recall, f1),\n rouge2: rouge_2 (precision, recall, f1),\n rougeL: rouge_l (precision, recall, f1),\n rougeLsum: rouge_lsum (precision, recall, f1)\nExamples:\n\n >>> rouge = datasets.load_metric(\'rouge\')\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> results = rouge.compute(predictions=predictions, references=references)\n >>> print(list(results.keys()))\n [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\']\n >>> print(results["rouge1"])\n AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0))\n >>> print(results["rouge1"].mid.fmeasure)\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ (datasets.Metric ): def __UpperCamelCase ( self) -> Optional[int]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence'), 'references': datasets.Value('string' , id='sequence'), }) , codebase_urls=['https://github.com/google-research/google-research/tree/master/rouge'] , reference_urls=[ 'https://en.wikipedia.org/wiki/ROUGE_(metric)', 'https://github.com/google-research/google-research/tree/master/rouge', ] , ) def __UpperCamelCase ( self , lowercase_ , lowercase_ , lowercase_=None , lowercase_=True , lowercase_=False) -> Any: if rouge_types is None: a__ =['rouge1', 'rouge2', 'rougeL', 'rougeLsum'] a__ =rouge_scorer.RougeScorer(rouge_types=lowercase_ , use_stemmer=lowercase_) if use_aggregator: a__ =scoring.BootstrapAggregator() else: a__ =[] for ref, pred in zip(lowercase_ , lowercase_): a__ =scorer.score(lowercase_ , lowercase_) if use_aggregator: aggregator.add_scores(lowercase_) else: scores.append(lowercase_) if use_aggregator: a__ =aggregator.aggregate() else: a__ ={} for key in scores[0]: a__ =[score[key] for score in scores] return result
20
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = { "microsoft/unispeech-large-1500h-cv": ( "https://huggingface.co/microsoft/unispeech-large-1500h-cv/resolve/main/config.json" ), # See all UniSpeech models at https://huggingface.co/models?filter=unispeech } class A (__UpperCAmelCase ): _SCREAMING_SNAKE_CASE = """unispeech""" def __init__( self , lowercase_=32 , lowercase_=768 , lowercase_=12 , lowercase_=12 , lowercase_=3072 , lowercase_="gelu" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=0.02 , lowercase_=1E-5 , lowercase_="group" , lowercase_="gelu" , lowercase_=(512, 512, 512, 512, 512, 512, 512) , lowercase_=(5, 2, 2, 2, 2, 2, 2) , lowercase_=(10, 3, 3, 3, 3, 2, 2) , lowercase_=False , lowercase_=128 , lowercase_=16 , lowercase_=False , lowercase_=True , lowercase_=0.05 , lowercase_=10 , lowercase_=2 , lowercase_=0.0 , lowercase_=10 , lowercase_=0 , lowercase_=320 , lowercase_=2 , lowercase_=0.1 , lowercase_=100 , lowercase_=256 , lowercase_=256 , lowercase_=0.1 , lowercase_="mean" , lowercase_=False , lowercase_=False , lowercase_=256 , lowercase_=80 , lowercase_=0 , lowercase_=1 , lowercase_=2 , lowercase_=0.5 , **lowercase_ , ) -> Tuple: '''simple docstring''' super().__init__(**lowercase_ , pad_token_id=lowercase_ , bos_token_id=lowercase_ , eos_token_id=lowercase_ ) _snake_case : Dict = hidden_size _snake_case : List[Any] = feat_extract_norm _snake_case : Any = feat_extract_activation _snake_case : str = list(lowercase_ ) _snake_case : Any = list(lowercase_ ) _snake_case : Dict = list(lowercase_ ) _snake_case : str = conv_bias _snake_case : Optional[int] = num_conv_pos_embeddings _snake_case : List[str] = num_conv_pos_embedding_groups _snake_case : int = len(self.conv_dim ) _snake_case : str = num_hidden_layers _snake_case : Union[str, Any] = intermediate_size _snake_case : Union[str, Any] = hidden_act _snake_case : int = num_attention_heads _snake_case : List[str] = hidden_dropout _snake_case : Tuple = attention_dropout _snake_case : List[str] = activation_dropout _snake_case : Dict = feat_proj_dropout _snake_case : Any = final_dropout _snake_case : List[Any] = layerdrop _snake_case : Optional[int] = layer_norm_eps _snake_case : Any = initializer_range _snake_case : Tuple = num_ctc_classes _snake_case : Dict = vocab_size _snake_case : List[str] = do_stable_layer_norm _snake_case : List[str] = use_weighted_layer_sum _snake_case : Optional[Any] = classifier_proj_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' F''' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel )}`.''' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _snake_case : Optional[Any] = apply_spec_augment _snake_case : Union[str, Any] = mask_time_prob _snake_case : Union[str, Any] = mask_time_length _snake_case : str = mask_time_min_masks _snake_case : Dict = mask_feature_prob _snake_case : List[str] = mask_feature_length _snake_case : int = mask_feature_min_masks # parameters for pretraining with codevector quantized representations _snake_case : List[Any] = num_codevectors_per_group _snake_case : Any = num_codevector_groups _snake_case : Dict = contrastive_logits_temperature _snake_case : str = feat_quantizer_dropout _snake_case : Optional[int] = num_negatives _snake_case : Optional[int] = codevector_dim _snake_case : List[Any] = proj_codevector_dim _snake_case : List[Any] = diversity_loss_weight # ctc loss _snake_case : Any = ctc_loss_reduction _snake_case : str = ctc_zero_infinity # pretraining loss _snake_case : int = replace_prob @property def __a ( self ) -> int: '''simple docstring''' return functools.reduce(operator.mul , self.conv_stride , 1 )
326
0
from typing import Union import fire import torch from tqdm import tqdm def _A (UpperCamelCase : str , UpperCamelCase : str = "cpu" , UpperCamelCase : Union[str, None] = None ) ->None: '''simple docstring''' lowerCamelCase__ : int = torch.load(UpperCamelCase , map_location=UpperCamelCase ) for k, v in tqdm(state_dict.items() ): if not isinstance(UpperCamelCase , torch.Tensor ): raise TypeError("""FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin""" ) lowerCamelCase__ : List[Any] = v.half() if save_path is None: # overwrite src_path lowerCamelCase__ : Optional[Any] = src_path torch.save(UpperCamelCase , UpperCamelCase ) if __name__ == "__main__": fire.Fire(convert)
96
from string import ascii_uppercase _lowercase = {char: i for i, char in enumerate(ascii_uppercase)} _lowercase = dict(enumerate(ascii_uppercase)) def _A (UpperCamelCase : str , UpperCamelCase : str ) ->str: '''simple docstring''' lowerCamelCase__ : List[str] = len(UpperCamelCase ) lowerCamelCase__ : int = 0 while True: if x == i: lowerCamelCase__ : Union[str, Any] = 0 if len(UpperCamelCase ) == len(UpperCamelCase ): break key += key[i] i += 1 return key def _A (UpperCamelCase : str , UpperCamelCase : str ) ->str: '''simple docstring''' lowerCamelCase__ : int = """""" lowerCamelCase__ : str = 0 for letter in message: if letter == " ": cipher_text += " " else: lowerCamelCase__ : Optional[int] = (dicta[letter] - dicta[key_new[i]]) % 26 i += 1 cipher_text += dicta[x] return cipher_text def _A (UpperCamelCase : str , UpperCamelCase : str ) ->str: '''simple docstring''' lowerCamelCase__ : Dict = """""" lowerCamelCase__ : Tuple = 0 for letter in cipher_text: if letter == " ": or_txt += " " else: lowerCamelCase__ : List[Any] = (dicta[letter] + dicta[key_new[i]] + 26) % 26 i += 1 or_txt += dicta[x] return or_txt def _A () ->None: '''simple docstring''' lowerCamelCase__ : Union[str, Any] = """THE GERMAN ATTACK""" lowerCamelCase__ : Union[str, Any] = """SECRET""" lowerCamelCase__ : List[str] = generate_key(UpperCamelCase , UpperCamelCase ) lowerCamelCase__ : List[Any] = cipher_text(UpperCamelCase , UpperCamelCase ) print(f"Encrypted Text = {s}" ) print(f"Original Text = {original_text(UpperCamelCase , UpperCamelCase )}" ) if __name__ == "__main__": import doctest doctest.testmod() main()
96
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL _UpperCAmelCase : Any = logging.get_logger(__name__) def __lowerCamelCase ( UpperCamelCase__ ): '''simple docstring''' if isinstance(_lowercase , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(_lowercase , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(_lowercase ): return [[videos]] raise ValueError(F'''Could not make batched video from {videos}''' ) class lowercase ( UpperCamelCase_ ): __SCREAMING_SNAKE_CASE : int = ['''pixel_values'''] def __init__( self , snake_case = True , snake_case = None , snake_case = PILImageResampling.BILINEAR , snake_case = True , snake_case = None , snake_case = True , snake_case = 1 / 255 , snake_case = True , snake_case = None , snake_case = None , **snake_case , ): super().__init__(**UpperCamelCase__ ) snake_case_ = size if size is not None else {'''shortest_edge''': 224} snake_case_ = get_size_dict(UpperCamelCase__ , default_to_square=UpperCamelCase__ ) snake_case_ = crop_size if crop_size is not None else {'''height''': 224, '''width''': 224} snake_case_ = get_size_dict(UpperCamelCase__ , param_name='crop_size' ) snake_case_ = do_resize snake_case_ = size snake_case_ = do_center_crop snake_case_ = crop_size snake_case_ = resample snake_case_ = do_rescale snake_case_ = rescale_factor snake_case_ = do_normalize snake_case_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN snake_case_ = image_std if image_std is not None else IMAGENET_STANDARD_STD def a ( self , snake_case , snake_case , snake_case = PILImageResampling.BILINEAR , snake_case = None , **snake_case , ): snake_case_ = get_size_dict(UpperCamelCase__ , default_to_square=UpperCamelCase__ ) if "shortest_edge" in size: snake_case_ = get_resize_output_image_size(UpperCamelCase__ , size['shortest_edge'] , default_to_square=UpperCamelCase__ ) elif "height" in size and "width" in size: snake_case_ = (size['''height'''], size['''width''']) else: raise ValueError(F'''Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}''' ) return resize(UpperCamelCase__ , size=UpperCamelCase__ , resample=UpperCamelCase__ , data_format=UpperCamelCase__ , **UpperCamelCase__ ) def a ( self , snake_case , snake_case , snake_case = None , **snake_case , ): snake_case_ = get_size_dict(UpperCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(F'''Size must have \'height\' and \'width\' as keys. Got {size.keys()}''' ) return center_crop(UpperCamelCase__ , size=(size['height'], size['width']) , data_format=UpperCamelCase__ , **UpperCamelCase__ ) def a ( self , snake_case , snake_case , snake_case = None , **snake_case , ): return rescale(UpperCamelCase__ , scale=UpperCamelCase__ , data_format=UpperCamelCase__ , **UpperCamelCase__ ) def a ( self , snake_case , snake_case , snake_case , snake_case = None , **snake_case , ): return normalize(UpperCamelCase__ , mean=UpperCamelCase__ , std=UpperCamelCase__ , data_format=UpperCamelCase__ , **UpperCamelCase__ ) def a ( self , snake_case , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = ChannelDimension.FIRST , ): if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. snake_case_ = to_numpy_array(UpperCamelCase__ ) if do_resize: snake_case_ = self.resize(image=UpperCamelCase__ , size=UpperCamelCase__ , resample=UpperCamelCase__ ) if do_center_crop: snake_case_ = self.center_crop(UpperCamelCase__ , size=UpperCamelCase__ ) if do_rescale: snake_case_ = self.rescale(image=UpperCamelCase__ , scale=UpperCamelCase__ ) if do_normalize: snake_case_ = self.normalize(image=UpperCamelCase__ , mean=UpperCamelCase__ , std=UpperCamelCase__ ) snake_case_ = to_channel_dimension_format(UpperCamelCase__ , UpperCamelCase__ ) return image def a ( self , snake_case , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = None , snake_case = ChannelDimension.FIRST , **snake_case , ): snake_case_ = do_resize if do_resize is not None else self.do_resize snake_case_ = resample if resample is not None else self.resample snake_case_ = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ = do_rescale if do_rescale is not None else self.do_rescale snake_case_ = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ = do_normalize if do_normalize is not None else self.do_normalize snake_case_ = image_mean if image_mean is not None else self.image_mean snake_case_ = image_std if image_std is not None else self.image_std snake_case_ = size if size is not None else self.size snake_case_ = get_size_dict(UpperCamelCase__ , default_to_square=UpperCamelCase__ ) snake_case_ = crop_size if crop_size is not None else self.crop_size snake_case_ = get_size_dict(UpperCamelCase__ , param_name='crop_size' ) if not valid_images(UpperCamelCase__ ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) snake_case_ = make_batched(UpperCamelCase__ ) snake_case_ = [ [ self._preprocess_image( image=UpperCamelCase__ , do_resize=UpperCamelCase__ , size=UpperCamelCase__ , resample=UpperCamelCase__ , do_center_crop=UpperCamelCase__ , crop_size=UpperCamelCase__ , do_rescale=UpperCamelCase__ , rescale_factor=UpperCamelCase__ , do_normalize=UpperCamelCase__ , image_mean=UpperCamelCase__ , image_std=UpperCamelCase__ , data_format=UpperCamelCase__ , ) for img in video ] for video in videos ] snake_case_ = {'''pixel_values''': videos} return BatchFeature(data=UpperCamelCase__ , tensor_type=UpperCamelCase__ )
362
def A ( _lowercase , _lowercase , _lowercase , _lowercase , _lowercase , _lowercase ): if index == r: for j in range(_lowercase ): print(data[j] , end=''' ''' ) print(''' ''' ) return # When no more elements are there to put in data[] if i >= n: return # current is included, put next at next location SCREAMING_SNAKE_CASE : Dict = arr[i] combination_util(_lowercase , _lowercase , _lowercase , index + 1 , _lowercase , i + 1 ) # current is excluded, replace it with # next (Note that i+1 is passed, but # index is not changed) combination_util(_lowercase , _lowercase , _lowercase , _lowercase , _lowercase , i + 1 ) # The main function that prints all combinations # of size r in arr[] of size n. This function # mainly uses combinationUtil() def A ( _lowercase , _lowercase , _lowercase ): # A temporary array to store all combination one by one SCREAMING_SNAKE_CASE : Any = [0] * r # Print all combination using temporary array 'data[]' combination_util(_lowercase , _lowercase , _lowercase , 0 , _lowercase , 0 ) if __name__ == "__main__": # Driver code to check the function above __UpperCamelCase : Optional[int] = [10, 20, 30, 40, 50] print_combination(arr, len(arr), 3) # This code is contributed by Ambuj sahu
248
0
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowerCAmelCase : Optional[int] = logging.get_logger(__name__) lowerCAmelCase : Union[str, Any] = '▁' lowerCAmelCase : int = {'vocab_file': 'sentencepiece.bpe.model', 'monolingual_vocab_file': 'dict.txt'} lowerCAmelCase : Any = { 'vocab_file': { 'vinai/bartpho-syllable': 'https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model', }, 'monolingual_vocab_file': { 'vinai/bartpho-syllable': 'https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt', }, } lowerCAmelCase : Any = {'vinai/bartpho-syllable': 1024} class a ( snake_case__ ): SCREAMING_SNAKE_CASE__ : Dict = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : str = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : Tuple = ["""input_ids""", """attention_mask"""] def __init__( self , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase="<s>" , _lowerCAmelCase="</s>" , _lowerCAmelCase="</s>" , _lowerCAmelCase="<s>" , _lowerCAmelCase="<unk>" , _lowerCAmelCase="<pad>" , _lowerCAmelCase="<mask>" , _lowerCAmelCase = None , **_lowerCAmelCase , ): """simple docstring""" __SCREAMING_SNAKE_CASE: Optional[Any] = AddedToken(UpperCAmelCase_ , lstrip=UpperCAmelCase_ , rstrip=UpperCAmelCase_ ) if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ) else mask_token __SCREAMING_SNAKE_CASE: Tuple = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=UpperCAmelCase_ , eos_token=UpperCAmelCase_ , unk_token=UpperCAmelCase_ , sep_token=UpperCAmelCase_ , cls_token=UpperCAmelCase_ , pad_token=UpperCAmelCase_ , mask_token=UpperCAmelCase_ , sp_model_kwargs=self.sp_model_kwargs , **UpperCAmelCase_ , ) __SCREAMING_SNAKE_CASE: List[str] = vocab_file __SCREAMING_SNAKE_CASE: Dict = monolingual_vocab_file __SCREAMING_SNAKE_CASE: Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(UpperCAmelCase_ ) ) # Load the reduced vocab # Keep order of special tokens for backward compatibility __SCREAMING_SNAKE_CASE: str = {} __SCREAMING_SNAKE_CASE: List[Any] = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(UpperCAmelCase_ ) not in self.fairseq_tokens_to_ids: __SCREAMING_SNAKE_CASE: Any = cnt cnt += 1 with open(UpperCAmelCase_ , '''r''' , encoding='''utf-8''' ) as f: for line in f.readlines(): __SCREAMING_SNAKE_CASE: List[Any] = line.strip().split()[0] __SCREAMING_SNAKE_CASE: Tuple = len(self.fairseq_tokens_to_ids ) if str(UpperCAmelCase_ ) not in self.fairseq_tokens_to_ids: __SCREAMING_SNAKE_CASE: Union[str, Any] = len(self.fairseq_tokens_to_ids ) __SCREAMING_SNAKE_CASE: Tuple = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ): """simple docstring""" __SCREAMING_SNAKE_CASE: List[str] = self.__dict__.copy() __SCREAMING_SNAKE_CASE: Optional[int] = None __SCREAMING_SNAKE_CASE: str = self.sp_model.serialized_model_proto() return state def __setstate__( self , _lowerCAmelCase ): """simple docstring""" __SCREAMING_SNAKE_CASE: List[str] = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): __SCREAMING_SNAKE_CASE: Tuple = {} __SCREAMING_SNAKE_CASE: Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def snake_case_ ( self , _lowerCAmelCase , _lowerCAmelCase = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] __SCREAMING_SNAKE_CASE: List[Any] = [self.cls_token_id] __SCREAMING_SNAKE_CASE: Optional[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def snake_case_ ( self , _lowerCAmelCase , _lowerCAmelCase = None , _lowerCAmelCase = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCAmelCase_ , token_ids_a=UpperCAmelCase_ , already_has_special_tokens=UpperCAmelCase_ ) if token_ids_a is None: return [1] + ([0] * len(UpperCAmelCase_ )) + [1] return [1] + ([0] * len(UpperCAmelCase_ )) + [1, 1] + ([0] * len(UpperCAmelCase_ )) + [1] def snake_case_ ( self , _lowerCAmelCase , _lowerCAmelCase = None ): """simple docstring""" __SCREAMING_SNAKE_CASE: int = [self.sep_token_id] __SCREAMING_SNAKE_CASE: List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def snake_case_ ( self ): """simple docstring""" return len(self.fairseq_ids_to_tokens ) def snake_case_ ( self ): """simple docstring""" __SCREAMING_SNAKE_CASE: int = {self.convert_ids_to_tokens(UpperCAmelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def snake_case_ ( self , _lowerCAmelCase ): """simple docstring""" return self.sp_model.encode(UpperCAmelCase_ , out_type=UpperCAmelCase_ ) def snake_case_ ( self , _lowerCAmelCase ): """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def snake_case_ ( self , _lowerCAmelCase ): """simple docstring""" return self.fairseq_ids_to_tokens[index] def snake_case_ ( self , _lowerCAmelCase ): """simple docstring""" __SCREAMING_SNAKE_CASE: str = ''''''.join(UpperCAmelCase_ ).replace(UpperCAmelCase_ , ''' ''' ).strip() return out_string def snake_case_ ( self , _lowerCAmelCase , _lowerCAmelCase = None ): """simple docstring""" if not os.path.isdir(UpperCAmelCase_ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return __SCREAMING_SNAKE_CASE: List[str] = os.path.join( UpperCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) __SCREAMING_SNAKE_CASE: Dict = os.path.join( UpperCAmelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''monolingual_vocab_file'''] , ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCAmelCase_ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCAmelCase_ , '''wb''' ) as fi: __SCREAMING_SNAKE_CASE: Any = self.sp_model.serialized_model_proto() fi.write(UpperCAmelCase_ ) if os.path.abspath(self.monolingual_vocab_file ) != os.path.abspath( UpperCAmelCase_ ) and os.path.isfile(self.monolingual_vocab_file ): copyfile(self.monolingual_vocab_file , UpperCAmelCase_ ) elif not os.path.isfile(self.monolingual_vocab_file ): with open(UpperCAmelCase_ , '''w''' , encoding='''utf-8''' ) as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(f"""{str(UpperCAmelCase_ )} \n""" ) return out_vocab_file, out_monolingual_vocab_file
702
from __future__ import annotations from math import pow, sqrt def lowerCAmelCase ( UpperCamelCase__ : float , UpperCamelCase__ : float , UpperCamelCase__ : float ) -> dict[str, float]: """simple docstring""" if (resistance, reactance, impedance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance == 0: return {"resistance": sqrt(pow(UpperCamelCase__ , 2 ) - pow(UpperCamelCase__ , 2 ) )} elif reactance == 0: return {"reactance": sqrt(pow(UpperCamelCase__ , 2 ) - pow(UpperCamelCase__ , 2 ) )} elif impedance == 0: return {"impedance": sqrt(pow(UpperCamelCase__ , 2 ) + pow(UpperCamelCase__ , 2 ) )} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
146
0
"""simple docstring""" import requests _a = """YOUR API KEY""" def lowerCamelCase__ ( __snake_case, __snake_case = giphy_api_key ) -> list: """simple docstring""" _UpperCamelCase = '''+'''.join(query.split() ) _UpperCamelCase = F'''https://api.giphy.com/v1/gifs/search?q={formatted_query}&api_key={api_key}''' _UpperCamelCase = requests.get(__snake_case ).json()['''data'''] return [gif["url"] for gif in gifs] if __name__ == "__main__": print("""\n""".join(get_gifs("""space ship""")))
19
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { """microsoft/wavlm-base""": """https://huggingface.co/microsoft/wavlm-base/resolve/main/config.json""", # See all WavLM models at https://huggingface.co/models?filter=wavlm } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'wavlm' def __init__( self , __a=32 , __a=7_68 , __a=12 , __a=12 , __a=30_72 , __a="gelu" , __a=0.1 , __a=0.1 , __a=0.1 , __a=0.0 , __a=0.1 , __a=0.1 , __a=0.02 , __a=1e-5 , __a="group" , __a="gelu" , __a=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , __a=(5, 2, 2, 2, 2, 2, 2) , __a=(10, 3, 3, 3, 3, 2, 2) , __a=False , __a=1_28 , __a=16 , __a=3_20 , __a=8_00 , __a=False , __a=True , __a=0.05 , __a=10 , __a=2 , __a=0.0 , __a=10 , __a=3_20 , __a=2 , __a=0.1 , __a=1_00 , __a=2_56 , __a=2_56 , __a=0.1 , __a="mean" , __a=False , __a=False , __a=2_56 , __a=(5_12, 5_12, 5_12, 5_12, 15_00) , __a=(5, 3, 3, 1, 1) , __a=(1, 2, 3, 1, 1) , __a=5_12 , __a=80 , __a=0 , __a=1 , __a=2 , __a=False , __a=3 , __a=2 , __a=3 , __a=None , **__a , ) -> Union[str, Any]: '''simple docstring''' super().__init__(**__a , pad_token_id=__a , bos_token_id=__a , eos_token_id=__a) _UpperCamelCase = hidden_size _UpperCamelCase = feat_extract_norm _UpperCamelCase = feat_extract_activation _UpperCamelCase = list(__a) _UpperCamelCase = list(__a) _UpperCamelCase = list(__a) _UpperCamelCase = conv_bias _UpperCamelCase = num_buckets _UpperCamelCase = max_bucket_distance _UpperCamelCase = num_conv_pos_embeddings _UpperCamelCase = num_conv_pos_embedding_groups _UpperCamelCase = len(self.conv_dim) _UpperCamelCase = num_hidden_layers _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_act _UpperCamelCase = num_attention_heads _UpperCamelCase = hidden_dropout _UpperCamelCase = attention_dropout _UpperCamelCase = activation_dropout _UpperCamelCase = feat_proj_dropout _UpperCamelCase = final_dropout _UpperCamelCase = layerdrop _UpperCamelCase = layer_norm_eps _UpperCamelCase = initializer_range _UpperCamelCase = num_ctc_classes _UpperCamelCase = vocab_size _UpperCamelCase = do_stable_layer_norm _UpperCamelCase = use_weighted_layer_sum _UpperCamelCase = classifier_proj_size if ( (len(self.conv_stride) != self.num_feat_extract_layers) or (len(self.conv_kernel) != self.num_feat_extract_layers) or (len(self.conv_dim) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' F''' {len(self.conv_dim)}`, `len(config.conv_stride) = {len(self.conv_stride)}`,''' F''' `len(config.conv_kernel) = {len(self.conv_kernel)}`.''') # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 _UpperCamelCase = apply_spec_augment _UpperCamelCase = mask_time_prob _UpperCamelCase = mask_time_length _UpperCamelCase = mask_time_min_masks _UpperCamelCase = mask_feature_prob _UpperCamelCase = mask_feature_length # parameters for pretraining with codevector quantized representations _UpperCamelCase = num_codevectors_per_group _UpperCamelCase = num_codevector_groups _UpperCamelCase = contrastive_logits_temperature _UpperCamelCase = num_negatives _UpperCamelCase = codevector_dim _UpperCamelCase = proj_codevector_dim _UpperCamelCase = diversity_loss_weight # ctc loss _UpperCamelCase = ctc_loss_reduction _UpperCamelCase = ctc_zero_infinity # adapter _UpperCamelCase = add_adapter _UpperCamelCase = adapter_kernel_size _UpperCamelCase = adapter_stride _UpperCamelCase = num_adapter_layers _UpperCamelCase = output_hidden_size or hidden_size # SequenceClassification-specific parameter. Feel free to ignore for other classes. _UpperCamelCase = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. _UpperCamelCase = list(__a) _UpperCamelCase = list(__a) _UpperCamelCase = list(__a) _UpperCamelCase = xvector_output_dim @property def UpperCAmelCase ( self) -> Optional[Any]: '''simple docstring''' return functools.reduce(operator.mul , self.conv_stride , 1)
19
1
'''simple docstring''' def _SCREAMING_SNAKE_CASE ( snake_case_ ): _lowercase = generate_pascal_triangle(snake_case_ ) for row_idx in range(snake_case_ ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=""" """ ) # Print row values for col_idx in range(row_idx + 1 ): if col_idx != row_idx: print(triangle[row_idx][col_idx] , end=""" """ ) else: print(triangle[row_idx][col_idx] , end="""""" ) print() def _SCREAMING_SNAKE_CASE ( snake_case_ ): if not isinstance(snake_case_ , snake_case_ ): raise TypeError("""The input value of 'num_rows' should be 'int'""" ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( """The input value of 'num_rows' should be greater than or equal to 0""" ) _lowercase = [] for current_row_idx in range(snake_case_ ): _lowercase = populate_current_row(snake_case_ , snake_case_ ) triangle.append(snake_case_ ) return triangle def _SCREAMING_SNAKE_CASE ( snake_case_ , snake_case_ ): _lowercase = [-1] * (current_row_idx + 1) # first and last elements of current row are equal to 1 _lowercase , _lowercase = 1, 1 for current_col_idx in range(1 , snake_case_ ): calculate_current_element( snake_case_ , snake_case_ , snake_case_ , snake_case_ ) return current_row def _SCREAMING_SNAKE_CASE ( snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): _lowercase = triangle[current_row_idx - 1][current_col_idx - 1] _lowercase = triangle[current_row_idx - 1][current_col_idx] _lowercase = above_to_left_elt + above_to_right_elt def _SCREAMING_SNAKE_CASE ( snake_case_ ): if not isinstance(snake_case_ , snake_case_ ): raise TypeError("""The input value of 'num_rows' should be 'int'""" ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( """The input value of 'num_rows' should be greater than or equal to 0""" ) _lowercase = [[1]] for row_index in range(1 , snake_case_ ): _lowercase = [0] + result[-1] + [0] _lowercase = row_index + 1 # Calculate the number of distinct elements in a row _lowercase = sum(divmod(snake_case_ , 2 ) ) _lowercase = [ temp_row[i - 1] + temp_row[i] for i in range(1 , distinct_elements + 1 ) ] _lowercase = row_first_half[: (row_index + 1) // 2] row_second_half.reverse() _lowercase = row_first_half + row_second_half result.append(snake_case_ ) return result def _SCREAMING_SNAKE_CASE ( ): from collections.abc import Callable from timeit import timeit def benchmark_a_function(snake_case_ , snake_case_ ) -> None: _lowercase = F"""{func.__name__}({value})""" _lowercase = timeit(F"""__main__.{call}""" , setup="""import __main__""" ) # print(f"{call:38} = {func(value)} -- {timing:.4f} seconds") print(F"""{call:38} -- {timing:.4f} seconds""" ) for value in range(15 ): # (1, 7, 14): for func in (generate_pascal_triangle, generate_pascal_triangle_optimized): benchmark_a_function(snake_case_ , snake_case_ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
572
'''simple docstring''' from collections import OrderedDict from typing import List, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'google/efficientnet-b7': 'https://huggingface.co/google/efficientnet-b7/resolve/main/config.json', } class __a ( _snake_case ): __SCREAMING_SNAKE_CASE : int = 'efficientnet' def __init__( self : Optional[int] , lowercase__ : int = 3 , lowercase__ : int = 6_00 , lowercase__ : float = 2.0 , lowercase__ : float = 3.1 , lowercase__ : int = 8 , lowercase__ : List[int] = [3, 3, 5, 3, 5, 5, 3] , lowercase__ : List[int] = [32, 16, 24, 40, 80, 1_12, 1_92] , lowercase__ : List[int] = [16, 24, 40, 80, 1_12, 1_92, 3_20] , lowercase__ : List[int] = [] , lowercase__ : List[int] = [1, 2, 2, 2, 1, 2, 1] , lowercase__ : List[int] = [1, 2, 2, 3, 3, 4, 1] , lowercase__ : List[int] = [1, 6, 6, 6, 6, 6, 6] , lowercase__ : float = 0.25 , lowercase__ : str = "swish" , lowercase__ : int = 25_60 , lowercase__ : str = "mean" , lowercase__ : float = 0.02 , lowercase__ : float = 0.001 , lowercase__ : float = 0.99 , lowercase__ : float = 0.5 , lowercase__ : float = 0.2 , **lowercase__ : List[Any] , ) ->Tuple: """simple docstring""" super().__init__(**lowercase__) _lowercase = num_channels _lowercase = image_size _lowercase = width_coefficient _lowercase = depth_coefficient _lowercase = depth_divisor _lowercase = kernel_sizes _lowercase = in_channels _lowercase = out_channels _lowercase = depthwise_padding _lowercase = strides _lowercase = num_block_repeats _lowercase = expand_ratios _lowercase = squeeze_expansion_ratio _lowercase = hidden_act _lowercase = hidden_dim _lowercase = pooling_type _lowercase = initializer_range _lowercase = batch_norm_eps _lowercase = batch_norm_momentum _lowercase = dropout_rate _lowercase = drop_connect_rate _lowercase = sum(lowercase__) * 4 class __a ( _snake_case ): __SCREAMING_SNAKE_CASE : List[str] = version.parse('1.11' ) @property def _UpperCAmelCase ( self : Union[str, Any]) ->Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ]) @property def _UpperCAmelCase ( self : str) ->float: """simple docstring""" return 1e-5
572
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging snake_case = logging.get_logger(__name__) snake_case = { """xlm-mlm-en-2048""": """https://huggingface.co/xlm-mlm-en-2048/resolve/main/config.json""", """xlm-mlm-ende-1024""": """https://huggingface.co/xlm-mlm-ende-1024/resolve/main/config.json""", """xlm-mlm-enfr-1024""": """https://huggingface.co/xlm-mlm-enfr-1024/resolve/main/config.json""", """xlm-mlm-enro-1024""": """https://huggingface.co/xlm-mlm-enro-1024/resolve/main/config.json""", """xlm-mlm-tlm-xnli15-1024""": """https://huggingface.co/xlm-mlm-tlm-xnli15-1024/resolve/main/config.json""", """xlm-mlm-xnli15-1024""": """https://huggingface.co/xlm-mlm-xnli15-1024/resolve/main/config.json""", """xlm-clm-enfr-1024""": """https://huggingface.co/xlm-clm-enfr-1024/resolve/main/config.json""", """xlm-clm-ende-1024""": """https://huggingface.co/xlm-clm-ende-1024/resolve/main/config.json""", """xlm-mlm-17-1280""": """https://huggingface.co/xlm-mlm-17-1280/resolve/main/config.json""", """xlm-mlm-100-1280""": """https://huggingface.co/xlm-mlm-100-1280/resolve/main/config.json""", } class SCREAMING_SNAKE_CASE ( _lowerCAmelCase ): '''simple docstring''' UpperCamelCase_ : Optional[Any] = '''xlm''' UpperCamelCase_ : int = { '''hidden_size''': '''emb_dim''', '''num_attention_heads''': '''n_heads''', '''num_hidden_layers''': '''n_layers''', '''n_words''': '''vocab_size''', # For backward compatibility } def __init__( self : List[Any] , UpperCAmelCase_ : Optional[Any]=3_0145 , UpperCAmelCase_ : Any=2048 , UpperCAmelCase_ : int=12 , UpperCAmelCase_ : int=16 , UpperCAmelCase_ : List[Any]=0.1 , UpperCAmelCase_ : Tuple=0.1 , UpperCAmelCase_ : Union[str, Any]=True , UpperCAmelCase_ : int=False , UpperCAmelCase_ : Any=False , UpperCAmelCase_ : Dict=False , UpperCAmelCase_ : int=1 , UpperCAmelCase_ : int=True , UpperCAmelCase_ : List[Any]=512 , UpperCAmelCase_ : Dict=2048**-0.5 , UpperCAmelCase_ : Union[str, Any]=1E-12 , UpperCAmelCase_ : List[Any]=0.02 , UpperCAmelCase_ : List[Any]=0 , UpperCAmelCase_ : List[Any]=1 , UpperCAmelCase_ : Tuple=2 , UpperCAmelCase_ : str=3 , UpperCAmelCase_ : str=5 , UpperCAmelCase_ : Tuple=True , UpperCAmelCase_ : Dict="first" , UpperCAmelCase_ : List[Any]=True , UpperCAmelCase_ : Any=None , UpperCAmelCase_ : Optional[Any]=True , UpperCAmelCase_ : Dict=0.1 , UpperCAmelCase_ : Tuple=5 , UpperCAmelCase_ : Optional[Any]=5 , UpperCAmelCase_ : Tuple=0 , UpperCAmelCase_ : List[str]=0 , UpperCAmelCase_ : Dict=2 , UpperCAmelCase_ : str=0 , **UpperCAmelCase_ : Union[str, Any] , ): SCREAMING_SNAKE_CASE : str = vocab_size SCREAMING_SNAKE_CASE : int = emb_dim SCREAMING_SNAKE_CASE : Union[str, Any] = n_layers SCREAMING_SNAKE_CASE : List[Any] = n_heads SCREAMING_SNAKE_CASE : Optional[Any] = dropout SCREAMING_SNAKE_CASE : Any = attention_dropout SCREAMING_SNAKE_CASE : str = gelu_activation SCREAMING_SNAKE_CASE : int = sinusoidal_embeddings SCREAMING_SNAKE_CASE : Optional[Any] = causal SCREAMING_SNAKE_CASE : List[str] = asm SCREAMING_SNAKE_CASE : Optional[Any] = n_langs SCREAMING_SNAKE_CASE : Dict = use_lang_emb SCREAMING_SNAKE_CASE : Dict = layer_norm_eps SCREAMING_SNAKE_CASE : Optional[Any] = bos_index SCREAMING_SNAKE_CASE : str = eos_index SCREAMING_SNAKE_CASE : int = pad_index SCREAMING_SNAKE_CASE : Dict = unk_index SCREAMING_SNAKE_CASE : Dict = mask_index SCREAMING_SNAKE_CASE : Tuple = is_encoder SCREAMING_SNAKE_CASE : Dict = max_position_embeddings SCREAMING_SNAKE_CASE : List[str] = embed_init_std SCREAMING_SNAKE_CASE : Dict = init_std SCREAMING_SNAKE_CASE : Any = summary_type SCREAMING_SNAKE_CASE : List[str] = summary_use_proj SCREAMING_SNAKE_CASE : int = summary_activation SCREAMING_SNAKE_CASE : Optional[Any] = summary_proj_to_labels SCREAMING_SNAKE_CASE : str = summary_first_dropout SCREAMING_SNAKE_CASE : str = start_n_top SCREAMING_SNAKE_CASE : Optional[int] = end_n_top SCREAMING_SNAKE_CASE : Optional[Any] = mask_token_id SCREAMING_SNAKE_CASE : int = lang_id if "n_words" in kwargs: SCREAMING_SNAKE_CASE : List[Any] = kwargs["n_words"] super().__init__(pad_token_id=UpperCAmelCase_ , bos_token_id=UpperCAmelCase_ , **UpperCAmelCase_ ) class SCREAMING_SNAKE_CASE ( _lowerCAmelCase ): '''simple docstring''' @property def _A ( self : List[str] ): if self.task == "multiple-choice": SCREAMING_SNAKE_CASE : Dict = {0: "batch", 1: "choice", 2: "sequence"} else: SCREAMING_SNAKE_CASE : int = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis), ] )
62
'''simple docstring''' import argparse import datetime import json import time import warnings from logging import getLogger from pathlib import Path from typing import Dict, List import torch from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import calculate_bleu, calculate_rouge, chunks, parse_numeric_n_bool_cl_kwargs, use_task_specific_params __UpperCamelCase : int = getLogger(__name__) __UpperCamelCase : int = """cuda""" if torch.cuda.is_available() else """cpu""" def __UpperCAmelCase ( SCREAMING_SNAKE_CASE__: List[str], SCREAMING_SNAKE_CASE__: str, SCREAMING_SNAKE_CASE__: str, SCREAMING_SNAKE_CASE__: int = 8, SCREAMING_SNAKE_CASE__: str = DEFAULT_DEVICE, SCREAMING_SNAKE_CASE__: Any=False, SCREAMING_SNAKE_CASE__: Tuple="summarization", SCREAMING_SNAKE_CASE__: List[Any]=None, **SCREAMING_SNAKE_CASE__: int, ) -> Dict: """simple docstring""" __a = Path(SCREAMING_SNAKE_CASE__ ).open('w', encoding='utf-8' ) __a = str(SCREAMING_SNAKE_CASE__ ) __a = AutoModelForSeqaSeqLM.from_pretrained(SCREAMING_SNAKE_CASE__ ).to(SCREAMING_SNAKE_CASE__ ) if fpaa: __a = model.half() __a = AutoTokenizer.from_pretrained(SCREAMING_SNAKE_CASE__ ) logger.info(f"""Inferred tokenizer type: {tokenizer.__class__}""" ) # if this is wrong, check config.model_type. __a = time.time() # update config with task specific params use_task_specific_params(SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) if prefix is None: __a = prefix or getattr(model.config, 'prefix', '' ) or '' for examples_chunk in tqdm(list(chunks(SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) ) ): __a = [prefix + text for text in examples_chunk] __a = tokenizer(SCREAMING_SNAKE_CASE__, return_tensors='pt', truncation=SCREAMING_SNAKE_CASE__, padding='longest' ).to(SCREAMING_SNAKE_CASE__ ) __a = model.generate( input_ids=batch.input_ids, attention_mask=batch.attention_mask, **SCREAMING_SNAKE_CASE__, ) __a = tokenizer.batch_decode(SCREAMING_SNAKE_CASE__, skip_special_tokens=SCREAMING_SNAKE_CASE__, clean_up_tokenization_spaces=SCREAMING_SNAKE_CASE__ ) for hypothesis in dec: fout.write(hypothesis + '\n' ) fout.flush() fout.close() __a = int(time.time() - start_time ) # seconds __a = len(SCREAMING_SNAKE_CASE__ ) return {"n_obs": n_obs, "runtime": runtime, "seconds_per_sample": round(runtime / n_obs, 4 )} def __UpperCAmelCase ( ) -> List[str]: """simple docstring""" return datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S' ) def __UpperCAmelCase ( SCREAMING_SNAKE_CASE__: Optional[int]=True ) -> List[Any]: """simple docstring""" __a = argparse.ArgumentParser() parser.add_argument('model_name', type=SCREAMING_SNAKE_CASE__, help='like facebook/bart-large-cnn,t5-base, etc.' ) parser.add_argument('input_path', type=SCREAMING_SNAKE_CASE__, help='like cnn_dm/test.source' ) parser.add_argument('save_path', type=SCREAMING_SNAKE_CASE__, help='where to save summaries' ) parser.add_argument('--reference_path', type=SCREAMING_SNAKE_CASE__, required=SCREAMING_SNAKE_CASE__, help='like cnn_dm/test.target' ) parser.add_argument('--score_path', type=SCREAMING_SNAKE_CASE__, required=SCREAMING_SNAKE_CASE__, default='metrics.json', help='where to save metrics' ) parser.add_argument('--device', type=SCREAMING_SNAKE_CASE__, required=SCREAMING_SNAKE_CASE__, default=SCREAMING_SNAKE_CASE__, help='cuda, cuda:1, cpu etc.' ) parser.add_argument( '--prefix', type=SCREAMING_SNAKE_CASE__, required=SCREAMING_SNAKE_CASE__, default=SCREAMING_SNAKE_CASE__, help='will be added to the begininng of src examples' ) parser.add_argument('--task', type=SCREAMING_SNAKE_CASE__, default='summarization', help='used for task_specific_params + metrics' ) parser.add_argument('--bs', type=SCREAMING_SNAKE_CASE__, default=8, required=SCREAMING_SNAKE_CASE__, help='batch size' ) parser.add_argument( '--n_obs', type=SCREAMING_SNAKE_CASE__, default=-1, required=SCREAMING_SNAKE_CASE__, help='How many observations. Defaults to all.' ) parser.add_argument('--fp16', action='store_true' ) parser.add_argument('--dump-args', action='store_true', help='print the custom hparams with the results' ) parser.add_argument( '--info', nargs='?', type=SCREAMING_SNAKE_CASE__, const=datetime_now(), help=( 'use in conjunction w/ --dump-args to print with the results whatever other info you\'d like, e.g.' ' lang=en-ru. If no value is passed, the current datetime string will be used.' ), ) # Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate __a , __a = parser.parse_known_args() __a = parse_numeric_n_bool_cl_kwargs(SCREAMING_SNAKE_CASE__ ) if parsed_args and verbose: print(f"""parsed the following generate kwargs: {parsed_args}""" ) __a = [' ' + x.rstrip() if 't5' in args.model_name else x.rstrip() for x in open(args.input_path ).readlines()] if args.n_obs > 0: __a = examples[: args.n_obs] Path(args.save_path ).parent.mkdir(exist_ok=SCREAMING_SNAKE_CASE__ ) if args.reference_path is None and Path(args.score_path ).exists(): warnings.warn(f"""score_path {args.score_path} will be overwritten unless you type ctrl-c.""" ) if args.device == "cpu" and args.fpaa: # this mix leads to RuntimeError: "threshold_cpu" not implemented for 'Half' raise ValueError('Can\'t mix --fp16 and --device cpu' ) __a = generate_summaries_or_translations( SCREAMING_SNAKE_CASE__, args.save_path, args.model_name, batch_size=args.bs, device=args.device, fpaa=args.fpaa, task=args.task, prefix=args.prefix, **SCREAMING_SNAKE_CASE__, ) if args.reference_path is None: return {} # Compute scores __a = calculate_bleu if 'translation' in args.task else calculate_rouge __a = [x.rstrip() for x in open(args.save_path ).readlines()] __a = [x.rstrip() for x in open(args.reference_path ).readlines()][: len(SCREAMING_SNAKE_CASE__ )] __a = score_fn(SCREAMING_SNAKE_CASE__, SCREAMING_SNAKE_CASE__ ) scores.update(SCREAMING_SNAKE_CASE__ ) if args.dump_args: scores.update(SCREAMING_SNAKE_CASE__ ) if args.info: __a = args.info if verbose: print(SCREAMING_SNAKE_CASE__ ) if args.score_path is not None: json.dump(SCREAMING_SNAKE_CASE__, open(args.score_path, 'w' ) ) return scores if __name__ == "__main__": # Usage for MT: # python run_eval.py MODEL_NAME $DATA_DIR/test.source $save_dir/test_translations.txt --reference_path $DATA_DIR/test.target --score_path $save_dir/test_bleu.json --task translation $@ run_generate(verbose=True)
448
0
"""simple docstring""" import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class snake_case_ ( _lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_: str = ["""image_processor""", """tokenizer"""] SCREAMING_SNAKE_CASE_: Union[str, Any] = """CLIPImageProcessor""" SCREAMING_SNAKE_CASE_: Union[str, Any] = ("""XLMRobertaTokenizer""", """XLMRobertaTokenizerFast""") def __init__( self , __a=None , __a=None , **__a ): """simple docstring""" A__ = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , __a , ) A__ = kwargs.pop('feature_extractor' ) A__ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(__a , __a ) def __call__( self , __a=None , __a=None , __a=None , **__a ): """simple docstring""" if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: A__ = self.tokenizer(__a , return_tensors=__a , **__a ) if images is not None: A__ = self.image_processor(__a , return_tensors=__a , **__a ) if text is not None and images is not None: A__ = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__a ) , tensor_type=__a ) def _UpperCAmelCase ( self , *__a , **__a ): """simple docstring""" return self.tokenizer.batch_decode(*__a , **__a ) def _UpperCAmelCase ( self , *__a , **__a ): """simple docstring""" return self.tokenizer.decode(*__a , **__a ) @property def _UpperCAmelCase ( self ): """simple docstring""" A__ = self.tokenizer.model_input_names A__ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
554
"""simple docstring""" import io import math from typing import Dict, Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import convert_to_rgb, normalize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, get_image_size, infer_channel_dimension_format, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_torch_available, is_vision_available, logging from ...utils.import_utils import requires_backends if is_vision_available(): import textwrap from PIL import Image, ImageDraw, ImageFont if is_torch_available(): import torch from transformers.pytorch_utils import is_torch_greater_or_equal_than_1_11 else: SCREAMING_SNAKE_CASE : str = False SCREAMING_SNAKE_CASE : str = logging.get_logger(__name__) SCREAMING_SNAKE_CASE : Optional[Any] = '''ybelkada/fonts''' def __lowerCamelCase ( ): if is_torch_available() and not is_torch_greater_or_equal_than_1_11: raise ImportError( f'''You are using torch=={torch.__version__}, but torch>=1.11.0 is required to use ''' 'Pix2StructImageProcessor. Please upgrade torch.' ) def __lowerCamelCase ( lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ): requires_backends(lowerCAmelCase__ ,['torch'] ) _check_torch_version() A__ = image_tensor.unsqueeze(0 ) A__ = torch.nn.functional.unfold(lowerCAmelCase__ ,(patch_height, patch_width) ,stride=(patch_height, patch_width) ) A__ = patches.reshape(image_tensor.size(0 ) ,image_tensor.size(1 ) ,lowerCAmelCase__ ,lowerCAmelCase__ ,-1 ) A__ = patches.permute(0 ,4 ,2 ,3 ,1 ).reshape( image_tensor.size(2 ) // patch_height ,image_tensor.size(3 ) // patch_width ,image_tensor.size(1 ) * patch_height * patch_width ,) return patches.unsqueeze(0 ) def __lowerCamelCase ( lowerCAmelCase__ ,lowerCAmelCase__ = 36 ,lowerCAmelCase__ = "black" ,lowerCAmelCase__ = "white" ,lowerCAmelCase__ = 5 ,lowerCAmelCase__ = 5 ,lowerCAmelCase__ = 5 ,lowerCAmelCase__ = 5 ,lowerCAmelCase__ = None ,lowerCAmelCase__ = None ,): requires_backends(lowerCAmelCase__ ,'vision' ) # Add new lines so that each line is no more than 80 characters. A__ = textwrap.TextWrapper(width=80 ) A__ = wrapper.wrap(text=lowerCAmelCase__ ) A__ = '\n'.join(lowerCAmelCase__ ) if font_bytes is not None and font_path is None: A__ = io.BytesIO(lowerCAmelCase__ ) elif font_path is not None: A__ = font_path else: A__ = hf_hub_download(lowerCAmelCase__ ,'Arial.TTF' ) A__ = ImageFont.truetype(lowerCAmelCase__ ,encoding='UTF-8' ,size=lowerCAmelCase__ ) # Use a temporary canvas to determine the width and height in pixels when # rendering the text. A__ = ImageDraw.Draw(Image.new('RGB' ,(1, 1) ,lowerCAmelCase__ ) ) A__ , A__ , A__ , A__ = temp_draw.textbbox((0, 0) ,lowerCAmelCase__ ,lowerCAmelCase__ ) # Create the actual image with a bit of padding around the text. A__ = text_width + left_padding + right_padding A__ = text_height + top_padding + bottom_padding A__ = Image.new('RGB' ,(image_width, image_height) ,lowerCAmelCase__ ) A__ = ImageDraw.Draw(lowerCAmelCase__ ) draw.text(xy=(left_padding, top_padding) ,text=lowerCAmelCase__ ,fill=lowerCAmelCase__ ,font=lowerCAmelCase__ ) return image def __lowerCamelCase ( lowerCAmelCase__ ,lowerCAmelCase__ ,**lowerCAmelCase__ ): requires_backends(lowerCAmelCase__ ,'vision' ) # Convert to PIL image if necessary A__ = to_pil_image(lowerCAmelCase__ ) A__ = render_text(lowerCAmelCase__ ,**lowerCAmelCase__ ) A__ = max(header_image.width ,image.width ) A__ = int(image.height * (new_width / image.width) ) A__ = int(header_image.height * (new_width / header_image.width) ) A__ = Image.new('RGB' ,(new_width, new_height + new_header_height) ,'white' ) new_image.paste(header_image.resize((new_width, new_header_height) ) ,(0, 0) ) new_image.paste(image.resize((new_width, new_height) ) ,(0, new_header_height) ) # Convert back to the original framework if necessary A__ = to_numpy_array(lowerCAmelCase__ ) if infer_channel_dimension_format(lowerCAmelCase__ ) == ChannelDimension.LAST: A__ = to_channel_dimension_format(lowerCAmelCase__ ,ChannelDimension.LAST ) return new_image class snake_case_ ( _lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_: Optional[int] = ["""flattened_patches"""] def __init__( self , __a = True , __a = True , __a = None , __a = 2048 , __a = False , **__a , ): """simple docstring""" super().__init__(**__a ) A__ = patch_size if patch_size is not None else {'height': 16, 'width': 16} A__ = do_normalize A__ = do_convert_rgb A__ = max_patches A__ = is_vqa def _UpperCAmelCase ( self , __a , __a , __a , **__a ): """simple docstring""" requires_backends(self.extract_flattened_patches , 'torch' ) _check_torch_version() # convert to torch A__ = to_channel_dimension_format(__a , ChannelDimension.FIRST ) A__ = torch.from_numpy(__a ) A__ , A__ = patch_size['height'], patch_size['width'] A__ , A__ = get_image_size(__a ) # maximize scale s.t. A__ = math.sqrt(max_patches * (patch_height / image_height) * (patch_width / image_width) ) A__ = max(min(math.floor(scale * image_height / patch_height ) , __a ) , 1 ) A__ = max(min(math.floor(scale * image_width / patch_width ) , __a ) , 1 ) A__ = max(num_feasible_rows * patch_height , 1 ) A__ = max(num_feasible_cols * patch_width , 1 ) A__ = torch.nn.functional.interpolate( image.unsqueeze(0 ) , size=(resized_height, resized_width) , mode='bilinear' , align_corners=__a , antialias=__a , ).squeeze(0 ) # [1, rows, columns, patch_height * patch_width * image_channels] A__ = torch_extract_patches(__a , __a , __a ) A__ = patches.shape A__ = patches_shape[1] A__ = patches_shape[2] A__ = patches_shape[3] # [rows * columns, patch_height * patch_width * image_channels] A__ = patches.reshape([rows * columns, depth] ) # [rows * columns, 1] A__ = torch.arange(__a ).reshape([rows, 1] ).repeat(1 , __a ).reshape([rows * columns, 1] ) A__ = torch.arange(__a ).reshape([1, columns] ).repeat(__a , 1 ).reshape([rows * columns, 1] ) # Offset by 1 so the ids do not contain zeros, which represent padding. row_ids += 1 col_ids += 1 # Prepare additional patch features. # [rows * columns, 1] A__ = row_ids.to(torch.floataa ) A__ = col_ids.to(torch.floataa ) # [rows * columns, 2 + patch_height * patch_width * image_channels] A__ = torch.cat([row_ids, col_ids, patches] , -1 ) # [max_patches, 2 + patch_height * patch_width * image_channels] A__ = torch.nn.functional.pad(__a , [0, 0, 0, max_patches - (rows * columns)] ).float() A__ = to_numpy_array(__a ) return result def _UpperCAmelCase ( self , __a , __a = None , **__a ): """simple docstring""" if image.dtype == np.uinta: A__ = image.astype(np.floataa ) # take mean across the whole `image` A__ = np.mean(__a ) A__ = np.std(__a ) A__ = max(__a , 1.0 / math.sqrt(np.prod(image.shape ) ) ) return normalize(__a , mean=__a , std=__a , **__a ) def _UpperCAmelCase ( self , __a , __a = None , __a = None , __a = None , __a = None , __a = None , __a = None , __a = ChannelDimension.FIRST , **__a , ): """simple docstring""" A__ = do_normalize if do_normalize is not None else self.do_normalize A__ = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb A__ = patch_size if patch_size is not None else self.patch_size A__ = max_patches if max_patches is not None else self.max_patches A__ = self.is_vqa if kwargs.get('data_format' , __a ) is not None: raise ValueError('data_format is not an accepted input as the outputs are ' ) A__ = make_list_of_images(__a ) if not valid_images(__a ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) # PIL RGBA images are converted to RGB if do_convert_rgb: A__ = [convert_to_rgb(__a ) for image in images] # All transformations expect numpy arrays. A__ = [to_numpy_array(__a ) for image in images] if is_vqa: if header_text is None: raise ValueError('A header text must be provided for VQA models.' ) A__ = kwargs.pop('font_bytes' , __a ) A__ = kwargs.pop('font_path' , __a ) if isinstance(__a , __a ): A__ = [header_text] * len(__a ) A__ = [ render_header(__a , header_text[i] , font_bytes=__a , font_path=__a ) for i, image in enumerate(__a ) ] if do_normalize: A__ = [self.normalize(image=__a ) for image in images] # convert to torch tensor and permute A__ = [ self.extract_flattened_patches(image=__a , max_patches=__a , patch_size=__a ) for image in images ] # create attention mask in numpy A__ = [(image.sum(axis=-1 ) != 0).astype(np.floataa ) for image in images] A__ = BatchFeature( data={'flattened_patches': images, 'attention_mask': attention_masks} , tensor_type=__a ) return encoded_outputs
554
1
from itertools import product def lowerCamelCase__ ( __lowerCamelCase : int , __lowerCamelCase : int ): __UpperCAmelCase : Tuple = sides_number __UpperCAmelCase : int = max_face_number * dice_number __UpperCAmelCase : List[str] = [0] * (max_total + 1) __UpperCAmelCase : Union[str, Any] = 1 __UpperCAmelCase : List[str] = range(__lowerCamelCase , max_face_number + 1 ) for dice_numbers in product(__lowerCamelCase , repeat=__lowerCamelCase ): __UpperCAmelCase : Optional[Any] = sum(__lowerCamelCase ) totals_frequencies[total] += 1 return totals_frequencies def lowerCamelCase__ ( ): __UpperCAmelCase : List[Any] = total_frequency_distribution( sides_number=4 , dice_number=9 ) __UpperCAmelCase : Optional[Any] = total_frequency_distribution( sides_number=6 , dice_number=6 ) __UpperCAmelCase : int = 0 __UpperCAmelCase : Dict = 9 __UpperCAmelCase : int = 4 * 9 __UpperCAmelCase : str = 6 for peter_total in range(__lowerCamelCase , max_peter_total + 1 ): peter_wins_count += peter_totals_frequencies[peter_total] * sum( colin_totals_frequencies[min_colin_total:peter_total] ) __UpperCAmelCase : List[str] = (4**9) * (6**6) __UpperCAmelCase : str = peter_wins_count / total_games_number __UpperCAmelCase : List[str] = round(__lowerCamelCase , ndigits=7 ) return rounded_peter_win_probability if __name__ == "__main__": print(f"""{solution() = }""")
63
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class __UpperCamelCase : def __init__( self : int , lowerCAmelCase : Optional[int] , lowerCAmelCase : Optional[Any]=13 , lowerCAmelCase : Optional[int]=2 , lowerCAmelCase : Dict=24 , lowerCAmelCase : Dict=16 , lowerCAmelCase : Union[str, Any]=True , lowerCAmelCase : Union[str, Any]=True , lowerCAmelCase : Tuple=32 , lowerCAmelCase : Optional[Any]=5 , lowerCAmelCase : Any=4 , lowerCAmelCase : Tuple=37 , lowerCAmelCase : str="gelu" , lowerCAmelCase : List[Any]=0.1 , lowerCAmelCase : List[str]=0.1 , lowerCAmelCase : int=10 , lowerCAmelCase : str=0.02 , lowerCAmelCase : Any=None , lowerCAmelCase : List[str]=2 , lowerCAmelCase : str=2 , ): '''simple docstring''' UpperCAmelCase_ = parent UpperCAmelCase_ = batch_size UpperCAmelCase_ = patch_size UpperCAmelCase_ = max_length UpperCAmelCase_ = num_mel_bins UpperCAmelCase_ = is_training UpperCAmelCase_ = use_labels UpperCAmelCase_ = hidden_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = hidden_act UpperCAmelCase_ = hidden_dropout_prob UpperCAmelCase_ = attention_probs_dropout_prob UpperCAmelCase_ = type_sequence_label_size UpperCAmelCase_ = initializer_range UpperCAmelCase_ = scope UpperCAmelCase_ = frequency_stride UpperCAmelCase_ = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) UpperCAmelCase_ = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 UpperCAmelCase_ = (self.max_length - self.patch_size) // self.time_stride + 1 UpperCAmelCase_ = frequency_out_dimension * time_out_dimension UpperCAmelCase_ = num_patches + 2 def __A ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase_ = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) UpperCAmelCase_ = None if self.use_labels: UpperCAmelCase_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase_ = self.get_config() return config, input_values, labels def __A ( self : List[str] ): '''simple docstring''' return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCAmelCase , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def __A ( self : List[Any] , lowerCAmelCase : List[Any] , lowerCAmelCase : int , lowerCAmelCase : Optional[int] ): '''simple docstring''' UpperCAmelCase_ = ASTModel(config=lowerCAmelCase ) model.to(lowerCAmelCase ) model.eval() UpperCAmelCase_ = model(lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __A ( self : Any ): '''simple docstring''' UpperCAmelCase_ = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ) = config_and_inputs UpperCAmelCase_ = {"input_values": input_values} return config, inputs_dict @require_torch class __UpperCamelCase ( lowercase , lowercase , unittest.TestCase ): SCREAMING_SNAKE_CASE__ = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE__ = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False SCREAMING_SNAKE_CASE__ = False def __A ( self : Optional[Any] , lowerCAmelCase : Dict , lowerCAmelCase : Union[str, Any] , lowerCAmelCase : Tuple , lowerCAmelCase : List[str] , lowerCAmelCase : int ): '''simple docstring''' if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def __A ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase_ = ASTModelTester(self ) UpperCAmelCase_ = ConfigTester(self , config_class=lowerCAmelCase , has_text_modality=lowerCAmelCase , hidden_size=37 ) def __A ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="AST does not use inputs_embeds" ) def __A ( self : Union[str, Any] ): '''simple docstring''' pass def __A ( self : Optional[Any] ): '''simple docstring''' UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCAmelCase_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase , nn.Linear ) ) def __A ( self : Any ): '''simple docstring''' UpperCAmelCase_ , UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ = model_class(lowerCAmelCase ) UpperCAmelCase_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ = [*signature.parameters.keys()] UpperCAmelCase_ = ["input_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase ) def __A ( self : List[Any] ): '''simple docstring''' UpperCAmelCase_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase ) @slow def __A ( self : Union[str, Any] ): '''simple docstring''' for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ = ASTModel.from_pretrained(lowerCAmelCase ) self.assertIsNotNone(lowerCAmelCase ) def __lowerCAmelCase ( ): UpperCAmelCase_ = hf_hub_download( repo_id="nielsr/audio-spectogram-transformer-checkpoint" , filename="sample_audio.flac" , repo_type="dataset" ) UpperCAmelCase_ , UpperCAmelCase_ = torchaudio.load(A ) return audio, sampling_rate @require_torch @require_torchaudio class __UpperCamelCase ( unittest.TestCase ): @cached_property def __A ( self : Optional[Any] ): '''simple docstring''' return ( ASTFeatureExtractor.from_pretrained("MIT/ast-finetuned-audioset-10-10-0.4593" ) if is_torchaudio_available() else None ) @slow def __A ( self : Union[str, Any] ): '''simple docstring''' UpperCAmelCase_ = self.default_feature_extractor UpperCAmelCase_ = ASTForAudioClassification.from_pretrained("MIT/ast-finetuned-audioset-10-10-0.4593" ).to(lowerCAmelCase ) UpperCAmelCase_ = self.default_feature_extractor UpperCAmelCase_ , UpperCAmelCase_ = prepare_audio() UpperCAmelCase_ = audio.squeeze().numpy() UpperCAmelCase_ = feature_extractor(lowerCAmelCase , sampling_rate=lowerCAmelCase , return_tensors="pt" ).to(lowerCAmelCase ) # forward pass with torch.no_grad(): UpperCAmelCase_ = model(**lowerCAmelCase ) # verify the logits UpperCAmelCase_ = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , lowerCAmelCase ) UpperCAmelCase_ = torch.tensor([-0.8_760, -7.0_042, -8.6_602] ).to(lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase , atol=1e-4 ) )
162
0
from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar _A = TypeVar('''KEY''') _A = TypeVar('''VAL''') @dataclass(frozen=a__ , slots=a__ ) class A ( Generic[KEY, VAL] ): __snake_case = 42 __snake_case = 42 class A ( _Item ): def __init__( self ): """simple docstring""" super().__init__(lowerCAmelCase__, lowerCAmelCase__ ) def __bool__( self ): """simple docstring""" return False _A = _DeletedItem() class A ( MutableMapping[KEY, VAL] ): def __init__( self, UpperCamelCase__ = 8, UpperCamelCase__ = 0.75 ): """simple docstring""" lowerCAmelCase_ = initial_block_size lowerCAmelCase_ = [None] * initial_block_size assert 0.0 < capacity_factor < 1.0 lowerCAmelCase_ = capacity_factor lowerCAmelCase_ = 0 def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__ ): """simple docstring""" return hash(lowerCAmelCase__ ) % len(self._buckets ) def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__ ): """simple docstring""" return (ind + 1) % len(self._buckets ) def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__, UpperCamelCase__, UpperCamelCase__ ): """simple docstring""" lowerCAmelCase_ = self._buckets[ind] if not stored: lowerCAmelCase_ = _Item(lowerCAmelCase__, lowerCAmelCase__ ) self._len += 1 return True elif stored.key == key: lowerCAmelCase_ = _Item(lowerCAmelCase__, lowerCAmelCase__ ) return True else: return False def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = len(self._buckets ) * self._capacity_factor return len(self ) >= int(lowerCAmelCase__ ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" if len(self._buckets ) <= self._initial_block_size: return False lowerCAmelCase_ = len(self._buckets ) * self._capacity_factor / 2 return len(self ) < limit def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__ ): """simple docstring""" lowerCAmelCase_ = self._buckets lowerCAmelCase_ = [None] * new_size lowerCAmelCase_ = 0 for item in old_buckets: if item: self._add_item(item.key, item.val ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" self._resize(len(self._buckets ) * 2 ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" self._resize(len(self._buckets ) // 2 ) def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__ ): """simple docstring""" lowerCAmelCase_ = self._get_bucket_index(lowerCAmelCase__ ) for _ in range(len(self._buckets ) ): yield ind lowerCAmelCase_ = self._get_next_ind(lowerCAmelCase__ ) def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__, UpperCamelCase__ ): """simple docstring""" for ind in self._iterate_buckets(lowerCAmelCase__ ): if self._try_set(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__ ): break def __setitem__( self, UpperCamelCase__, UpperCamelCase__ ): """simple docstring""" if self._is_full(): self._size_up() self._add_item(lowerCAmelCase__, lowerCAmelCase__ ) def __delitem__( self, UpperCamelCase__ ): """simple docstring""" for ind in self._iterate_buckets(lowerCAmelCase__ ): lowerCAmelCase_ = self._buckets[ind] if item is None: raise KeyError(lowerCAmelCase__ ) if item is _deleted: continue if item.key == key: lowerCAmelCase_ = _deleted self._len -= 1 break if self._is_sparse(): self._size_down() def __getitem__( self, UpperCamelCase__ ): """simple docstring""" for ind in self._iterate_buckets(lowerCAmelCase__ ): lowerCAmelCase_ = self._buckets[ind] if item is None: break if item is _deleted: continue if item.key == key: return item.val raise KeyError(lowerCAmelCase__ ) def __len__( self ): """simple docstring""" return self._len def __iter__( self ): """simple docstring""" yield from (item.key for item in self._buckets if item) def __repr__( self ): """simple docstring""" lowerCAmelCase_ = " ,".join( f"{item.key}: {item.val}" for item in self._buckets if item ) return f"HashMap({val_string})"
709
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _A = logging.get_logger(__name__) _A = '''▁''' _A = {'''vocab_file''': '''spiece.model'''} _A = { '''vocab_file''': { '''google/reformer-crime-and-punishment''': ( '''https://huggingface.co/google/reformer-crime-and-punishment/resolve/main/spiece.model''' ) } } _A = { '''google/reformer-crime-and-punishment''': 524_288, } class A ( __UpperCAmelCase ): __snake_case = VOCAB_FILES_NAMES __snake_case = PRETRAINED_VOCAB_FILES_MAP __snake_case = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __snake_case = ['input_ids', 'attention_mask'] def __init__( self, UpperCamelCase__, UpperCamelCase__="</s>", UpperCamelCase__="<unk>", UpperCamelCase__=[], UpperCamelCase__ = None, **UpperCamelCase__, ): """simple docstring""" lowerCAmelCase_ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=UpperCamelCase__, unk_token=UpperCamelCase__, additional_special_tokens=UpperCamelCase__, sp_model_kwargs=self.sp_model_kwargs, **UpperCamelCase__, ) lowerCAmelCase_ = vocab_file lowerCAmelCase_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCamelCase__ ) @property def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" return self.sp_model.get_piece_size() def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = {self.convert_ids_to_tokens(UpperCamelCase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): """simple docstring""" lowerCAmelCase_ = self.__dict__.copy() lowerCAmelCase_ = None return state def __setstate__( self, UpperCamelCase__ ): """simple docstring""" lowerCAmelCase_ = d # for backward compatibility if not hasattr(self, '''sp_model_kwargs''' ): lowerCAmelCase_ = {} lowerCAmelCase_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__ ): """simple docstring""" return self.sp_model.encode(UpperCamelCase__, out_type=UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__ ): """simple docstring""" return self.sp_model.piece_to_id(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__ ): """simple docstring""" if index < self.sp_model.get_piece_size(): lowerCAmelCase_ = self.sp_model.IdToPiece(UpperCamelCase__ ) return token def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__ ): """simple docstring""" lowerCAmelCase_ = [] lowerCAmelCase_ = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(UpperCamelCase__ ) + token lowerCAmelCase_ = [] else: current_sub_tokens.append(UpperCamelCase__ ) out_string += self.sp_model.decode(UpperCamelCase__ ) return out_string.strip() def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__, UpperCamelCase__ = None ): """simple docstring""" if not os.path.isdir(UpperCamelCase__ ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return lowerCAmelCase_ = os.path.join( UpperCamelCase__, (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file, UpperCamelCase__ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase__, '''wb''' ) as fi: lowerCAmelCase_ = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase__ ) return (out_vocab_file,)
325
0
'''simple docstring''' import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets SCREAMING_SNAKE_CASE = '\\n@inproceedings{lin-2004-rouge,\n title = "{ROUGE}: A Package for Automatic Evaluation of Summaries",\n author = "Lin, Chin-Yew",\n booktitle = "Text Summarization Branches Out",\n month = jul,\n year = "2004",\n address = "Barcelona, Spain",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/W04-1013",\n pages = "74--81",\n}\n' SCREAMING_SNAKE_CASE = '\\nROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for\nevaluating automatic summarization and machine translation software in natural language processing.\nThe metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation.\n\nNote that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters.\n\nThis metrics is a wrapper around Google Research reimplementation of ROUGE:\nhttps://github.com/google-research/google-research/tree/master/rouge\n' SCREAMING_SNAKE_CASE = '\nCalculates average rouge scores for a list of hypotheses and references\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n rouge_types: A list of rouge types to calculate.\n Valid names:\n `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring,\n `"rougeL"`: Longest common subsequence based scoring.\n `"rougeLSum"`: rougeLsum splits text using `"\n"`.\n See details in https://github.com/huggingface/datasets/issues/617\n use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes.\n use_aggregator: Return aggregates if this is set to True\nReturns:\n rouge1: rouge_1 (precision, recall, f1),\n rouge2: rouge_2 (precision, recall, f1),\n rougeL: rouge_l (precision, recall, f1),\n rougeLsum: rouge_lsum (precision, recall, f1)\nExamples:\n\n >>> rouge = datasets.load_metric(\'rouge\')\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> results = rouge.compute(predictions=predictions, references=references)\n >>> print(list(results.keys()))\n [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\']\n >>> print(results["rouge1"])\n AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0))\n >>> print(results["rouge1"].mid.fmeasure)\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class UpperCAmelCase_ ( datasets.Metric ): """simple docstring""" def A__ ( self : Union[str, Any] ) -> str: '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/google-research/google-research/tree/master/rouge'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/ROUGE_(metric)''', '''https://github.com/google-research/google-research/tree/master/rouge''', ] , ) def A__ ( self : Tuple , UpperCAmelCase : List[Any] , UpperCAmelCase : List[str] , UpperCAmelCase : Any=None , UpperCAmelCase : Tuple=True , UpperCAmelCase : List[str]=False ) -> Union[str, Any]: '''simple docstring''' if rouge_types is None: lowercase : Dict =['''rouge1''', '''rouge2''', '''rougeL''', '''rougeLsum'''] lowercase : Optional[Any] =rouge_scorer.RougeScorer(rouge_types=UpperCAmelCase , use_stemmer=UpperCAmelCase ) if use_aggregator: lowercase : int =scoring.BootstrapAggregator() else: lowercase : Any =[] for ref, pred in zip(UpperCAmelCase , UpperCAmelCase ): lowercase : Any =scorer.score(UpperCAmelCase , UpperCAmelCase ) if use_aggregator: aggregator.add_scores(UpperCAmelCase ) else: scores.append(UpperCAmelCase ) if use_aggregator: lowercase : int =aggregator.aggregate() else: lowercase : Dict ={} for key in scores[0]: lowercase : Any =[score[key] for score in scores] return result
94
import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, PLBartTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase__ =get_tests_dir('fixtures/test_sentencepiece.model') if is_torch_available(): from transformers.models.plbart.modeling_plbart import shift_tokens_right UpperCamelCase__ =5_0003 UpperCamelCase__ =5_0002 @require_sentencepiece @require_tokenizers class lowerCAmelCase__( __lowercase , unittest.TestCase ): '''simple docstring''' __snake_case = PLBartTokenizer __snake_case = None __snake_case = False def UpperCamelCase_ ( self ) -> str: super().setUp() # We have a SentencePiece fixture for testing _SCREAMING_SNAKE_CASE : Union[str, Any] = PLBartTokenizer(__lowerCamelCase , language_codes="base" , keep_accents=__lowerCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self ) -> List[str]: _SCREAMING_SNAKE_CASE : str = PLBartTokenizer(__lowerCamelCase , language_codes="base" , keep_accents=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Union[str, Any] = tokenizer.tokenize("This is a test" ) self.assertListEqual(__lowerCamelCase , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) _SCREAMING_SNAKE_CASE : Dict = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) _SCREAMING_SNAKE_CASE : Union[str, Any] = tokenizer.convert_tokens_to_ids(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] ] , ) _SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer.convert_ids_to_tokens(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) _SCREAMING_SNAKE_CASE : List[str] = tokenizer.vocab_size _SCREAMING_SNAKE_CASE : List[str] = [tokenizer.convert_ids_to_tokens(__lowerCamelCase ) for x in range(end - 4 , __lowerCamelCase )] self.assertListEqual(__lowerCamelCase , ["__java__", "__python__", "__en_XX__", "<mask>"] ) _SCREAMING_SNAKE_CASE : Dict = "java.lang.Exception, python.lang.Exception, javascript, php, ruby, go" _SCREAMING_SNAKE_CASE : Any = tokenizer(__lowerCamelCase ).input_ids self.assertEqual( tokenizer.decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase , clean_up_tokenization_spaces=__lowerCamelCase ) , __lowerCamelCase , ) def UpperCamelCase_ ( self ) -> int: _SCREAMING_SNAKE_CASE : Any = PLBartTokenizer(__lowerCamelCase , language_codes="multi" , keep_accents=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[int] = tokenizer.tokenize("This is a test" ) self.assertListEqual(__lowerCamelCase , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) _SCREAMING_SNAKE_CASE : List[str] = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", ".", ] , ) _SCREAMING_SNAKE_CASE : List[str] = tokenizer.convert_tokens_to_ids(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] ] , ) _SCREAMING_SNAKE_CASE : int = tokenizer.convert_ids_to_tokens(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", ".", ] , ) _SCREAMING_SNAKE_CASE : Union[str, Any] = tokenizer.vocab_size _SCREAMING_SNAKE_CASE : Optional[int] = [tokenizer.convert_ids_to_tokens(__lowerCamelCase ) for x in range(end - 7 , __lowerCamelCase )] self.assertListEqual( __lowerCamelCase , ["__java__", "__python__", "__en_XX__", "__javascript__", "__php__", "__ruby__", "__go__"] ) _SCREAMING_SNAKE_CASE : List[Any] = "java.lang.Exception, python.lang.Exception, javascript, php, ruby, go" _SCREAMING_SNAKE_CASE : int = tokenizer(__lowerCamelCase ).input_ids self.assertEqual( tokenizer.decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase , clean_up_tokenization_spaces=__lowerCamelCase ) , __lowerCamelCase , ) @require_torch @require_sentencepiece @require_tokenizers class lowerCAmelCase__( unittest.TestCase ): '''simple docstring''' __snake_case = 'uclanlp/plbart-python-en_XX' __snake_case = [ 'def maximum(a,b,c):NEW_LINE_INDENTreturn max([a,b,c])', 'def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])', ] __snake_case = [ 'Returns the maximum value of a b c.', 'Sums the values of a b c.', ] __snake_case = [ 1_3_4, 5_4_5_2, 3_3_4_6_0, 3_3_4_4_1, 3_3_4_6_3, 3_3_4_6_5, 3_3_4_6_3, 3_3_4_4_9, 9_8_8, 2_0, 3_3_4_5_6, 1_9, 3_3_4_5_6, 7_7_1, 3_9, 4_2_5_8, 8_8_9, 3_3_1_8, 3_3_4_4_1, 3_3_4_6_3, 3_3_4_6_5, 3_3_4_6_3, 3_3_4_4_9, 2_4_7_1, 2, PYTHON_CODE, ] @classmethod def UpperCamelCase_ ( cls ) -> Dict: _SCREAMING_SNAKE_CASE : PLBartTokenizer = PLBartTokenizer.from_pretrained( cls.checkpoint_name , language_codes="base" , src_lang="python" , tgt_lang="en_XX" ) _SCREAMING_SNAKE_CASE : Tuple = 1 return cls def UpperCamelCase_ ( self ) -> int: self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["__java__"] , 5_0_0_0_1 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["__python__"] , 5_0_0_0_2 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["__en_XX__"] , 5_0_0_0_3 ) def UpperCamelCase_ ( self ) -> Any: _SCREAMING_SNAKE_CASE : List[Any] = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , __lowerCamelCase ) def UpperCamelCase_ ( self ) -> Any: self.assertIn(__lowerCamelCase , self.tokenizer.all_special_ids ) _SCREAMING_SNAKE_CASE : Dict = [EN_CODE, 9_0_3_7, 3_3_4_4_2, 5_7, 7_5_2, 1_5_3, 1_4, 5_6, 1_8, 9, 2] _SCREAMING_SNAKE_CASE : int = self.tokenizer.decode(__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) _SCREAMING_SNAKE_CASE : int = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=__lowerCamelCase ) self.assertEqual(__lowerCamelCase , __lowerCamelCase ) self.assertNotIn(self.tokenizer.eos_token , __lowerCamelCase ) def UpperCamelCase_ ( self ) -> Any: _SCREAMING_SNAKE_CASE : str = ["def sum(a,b,c):NEW_LINE_INDENTreturn sum([a,b,c])" * 2_0] self.assertIsInstance(src_text[0] , __lowerCamelCase ) _SCREAMING_SNAKE_CASE : Any = 1_0 _SCREAMING_SNAKE_CASE : List[Any] = self.tokenizer(__lowerCamelCase , max_length=__lowerCamelCase , truncation=__lowerCamelCase ).input_ids[0] self.assertEqual(ids[-2] , 2 ) self.assertEqual(ids[-1] , __lowerCamelCase ) self.assertEqual(len(__lowerCamelCase ) , __lowerCamelCase ) def UpperCamelCase_ ( self ) -> str: self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "__java__"] ) , [5_0_0_0_4, 5_0_0_0_1] ) def UpperCamelCase_ ( self ) -> Union[str, Any]: _SCREAMING_SNAKE_CASE : Optional[Any] = tempfile.mkdtemp() _SCREAMING_SNAKE_CASE : Union[str, Any] = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(__lowerCamelCase ) _SCREAMING_SNAKE_CASE : Dict = PLBartTokenizer.from_pretrained(__lowerCamelCase ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , __lowerCamelCase ) @require_torch def UpperCamelCase_ ( self ) -> List[str]: _SCREAMING_SNAKE_CASE : Dict = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=__lowerCamelCase , return_tensors="pt" ) _SCREAMING_SNAKE_CASE : str = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 self.assertEqual(batch.input_ids[1][-2:].tolist() , [2, PYTHON_CODE] ) self.assertEqual(batch.decoder_input_ids[1][0] , __lowerCamelCase ) self.assertEqual(batch.decoder_input_ids[1][-1] , 2 ) self.assertEqual(batch.labels[1][-2:].tolist() , [2, EN_CODE] ) @require_torch def UpperCamelCase_ ( self ) -> Optional[Any]: _SCREAMING_SNAKE_CASE : Dict = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=__lowerCamelCase , truncation=__lowerCamelCase , max_length=len(self.expected_src_tokens ) , return_tensors="pt" , ) _SCREAMING_SNAKE_CASE : List[str] = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id ) self.assertIsInstance(__lowerCamelCase , __lowerCamelCase ) self.assertEqual((2, 2_6) , batch.input_ids.shape ) self.assertEqual((2, 2_6) , batch.attention_mask.shape ) _SCREAMING_SNAKE_CASE : str = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , __lowerCamelCase ) self.assertEqual(2 , batch.decoder_input_ids[0, -1] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id, PYTHON_CODE] ) def UpperCamelCase_ ( self ) -> Any: _SCREAMING_SNAKE_CASE : Tuple = self.tokenizer(self.src_text , padding=__lowerCamelCase , truncation=__lowerCamelCase , max_length=3 , return_tensors="pt" ) _SCREAMING_SNAKE_CASE : List[Any] = self.tokenizer( text_target=self.tgt_text , padding=__lowerCamelCase , truncation=__lowerCamelCase , max_length=1_0 , return_tensors="pt" ) _SCREAMING_SNAKE_CASE : List[Any] = targets["input_ids"] _SCREAMING_SNAKE_CASE : Any = shift_tokens_right(__lowerCamelCase , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 1_0 ) @require_torch def UpperCamelCase_ ( self ) -> Any: _SCREAMING_SNAKE_CASE : Dict = self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="en_XX" , tgt_lang="java" ) self.assertEqual( nested_simplify(__lowerCamelCase ) , { # A, test, EOS, en_XX "input_ids": [[1_5_0, 2_4_2, 2, 5_0_0_0_3]], "attention_mask": [[1, 1, 1, 1]], # java "forced_bos_token_id": 5_0_0_0_1, } , )
249
0
"""simple docstring""" import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __snake_case : List[Any] = logging.get_logger(__name__) __snake_case : Dict = { 'nvidia/segformer-b0-finetuned-ade-512-512': ( 'https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json' ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class A__ ( __lowerCamelCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = 'segformer' def __init__( self: Tuple , _SCREAMING_SNAKE_CASE: Optional[Any]=3 , _SCREAMING_SNAKE_CASE: Optional[int]=4 , _SCREAMING_SNAKE_CASE: Tuple=[2, 2, 2, 2] , _SCREAMING_SNAKE_CASE: List[str]=[8, 4, 2, 1] , _SCREAMING_SNAKE_CASE: Union[str, Any]=[32, 64, 160, 256] , _SCREAMING_SNAKE_CASE: Any=[7, 3, 3, 3] , _SCREAMING_SNAKE_CASE: Any=[4, 2, 2, 2] , _SCREAMING_SNAKE_CASE: Union[str, Any]=[1, 2, 5, 8] , _SCREAMING_SNAKE_CASE: Tuple=[4, 4, 4, 4] , _SCREAMING_SNAKE_CASE: str="gelu" , _SCREAMING_SNAKE_CASE: List[Any]=0.0 , _SCREAMING_SNAKE_CASE: int=0.0 , _SCREAMING_SNAKE_CASE: int=0.1 , _SCREAMING_SNAKE_CASE: List[str]=0.02 , _SCREAMING_SNAKE_CASE: Dict=0.1 , _SCREAMING_SNAKE_CASE: Dict=1e-6 , _SCREAMING_SNAKE_CASE: int=256 , _SCREAMING_SNAKE_CASE: Optional[int]=255 , **_SCREAMING_SNAKE_CASE: Tuple , ) -> Optional[Any]: """simple docstring""" super().__init__(**UpperCAmelCase_) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( "Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be" " removed, as the behaviour will default to that of reshape_last_stage = True." , UpperCAmelCase_ , ) __lowerCAmelCase : List[Any] = num_channels __lowerCAmelCase : Any = num_encoder_blocks __lowerCAmelCase : Dict = depths __lowerCAmelCase : int = sr_ratios __lowerCAmelCase : str = hidden_sizes __lowerCAmelCase : List[str] = patch_sizes __lowerCAmelCase : Optional[int] = strides __lowerCAmelCase : Dict = mlp_ratios __lowerCAmelCase : List[str] = num_attention_heads __lowerCAmelCase : int = hidden_act __lowerCAmelCase : Any = hidden_dropout_prob __lowerCAmelCase : str = attention_probs_dropout_prob __lowerCAmelCase : List[str] = classifier_dropout_prob __lowerCAmelCase : List[Any] = initializer_range __lowerCAmelCase : Union[str, Any] = drop_path_rate __lowerCAmelCase : int = layer_norm_eps __lowerCAmelCase : Dict = decoder_hidden_size __lowerCAmelCase : List[Any] = kwargs.get("reshape_last_stage" , UpperCAmelCase_) __lowerCAmelCase : List[str] = semantic_loss_ignore_index class A__ ( __lowerCamelCase ): '''simple docstring''' SCREAMING_SNAKE_CASE = version.parse('1.11' ) @property def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> str: """simple docstring""" return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ]) @property def _SCREAMING_SNAKE_CASE ( self: Optional[Any]) -> Optional[Any]: """simple docstring""" return 1e-4 @property def _SCREAMING_SNAKE_CASE ( self: Any) -> Optional[Any]: """simple docstring""" return 12
706
"""simple docstring""" import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging __snake_case : Tuple = logging.get_logger(__name__) def _lowercase ( __snake_case ,__snake_case ) -> Dict: __lowerCAmelCase : Optional[Any] = set() __lowerCAmelCase : List[Any] = [] def parse_line(__snake_case ): for line in fp: if isinstance(__snake_case ,__snake_case ): __lowerCAmelCase : Tuple = line.decode("UTF-8" ) if "warnings summary (final)" in line: continue # This means we are outside the body of a warning elif not line.startswith(" " ): # process a single warning and move it to `selected_warnings`. if len(__snake_case ) > 0: __lowerCAmelCase : List[Any] = "\n".join(__snake_case ) # Only keep the warnings specified in `targets` if any(F""": {x}: """ in warning for x in targets ): selected_warnings.add(__snake_case ) buffer.clear() continue else: __lowerCAmelCase : List[str] = line.strip() buffer.append(__snake_case ) if from_gh: for filename in os.listdir(__snake_case ): __lowerCAmelCase : List[str] = os.path.join(__snake_case ,__snake_case ) if not os.path.isdir(__snake_case ): # read the file if filename != "warnings.txt": continue with open(__snake_case ) as fp: parse_line(__snake_case ) else: try: with zipfile.ZipFile(__snake_case ) as z: for filename in z.namelist(): if not os.path.isdir(__snake_case ): # read the file if filename != "warnings.txt": continue with z.open(__snake_case ) as fp: parse_line(__snake_case ) except Exception: logger.warning( F"""{artifact_path} is either an invalid zip file or something else wrong. This file is skipped.""" ) return selected_warnings def _lowercase ( __snake_case ,__snake_case ) -> Any: __lowerCAmelCase : Any = set() __lowerCAmelCase : str = [os.path.join(__snake_case ,__snake_case ) for p in os.listdir(__snake_case ) if (p.endswith(".zip" ) or from_gh)] for p in paths: selected_warnings.update(extract_warnings_from_single_artifact(__snake_case ,__snake_case ) ) return selected_warnings if __name__ == "__main__": def _lowercase ( __snake_case ) -> Optional[Any]: return values.split("," ) __snake_case : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument('--workflow_run_id', type=str, required=True, help='A GitHub Actions workflow run id.') parser.add_argument( '--output_dir', type=str, required=True, help='Where to store the downloaded artifacts and other result files.', ) parser.add_argument('--token', default=None, type=str, help='A token that has actions:read permission.') # optional parameters parser.add_argument( '--targets', default='DeprecationWarning,UserWarning,FutureWarning', type=list_str, help='Comma-separated list of target warning(s) which we want to extract.', ) parser.add_argument( '--from_gh', action='store_true', help='If running from a GitHub action workflow and collecting warnings from its artifacts.', ) __snake_case : Tuple = parser.parse_args() __snake_case : Any = args.from_gh if from_gh: # The artifacts have to be downloaded using `actions/download-artifact@v3` pass else: os.makedirs(args.output_dir, exist_ok=True) # get download links __snake_case : str = get_artifacts_links(args.workflow_run_id, token=args.token) with open(os.path.join(args.output_dir, 'artifacts.json'), 'w', encoding='UTF-8') as fp: json.dump(artifacts, fp, ensure_ascii=False, indent=4) # download artifacts for idx, (name, url) in enumerate(artifacts.items()): print(name) print(url) print('=' * 80) download_artifact(name, url, args.output_dir, args.token) # Be gentle to GitHub time.sleep(1) # extract warnings from artifacts __snake_case : Tuple = extract_warnings(args.output_dir, args.targets) __snake_case : List[str] = sorted(selected_warnings) with open(os.path.join(args.output_dir, 'selected_warnings.json'), 'w', encoding='UTF-8') as fp: json.dump(selected_warnings, fp, ensure_ascii=False, indent=4)
615
0
import numpy as np class A : def __init__(self : Any ) -> List[str]: """simple docstring""" UpperCAmelCase__ = (0, 0) UpperCAmelCase__ = None UpperCAmelCase__ = 0 UpperCAmelCase__ = 0 UpperCAmelCase__ = 0 def __eq__(self : Tuple , __UpperCAmelCase : Optional[int] ) -> str: """simple docstring""" return self.position == cell.position def lowercase_ (self : str ) -> str: """simple docstring""" print(self.position ) class A : def __init__(self : str , __UpperCAmelCase : Tuple=(5, 5) ) -> int: """simple docstring""" UpperCAmelCase__ = np.zeros(__SCREAMING_SNAKE_CASE ) UpperCAmelCase__ = world_size[0] UpperCAmelCase__ = world_size[1] def lowercase_ (self : Union[str, Any] ) -> List[Any]: """simple docstring""" print(self.w ) def lowercase_ (self : List[Any] , __UpperCAmelCase : List[Any] ) -> Dict: """simple docstring""" UpperCAmelCase__ = [ (-1, -1), (-1, 0), (-1, 1), (0, -1), (0, 1), (1, -1), (1, 0), (1, 1), ] UpperCAmelCase__ = cell.position[0] UpperCAmelCase__ = cell.position[1] UpperCAmelCase__ = [] for n in neughbour_cord: UpperCAmelCase__ = current_x + n[0] UpperCAmelCase__ = current_y + n[1] if 0 <= x < self.world_x_limit and 0 <= y < self.world_y_limit: UpperCAmelCase__ = Cell() UpperCAmelCase__ = (x, y) UpperCAmelCase__ = cell neighbours.append(__SCREAMING_SNAKE_CASE ) return neighbours def lowerCAmelCase_ ( __A, __A, __A ) -> Any: '''simple docstring''' UpperCAmelCase__ = [] UpperCAmelCase__ = [] _open.append(a__ ) while _open: UpperCAmelCase__ = np.argmin([n.f for n in _open] ) UpperCAmelCase__ = _open[min_f] _closed.append(_open.pop(a__ ) ) if current == goal: break for n in world.get_neigbours(a__ ): for c in _closed: if c == n: continue UpperCAmelCase__ = current.g + 1 UpperCAmelCase__ , UpperCAmelCase__ = n.position UpperCAmelCase__ , UpperCAmelCase__ = goal.position UpperCAmelCase__ = (ya - ya) ** 2 + (xa - xa) ** 2 UpperCAmelCase__ = n.h + n.g for c in _open: if c == n and c.f < n.f: continue _open.append(a__ ) UpperCAmelCase__ = [] while current.parent is not None: path.append(current.position ) UpperCAmelCase__ = current.parent path.append(current.position ) return path[::-1] if __name__ == "__main__": UpperCamelCase__ = Gridworld() # Start position and goal UpperCamelCase__ = Cell() UpperCamelCase__ = (0, 0) UpperCamelCase__ = Cell() UpperCamelCase__ = (4, 4) print(f'''path from {start.position} to {goal.position}''') UpperCamelCase__ = astar(world, start, goal) # Just for visual reasons. for i in s: UpperCamelCase__ = 1 print(world.w)
486
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class lowerCAmelCase__ ( a ): """simple docstring""" lowerCAmelCase__ = ["image_processor", "tokenizer"] lowerCAmelCase__ = "CLIPImageProcessor" lowerCAmelCase__ = ("XLMRobertaTokenizer", "XLMRobertaTokenizerFast") def __init__( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Any=None , __SCREAMING_SNAKE_CASE : str=None , **__SCREAMING_SNAKE_CASE : Tuple ) -> Dict: """simple docstring""" __SCREAMING_SNAKE_CASE = None if "feature_extractor" in kwargs: warnings.warn( """The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`""" """ instead.""" , __SCREAMING_SNAKE_CASE , ) __SCREAMING_SNAKE_CASE = kwargs.pop("""feature_extractor""" ) __SCREAMING_SNAKE_CASE = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("""You need to specify an `image_processor`.""" ) if tokenizer is None: raise ValueError("""You need to specify a `tokenizer`.""" ) super().__init__(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def __call__( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : Union[str, Any]=None , __SCREAMING_SNAKE_CASE : Any=None , __SCREAMING_SNAKE_CASE : Optional[Any]=None , **__SCREAMING_SNAKE_CASE : str ) -> int: """simple docstring""" if text is None and images is None: raise ValueError("""You have to specify either text or images. Both cannot be none.""" ) if text is not None: __SCREAMING_SNAKE_CASE = self.tokenizer(__SCREAMING_SNAKE_CASE , return_tensors=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) if images is not None: __SCREAMING_SNAKE_CASE = self.image_processor(__SCREAMING_SNAKE_CASE , return_tensors=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) if text is not None and images is not None: __SCREAMING_SNAKE_CASE = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**__SCREAMING_SNAKE_CASE ) , tensor_type=__SCREAMING_SNAKE_CASE ) def UpperCAmelCase__ ( self : List[str] , *__SCREAMING_SNAKE_CASE : Tuple , **__SCREAMING_SNAKE_CASE : List[Any] ) -> Optional[int]: """simple docstring""" return self.tokenizer.batch_decode(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) def UpperCAmelCase__ ( self : List[Any] , *__SCREAMING_SNAKE_CASE : Union[str, Any] , **__SCREAMING_SNAKE_CASE : str ) -> Optional[Any]: """simple docstring""" return self.tokenizer.decode(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) @property def UpperCAmelCase__ ( self : List[Any] ) -> List[Any]: """simple docstring""" __SCREAMING_SNAKE_CASE = self.tokenizer.model_input_names __SCREAMING_SNAKE_CASE = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
627
0
import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMAEForPreTraining, ViTMAEModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class __SCREAMING_SNAKE_CASE : def __init__( self , __lowerCAmelCase , __lowerCAmelCase=13 , __lowerCAmelCase=30 , __lowerCAmelCase=2 , __lowerCAmelCase=3 , __lowerCAmelCase=True , __lowerCAmelCase=True , __lowerCAmelCase=32 , __lowerCAmelCase=5 , __lowerCAmelCase=4 , __lowerCAmelCase=37 , __lowerCAmelCase="gelu" , __lowerCAmelCase=0.1 , __lowerCAmelCase=0.1 , __lowerCAmelCase=10 , __lowerCAmelCase=0.02 , __lowerCAmelCase=3 , __lowerCAmelCase=0.6 , __lowerCAmelCase=None , ): UpperCamelCase__ = parent UpperCamelCase__ = batch_size UpperCamelCase__ = image_size UpperCamelCase__ = patch_size UpperCamelCase__ = num_channels UpperCamelCase__ = is_training UpperCamelCase__ = use_labels UpperCamelCase__ = hidden_size UpperCamelCase__ = num_hidden_layers UpperCamelCase__ = num_attention_heads UpperCamelCase__ = intermediate_size UpperCamelCase__ = hidden_act UpperCamelCase__ = hidden_dropout_prob UpperCamelCase__ = attention_probs_dropout_prob UpperCamelCase__ = type_sequence_label_size UpperCamelCase__ = initializer_range UpperCamelCase__ = mask_ratio UpperCamelCase__ = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) UpperCamelCase__ = (image_size // patch_size) ** 2 UpperCamelCase__ = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def _lowerCamelCase ( self ): UpperCamelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase__ = None if self.use_labels: UpperCamelCase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase__ = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCAmelCase , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): UpperCamelCase__ = ViTMAEModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() UpperCamelCase__ = model(__lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): UpperCamelCase__ = ViTMAEForPreTraining(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() UpperCamelCase__ = model(__lowerCAmelCase ) UpperCamelCase__ = (self.image_size // self.patch_size) ** 2 UpperCamelCase__ = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images UpperCamelCase__ = 1 UpperCamelCase__ = ViTMAEForPreTraining(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() UpperCamelCase__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCamelCase__ = model(__lowerCAmelCase ) UpperCamelCase__ = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def _lowerCamelCase ( self ): UpperCamelCase__ = self.prepare_config_and_inputs() UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = config_and_inputs UpperCamelCase__ = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( _a , _a , unittest.TestCase ): snake_case : Union[str, Any] = (ViTMAEModel, ViTMAEForPreTraining) if is_torch_available() else () snake_case : Optional[int] = {"""feature-extraction""": ViTMAEModel} if is_torch_available() else {} snake_case : Tuple = False snake_case : Optional[int] = False snake_case : int = False snake_case : Dict = False def _lowerCamelCase ( self ): UpperCamelCase__ = ViTMAEModelTester(self ) UpperCamelCase__ = ConfigTester(self , config_class=__lowerCAmelCase , has_text_modality=__lowerCAmelCase , hidden_size=37 ) def _lowerCamelCase ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="""ViTMAE does not use inputs_embeds""" ) def _lowerCamelCase ( self ): pass def _lowerCamelCase ( self ): UpperCamelCase__ , UpperCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase__ = model_class(__lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCamelCase__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCAmelCase , nn.Linear ) ) def _lowerCamelCase ( self ): UpperCamelCase__ , UpperCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase__ = model_class(__lowerCAmelCase ) UpperCamelCase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase__ = [*signature.parameters.keys()] UpperCamelCase__ = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , __lowerCAmelCase ) def _lowerCamelCase ( self ): UpperCamelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCAmelCase ) def _lowerCamelCase ( self ): UpperCamelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__lowerCAmelCase ) def _lowerCamelCase ( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ): # make masks reproducible np.random.seed(2 ) UpperCamelCase__ = int((pt_model.config.image_size // pt_model.config.patch_size) ** 2 ) UpperCamelCase__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) UpperCamelCase__ = torch.from_numpy(__lowerCAmelCase ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument UpperCamelCase__ = pt_noise super().check_pt_tf_models(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def _lowerCamelCase ( self ): UpperCamelCase__ , UpperCamelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase__ = model_class(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): UpperCamelCase__ = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) UpperCamelCase__ = outputs[0].cpu().numpy() UpperCamelCase__ = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__lowerCAmelCase ) UpperCamelCase__ = model_class.from_pretrained(__lowerCAmelCase ) model.to(__lowerCAmelCase ) # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): UpperCamelCase__ = model(**self._prepare_for_class(__lowerCAmelCase , __lowerCAmelCase ) ) # Make sure we don't have nans UpperCamelCase__ = after_outputs[0].cpu().numpy() UpperCamelCase__ = 0 UpperCamelCase__ = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(__lowerCAmelCase , 1E-5 ) @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.""" ) def _lowerCamelCase ( self ): pass @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.""" ) def _lowerCamelCase ( self ): pass @unittest.skip( reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load to get deterministic results.""" ) def _lowerCamelCase ( self ): pass @unittest.skip(reason="""ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load""" ) def _lowerCamelCase ( self ): pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def _lowerCamelCase ( self ): pass @slow def _lowerCamelCase ( self ): for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase__ = ViTMAEModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) def _UpperCamelCase (): """simple docstring""" UpperCamelCase__ = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self ): return ViTImageProcessor.from_pretrained("""facebook/vit-mae-base""" ) if is_vision_available() else None @slow def _lowerCamelCase ( self ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) UpperCamelCase__ = ViTMAEForPreTraining.from_pretrained("""facebook/vit-mae-base""" ).to(__lowerCAmelCase ) UpperCamelCase__ = self.default_image_processor UpperCamelCase__ = prepare_img() UpperCamelCase__ = image_processor(images=__lowerCAmelCase , return_tensors="""pt""" ).to(__lowerCAmelCase ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) UpperCamelCase__ = ViTMAEConfig() UpperCamelCase__ = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) UpperCamelCase__ = np.random.uniform(size=(1, num_patches) ) # forward pass with torch.no_grad(): UpperCamelCase__ = model(**__lowerCAmelCase , noise=torch.from_numpy(__lowerCAmelCase ).to(device=__lowerCAmelCase ) ) # verify the logits UpperCamelCase__ = torch.Size((1, 196, 768) ) self.assertEqual(outputs.logits.shape , __lowerCAmelCase ) UpperCamelCase__ = torch.tensor( [[-0.0548, -1.7023, -0.9325], [0.3721, -0.5670, -0.2233], [0.8235, -1.3878, -0.3524]] ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , expected_slice.to(__lowerCAmelCase ) , atol=1E-4 ) )
548
import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( """files""" , [ ["""full:README.md""", """dataset_infos.json"""], ["""empty:README.md""", """dataset_infos.json"""], ["""dataset_infos.json"""], ["""full:README.md"""], ] , ) def _UpperCamelCase (a__ :Any , a__ :Union[str, Any] ): """simple docstring""" UpperCamelCase__ = tmp_path_factory.mktemp("""dset_infos_dir""" ) if "full:README.md" in files: with open(dataset_infos_dir / """README.md""" , """w""" ) as f: f.write("""---\ndataset_info:\n dataset_size: 42\n---""" ) if "empty:README.md" in files: with open(dataset_infos_dir / """README.md""" , """w""" ) as f: f.write("""""" ) # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / """dataset_infos.json""" , """w""" ) as f: f.write("""{\"default\": {\"dataset_size\": 42}}""" ) UpperCamelCase__ = DatasetInfosDict.from_directory(a__ ) assert dataset_infos assert dataset_infos["default"].dataset_size == 42 @pytest.mark.parametrize( """dataset_info""" , [ DatasetInfo(), DatasetInfo( description="""foo""" , features=Features({"""a""": Value("""int32""" )} ) , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train"""}] , download_size=42 , ), ] , ) def _UpperCamelCase (a__ :Optional[int] , a__ :DatasetInfo ): """simple docstring""" UpperCamelCase__ = str(a__ ) dataset_info.write_to_directory(a__ ) UpperCamelCase__ = DatasetInfo.from_directory(a__ ) assert dataset_info == reloaded assert os.path.exists(os.path.join(a__ , """dataset_info.json""" ) ) def _UpperCamelCase (): """simple docstring""" UpperCamelCase__ = DatasetInfo( description="""foo""" , citation="""bar""" , homepage="""https://foo.bar""" , license="""CC0""" , features=Features({"""a""": Value("""int32""" )} ) , post_processed={} , supervised_keys=() , task_templates=[] , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train""", """num_examples""": 42}] , download_checksums={} , download_size=1337 , post_processing_size=442 , dataset_size=1234 , size_in_bytes=1337 + 442 + 1234 , ) UpperCamelCase__ = dataset_info._to_yaml_dict() assert sorted(a__ ) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML ) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key] , (list, dict, int, str) ) UpperCamelCase__ = yaml.safe_dump(a__ ) UpperCamelCase__ = yaml.safe_load(a__ ) assert dataset_info_yaml_dict == reloaded def _UpperCamelCase (): """simple docstring""" UpperCamelCase__ = DatasetInfo() UpperCamelCase__ = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( """dataset_infos_dict""" , [ DatasetInfosDict(), DatasetInfosDict({"""default""": DatasetInfo()} ), DatasetInfosDict({"""my_config_name""": DatasetInfo()} ), DatasetInfosDict( { """default""": DatasetInfo( description="""foo""" , features=Features({"""a""": Value("""int32""" )} ) , builder_name="""builder""" , config_name="""config""" , version="""1.0.0""" , splits=[{"""name""": """train"""}] , download_size=42 , ) } ), DatasetInfosDict( { """v1""": DatasetInfo(dataset_size=42 ), """v2""": DatasetInfo(dataset_size=1337 ), } ), ] , ) def _UpperCamelCase (a__ :int , a__ :DatasetInfosDict ): """simple docstring""" UpperCamelCase__ = str(a__ ) dataset_infos_dict.write_to_directory(a__ ) UpperCamelCase__ = DatasetInfosDict.from_directory(a__ ) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): UpperCamelCase__ = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml UpperCamelCase__ = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict() ) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(a__ , """README.md""" ) )
548
1
def lowercase ( __A : int ) -> "list[int]": '''simple docstring''' if upper_limit < 0: raise ValueError("""Limit for the Catalan sequence must be ≥ 0""" ) snake_case : Dict = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 snake_case : Dict = 1 if upper_limit > 0: snake_case : Optional[Any] = 1 # Recurrence relation: C(i) = sum(C(j).C(i-j-1)), from j = 0 to i for i in range(2 , upper_limit + 1 ): for j in range(__A ): catalan_list[i] += catalan_list[j] * catalan_list[i - j - 1] return catalan_list if __name__ == "__main__": print('''\n********* Catalan Numbers Using Dynamic Programming ************\n''') print('''\n*** Enter -1 at any time to quit ***''') print('''\nEnter the upper limit (≥ 0) for the Catalan number sequence: ''', end='''''') try: while True: __lowercase : Union[str, Any] = int(input().strip()) if N < 0: print('''\n********* Goodbye!! ************''') break else: print(f'''The Catalan numbers from 0 through {N} are:''') print(catalan_numbers(N)) print('''Try another upper limit for the sequence: ''', end='''''') except (NameError, ValueError): print('''\n********* Invalid input, goodbye! ************\n''') import doctest doctest.testmod()
36
'''simple docstring''' from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING UpperCamelCase =logging.get_logger(__name__) @add_end_docstrings(SCREAMING_SNAKE_CASE__ ) class A ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" def __init__( self , **__lowerCAmelCase ): super().__init__(**__lowerCAmelCase ) requires_backends(self , """vision""" ) requires_backends(self , """torch""" ) if self.framework != "pt": raise ValueError(F"The {self.__class__} is only available in PyTorch." ) self.check_model_type(__lowerCAmelCase ) def _UpperCAmelCase ( self , **__lowerCAmelCase ): UpperCamelCase_ : int = {} UpperCamelCase_ : int = {} UpperCamelCase_ : Any = {} # preprocess args if "points_per_batch" in kwargs: UpperCamelCase_ : Dict = kwargs["""points_per_batch"""] if "points_per_crop" in kwargs: UpperCamelCase_ : str = kwargs["""points_per_crop"""] if "crops_n_layers" in kwargs: UpperCamelCase_ : Any = kwargs["""crops_n_layers"""] if "crop_overlap_ratio" in kwargs: UpperCamelCase_ : str = kwargs["""crop_overlap_ratio"""] if "crop_n_points_downscale_factor" in kwargs: UpperCamelCase_ : Tuple = kwargs["""crop_n_points_downscale_factor"""] # postprocess args if "pred_iou_thresh" in kwargs: UpperCamelCase_ : Any = kwargs["""pred_iou_thresh"""] if "stability_score_offset" in kwargs: UpperCamelCase_ : List[str] = kwargs["""stability_score_offset"""] if "mask_threshold" in kwargs: UpperCamelCase_ : Tuple = kwargs["""mask_threshold"""] if "stability_score_thresh" in kwargs: UpperCamelCase_ : Optional[Any] = kwargs["""stability_score_thresh"""] if "crops_nms_thresh" in kwargs: UpperCamelCase_ : Dict = kwargs["""crops_nms_thresh"""] if "output_rle_mask" in kwargs: UpperCamelCase_ : Optional[int] = kwargs["""output_rle_mask"""] if "output_bboxes_mask" in kwargs: UpperCamelCase_ : List[Any] = kwargs["""output_bboxes_mask"""] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self , __lowerCAmelCase , *__lowerCAmelCase , __lowerCAmelCase=None , __lowerCAmelCase=None , **__lowerCAmelCase ): return super().__call__(__lowerCAmelCase , *__lowerCAmelCase , num_workers=__lowerCAmelCase , batch_size=__lowerCAmelCase , **__lowerCAmelCase ) def _UpperCAmelCase ( self , __lowerCAmelCase , __lowerCAmelCase=64 , __lowerCAmelCase = 0 , __lowerCAmelCase = 5_12 / 15_00 , __lowerCAmelCase = 32 , __lowerCAmelCase = 1 , ): UpperCamelCase_ : Tuple = load_image(__lowerCAmelCase ) UpperCamelCase_ : Tuple = self.image_processor.size["""longest_edge"""] UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ : Dict = self.image_processor.generate_crop_boxes( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) UpperCamelCase_ : Any = self.image_processor(images=__lowerCAmelCase , return_tensors="""pt""" ) with self.device_placement(): if self.framework == "pt": UpperCamelCase_ : str = self.get_inference_context() with inference_context(): UpperCamelCase_ : Optional[int] = self._ensure_tensor_on_device(__lowerCAmelCase , device=self.device ) UpperCamelCase_ : Tuple = self.model.get_image_embeddings(model_inputs.pop("""pixel_values""" ) ) UpperCamelCase_ : List[Any] = image_embeddings UpperCamelCase_ : Any = grid_points.shape[1] UpperCamelCase_ : List[str] = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( """Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. """ """To return all points at once, set points_per_batch to None""" ) for i in range(0 , __lowerCAmelCase , __lowerCAmelCase ): UpperCamelCase_ : Dict = grid_points[:, i : i + points_per_batch, :, :] UpperCamelCase_ : List[Any] = input_labels[:, i : i + points_per_batch] UpperCamelCase_ : Dict = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def _UpperCAmelCase ( self , __lowerCAmelCase , __lowerCAmelCase=0.88 , __lowerCAmelCase=0.95 , __lowerCAmelCase=0 , __lowerCAmelCase=1 , ): UpperCamelCase_ : Union[str, Any] = model_inputs.pop("""input_boxes""" ) UpperCamelCase_ : Tuple = model_inputs.pop("""is_last""" ) UpperCamelCase_ : Dict = model_inputs.pop("""original_sizes""" ).tolist() UpperCamelCase_ : int = model_inputs.pop("""reshaped_input_sizes""" ).tolist() UpperCamelCase_ : Optional[Any] = self.model(**__lowerCAmelCase ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks UpperCamelCase_ : List[str] = model_outputs["""pred_masks"""] UpperCamelCase_ : Optional[int] = self.image_processor.post_process_masks( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , binarize=__lowerCAmelCase ) UpperCamelCase_ : List[str] = model_outputs["""iou_scores"""] UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ : Optional[Any] = self.image_processor.filter_masks( masks[0] , iou_scores[0] , original_sizes[0] , input_boxes[0] , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def _UpperCAmelCase ( self , __lowerCAmelCase , __lowerCAmelCase=False , __lowerCAmelCase=False , __lowerCAmelCase=0.7 , ): UpperCamelCase_ : Tuple = [] UpperCamelCase_ : Optional[Any] = [] UpperCamelCase_ : Tuple = [] for model_output in model_outputs: all_scores.append(model_output.pop("""iou_scores""" ) ) all_masks.extend(model_output.pop("""masks""" ) ) all_boxes.append(model_output.pop("""boxes""" ) ) UpperCamelCase_ : Any = torch.cat(__lowerCAmelCase ) UpperCamelCase_ : Union[str, Any] = torch.cat(__lowerCAmelCase ) UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ : Any = self.image_processor.post_process_for_mask_generation( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) UpperCamelCase_ : List[Any] = defaultdict(__lowerCAmelCase ) for output in model_outputs: for k, v in output.items(): extra[k].append(__lowerCAmelCase ) UpperCamelCase_ : Optional[Any] = {} if output_rle_mask: UpperCamelCase_ : int = rle_mask if output_bboxes_mask: UpperCamelCase_ : int = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
208
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _A : Any = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A : Any = ['GPTSw3Tokenizer'] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys _A : Optional[int] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
130
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _A : Optional[Any] = logging.get_logger(__name__) def _a ( UpperCAmelCase ) -> Any: """simple docstring""" lowerCamelCase__ : Tuple = DPTConfig(embedding_type='''hybrid''' ) if "large" in checkpoint_url: lowerCamelCase__ : Tuple = 1024 lowerCamelCase__ : Any = 4096 lowerCamelCase__ : Optional[Any] = 24 lowerCamelCase__ : Dict = 16 lowerCamelCase__ : Optional[Any] = [5, 11, 17, 23] lowerCamelCase__ : str = [256, 512, 1024, 1024] lowerCamelCase__ : List[str] = (1, 384, 384) if "nyu" or "midas" in checkpoint_url: lowerCamelCase__ : List[str] = 768 lowerCamelCase__ : Any = [1, 1, 1, 0.5] lowerCamelCase__ : Dict = [256, 512, 768, 768] lowerCamelCase__ : Dict = 150 lowerCamelCase__ : str = 16 lowerCamelCase__ : List[Any] = (1, 384, 384) lowerCamelCase__ : Any = False lowerCamelCase__ : int = '''project''' if "ade" in checkpoint_url: lowerCamelCase__ : Optional[int] = True lowerCamelCase__ : List[Any] = 768 lowerCamelCase__ : int = [1, 1, 1, 0.5] lowerCamelCase__ : Any = 150 lowerCamelCase__ : Dict = 16 lowerCamelCase__ : Optional[Any] = '''huggingface/label-files''' lowerCamelCase__ : Any = '''ade20k-id2label.json''' lowerCamelCase__ : Optional[Any] = json.load(open(cached_download(hf_hub_url(UpperCAmelCase , UpperCAmelCase , repo_type='''dataset''' ) ) , '''r''' ) ) lowerCamelCase__ : Any = {int(UpperCAmelCase ): v for k, v in idalabel.items()} lowerCamelCase__ : Any = idalabel lowerCamelCase__ : List[Any] = {v: k for k, v in idalabel.items()} lowerCamelCase__ : Optional[Any] = [1, 150, 480, 480] return config, expected_shape def _a ( UpperCAmelCase ) -> Optional[Any]: """simple docstring""" lowerCamelCase__ : Optional[int] = ['''pretrained.model.head.weight''', '''pretrained.model.head.bias'''] for k in ignore_keys: state_dict.pop(UpperCAmelCase , UpperCAmelCase ) def _a ( UpperCAmelCase ) -> Dict: """simple docstring""" if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): lowerCamelCase__ : Optional[int] = name.replace('''pretrained.model''' , '''dpt.encoder''' ) if "pretrained.model" in name: lowerCamelCase__ : Tuple = name.replace('''pretrained.model''' , '''dpt.embeddings''' ) if "patch_embed" in name: lowerCamelCase__ : int = name.replace('''patch_embed''' , '''''' ) if "pos_embed" in name: lowerCamelCase__ : List[Any] = name.replace('''pos_embed''' , '''position_embeddings''' ) if "attn.proj" in name: lowerCamelCase__ : str = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "proj" in name and "project" not in name: lowerCamelCase__ : Any = name.replace('''proj''' , '''projection''' ) if "blocks" in name: lowerCamelCase__ : List[Any] = name.replace('''blocks''' , '''layer''' ) if "mlp.fc1" in name: lowerCamelCase__ : Dict = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: lowerCamelCase__ : List[Any] = name.replace('''mlp.fc2''' , '''output.dense''' ) if "norm1" in name and "backbone" not in name: lowerCamelCase__ : List[str] = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name and "backbone" not in name: lowerCamelCase__ : Any = name.replace('''norm2''' , '''layernorm_after''' ) if "scratch.output_conv" in name: lowerCamelCase__ : Tuple = name.replace('''scratch.output_conv''' , '''head''' ) if "scratch" in name: lowerCamelCase__ : int = name.replace('''scratch''' , '''neck''' ) if "layer1_rn" in name: lowerCamelCase__ : Any = name.replace('''layer1_rn''' , '''convs.0''' ) if "layer2_rn" in name: lowerCamelCase__ : Union[str, Any] = name.replace('''layer2_rn''' , '''convs.1''' ) if "layer3_rn" in name: lowerCamelCase__ : Optional[int] = name.replace('''layer3_rn''' , '''convs.2''' ) if "layer4_rn" in name: lowerCamelCase__ : Tuple = name.replace('''layer4_rn''' , '''convs.3''' ) if "refinenet" in name: lowerCamelCase__ : Optional[Any] = int(name[len('''neck.refinenet''' ) : len('''neck.refinenet''' ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 lowerCamelCase__ : List[str] = name.replace(f"refinenet{layer_idx}" , f"fusion_stage.layers.{abs(layer_idx-4 )}" ) if "out_conv" in name: lowerCamelCase__ : str = name.replace('''out_conv''' , '''projection''' ) if "resConfUnit1" in name: lowerCamelCase__ : List[str] = name.replace('''resConfUnit1''' , '''residual_layer1''' ) if "resConfUnit2" in name: lowerCamelCase__ : Optional[int] = name.replace('''resConfUnit2''' , '''residual_layer2''' ) if "conv1" in name: lowerCamelCase__ : int = name.replace('''conv1''' , '''convolution1''' ) if "conv2" in name: lowerCamelCase__ : List[str] = name.replace('''conv2''' , '''convolution2''' ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: lowerCamelCase__ : Optional[Any] = name.replace('''pretrained.act_postprocess1.0.project.0''' , '''neck.reassemble_stage.readout_projects.0.0''' ) if "pretrained.act_postprocess2.0.project.0" in name: lowerCamelCase__ : Dict = name.replace('''pretrained.act_postprocess2.0.project.0''' , '''neck.reassemble_stage.readout_projects.1.0''' ) if "pretrained.act_postprocess3.0.project.0" in name: lowerCamelCase__ : str = name.replace('''pretrained.act_postprocess3.0.project.0''' , '''neck.reassemble_stage.readout_projects.2.0''' ) if "pretrained.act_postprocess4.0.project.0" in name: lowerCamelCase__ : Optional[int] = name.replace('''pretrained.act_postprocess4.0.project.0''' , '''neck.reassemble_stage.readout_projects.3.0''' ) # resize blocks if "pretrained.act_postprocess1.3" in name: lowerCamelCase__ : Dict = name.replace('''pretrained.act_postprocess1.3''' , '''neck.reassemble_stage.layers.0.projection''' ) if "pretrained.act_postprocess1.4" in name: lowerCamelCase__ : str = name.replace('''pretrained.act_postprocess1.4''' , '''neck.reassemble_stage.layers.0.resize''' ) if "pretrained.act_postprocess2.3" in name: lowerCamelCase__ : int = name.replace('''pretrained.act_postprocess2.3''' , '''neck.reassemble_stage.layers.1.projection''' ) if "pretrained.act_postprocess2.4" in name: lowerCamelCase__ : Union[str, Any] = name.replace('''pretrained.act_postprocess2.4''' , '''neck.reassemble_stage.layers.1.resize''' ) if "pretrained.act_postprocess3.3" in name: lowerCamelCase__ : Optional[int] = name.replace('''pretrained.act_postprocess3.3''' , '''neck.reassemble_stage.layers.2.projection''' ) if "pretrained.act_postprocess4.3" in name: lowerCamelCase__ : Dict = name.replace('''pretrained.act_postprocess4.3''' , '''neck.reassemble_stage.layers.3.projection''' ) if "pretrained.act_postprocess4.4" in name: lowerCamelCase__ : Tuple = name.replace('''pretrained.act_postprocess4.4''' , '''neck.reassemble_stage.layers.3.resize''' ) if "pretrained" in name: lowerCamelCase__ : Any = name.replace('''pretrained''' , '''dpt''' ) if "bn" in name: lowerCamelCase__ : List[str] = name.replace('''bn''' , '''batch_norm''' ) if "head" in name: lowerCamelCase__ : Optional[Any] = name.replace('''head''' , '''head.head''' ) if "encoder.norm" in name: lowerCamelCase__ : List[Any] = name.replace('''encoder.norm''' , '''layernorm''' ) if "auxlayer" in name: lowerCamelCase__ : List[str] = name.replace('''auxlayer''' , '''auxiliary_head.head''' ) if "backbone" in name: lowerCamelCase__ : Union[str, Any] = name.replace('''backbone''' , '''backbone.bit.encoder''' ) if ".." in name: lowerCamelCase__ : Optional[Any] = name.replace('''..''' , '''.''' ) if "stem.conv" in name: lowerCamelCase__ : str = name.replace('''stem.conv''' , '''bit.embedder.convolution''' ) if "blocks" in name: lowerCamelCase__ : List[Any] = name.replace('''blocks''' , '''layers''' ) if "convolution" in name and "backbone" in name: lowerCamelCase__ : Tuple = name.replace('''convolution''' , '''conv''' ) if "layer" in name and "backbone" in name: lowerCamelCase__ : Union[str, Any] = name.replace('''layer''' , '''layers''' ) if "backbone.bit.encoder.bit" in name: lowerCamelCase__ : Union[str, Any] = name.replace('''backbone.bit.encoder.bit''' , '''backbone.bit''' ) if "embedder.conv" in name: lowerCamelCase__ : int = name.replace('''embedder.conv''' , '''embedder.convolution''' ) if "backbone.bit.encoder.stem.norm" in name: lowerCamelCase__ : int = name.replace('''backbone.bit.encoder.stem.norm''' , '''backbone.bit.embedder.norm''' ) return name def _a ( UpperCAmelCase , UpperCAmelCase ) -> Tuple: """simple docstring""" for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) lowerCamelCase__ : Dict = state_dict.pop(f"dpt.encoder.layer.{i}.attn.qkv.weight" ) lowerCamelCase__ : Optional[int] = state_dict.pop(f"dpt.encoder.layer.{i}.attn.qkv.bias" ) # next, add query, keys and values (in that order) to the state dict lowerCamelCase__ : int = in_proj_weight[: config.hidden_size, :] lowerCamelCase__ : str = in_proj_bias[: config.hidden_size] lowerCamelCase__ : int = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] lowerCamelCase__ : Union[str, Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] lowerCamelCase__ : Union[str, Any] = in_proj_weight[ -config.hidden_size :, : ] lowerCamelCase__ : List[Any] = in_proj_bias[-config.hidden_size :] def _a ( ) -> str: """simple docstring""" lowerCamelCase__ : Tuple = '''http://images.cocodataset.org/val2017/000000039769.jpg''' lowerCamelCase__ : Optional[Any] = Image.open(requests.get(UpperCAmelCase , stream=UpperCAmelCase ).raw ) return im @torch.no_grad() def _a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Optional[Any]: """simple docstring""" lowerCamelCase__ , lowerCamelCase__ : str = get_dpt_config(UpperCAmelCase ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") lowerCamelCase__ : int = torch.load(UpperCAmelCase , map_location='''cpu''' ) # remove certain keys remove_ignore_keys_(UpperCAmelCase ) # rename keys for key in state_dict.copy().keys(): lowerCamelCase__ : Union[str, Any] = state_dict.pop(UpperCAmelCase ) lowerCamelCase__ : List[str] = val # read in qkv matrices read_in_q_k_v(UpperCAmelCase , UpperCAmelCase ) # load HuggingFace model lowerCamelCase__ : Optional[Any] = DPTForSemanticSegmentation(UpperCAmelCase ) if '''ade''' in checkpoint_url else DPTForDepthEstimation(UpperCAmelCase ) model.load_state_dict(UpperCAmelCase ) model.eval() # Check outputs on an image lowerCamelCase__ : List[str] = 480 if '''ade''' in checkpoint_url else 384 lowerCamelCase__ : List[Any] = DPTImageProcessor(size=UpperCAmelCase ) lowerCamelCase__ : Optional[int] = prepare_img() lowerCamelCase__ : List[str] = image_processor(UpperCAmelCase , return_tensors='''pt''' ) # forward pass lowerCamelCase__ : Tuple = model(**UpperCAmelCase ).logits if '''ade''' in checkpoint_url else model(**UpperCAmelCase ).predicted_depth if show_prediction: lowerCamelCase__ : Union[str, Any] = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode='''bicubic''' , align_corners=UpperCAmelCase , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 255 ).show() if pytorch_dump_folder_path is not None: Path(UpperCAmelCase ).mkdir(exist_ok=UpperCAmelCase ) print(f"Saving model to {pytorch_dump_folder_path}" ) model.save_pretrained(UpperCAmelCase ) print(f"Saving image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(UpperCAmelCase ) if push_to_hub: model.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) image_processor.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) if __name__ == "__main__": _A : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt', type=str, help='URL of the original DPT checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=False, help='Path to the output PyTorch model directory.', ) parser.add_argument( '--push_to_hub', action='store_true', ) parser.add_argument( '--model_name', default='dpt-large', type=str, help='Name of the model, in case you\'re pushing to the hub.', ) parser.add_argument( '--show_prediction', action='store_true', ) _A : List[Any] = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
130
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json''' ), } class lowerCAmelCase ( __lowerCAmelCase ): lowerCAmelCase_ = '''dpr''' def __init__( self : Tuple , __lowercase : Optional[Any]=30522 , __lowercase : str=768 , __lowercase : List[Any]=12 , __lowercase : Any=12 , __lowercase : Optional[int]=3072 , __lowercase : Dict="gelu" , __lowercase : Optional[Any]=0.1 , __lowercase : Dict=0.1 , __lowercase : Optional[Any]=512 , __lowercase : Dict=2 , __lowercase : Optional[int]=0.0_2 , __lowercase : int=1E-12 , __lowercase : Tuple=0 , __lowercase : Union[str, Any]="absolute" , __lowercase : Dict = 0 , **__lowercase : Dict , ): """simple docstring""" super().__init__(pad_token_id=lowerCAmelCase_ , **lowerCAmelCase_ ) __lowercase =vocab_size __lowercase =hidden_size __lowercase =num_hidden_layers __lowercase =num_attention_heads __lowercase =hidden_act __lowercase =intermediate_size __lowercase =hidden_dropout_prob __lowercase =attention_probs_dropout_prob __lowercase =max_position_embeddings __lowercase =type_vocab_size __lowercase =initializer_range __lowercase =layer_norm_eps __lowercase =projection_dim __lowercase =position_embedding_type
119
import argparse import os import torch from transformers import FlavaConfig, FlavaForPreTraining from transformers.models.flava.convert_dalle_to_flava_codebook import convert_dalle_checkpoint def snake_case ( snake_case__ :Any) -> Union[str, Any]: # encoder.embeddings are double copied in original FLAVA return sum(param.float().sum() if """encoder.embeddings""" not in key else 0 for key, param in state_dict.items()) def snake_case ( snake_case__ :Tuple , snake_case__ :Dict) -> Union[str, Any]: _A = {} for key, value in state_dict.items(): if "text_encoder.embeddings" in key or "image_encoder.embeddings" in key: continue _A = key.replace("""heads.cmd.mim_head.cls.predictions""" , """mmm_image_head""") _A = key.replace("""heads.cmd.mlm_head.cls.predictions""" , """mmm_text_head""") _A = key.replace("""heads.cmd.itm_head.cls""" , """itm_head""") _A = key.replace("""heads.cmd.itm_head.pooler""" , """itm_head.pooler""") _A = key.replace("""heads.cmd.clip_head.logit_scale""" , """flava.logit_scale""") _A = key.replace("""heads.fairseq_mlm.cls.predictions""" , """mlm_head""") _A = key.replace("""heads.imagenet.mim_head.cls.predictions""" , """mim_head""") _A = key.replace("""mm_text_projection""" , """flava.text_to_mm_projection""") _A = key.replace("""mm_image_projection""" , """flava.image_to_mm_projection""") _A = key.replace("""image_encoder.module""" , """flava.image_model""") _A = key.replace("""text_encoder.module""" , """flava.text_model""") _A = key.replace("""mm_encoder.module.encoder.cls_token""" , """flava.multimodal_model.cls_token""") _A = key.replace("""mm_encoder.module""" , """flava.multimodal_model""") _A = key.replace("""text_projection""" , """flava.text_projection""") _A = key.replace("""image_projection""" , """flava.image_projection""") _A = value.float() for key, value in codebook_state_dict.items(): _A = value return upgrade @torch.no_grad() def snake_case ( snake_case__ :str , snake_case__ :str , snake_case__ :List[Any] , snake_case__ :Dict=None) -> Any: if config_path is not None: _A = FlavaConfig.from_pretrained(snake_case__) else: _A = FlavaConfig() _A = FlavaForPreTraining(snake_case__).eval() _A = convert_dalle_checkpoint(snake_case__ , snake_case__ , save_checkpoint=snake_case__) if os.path.exists(snake_case__): _A = torch.load(snake_case__ , map_location="""cpu""") else: _A = torch.hub.load_state_dict_from_url(snake_case__ , map_location="""cpu""") _A = upgrade_state_dict(snake_case__ , snake_case__) hf_model.load_state_dict(snake_case__) _A = hf_model.state_dict() _A = count_parameters(snake_case__) _A = count_parameters(snake_case__) + count_parameters(snake_case__) assert torch.allclose(snake_case__ , snake_case__ , atol=1E-3) hf_model.save_pretrained(snake_case__) if __name__ == "__main__": _SCREAMING_SNAKE_CASE = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to flava checkpoint') parser.add_argument('--codebook_path', default=None, type=str, help='Path to flava codebook checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') _SCREAMING_SNAKE_CASE = parser.parse_args() convert_flava_checkpoint(args.checkpoint_path, args.codebook_path, args.pytorch_dump_folder_path, args.config_path)
401
0
from __future__ import annotations def snake_case__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->list: UpperCAmelCase__ = [] UpperCAmelCase__ , UpperCAmelCase__ = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) UpperCAmelCase__ = result + left + right return input_list def snake_case__ ( _SCREAMING_SNAKE_CASE ) ->list: if len(_SCREAMING_SNAKE_CASE ) <= 1: return input_list UpperCAmelCase__ = list(_SCREAMING_SNAKE_CASE ) # iteration for two-way merging UpperCAmelCase__ = 2 while p <= len(_SCREAMING_SNAKE_CASE ): # getting low, high and middle value for merge-sort of single list for i in range(0 , len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ): UpperCAmelCase__ = i UpperCAmelCase__ = i + p - 1 UpperCAmelCase__ = (low + high + 1) // 2 UpperCAmelCase__ = merge(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # final merge of last two parts if p * 2 >= len(_SCREAMING_SNAKE_CASE ): UpperCAmelCase__ = i UpperCAmelCase__ = merge(_SCREAMING_SNAKE_CASE , 0 , _SCREAMING_SNAKE_CASE , len(_SCREAMING_SNAKE_CASE ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": a : Dict = input('''Enter numbers separated by a comma:\n''').strip() if user_input == "": a : List[Any] = [] else: a : List[Any] = [int(item.strip()) for item in user_input.split(''',''')] print(iter_merge_sort(unsorted))
701
"""simple docstring""" import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def snake_case__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ) ->str: UpperCAmelCase__ = OmegaConf.load(_SCREAMING_SNAKE_CASE ) if display: print(yaml.dump(OmegaConf.to_container(_SCREAMING_SNAKE_CASE ) ) ) return config def snake_case__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None ) ->List[Any]: if conf_path is None: UpperCAmelCase__ = """./model_checkpoints/vqgan_only.yaml""" UpperCAmelCase__ = load_config(_SCREAMING_SNAKE_CASE , display=_SCREAMING_SNAKE_CASE ) UpperCAmelCase__ = VQModel(**config.model.params ) if ckpt_path is None: UpperCAmelCase__ = """./model_checkpoints/vqgan_only.pt""" UpperCAmelCase__ = torch.load(_SCREAMING_SNAKE_CASE , map_location=_SCREAMING_SNAKE_CASE ) if ".ckpt" in ckpt_path: UpperCAmelCase__ = sd["""state_dict"""] model.load_state_dict(_SCREAMING_SNAKE_CASE , strict=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) del sd return model def snake_case__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->str: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = model.encode(_SCREAMING_SNAKE_CASE ) print(F'''VQGAN --- {model.__class__.__name__}: latent shape: {z.shape[2:]}''' ) UpperCAmelCase__ = model.decode(_SCREAMING_SNAKE_CASE ) return xrec def snake_case__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ) ->int: UpperCAmelCase__ , UpperCAmelCase__ = string.rsplit(""".""" , 1 ) if reload: UpperCAmelCase__ = importlib.import_module(_SCREAMING_SNAKE_CASE ) importlib.reload(_SCREAMING_SNAKE_CASE ) return getattr(importlib.import_module(_SCREAMING_SNAKE_CASE , package=_SCREAMING_SNAKE_CASE ) , cls ) def snake_case__ ( _SCREAMING_SNAKE_CASE ) ->str: if "target" not in config: raise KeyError("""Expected key `target` to instantiate.""" ) return get_obj_from_str(config["""target"""] )(**config.get("""params""" , {} ) ) def snake_case__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True ) ->str: UpperCAmelCase__ = instantiate_from_config(_SCREAMING_SNAKE_CASE ) if sd is not None: model.load_state_dict(_SCREAMING_SNAKE_CASE ) if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def snake_case__ ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->List[str]: # load the specified checkpoint if ckpt: UpperCAmelCase__ = torch.load(_SCREAMING_SNAKE_CASE , map_location="""cpu""" ) UpperCAmelCase__ = pl_sd["""global_step"""] print(F'''loaded model from global step {global_step}.''' ) else: UpperCAmelCase__ = {"""state_dict""": None} UpperCAmelCase__ = None UpperCAmelCase__ = load_model_from_config(config.model , pl_sd["""state_dict"""] , gpu=_SCREAMING_SNAKE_CASE , eval_mode=_SCREAMING_SNAKE_CASE )["""model"""] return model, global_step
422
0
'''simple docstring''' import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class UpperCAmelCase : '''simple docstring''' @staticmethod def _lowerCAmelCase( *__lowerCAmelCase , **__lowerCAmelCase ) -> Optional[Any]: pass @is_pipeline_test @require_vision class UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @require_torch def _lowerCAmelCase( self ) -> str: lowercase__ : List[str] = pipeline( model='''hf-internal-testing/tiny-random-clip-zero-shot-image-classification''' , ) lowercase__ : Union[str, Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) lowercase__ : Any = image_classifier(__lowerCAmelCase , candidate_labels=['''a''', '''b''', '''c'''] ) # The floating scores are so close, we enter floating error approximation and the order is not guaranteed across # python and torch versions. self.assertIn( nested_simplify(__lowerCAmelCase ) , [ [{'''score''': 0.3_3_3, '''label''': '''a'''}, {'''score''': 0.3_3_3, '''label''': '''b'''}, {'''score''': 0.3_3_3, '''label''': '''c'''}], [{'''score''': 0.3_3_3, '''label''': '''a'''}, {'''score''': 0.3_3_3, '''label''': '''c'''}, {'''score''': 0.3_3_3, '''label''': '''b'''}], ] , ) lowercase__ : str = image_classifier([image] * 5 , candidate_labels=['''A''', '''B''', '''C'''] , batch_size=2 ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [ [ {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, ], [ {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, ], [ {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, ], [ {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, ], [ {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, ], ] , ) @require_tf def _lowerCAmelCase( self ) -> Dict: lowercase__ : List[Any] = pipeline( model='''hf-internal-testing/tiny-random-clip-zero-shot-image-classification''' , framework='''tf''' ) lowercase__ : List[Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) lowercase__ : Dict = image_classifier(__lowerCAmelCase , candidate_labels=['''a''', '''b''', '''c'''] ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [{'''score''': 0.3_3_3, '''label''': '''a'''}, {'''score''': 0.3_3_3, '''label''': '''b'''}, {'''score''': 0.3_3_3, '''label''': '''c'''}] , ) lowercase__ : int = image_classifier([image] * 5 , candidate_labels=['''A''', '''B''', '''C'''] , batch_size=2 ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [ [ {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, ], [ {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, ], [ {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, ], [ {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, ], [ {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, {'''score''': 0.3_3_3, '''label''': ANY(__lowerCAmelCase )}, ], ] , ) @slow @require_torch def _lowerCAmelCase( self ) -> str: lowercase__ : Optional[int] = pipeline( task='''zero-shot-image-classification''' , model='''openai/clip-vit-base-patch32''' , ) # This is an image of 2 cats with remotes and no planes lowercase__ : Union[str, Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) lowercase__ : int = image_classifier(__lowerCAmelCase , candidate_labels=['''cat''', '''plane''', '''remote'''] ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [ {'''score''': 0.5_1_1, '''label''': '''remote'''}, {'''score''': 0.4_8_5, '''label''': '''cat'''}, {'''score''': 0.0_0_4, '''label''': '''plane'''}, ] , ) lowercase__ : Optional[int] = image_classifier([image] * 5 , candidate_labels=['''cat''', '''plane''', '''remote'''] , batch_size=2 ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [ [ {'''score''': 0.5_1_1, '''label''': '''remote'''}, {'''score''': 0.4_8_5, '''label''': '''cat'''}, {'''score''': 0.0_0_4, '''label''': '''plane'''}, ], ] * 5 , ) @slow @require_tf def _lowerCAmelCase( self ) -> Optional[int]: lowercase__ : Any = pipeline( task='''zero-shot-image-classification''' , model='''openai/clip-vit-base-patch32''' , framework='''tf''' ) # This is an image of 2 cats with remotes and no planes lowercase__ : Optional[Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) lowercase__ : Optional[Any] = image_classifier(__lowerCAmelCase , candidate_labels=['''cat''', '''plane''', '''remote'''] ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [ {'''score''': 0.5_1_1, '''label''': '''remote'''}, {'''score''': 0.4_8_5, '''label''': '''cat'''}, {'''score''': 0.0_0_4, '''label''': '''plane'''}, ] , ) lowercase__ : int = image_classifier([image] * 5 , candidate_labels=['''cat''', '''plane''', '''remote'''] , batch_size=2 ) self.assertEqual( nested_simplify(__lowerCAmelCase ) , [ [ {'''score''': 0.5_1_1, '''label''': '''remote'''}, {'''score''': 0.4_8_5, '''label''': '''cat'''}, {'''score''': 0.0_0_4, '''label''': '''plane'''}, ], ] * 5 , )
152
'''simple docstring''' from __future__ import annotations def _lowercase ( __A ,__A ,__A ,__A ,__A ,): '''simple docstring''' __UpperCamelCase = len(__A ) # If row is equal to the size of the board it means there are a queen in each row in # the current board (possible_board) if row == n: # We convert the variable possible_board that looks like this: [1, 3, 0, 2] to # this: ['. Q . . ', '. . . Q ', 'Q . . . ', '. . Q . '] boards.append([""". """ * i + """Q """ + """. """ * (n - 1 - i) for i in possible_board] ) return # We iterate each column in the row to find all possible results in each row for col in range(__A ): # We apply that we learned previously. First we check that in the current board # (possible_board) there are not other same value because if there is it means # that there are a collision in vertical. Then we apply the two formulas we # learned before: # # 45º: y - x = b or 45: row - col = b # 135º: y + x = b or row + col = b. # # And we verify if the results of this two formulas not exist in their variables # respectively. (diagonal_right_collisions, diagonal_left_collisions) # # If any or these are True it means there is a collision so we continue to the # next value in the for loop. if ( col in possible_board or row - col in diagonal_right_collisions or row + col in diagonal_left_collisions ): continue # If it is False we call dfs function again and we update the inputs depth_first_search( [*possible_board, col] ,[*diagonal_right_collisions, row - col] ,[*diagonal_left_collisions, row + col] ,__A ,__A ,) def _lowercase ( __A ): '''simple docstring''' __UpperCamelCase = [] depth_first_search([] ,[] ,[] ,__A ,__A ) # Print all the boards for board in boards: for column in board: print(__A ) print("""""" ) print(len(__A ) ,"""solutions were found.""" ) if __name__ == "__main__": import doctest doctest.testmod() n_queens_solution(4)
601
0
"""simple docstring""" from __future__ import annotations def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = position __lowerCAmelCase = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] __lowerCAmelCase = [] for position in positions: __lowerCAmelCase = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(_lowerCamelCase ) return permissible_positions def _lowerCamelCase ( _UpperCamelCase ): '''simple docstring''' return not any(elem == 0 for row in board for elem in row ) def _lowerCamelCase ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ): '''simple docstring''' if is_complete(_lowerCamelCase ): return True for position in get_valid_pos(_lowerCamelCase , len(_lowerCamelCase ) ): __lowerCAmelCase = position if board[y][x] == 0: __lowerCAmelCase = curr + 1 if open_knight_tour_helper(_lowerCamelCase , _lowerCamelCase , curr + 1 ): return True __lowerCAmelCase = 0 return False def _lowerCamelCase ( _UpperCamelCase ): '''simple docstring''' __lowerCAmelCase = [[0 for i in range(_lowerCamelCase )] for j in range(_lowerCamelCase )] for i in range(_lowerCamelCase ): for j in range(_lowerCamelCase ): __lowerCAmelCase = 1 if open_knight_tour_helper(_lowerCamelCase , (i, j) , 1 ): return board __lowerCAmelCase = 0 __lowerCAmelCase = f"Open Kight Tour cannot be performed on a board of size {n}" raise ValueError(_lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod()
706
"""simple docstring""" import tensorflow as tf from ...tf_utils import shape_list class _UpperCamelCase ( tf.keras.layers.Layer ): '''simple docstring''' def __init__( self , __a , __a , __a , __a , __a=1 , __a=False , **__a ): super().__init__(**__a ) __lowerCAmelCase = vocab_size __lowerCAmelCase = d_embed __lowerCAmelCase = d_proj __lowerCAmelCase = cutoffs + [vocab_size] __lowerCAmelCase = [0] + self.cutoffs __lowerCAmelCase = div_val __lowerCAmelCase = self.cutoffs[0] __lowerCAmelCase = len(self.cutoffs ) - 1 __lowerCAmelCase = self.shortlist_size + self.n_clusters __lowerCAmelCase = keep_order __lowerCAmelCase = [] __lowerCAmelCase = [] def snake_case ( self , __a ): if self.n_clusters > 0: __lowerCAmelCase = self.add_weight( shape=(self.n_clusters, self.d_embed) , initializer="zeros" , trainable=__a , name="cluster_weight" ) __lowerCAmelCase = self.add_weight( shape=(self.n_clusters,) , initializer="zeros" , trainable=__a , name="cluster_bias" ) if self.div_val == 1: for i in range(len(self.cutoffs ) ): if self.d_proj != self.d_embed: __lowerCAmelCase = self.add_weight( shape=(self.d_embed, self.d_proj) , initializer="zeros" , trainable=__a , name=f"out_projs_._{i}" , ) self.out_projs.append(__a ) else: self.out_projs.append(__a ) __lowerCAmelCase = self.add_weight( shape=(self.vocab_size, self.d_embed) , initializer="zeros" , trainable=__a , name=f"out_layers_._{i}_._weight" , ) __lowerCAmelCase = self.add_weight( shape=(self.vocab_size,) , initializer="zeros" , trainable=__a , name=f"out_layers_._{i}_._bias" , ) self.out_layers.append((weight, bias) ) else: for i in range(len(self.cutoffs ) ): __lowerCAmelCase , __lowerCAmelCase = self.cutoff_ends[i], self.cutoff_ends[i + 1] __lowerCAmelCase = self.d_embed // (self.div_val**i) __lowerCAmelCase = self.add_weight( shape=(d_emb_i, self.d_proj) , initializer="zeros" , trainable=__a , name=f"out_projs_._{i}" ) self.out_projs.append(__a ) __lowerCAmelCase = self.add_weight( shape=(r_idx - l_idx, d_emb_i) , initializer="zeros" , trainable=__a , name=f"out_layers_._{i}_._weight" , ) __lowerCAmelCase = self.add_weight( shape=(r_idx - l_idx,) , initializer="zeros" , trainable=__a , name=f"out_layers_._{i}_._bias" , ) self.out_layers.append((weight, bias) ) super().build(__a ) @staticmethod def snake_case ( __a , __a , __a , __a=None ): __lowerCAmelCase = x if proj is not None: __lowerCAmelCase = tf.einsum("ibd,ed->ibe" , __a , __a ) return tf.einsum("ibd,nd->ibn" , __a , __a ) + b @staticmethod def snake_case ( __a , __a ): __lowerCAmelCase = shape_list(__a ) __lowerCAmelCase = tf.range(lp_size[0] , dtype=target.dtype ) __lowerCAmelCase = tf.stack([r, target] , 1 ) return tf.gather_nd(__a , __a ) def snake_case ( self , __a , __a , __a=True , __a=False ): __lowerCAmelCase = 0 if self.n_clusters == 0: __lowerCAmelCase = self._logit(__a , self.out_layers[0][0] , self.out_layers[0][1] , self.out_projs[0] ) if target is not None: __lowerCAmelCase = tf.nn.sparse_softmax_cross_entropy_with_logits(labels=__a , logits=__a ) __lowerCAmelCase = tf.nn.log_softmax(__a , axis=-1 ) else: __lowerCAmelCase = shape_list(__a ) __lowerCAmelCase = [] __lowerCAmelCase = tf.zeros(hidden_sizes[:2] ) for i in range(len(self.cutoffs ) ): __lowerCAmelCase , __lowerCAmelCase = self.cutoff_ends[i], self.cutoff_ends[i + 1] if target is not None: __lowerCAmelCase = (target >= l_idx) & (target < r_idx) __lowerCAmelCase = tf.where(__a ) __lowerCAmelCase = tf.boolean_mask(__a , __a ) - l_idx if self.div_val == 1: __lowerCAmelCase = self.out_layers[0][0][l_idx:r_idx] __lowerCAmelCase = self.out_layers[0][1][l_idx:r_idx] else: __lowerCAmelCase = self.out_layers[i][0] __lowerCAmelCase = self.out_layers[i][1] if i == 0: __lowerCAmelCase = tf.concat([cur_W, self.cluster_weight] , 0 ) __lowerCAmelCase = tf.concat([cur_b, self.cluster_bias] , 0 ) __lowerCAmelCase = self._logit(__a , __a , __a , self.out_projs[0] ) __lowerCAmelCase = tf.nn.log_softmax(__a ) out.append(head_logprob[..., : self.cutoffs[0]] ) if target is not None: __lowerCAmelCase = tf.boolean_mask(__a , __a ) __lowerCAmelCase = self._gather_logprob(__a , __a ) else: __lowerCAmelCase = self._logit(__a , __a , __a , self.out_projs[i] ) __lowerCAmelCase = tf.nn.log_softmax(__a ) __lowerCAmelCase = self.cutoffs[0] + i - 1 # No probability for the head cluster __lowerCAmelCase = head_logprob[..., cluster_prob_idx, None] + tail_logprob out.append(__a ) if target is not None: __lowerCAmelCase = tf.boolean_mask(__a , __a ) __lowerCAmelCase = tf.boolean_mask(__a , __a ) __lowerCAmelCase = self._gather_logprob(__a , __a ) cur_logprob += cur_head_logprob[:, self.cutoff_ends[1] + i - 1] if target is not None: loss += tf.scatter_nd(__a , -cur_logprob , shape_list(__a ) ) __lowerCAmelCase = tf.concat(__a , axis=-1 ) if target is not None: if return_mean: __lowerCAmelCase = tf.reduce_mean(__a ) # Add the training-time loss value to the layer using `self.add_loss()`. self.add_loss(__a ) # Log the loss as a metric (we could log arbitrary metrics, # including different metrics for training and inference. self.add_metric(__a , name=self.name , aggregation="mean" if return_mean else "" ) return out
282
0
from collections import deque class __lowercase : """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ : int = process_name # process name SCREAMING_SNAKE_CASE_ : str = arrival_time # arrival time of the process # completion time of finished process or last interrupted time SCREAMING_SNAKE_CASE_ : Optional[Any] = arrival_time SCREAMING_SNAKE_CASE_ : Tuple = burst_time # remaining burst time SCREAMING_SNAKE_CASE_ : Optional[int] = 0 # total time of the process wait in ready queue SCREAMING_SNAKE_CASE_ : Any = 0 # time from arrival time to completion time class __lowercase : """simple docstring""" def __init__( self , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = number_of_queues # time slice of queues that round robin algorithm applied SCREAMING_SNAKE_CASE_ : int = time_slices # unfinished process is in this ready_queue SCREAMING_SNAKE_CASE_ : Union[str, Any] = queue # current time SCREAMING_SNAKE_CASE_ : List[str] = current_time # finished process is in this sequence queue SCREAMING_SNAKE_CASE_ : deque[Process] = deque() def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[int] = [] for i in range(len(self.finish_queue ) ): sequence.append(self.finish_queue[i].process_name ) return sequence def UpperCamelCase__ ( self , lowerCAmelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = [] for i in range(len(lowerCAmelCase__ ) ): waiting_times.append(queue[i].waiting_time ) return waiting_times def UpperCamelCase__ ( self , lowerCAmelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[str] = [] for i in range(len(lowerCAmelCase__ ) ): turnaround_times.append(queue[i].turnaround_time ) return turnaround_times def UpperCamelCase__ ( self , lowerCAmelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = [] for i in range(len(lowerCAmelCase__ ) ): completion_times.append(queue[i].stop_time ) return completion_times def UpperCamelCase__ ( self , lowerCAmelCase__ ): """simple docstring""" return [q.burst_time for q in queue] def UpperCamelCase__ ( self , lowerCAmelCase__ ): """simple docstring""" process.waiting_time += self.current_time - process.stop_time return process.waiting_time def UpperCamelCase__ ( self , lowerCAmelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ : deque[Process] = deque() # sequence deque of finished process while len(lowerCAmelCase__ ) != 0: SCREAMING_SNAKE_CASE_ : Dict = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of current process self.update_waiting_time(lowerCAmelCase__ ) # update current time self.current_time += cp.burst_time # finish the process and set the process's burst-time 0 SCREAMING_SNAKE_CASE_ : int = 0 # set the process's turnaround time because it is finished SCREAMING_SNAKE_CASE_ : str = self.current_time - cp.arrival_time # set the completion time SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.current_time # add the process to queue that has finished queue finished.append(lowerCAmelCase__ ) self.finish_queue.extend(lowerCAmelCase__ ) # add finished process to finish queue # FCFS will finish all remaining processes return finished def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE_ : deque[Process] = deque() # sequence deque of terminated process # just for 1 cycle and unfinished processes will go back to queue for _ in range(len(lowerCAmelCase__ ) ): SCREAMING_SNAKE_CASE_ : List[Any] = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of unfinished processes self.update_waiting_time(lowerCAmelCase__ ) # if the burst time of process is bigger than time-slice if cp.burst_time > time_slice: # use CPU for only time-slice self.current_time += time_slice # update remaining burst time cp.burst_time -= time_slice # update end point time SCREAMING_SNAKE_CASE_ : Optional[int] = self.current_time # locate the process behind the queue because it is not finished ready_queue.append(lowerCAmelCase__ ) else: # use CPU for remaining burst time self.current_time += cp.burst_time # set burst time 0 because the process is finished SCREAMING_SNAKE_CASE_ : Optional[Any] = 0 # set the finish time SCREAMING_SNAKE_CASE_ : str = self.current_time # update the process' turnaround time because it is finished SCREAMING_SNAKE_CASE_ : Any = self.current_time - cp.arrival_time # add the process to queue that has finished queue finished.append(lowerCAmelCase__ ) self.finish_queue.extend(lowerCAmelCase__ ) # add finished process to finish queue # return finished processes queue and remaining processes queue return finished, ready_queue def UpperCamelCase__ ( self ): """simple docstring""" for i in range(self.number_of_queues - 1 ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.round_robin( self.ready_queue , self.time_slices[i] ) # the last queue has first_come_first_served algorithm self.first_come_first_served(self.ready_queue ) return self.finish_queue if __name__ == "__main__": import doctest lowerCAmelCase__ : Tuple =Process('P1', 0, 53) lowerCAmelCase__ : str =Process('P2', 0, 17) lowerCAmelCase__ : int =Process('P3', 0, 68) lowerCAmelCase__ : List[Any] =Process('P4', 0, 24) lowerCAmelCase__ : Tuple =3 lowerCAmelCase__ : Any =[17, 25] lowerCAmelCase__ : Tuple =deque([Pa, Pa, Pa, Pa]) if len(time_slices) != number_of_queues - 1: raise SystemExit(0) doctest.testmod(extraglobs={'queue': deque([Pa, Pa, Pa, Pa])}) lowerCAmelCase__ : Optional[int] =Process('P1', 0, 53) lowerCAmelCase__ : List[str] =Process('P2', 0, 17) lowerCAmelCase__ : List[Any] =Process('P3', 0, 68) lowerCAmelCase__ : Optional[int] =Process('P4', 0, 24) lowerCAmelCase__ : List[Any] =3 lowerCAmelCase__ : Optional[int] =[17, 25] lowerCAmelCase__ : Optional[int] =deque([Pa, Pa, Pa, Pa]) lowerCAmelCase__ : List[Any] =MLFQ(number_of_queues, time_slices, queue, 0) lowerCAmelCase__ : Tuple =mlfq.multi_level_feedback_queue() # print total waiting times of processes(P1, P2, P3, P4) print( F"""waiting time:\ \t\t\t{MLFQ.calculate_waiting_time(mlfq, [Pa, Pa, Pa, Pa])}""" ) # print completion times of processes(P1, P2, P3, P4) print( F"""completion time:\ \t\t{MLFQ.calculate_completion_time(mlfq, [Pa, Pa, Pa, Pa])}""" ) # print total turnaround times of processes(P1, P2, P3, P4) print( F"""turnaround time:\ \t\t{MLFQ.calculate_turnaround_time(mlfq, [Pa, Pa, Pa, Pa])}""" ) # print sequence of finished processes print( F"""sequence of finished processes:\ {mlfq.calculate_sequence_of_finish_queue()}""" )
101
"""simple docstring""" import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import BatchEncoding, MarianTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import is_sentencepiece_available, is_tf_available, is_torch_available if is_sentencepiece_available(): from transformers.models.marian.tokenization_marian import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin UpperCAmelCase__ =get_tests_dir("fixtures/test_sentencepiece.model") UpperCAmelCase__ ={"target_lang": "fi", "source_lang": "en"} UpperCAmelCase__ =">>zh<<" UpperCAmelCase__ ="Helsinki-NLP/" if is_torch_available(): UpperCAmelCase__ ="pt" elif is_tf_available(): UpperCAmelCase__ ="tf" else: UpperCAmelCase__ ="jax" @require_sentencepiece class lowerCamelCase__ ( _a , unittest.TestCase ): a : Dict = MarianTokenizer a : Optional[int] = False a : Any = True def SCREAMING_SNAKE_CASE_ ( self : int ): '''simple docstring''' super().setUp() __lowercase = ["""</s>""", """<unk>""", """▁This""", """▁is""", """▁a""", """▁t""", """est""", """\u0120""", """<pad>"""] __lowercase = dict(zip(A_ , range(len(A_ ) ) ) ) __lowercase = Path(self.tmpdirname ) save_json(A_ , save_dir / VOCAB_FILES_NAMES["""vocab"""] ) save_json(A_ , save_dir / VOCAB_FILES_NAMES["""tokenizer_config_file"""] ) if not (save_dir / VOCAB_FILES_NAMES["source_spm"]).exists(): copyfile(A_ , save_dir / VOCAB_FILES_NAMES["""source_spm"""] ) copyfile(A_ , save_dir / VOCAB_FILES_NAMES["""target_spm"""] ) __lowercase = MarianTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def SCREAMING_SNAKE_CASE_ ( self : Dict , **A_ : Any ): '''simple docstring''' return MarianTokenizer.from_pretrained(self.tmpdirname , **A_ ) def SCREAMING_SNAKE_CASE_ ( self : int , A_ : List[str] ): '''simple docstring''' return ( "This is a test", "This is a test", ) def SCREAMING_SNAKE_CASE_ ( self : Any ): '''simple docstring''' __lowercase = """</s>""" __lowercase = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A_ ) , A_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A_ ) , A_ ) def SCREAMING_SNAKE_CASE_ ( self : Dict ): '''simple docstring''' __lowercase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """</s>""" ) self.assertEqual(vocab_keys[1] , """<unk>""" ) self.assertEqual(vocab_keys[-1] , """<pad>""" ) self.assertEqual(len(A_ ) , 9 ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 9 ) def SCREAMING_SNAKE_CASE_ ( self : Dict ): '''simple docstring''' __lowercase = MarianTokenizer.from_pretrained(F'''{ORG_NAME}opus-mt-en-de''' ) __lowercase = en_de_tokenizer(["""I am a small frog"""] , return_tensors=A_ ) self.assertIsInstance(A_ , A_ ) __lowercase = [3_8, 1_2_1, 1_4, 6_9_7, 3_8_8_4_8, 0] self.assertListEqual(A_ , batch.input_ids[0] ) __lowercase = tempfile.mkdtemp() en_de_tokenizer.save_pretrained(A_ ) __lowercase = [x.name for x in Path(A_ ).glob("""*""" )] self.assertIn("""source.spm""" , A_ ) MarianTokenizer.from_pretrained(A_ ) def SCREAMING_SNAKE_CASE_ ( self : Dict ): '''simple docstring''' __lowercase = self.get_tokenizer() __lowercase = tok( ["""I am a small frog""" * 1_0_0_0, """I am a small frog"""] , padding=A_ , truncation=A_ , return_tensors=A_ ) self.assertIsInstance(A_ , A_ ) self.assertEqual(batch.input_ids.shape , (2, 5_1_2) ) def SCREAMING_SNAKE_CASE_ ( self : Dict ): '''simple docstring''' __lowercase = self.get_tokenizer() __lowercase = tok(["""I am a tiny frog""", """I am a small frog"""] , padding=A_ , return_tensors=A_ ) self.assertIsInstance(A_ , A_ ) self.assertEqual(batch_smaller.input_ids.shape , (2, 1_0) ) @slow def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): '''simple docstring''' __lowercase = {"""input_ids""": [[4_3_4_9_5, 4_6_2, 2_0, 4_2_1_6_4, 1_3_6_9, 5_2, 4_6_4, 1_3_2, 1_7_0_3, 4_9_2, 1_3, 7_4_9_1, 3_8_9_9_9, 6, 8, 4_6_4, 1_3_2, 1_7_0_3, 4_9_2, 1_3, 4_6_6_9, 3_7_8_6_7, 1_3, 7_5_2_5, 2_7, 1_5_9_3, 9_8_8, 1_3, 3_3_9_7_2, 7_0_2_9, 6, 2_0, 8_2_5_1, 3_8_3, 2, 2_7_0, 5_8_6_6, 3_7_8_8, 2, 2_3_5_3, 8_2_5_1, 1_2_3_3_8, 2, 1_3_9_5_8, 3_8_7, 2, 3_6_2_9, 6_9_5_3, 1_8_8, 2_9_0_0, 2, 1_3_9_5_8, 8_0_1_1, 1_1_5_0_1, 2_3, 8_4_6_0, 4_0_7_3, 3_4_0_0_9, 2_0, 4_3_5, 1_1_4_3_9, 2_7, 8, 8_4_6_0, 4_0_7_3, 6_0_0_4, 2_0, 9_9_8_8, 3_7_5, 2_7, 3_3, 2_6_6, 1_9_4_5, 1_0_7_6, 1_3_5_0, 3_7_8_6_7, 3_2_8_8, 5, 5_7_7, 1_0_7_6, 4_3_7_4, 8, 5_0_8_2, 5, 2_6_4_5_3, 2_5_7, 5_5_6, 4_0_3, 2, 2_4_2, 1_3_2, 3_8_3, 3_1_6, 4_9_2, 8, 1_0_7_6_7, 6, 3_1_6, 3_0_4, 4_2_3_9, 3, 0], [1_4_8, 1_5_7_2_2, 1_9, 1_8_3_9, 1_2, 1_3_5_0, 1_3, 2_2_3_2_7, 5_0_8_2, 5_4_1_8, 4_7_5_6_7, 3_5_9_3_8, 5_9, 3_1_8, 1_9_5_5_2, 1_0_8, 2_1_8_3, 5_4, 1_4_9_7_6, 4_8_3_5, 3_2, 5_4_7, 1_1_1_4, 8, 3_1_5, 2_4_1_7, 5, 9_2, 1_9_0_8_8, 3, 0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0], [3_6, 6_3_9_5, 1_2_5_7_0, 3_9_1_4_7, 1_1_5_9_7, 6, 2_6_6, 4, 4_5_4_0_5, 7_2_9_6, 3, 0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0, 5_8_1_0_0]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A_ , model_name="""Helsinki-NLP/opus-mt-en-de""" , revision="""1a8c2263da11e68e50938f97e10cd57820bd504c""" , decode_kwargs={"""use_source_tokenizer""": True} , ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ): '''simple docstring''' __lowercase = MarianTokenizer.from_pretrained("""hf-internal-testing/test-marian-two-vocabs""" ) __lowercase = """Tämä on testi""" __lowercase = """This is a test""" __lowercase = [7_6, 7, 2_0_4_7, 2] __lowercase = [6_9, 1_2, 1_1, 9_4_0, 2] __lowercase = tokenizer(A_ ).input_ids self.assertListEqual(A_ , A_ ) __lowercase = tokenizer(text_target=A_ ).input_ids self.assertListEqual(A_ , A_ ) __lowercase = tokenizer.decode(A_ , skip_special_tokens=A_ ) self.assertEqual(A_ , A_ )
616
0
def SCREAMING_SNAKE_CASE__ ( lowerCAmelCase_ : str ) -> str: """simple docstring""" return " ".join(input_str.split()[::-1] ) if __name__ == "__main__": import doctest doctest.testmod()
702
from __future__ import annotations __SCREAMING_SNAKE_CASE = '#' class lowerCAmelCase_ : '''simple docstring''' def __init__( self ): SCREAMING_SNAKE_CASE_ : dict ={} def __lowerCamelCase ( self , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : Tuple =self._trie for char in text: if char not in trie: SCREAMING_SNAKE_CASE_ : Optional[int] ={} SCREAMING_SNAKE_CASE_ : Any =trie[char] SCREAMING_SNAKE_CASE_ : Optional[Any] =True def __lowerCamelCase ( self , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : Tuple =self._trie for char in prefix: if char in trie: SCREAMING_SNAKE_CASE_ : Tuple =trie[char] else: return [] return self._elements(__UpperCAmelCase ) def __lowerCamelCase ( self , __UpperCAmelCase ): SCREAMING_SNAKE_CASE_ : Optional[int] =[] for c, v in d.items(): SCREAMING_SNAKE_CASE_ : List[Any] =[' '] if c == END else [(c + s) for s in self._elements(__UpperCAmelCase )] result.extend(__UpperCAmelCase ) return tuple(__UpperCAmelCase ) __SCREAMING_SNAKE_CASE = Trie() __SCREAMING_SNAKE_CASE = ('depart', 'detergent', 'daring', 'dog', 'deer', 'deal') for word in words: trie.insert_word(word) def SCREAMING_SNAKE_CASE__ ( lowerCAmelCase_ : str ) -> tuple: """simple docstring""" SCREAMING_SNAKE_CASE_ : str =trie.find_word(lowerCAmelCase_ ) return tuple(string + word for word in suffixes ) def SCREAMING_SNAKE_CASE__ ( ) -> None: """simple docstring""" print(autocomplete_using_trie('de' ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
153
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available __A : Tuple = {"configuration_speech_encoder_decoder": ["SpeechEncoderDecoderConfig"]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : str = ["SpeechEncoderDecoderModel"] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : int = ["FlaxSpeechEncoderDecoderModel"] if TYPE_CHECKING: from .configuration_speech_encoder_decoder import SpeechEncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_encoder_decoder import SpeechEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_speech_encoder_decoder import FlaxSpeechEncoderDecoderModel else: import sys __A : str = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
130
__A : Tuple = {str(digit): digit**5 for digit in range(10)} def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> int: '''simple docstring''' return sum(DIGITS_FIFTH_POWER[digit] for digit in str(UpperCamelCase__ ) ) def __SCREAMING_SNAKE_CASE ( ) -> int: '''simple docstring''' return sum( number for number in range(1000 , 100_0000 ) if number == digits_fifth_powers_sum(UpperCamelCase__ ) ) if __name__ == "__main__": print(solution())
130
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = { 'andreasmadsen/efficient_mlm_m0.40': ( 'https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json' ), } class _A ( _UpperCAmelCase ): _UpperCamelCase : Union[str, Any] = '''roberta-prelayernorm''' def __init__( self : Dict , _A : int=50_265 , _A : Optional[int]=768 , _A : Optional[int]=12 , _A : int=12 , _A : Optional[int]=3_072 , _A : Optional[int]="gelu" , _A : Optional[int]=0.1 , _A : Optional[Any]=0.1 , _A : int=512 , _A : Any=2 , _A : Any=0.02 , _A : str=1E-12 , _A : List[Any]=1 , _A : Any=0 , _A : List[Any]=2 , _A : Tuple="absolute" , _A : str=True , _A : Union[str, Any]=None , **_A : Any , ) -> Dict: """simple docstring""" super().__init__(pad_token_id=lowercase__ , bos_token_id=lowercase__ , eos_token_id=lowercase__ , **lowercase__ ) lowercase : Dict = vocab_size lowercase : Optional[int] = hidden_size lowercase : Tuple = num_hidden_layers lowercase : Tuple = num_attention_heads lowercase : Any = hidden_act lowercase : Optional[int] = intermediate_size lowercase : str = hidden_dropout_prob lowercase : List[str] = attention_probs_dropout_prob lowercase : List[str] = max_position_embeddings lowercase : List[str] = type_vocab_size lowercase : Dict = initializer_range lowercase : Tuple = layer_norm_eps lowercase : str = position_embedding_type lowercase : Optional[int] = use_cache lowercase : List[str] = classifier_dropout class _A ( _UpperCAmelCase ): @property def __a ( self : int ) -> Optional[Any]: """simple docstring""" if self.task == "multiple-choice": lowercase : Dict = {0: "batch", 1: "choice", 2: "sequence"} else: lowercase : List[str] = {0: "batch", 1: "sequence"} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
702
# A Bipartite Graph is a graph whose vertices can be divided into two independent sets, # U and V such that every edge (u, v) either connects a vertex from U to V or a vertex # from V to U. In other words, for every edge (u, v), either u belongs to U and v to V, # or u belongs to V and v to U. We can also say that there is no edge that connects # vertices of same set. def snake_case( __magic_name__ ) -> Any: '''simple docstring''' lowercase : Union[str, Any] = [False] * len(__magic_name__ ) lowercase : int = [-1] * len(__magic_name__ ) def dfs(__magic_name__ , __magic_name__ ): lowercase : str = True lowercase : Tuple = c for u in graph[v]: if not visited[u]: dfs(__magic_name__ , 1 - c ) for i in range(len(__magic_name__ ) ): if not visited[i]: dfs(__magic_name__ , 0 ) for i in range(len(__magic_name__ ) ): for j in graph[i]: if color[i] == color[j]: return False return True # Adjacency list of graph lowerCAmelCase_ = {0: [1, 3], 1: [0, 2], 2: [1, 3], 3: [0, 2], 4: []} print(check_bipartite_dfs(graph))
596
0
'''simple docstring''' import argparse import os import torch from transformers.utils import WEIGHTS_NAME SCREAMING_SNAKE_CASE_: Optional[Any] =['small', 'medium', 'large'] SCREAMING_SNAKE_CASE_: Any ='lm_head.decoder.weight' SCREAMING_SNAKE_CASE_: List[Any] ='lm_head.weight' def lowerCAmelCase_ ( snake_case_ : str , snake_case_ : str ) -> Dict: '''simple docstring''' UpperCAmelCase_ = torch.load(snake_case_ ) UpperCAmelCase_ = d.pop(snake_case_ ) os.makedirs(snake_case_ , exist_ok=snake_case_ ) torch.save(snake_case_ , os.path.join(snake_case_ , snake_case_ ) ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_: str =argparse.ArgumentParser() parser.add_argument('--dialogpt_path', default='.', type=str) SCREAMING_SNAKE_CASE_: Optional[int] =parser.parse_args() for MODEL in DIALOGPT_MODELS: SCREAMING_SNAKE_CASE_: Dict =os.path.join(args.dialogpt_path, f"{MODEL}_ft.pkl") SCREAMING_SNAKE_CASE_: Tuple =f"./DialoGPT-{MODEL}" convert_dialogpt_checkpoint( checkpoint_path, pytorch_dump_folder_path, )
78
import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __magic_name__ ( A__ ): def SCREAMING_SNAKE_CASE_ ( self : Any ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "embed_dim" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , "num_heads" ) ) class __magic_name__ : def __init__( self : List[str] , UpperCamelCase__ : Any , UpperCamelCase__ : List[Any]=13 , UpperCamelCase__ : Dict=64 , UpperCamelCase__ : Optional[Any]=3 , UpperCamelCase__ : Optional[Any]=[16, 48, 96] , UpperCamelCase__ : int=[1, 3, 6] , UpperCamelCase__ : int=[1, 2, 10] , UpperCamelCase__ : List[str]=[7, 3, 3] , UpperCamelCase__ : Any=[4, 2, 2] , UpperCamelCase__ : List[Any]=[2, 1, 1] , UpperCamelCase__ : Optional[int]=[2, 2, 2] , UpperCamelCase__ : Optional[Any]=[False, False, True] , UpperCamelCase__ : List[str]=[0.0, 0.0, 0.0] , UpperCamelCase__ : List[Any]=0.02 , UpperCamelCase__ : int=1e-1_2 , UpperCamelCase__ : str=True , UpperCamelCase__ : Optional[int]=True , UpperCamelCase__ : Optional[Any]=2 , ) -> Optional[int]: '''simple docstring''' UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = image_size UpperCAmelCase = patch_sizes UpperCAmelCase = patch_stride UpperCAmelCase = patch_padding UpperCAmelCase = is_training UpperCAmelCase = use_labels UpperCAmelCase = num_labels UpperCAmelCase = num_channels UpperCAmelCase = embed_dim UpperCAmelCase = num_heads UpperCAmelCase = stride_kv UpperCAmelCase = depth UpperCAmelCase = cls_token UpperCAmelCase = attention_drop_rate UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps def SCREAMING_SNAKE_CASE_ ( self : Dict ) -> Tuple: '''simple docstring''' UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase = None if self.use_labels: UpperCAmelCase = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE_ ( self : Any ) -> Union[str, Any]: '''simple docstring''' return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def SCREAMING_SNAKE_CASE_ ( self : Tuple , UpperCamelCase__ : int , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Any ) -> List[str]: '''simple docstring''' UpperCAmelCase = CvtModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() UpperCAmelCase = model(UpperCamelCase__ ) UpperCAmelCase = (self.image_size, self.image_size) UpperCAmelCase , UpperCAmelCase = image_size[0], image_size[1] for i in range(len(self.depth ) ): UpperCAmelCase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) UpperCAmelCase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width) ) def SCREAMING_SNAKE_CASE_ ( self : Dict , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[int] ) -> Dict: '''simple docstring''' UpperCAmelCase = self.num_labels UpperCAmelCase = CvtForImageClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() UpperCAmelCase = model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ) -> Dict: '''simple docstring''' UpperCAmelCase = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase = config_and_inputs UpperCAmelCase = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __magic_name__ ( A__, A__, unittest.TestCase ): lowercase : Union[str, Any] =(CvtModel, CvtForImageClassification) if is_torch_available() else () lowercase : List[str] =( {'''feature-extraction''': CvtModel, '''image-classification''': CvtForImageClassification} if is_torch_available() else {} ) lowercase : Dict =False lowercase : Optional[Any] =False lowercase : Union[str, Any] =False lowercase : List[Any] =False lowercase : Optional[int] =False def SCREAMING_SNAKE_CASE_ ( self : List[Any] ) -> Tuple: '''simple docstring''' UpperCAmelCase = CvtModelTester(self ) UpperCAmelCase = ConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ , hidden_size=37 ) def SCREAMING_SNAKE_CASE_ ( self : str ) -> List[str]: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ) -> List[str]: '''simple docstring''' return @unittest.skip(reason="Cvt does not output attentions" ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ) -> int: '''simple docstring''' pass @unittest.skip(reason="Cvt does not use inputs_embeds" ) def SCREAMING_SNAKE_CASE_ ( self : Any ) -> int: '''simple docstring''' pass @unittest.skip(reason="Cvt does not support input and output embeddings" ) def SCREAMING_SNAKE_CASE_ ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' pass def SCREAMING_SNAKE_CASE_ ( self : List[Any] ) -> Dict: '''simple docstring''' UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = model_class(UpperCamelCase__ ) UpperCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase = [*signature.parameters.keys()] UpperCAmelCase = ["pixel_values"] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ ( self : Any ) -> Optional[int]: '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ) -> Union[str, Any]: '''simple docstring''' def check_hidden_states_output(UpperCamelCase__ : Optional[int] , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Union[str, Any] ): UpperCAmelCase = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): UpperCAmelCase = model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) UpperCAmelCase = outputs.hidden_states UpperCAmelCase = len(self.model_tester.depth ) self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) UpperCAmelCase , UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ) -> Optional[int]: '''simple docstring''' UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase__ ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ) -> List[str]: '''simple docstring''' pass @slow def SCREAMING_SNAKE_CASE_ ( self : Any ) -> Optional[int]: '''simple docstring''' for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase = CvtModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def lowerCamelCase_() -> Union[str, Any]: UpperCAmelCase = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __magic_name__ ( unittest.TestCase ): @cached_property def SCREAMING_SNAKE_CASE_ ( self : Dict ) -> str: '''simple docstring''' return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE_ ( self : str ) -> Tuple: '''simple docstring''' UpperCAmelCase = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(UpperCamelCase__ ) UpperCAmelCase = self.default_image_processor UpperCAmelCase = prepare_img() UpperCAmelCase = image_processor(images=UpperCamelCase__ , return_tensors="pt" ).to(UpperCamelCase__ ) # forward pass with torch.no_grad(): UpperCAmelCase = model(**UpperCamelCase__ ) # verify the logits UpperCAmelCase = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) UpperCAmelCase = torch.tensor([0.92_85, 0.90_15, -0.31_50] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCamelCase__ , atol=1e-4 ) )
323
0
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __UpperCAmelCase = {"""configuration_van""": ["""VAN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """VanConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __UpperCAmelCase = [ """VAN_PRETRAINED_MODEL_ARCHIVE_LIST""", """VanForImageClassification""", """VanModel""", """VanPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_van import VAN_PRETRAINED_CONFIG_ARCHIVE_MAP, VanConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_van import ( VAN_PRETRAINED_MODEL_ARCHIVE_LIST, VanForImageClassification, VanModel, VanPreTrainedModel, ) else: import sys __UpperCAmelCase = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
218
from __future__ import annotations import csv import requests from bsa import BeautifulSoup def snake_case_ (__A : str = "" ) -> dict[str, float]: __lowerCAmelCase : str = url or """https://www.imdb.com/chart/top/?ref_=nv_mv_250""" __lowerCAmelCase : Union[str, Any] = BeautifulSoup(requests.get(__A ).text , """html.parser""" ) __lowerCAmelCase : int = soup.find_all("""td""" , attrs="""titleColumn""" ) __lowerCAmelCase : int = soup.find_all("""td""" , class_="""ratingColumn imdbRating""" ) return { title.a.text: float(rating.strong.text ) for title, rating in zip(__A , __A ) } def snake_case_ (__A : str = "IMDb_Top_250_Movies.csv" ) -> None: __lowerCAmelCase : int = get_imdb_top_aaa_movies() with open(__A , """w""" , newline="""""" ) as out_file: __lowerCAmelCase : Dict = csv.writer(__A ) writer.writerow(["""Movie title""", """IMDb rating"""] ) for title, rating in movies.items(): writer.writerow([title, rating] ) if __name__ == "__main__": write_movies()
218
1