code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' def snake_case_ ( _lowerCAmelCase : float , _lowerCAmelCase : float ) -> float: return price * (1 + tax_rate) if __name__ == "__main__": print(F"{price_plus_tax(100, 0.25) = }") print(F"{price_plus_tax(125.50, 0.05) = }")
23
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE( A__ , A__ , A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = AltDiffusionPipeline lowerCamelCase__ = TEXT_TO_IMAGE_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS def A ( self : Dict ) -> int: torch.manual_seed(0 ) UpperCAmelCase : str = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) UpperCAmelCase : Dict = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) UpperCAmelCase : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) UpperCAmelCase : Dict = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5002 , ) UpperCAmelCase : List[Any] = CLIPTextModel(__snake_case ) UpperCAmelCase : Optional[Any] = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) UpperCAmelCase : Optional[int] = 77 UpperCAmelCase : Optional[int] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def A ( self : Optional[Any] , __snake_case : Dict , __snake_case : List[str]=0 ) -> Union[str, Any]: if str(__snake_case ).startswith('''mps''' ): UpperCAmelCase : str = torch.manual_seed(__snake_case ) else: UpperCAmelCase : Tuple = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) UpperCAmelCase : Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def A ( self : Union[str, Any] ) -> List[str]: super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def A ( self : Tuple ) -> List[str]: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def A ( self : Tuple ) -> Optional[int]: UpperCAmelCase : List[str] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase : Any = self.get_dummy_components() torch.manual_seed(0 ) UpperCAmelCase : Optional[int] = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase : List[str] = RobertaSeriesModelWithTransformation(__snake_case ) UpperCAmelCase : str = text_encoder UpperCAmelCase : Optional[int] = AltDiffusionPipeline(**__snake_case ) UpperCAmelCase : str = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : Optional[int] = self.get_dummy_inputs(__snake_case ) UpperCAmelCase : Optional[int] = '''A photo of an astronaut''' UpperCAmelCase : List[Any] = alt_pipe(**__snake_case ) UpperCAmelCase : Optional[Any] = output.images UpperCAmelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : List[str] = np.array( [0.5_74_81_62, 0.60_44_71_45, 0.48_82_12_17, 0.50_10_06_36, 0.5_43_11_85, 0.45_76_36_83, 0.49_65_76_96, 0.48_13_27_33, 0.47_57_30_93] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A ( self : int ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase : int = self.get_dummy_components() UpperCAmelCase : int = PNDMScheduler(skip_prk_steps=__snake_case ) torch.manual_seed(0 ) UpperCAmelCase : int = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase : Union[str, Any] = RobertaSeriesModelWithTransformation(__snake_case ) UpperCAmelCase : Union[str, Any] = text_encoder UpperCAmelCase : Optional[int] = AltDiffusionPipeline(**__snake_case ) UpperCAmelCase : Dict = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : int = self.get_dummy_inputs(__snake_case ) UpperCAmelCase : Optional[int] = alt_pipe(**__snake_case ) UpperCAmelCase : Optional[int] = output.images UpperCAmelCase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : Optional[int] = np.array( [0.51_60_50_93, 0.5_70_72_41, 0.47_36_55_07, 0.50_57_88_86, 0.5_63_38_77, 0.4_64_25_03, 0.5_18_20_81, 0.48_76_34_84, 0.49_08_42_37] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : str ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : List[Any] ) -> Any: # make sure here that pndm scheduler skips prk UpperCAmelCase : List[Any] = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''' , safety_checker=__snake_case ) UpperCAmelCase : Tuple = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : List[Any] = '''A painting of a squirrel eating a burger''' UpperCAmelCase : Any = torch.manual_seed(0 ) UpperCAmelCase : Optional[int] = alt_pipe([prompt] , generator=__snake_case , guidance_scale=6.0 , num_inference_steps=20 , output_type='''np''' ) UpperCAmelCase : Dict = output.images UpperCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase : List[str] = np.array([0.10_10, 0.08_00, 0.07_94, 0.08_85, 0.08_43, 0.07_62, 0.07_69, 0.07_29, 0.05_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A ( self : Tuple ) -> int: UpperCAmelCase : int = DDIMScheduler.from_pretrained('''BAAI/AltDiffusion''' , subfolder='''scheduler''' ) UpperCAmelCase : Tuple = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''' , scheduler=__snake_case , safety_checker=__snake_case ) UpperCAmelCase : Dict = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : Tuple = '''A painting of a squirrel eating a burger''' UpperCAmelCase : Optional[int] = torch.manual_seed(0 ) UpperCAmelCase : List[Any] = alt_pipe([prompt] , generator=__snake_case , num_inference_steps=2 , output_type='''numpy''' ) UpperCAmelCase : Dict = output.images UpperCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase : Union[str, Any] = np.array([0.40_19, 0.40_52, 0.38_10, 0.41_19, 0.39_16, 0.39_82, 0.46_51, 0.41_95, 0.53_23] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
23
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__: int = logging.get_logger(__name__) UpperCamelCase__: Tuple = { "funnel-transformer/small": "https://huggingface.co/funnel-transformer/small/resolve/main/config.json", "funnel-transformer/small-base": "https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json", "funnel-transformer/medium": "https://huggingface.co/funnel-transformer/medium/resolve/main/config.json", "funnel-transformer/medium-base": "https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json", "funnel-transformer/intermediate": ( "https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json" ), "funnel-transformer/intermediate-base": ( "https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json" ), "funnel-transformer/large": "https://huggingface.co/funnel-transformer/large/resolve/main/config.json", "funnel-transformer/large-base": "https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json", "funnel-transformer/xlarge": "https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json", "funnel-transformer/xlarge-base": "https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json", } class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = """funnel""" lowerCamelCase__ = { """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", } def __init__( self : Dict , __snake_case : Optional[Any]=30522 , __snake_case : List[str]=[4, 4, 4] , __snake_case : Dict=None , __snake_case : Union[str, Any]=2 , __snake_case : Optional[int]=768 , __snake_case : Dict=12 , __snake_case : Any=64 , __snake_case : Any=3072 , __snake_case : Any="gelu_new" , __snake_case : Any=0.1 , __snake_case : Union[str, Any]=0.1 , __snake_case : str=0.0 , __snake_case : Tuple=0.1 , __snake_case : Dict=None , __snake_case : Optional[int]=1E-9 , __snake_case : Tuple="mean" , __snake_case : List[str]="relative_shift" , __snake_case : Dict=True , __snake_case : int=True , __snake_case : Optional[Any]=True , **__snake_case : Optional[Any] , ) -> Optional[int]: UpperCAmelCase : Optional[int] = vocab_size UpperCAmelCase : List[Any] = block_sizes UpperCAmelCase : Optional[int] = [1] * len(__snake_case ) if block_repeats is None else block_repeats assert len(__snake_case ) == len( self.block_repeats ), "`block_sizes` and `block_repeats` should have the same length." UpperCAmelCase : Optional[int] = num_decoder_layers UpperCAmelCase : List[Any] = d_model UpperCAmelCase : List[str] = n_head UpperCAmelCase : Optional[Any] = d_head UpperCAmelCase : Union[str, Any] = d_inner UpperCAmelCase : str = hidden_act UpperCAmelCase : int = hidden_dropout UpperCAmelCase : List[Any] = attention_dropout UpperCAmelCase : Any = activation_dropout UpperCAmelCase : str = initializer_range UpperCAmelCase : List[Any] = initializer_std UpperCAmelCase : Tuple = layer_norm_eps assert pooling_type in [ "mean", "max", ], F"""Got {pooling_type} for `pooling_type` but only 'mean' and 'max' are supported.""" UpperCAmelCase : Optional[Any] = pooling_type assert attention_type in [ "relative_shift", "factorized", ], F"""Got {attention_type} for `attention_type` but only 'relative_shift' and 'factorized' are supported.""" UpperCAmelCase : Optional[int] = attention_type UpperCAmelCase : str = separate_cls UpperCAmelCase : List[str] = truncate_seq UpperCAmelCase : Tuple = pool_q_only super().__init__(**__snake_case ) @property def A ( self : int ) -> Optional[int]: return sum(self.block_sizes ) @num_hidden_layers.setter def A ( self : List[Any] , __snake_case : List[str] ) -> Union[str, Any]: raise NotImplementedError( '''This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`.''' ) @property def A ( self : Dict ) -> int: return len(self.block_sizes ) @num_blocks.setter def A ( self : Dict , __snake_case : Optional[Any] ) -> Any: raise NotImplementedError('''This model does not support the setting of `num_blocks`. Please set `block_sizes`.''' )
23
'''simple docstring''' import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] ) -> Optional[int]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[int] ) -> Dict: UpperCAmelCase : Dict = tmp_path / '''cache''' UpperCAmelCase : List[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase : str = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict , _lowerCAmelCase : int ) -> Optional[int]: UpperCAmelCase : Any = tmp_path / '''cache''' UpperCAmelCase : Tuple = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Any = features.copy() if features else default_expected_features UpperCAmelCase : List[Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Dict = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Tuple ) -> Tuple: UpperCAmelCase : Optional[Any] = tmp_path / '''cache''' UpperCAmelCase : Optional[int] = {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''} UpperCAmelCase : int = features.copy() if features else default_expected_features UpperCAmelCase : Any = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict ) -> Union[str, Any]: # jsonl_312_path features are {"col_3": "float64", "col_1": "string", "col_2": "int64"} UpperCAmelCase : Tuple = {'''col_2''': '''int64''', '''col_3''': '''float64''', '''col_1''': '''string'''} UpperCAmelCase : List[str] = features.copy() UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = tmp_path / '''cache''' UpperCAmelCase : List[str] = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ) -> Optional[Any]: UpperCAmelCase : Any = tmp_path / '''cache''' UpperCAmelCase : List[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : List[Any] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , split=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('''path_type''' , [str, list] ) def snake_case_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> Dict: if issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase : str = jsonl_path elif issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase : Dict = [jsonl_path] UpperCAmelCase : int = tmp_path / '''cache''' UpperCAmelCase : Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[int] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : Dict=("train",) ) -> Union[str, Any]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) for split in splits: UpperCAmelCase : List[str] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ) -> Any: UpperCAmelCase : Optional[Any] = tmp_path / '''cache''' UpperCAmelCase : List[str] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase : Optional[int] = JsonDatasetReader({'''train''': jsonl_path} , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] ) -> int: UpperCAmelCase : Dict = tmp_path / '''cache''' UpperCAmelCase : Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[int] = features.copy() if features else default_expected_features UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = JsonDatasetReader({'''train''': jsonl_path} , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict ) -> Union[str, Any]: if split: UpperCAmelCase : Optional[int] = {split: jsonl_path} else: UpperCAmelCase : Any = '''train''' UpperCAmelCase : Any = {'''train''': jsonl_path, '''test''': jsonl_path} UpperCAmelCase : Tuple = tmp_path / '''cache''' UpperCAmelCase : int = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[Any] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def snake_case_ ( _lowerCAmelCase : List[str] ) -> str: return json.load(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Dict ) -> str: return [json.loads(_lowerCAmelCase ) for line in buffer] class SCREAMING_SNAKE_CASE: """simple docstring""" @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def A ( self : Union[str, Any] , __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : Optional[int] ) -> Dict: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case ).write() buffer.seek(0 ) UpperCAmelCase : Union[str, Any] = load_json_function(__snake_case ) assert isinstance(__snake_case , __snake_case ) assert isinstance(exported_content[0] , __snake_case ) assert len(__snake_case ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def A ( self : Optional[int] , __snake_case : Optional[Any] , __snake_case : str , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : Optional[Any] ) -> List[Any]: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , orient=__snake_case ).write() buffer.seek(0 ) UpperCAmelCase : Union[str, Any] = load_json(__snake_case ) assert isinstance(__snake_case , __snake_case ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(__snake_case , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(__snake_case ) == 10 @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def A ( self : str , __snake_case : str , __snake_case : str , __snake_case : int ) -> Any: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase : Any = load_json_function(__snake_case ) assert isinstance(__snake_case , __snake_case ) assert isinstance(exported_content[0] , __snake_case ) assert len(__snake_case ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def A ( self : Any , __snake_case : int , __snake_case : Union[str, Any] , __snake_case : Optional[int] , __snake_case : Optional[Any] , __snake_case : List[str] ) -> Any: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , orient=__snake_case , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase : List[str] = load_json(__snake_case ) assert isinstance(__snake_case , __snake_case ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(__snake_case , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(__snake_case ) == 10 def A ( self : List[Any] , __snake_case : str ) -> Dict: with pytest.raises(__snake_case ): with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , num_proc=0 ) @pytest.mark.parametrize('''compression, extension''' , [('''gzip''', '''gz'''), ('''bz2''', '''bz2'''), ('''xz''', '''xz''')] ) def A ( self : Optional[int] , __snake_case : Any , __snake_case : str , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Dict ) -> Union[str, Any]: UpperCAmelCase : List[str] = tmp_path_factory.mktemp('''data''' ) / F"""test.json.{extension}""" UpperCAmelCase : List[Any] = str(shared_datadir / F"""test_file.json.{extension}""" ) JsonDatasetWriter(__snake_case , __snake_case , compression=__snake_case ).write() with fsspec.open(__snake_case , '''rb''' , compression='''infer''' ) as f: UpperCAmelCase : str = f.read() with fsspec.open(__snake_case , '''rb''' , compression='''infer''' ) as f: UpperCAmelCase : Optional[int] = f.read() assert exported_content == original_content
23
1
'''simple docstring''' import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # and perform gradient accumulation # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## UpperCamelCase__: Union[str, Any] = 16 UpperCamelCase__: Any = 32 def snake_case_ ( _lowerCAmelCase : Accelerator , _lowerCAmelCase : int = 16 ) -> Optional[int]: UpperCAmelCase : Tuple = AutoTokenizer.from_pretrained('''bert-base-cased''' ) UpperCAmelCase : str = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(_lowerCAmelCase : str ): # max_length=None => use the model max length (it's actually the default) UpperCAmelCase : Optional[int] = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=_lowerCAmelCase , max_length=_lowerCAmelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): UpperCAmelCase : Tuple = datasets.map( _lowerCAmelCase , batched=_lowerCAmelCase , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library UpperCAmelCase : Dict = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(_lowerCAmelCase : List[str] ): # On TPU it's best to pad everything to the same length or training will be very slow. UpperCAmelCase : Tuple = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": UpperCAmelCase : Any = 16 elif accelerator.mixed_precision != "no": UpperCAmelCase : int = 8 else: UpperCAmelCase : Union[str, Any] = None return tokenizer.pad( _lowerCAmelCase , padding='''longest''' , max_length=_lowerCAmelCase , pad_to_multiple_of=_lowerCAmelCase , return_tensors='''pt''' , ) # Instantiate dataloaders. UpperCAmelCase : Optional[int] = DataLoader( tokenized_datasets['''train'''] , shuffle=_lowerCAmelCase , collate_fn=_lowerCAmelCase , batch_size=_lowerCAmelCase ) UpperCAmelCase : Dict = DataLoader( tokenized_datasets['''validation'''] , shuffle=_lowerCAmelCase , collate_fn=_lowerCAmelCase , batch_size=_lowerCAmelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("TESTING_MOCKED_DATALOADERS", None) == "1": from accelerate.test_utils.training import mocked_dataloaders UpperCamelCase__: Dict = mocked_dataloaders # noqa: F811 def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : int ) -> Any: # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , _lowerCAmelCase ) == "1": UpperCAmelCase : Tuple = 2 # New Code # UpperCAmelCase : Optional[int] = int(args.gradient_accumulation_steps ) # Initialize accelerator UpperCAmelCase : Tuple = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , gradient_accumulation_steps=_lowerCAmelCase ) if accelerator.distributed_type == DistributedType.TPU and gradient_accumulation_steps > 1: raise NotImplementedError( '''Gradient accumulation on TPUs is currently not supported. Pass `gradient_accumulation_steps=1`''' ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs UpperCAmelCase : List[Any] = config['''lr'''] UpperCAmelCase : Optional[int] = int(config['''num_epochs'''] ) UpperCAmelCase : List[str] = int(config['''seed'''] ) UpperCAmelCase : Union[str, Any] = int(config['''batch_size'''] ) UpperCAmelCase : Dict = evaluate.load('''glue''' , '''mrpc''' ) set_seed(_lowerCAmelCase ) UpperCAmelCase , UpperCAmelCase : int = get_dataloaders(_lowerCAmelCase , _lowerCAmelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) UpperCAmelCase : List[Any] = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=_lowerCAmelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). UpperCAmelCase : Optional[Any] = model.to(accelerator.device ) # Instantiate optimizer UpperCAmelCase : List[Any] = AdamW(params=model.parameters() , lr=_lowerCAmelCase ) # Instantiate scheduler UpperCAmelCase : int = get_linear_schedule_with_warmup( optimizer=_lowerCAmelCase , num_warmup_steps=100 , num_training_steps=(len(_lowerCAmelCase ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Optional[Any] = accelerator.prepare( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Now we train the model for epoch in range(_lowerCAmelCase ): model.train() for step, batch in enumerate(_lowerCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(_lowerCAmelCase ): UpperCAmelCase : Any = model(**_lowerCAmelCase ) UpperCAmelCase : Any = output.loss accelerator.backward(_lowerCAmelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(_lowerCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): UpperCAmelCase : List[str] = model(**_lowerCAmelCase ) UpperCAmelCase : int = outputs.logits.argmax(dim=-1 ) UpperCAmelCase , UpperCAmelCase : int = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=_lowerCAmelCase , references=_lowerCAmelCase , ) UpperCAmelCase : str = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , _lowerCAmelCase ) def snake_case_ ( ) -> Union[str, Any]: UpperCAmelCase : Any = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=_lowerCAmelCase , default=_lowerCAmelCase , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) # New Code # parser.add_argument( '''--gradient_accumulation_steps''' , type=_lowerCAmelCase , default=1 , help='''The number of minibatches to be ran before gradients are accumulated.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) UpperCAmelCase : Optional[int] = parser.parse_args() UpperCAmelCase : Dict = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(_lowerCAmelCase , _lowerCAmelCase ) if __name__ == "__main__": main()
23
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCamelCase__: Tuple = logging.get_logger(__name__) UpperCamelCase__: Optional[int] = {"vocab_file": "sentencepiece.bpe.model"} UpperCamelCase__: Optional[int] = { "vocab_file": { "moussaKam/mbarthez": "https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez": "https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez-orangesum-title": ( "https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model" ), }, } UpperCamelCase__: Dict = { "moussaKam/mbarthez": 1024, "moussaKam/barthez": 1024, "moussaKam/barthez-orangesum-title": 1024, } UpperCamelCase__: Tuple = "▁" class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = ["""input_ids""", """attention_mask"""] def __init__( self : List[Any] , __snake_case : List[Any] , __snake_case : Tuple="<s>" , __snake_case : List[Any]="</s>" , __snake_case : int="</s>" , __snake_case : Any="<s>" , __snake_case : Optional[int]="<unk>" , __snake_case : Union[str, Any]="<pad>" , __snake_case : Union[str, Any]="<mask>" , __snake_case : Optional[Dict[str, Any]] = None , **__snake_case : Dict , ) -> None: # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase : int = AddedToken(__snake_case , lstrip=__snake_case , rstrip=__snake_case ) if isinstance(__snake_case , __snake_case ) else mask_token UpperCAmelCase : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__snake_case , eos_token=__snake_case , unk_token=__snake_case , sep_token=__snake_case , cls_token=__snake_case , pad_token=__snake_case , mask_token=__snake_case , sp_model_kwargs=self.sp_model_kwargs , **__snake_case , ) UpperCAmelCase : Optional[int] = vocab_file UpperCAmelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__snake_case ) ) UpperCAmelCase : int = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} UpperCAmelCase : Optional[Any] = len(self.sp_model ) - 1 UpperCAmelCase : Optional[int] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def A ( self : Tuple , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase : Union[str, Any] = [self.cls_token_id] UpperCAmelCase : Optional[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A ( self : int , __snake_case : List[int] , __snake_case : Optional[List[int]] = None , __snake_case : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__snake_case , token_ids_a=__snake_case , already_has_special_tokens=__snake_case ) if token_ids_a is None: return [1] + ([0] * len(__snake_case )) + [1] return [1] + ([0] * len(__snake_case )) + [1, 1] + ([0] * len(__snake_case )) + [1] def A ( self : Optional[int] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: UpperCAmelCase : Tuple = [self.sep_token_id] UpperCAmelCase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def A ( self : Dict ) -> Optional[int]: return len(self.sp_model ) def A ( self : List[str] ) -> Dict: UpperCAmelCase : Optional[Any] = {self.convert_ids_to_tokens(__snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def A ( self : Optional[Any] , __snake_case : str ) -> List[str]: return self.sp_model.encode(__snake_case , out_type=__snake_case ) def A ( self : int , __snake_case : int ) -> int: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCAmelCase : Optional[Any] = self.sp_model.PieceToId(__snake_case ) return spm_id if spm_id else self.unk_token_id def A ( self : int , __snake_case : Any ) -> List[Any]: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(__snake_case ) def A ( self : List[Any] , __snake_case : Union[str, Any] ) -> List[str]: UpperCAmelCase : Optional[Any] = [] UpperCAmelCase : int = '''''' UpperCAmelCase : Union[str, Any] = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__snake_case ) + token UpperCAmelCase : str = True UpperCAmelCase : List[str] = [] else: current_sub_tokens.append(__snake_case ) UpperCAmelCase : Optional[int] = False out_string += self.sp_model.decode(__snake_case ) return out_string.strip() def __getstate__( self : Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase : Optional[Any] = self.__dict__.copy() UpperCAmelCase : Any = None return state def __setstate__( self : Optional[int] , __snake_case : Union[str, Any] ) -> List[Any]: UpperCAmelCase : Dict = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): UpperCAmelCase : Optional[Any] = {} UpperCAmelCase : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def A ( self : Optional[int] , __snake_case : str , __snake_case : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__snake_case ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase : Union[str, Any] = os.path.join( __snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __snake_case ) elif not os.path.isfile(self.vocab_file ): with open(__snake_case , '''wb''' ) as fi: UpperCAmelCase : Any = self.sp_model.serialized_model_proto() fi.write(__snake_case ) return (out_vocab_file,)
23
1
'''simple docstring''' import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging UpperCamelCase__: Tuple = logging.get_logger(__name__) UpperCamelCase__: Optional[Any] = "▁" UpperCamelCase__: Tuple = { "vocab_file": "vocab.json", "spm_file": "sentencepiece.bpe.model", "tokenizer_config_file": "tokenizer_config.json", } UpperCamelCase__: Any = { "vocab_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json", }, "spm_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model", }, "tokenizer_config_file": { "facebook/m2m100_418M": "https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json", "facebook/m2m100_1.2B": "https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json", }, } UpperCamelCase__: Dict = { "facebook/m2m100_418M": 1024, } # fmt: off UpperCamelCase__: str = { "m2m100": ["af", "am", "ar", "ast", "az", "ba", "be", "bg", "bn", "br", "bs", "ca", "ceb", "cs", "cy", "da", "de", "el", "en", "es", "et", "fa", "ff", "fi", "fr", "fy", "ga", "gd", "gl", "gu", "ha", "he", "hi", "hr", "ht", "hu", "hy", "id", "ig", "ilo", "is", "it", "ja", "jv", "ka", "kk", "km", "kn", "ko", "lb", "lg", "ln", "lo", "lt", "lv", "mg", "mk", "ml", "mn", "mr", "ms", "my", "ne", "nl", "no", "ns", "oc", "or", "pa", "pl", "ps", "pt", "ro", "ru", "sd", "si", "sk", "sl", "so", "sq", "sr", "ss", "su", "sv", "sw", "ta", "th", "tl", "tn", "tr", "uk", "ur", "uz", "vi", "wo", "xh", "yi", "yo", "zh", "zu"], "wmt21": ["en", "ha", "is", "ja", "cs", "ru", "zh", "de"] } class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = ["""input_ids""", """attention_mask"""] lowerCamelCase__ = [] lowerCamelCase__ = [] def __init__( self : Dict , __snake_case : Any , __snake_case : Tuple , __snake_case : Any=None , __snake_case : Any=None , __snake_case : Optional[int]="<s>" , __snake_case : List[Any]="</s>" , __snake_case : Optional[int]="</s>" , __snake_case : Dict="<pad>" , __snake_case : str="<unk>" , __snake_case : int="m2m100" , __snake_case : Optional[Dict[str, Any]] = None , __snake_case : int=8 , **__snake_case : Optional[Any] , ) -> None: UpperCAmelCase : List[Any] = {} if sp_model_kwargs is None else sp_model_kwargs UpperCAmelCase : Tuple = language_codes UpperCAmelCase : Union[str, Any] = FAIRSEQ_LANGUAGE_CODES[language_codes] UpperCAmelCase : Union[str, Any] = {lang_code: F"""__{lang_code}__""" for lang_code in fairseq_language_code} UpperCAmelCase : List[Any] = kwargs.get('''additional_special_tokens''' , [] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(__snake_case ) for lang_code in fairseq_language_code if self.get_lang_token(__snake_case ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=__snake_case , tgt_lang=__snake_case , bos_token=__snake_case , eos_token=__snake_case , sep_token=__snake_case , unk_token=__snake_case , pad_token=__snake_case , language_codes=__snake_case , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=__snake_case , **__snake_case , ) UpperCAmelCase : Any = vocab_file UpperCAmelCase : List[Any] = load_json(__snake_case ) UpperCAmelCase : Any = {v: k for k, v in self.encoder.items()} UpperCAmelCase : Tuple = spm_file UpperCAmelCase : Tuple = load_spm(__snake_case , self.sp_model_kwargs ) UpperCAmelCase : Optional[Any] = len(self.encoder ) UpperCAmelCase : Union[str, Any] = { self.get_lang_token(__snake_case ): self.encoder_size + i for i, lang_code in enumerate(__snake_case ) } UpperCAmelCase : Optional[Any] = {lang_code: self.encoder_size + i for i, lang_code in enumerate(__snake_case )} UpperCAmelCase : List[Any] = {v: k for k, v in self.lang_token_to_id.items()} UpperCAmelCase : Union[str, Any] = src_lang if src_lang is not None else '''en''' UpperCAmelCase : Union[str, Any] = tgt_lang UpperCAmelCase : List[str] = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) UpperCAmelCase : Optional[int] = num_madeup_words @property def A ( self : int ) -> int: return len(self.encoder ) + len(self.lang_token_to_id ) @property def A ( self : List[str] ) -> str: return self._src_lang @src_lang.setter def A ( self : str , __snake_case : str ) -> None: UpperCAmelCase : Union[str, Any] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def A ( self : List[str] , __snake_case : str ) -> List[str]: return self.sp_model.encode(__snake_case , out_type=__snake_case ) def A ( self : str , __snake_case : Tuple ) -> Tuple: if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(__snake_case , self.encoder[self.unk_token] ) def A ( self : Any , __snake_case : int ) -> str: if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(__snake_case , self.unk_token ) def A ( self : Optional[Any] , __snake_case : Tuple ) -> Tuple: UpperCAmelCase : Optional[int] = [] UpperCAmelCase : Dict = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__snake_case ) + token UpperCAmelCase : Optional[Any] = [] else: current_sub_tokens.append(__snake_case ) out_string += self.sp_model.decode(__snake_case ) return out_string.strip() def A ( self : Dict , __snake_case : List[int] , __snake_case : Optional[List[int]] = None , __snake_case : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__snake_case , token_ids_a=__snake_case , already_has_special_tokens=__snake_case ) UpperCAmelCase : Optional[Any] = [1] * len(self.prefix_tokens ) UpperCAmelCase : Union[str, Any] = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(__snake_case )) + suffix_ones return prefix_ones + ([0] * len(__snake_case )) + ([0] * len(__snake_case )) + suffix_ones def A ( self : Tuple , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def A ( self : Optional[Any] ) -> Dict: UpperCAmelCase : Optional[int] = {self.convert_ids_to_tokens(__snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Optional[Any] ) -> Dict: UpperCAmelCase : List[Any] = self.__dict__.copy() UpperCAmelCase : int = None return state def __setstate__( self : List[Any] , __snake_case : Dict ) -> None: UpperCAmelCase : Optional[int] = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): UpperCAmelCase : Optional[int] = {} UpperCAmelCase : Optional[Any] = load_spm(self.spm_file , self.sp_model_kwargs ) def A ( self : int , __snake_case : str , __snake_case : Optional[str] = None ) -> Tuple[str]: UpperCAmelCase : Optional[Any] = Path(__snake_case ) if not save_dir.is_dir(): raise OSError(F"""{save_directory} should be a directory""" ) UpperCAmelCase : List[Any] = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''vocab_file'''] ) UpperCAmelCase : Optional[int] = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''spm_file'''] ) save_json(self.encoder , __snake_case ) if os.path.abspath(self.spm_file ) != os.path.abspath(__snake_case ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , __snake_case ) elif not os.path.isfile(self.spm_file ): with open(__snake_case , '''wb''' ) as fi: UpperCAmelCase : Dict = self.sp_model.serialized_model_proto() fi.write(__snake_case ) return (str(__snake_case ), str(__snake_case )) def A ( self : int , __snake_case : List[str] , __snake_case : str = "en" , __snake_case : Optional[List[str]] = None , __snake_case : str = "ro" , **__snake_case : Optional[int] , ) -> BatchEncoding: UpperCAmelCase : List[Any] = src_lang UpperCAmelCase : List[Any] = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(__snake_case , __snake_case , **__snake_case ) def A ( self : int , __snake_case : Any , __snake_case : Optional[str] , __snake_case : Optional[str] , **__snake_case : Any ) -> Optional[int]: if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) UpperCAmelCase : Union[str, Any] = src_lang UpperCAmelCase : Dict = self(__snake_case , add_special_tokens=__snake_case , **__snake_case ) UpperCAmelCase : Any = self.get_lang_id(__snake_case ) UpperCAmelCase : Dict = tgt_lang_id return inputs def A ( self : str ) -> int: self.set_src_lang_special_tokens(self.src_lang ) def A ( self : Any ) -> int: self.set_tgt_lang_special_tokens(self.tgt_lang ) def A ( self : Any , __snake_case : str ) -> None: UpperCAmelCase : Dict = self.get_lang_token(__snake_case ) UpperCAmelCase : str = self.lang_token_to_id[lang_token] UpperCAmelCase : Any = [self.cur_lang_id] UpperCAmelCase : str = [self.eos_token_id] def A ( self : Dict , __snake_case : str ) -> None: UpperCAmelCase : Union[str, Any] = self.get_lang_token(__snake_case ) UpperCAmelCase : str = self.lang_token_to_id[lang_token] UpperCAmelCase : Optional[int] = [self.cur_lang_id] UpperCAmelCase : int = [self.eos_token_id] def A ( self : Optional[Any] , __snake_case : str ) -> str: return self.lang_code_to_token[lang] def A ( self : Any , __snake_case : str ) -> int: UpperCAmelCase : Any = self.get_lang_token(__snake_case ) return self.lang_token_to_id[lang_token] def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : Dict[str, Any] ) -> sentencepiece.SentencePieceProcessor: UpperCAmelCase : Any = sentencepiece.SentencePieceProcessor(**_lowerCAmelCase ) spm.Load(str(_lowerCAmelCase ) ) return spm def snake_case_ ( _lowerCAmelCase : str ) -> Union[Dict, List]: with open(_lowerCAmelCase , '''r''' ) as f: return json.load(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : str ) -> None: with open(_lowerCAmelCase , '''w''' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase , indent=2 )
23
'''simple docstring''' import tempfile import unittest import numpy as np from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import BertConfig, is_flax_available from transformers.testing_utils import TOKEN, USER, is_staging_test, require_flax if is_flax_available(): import os from flax.core.frozen_dict import unfreeze from flax.traverse_util import flatten_dict from transformers import FlaxBertModel UpperCamelCase__: Tuple = "0.12" # assumed parallelism: 8 @require_flax @is_staging_test class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" @classmethod def A ( cls : Union[str, Any] ) -> int: UpperCAmelCase : Optional[Any] = TOKEN HfFolder.save_token(__snake_case ) @classmethod def A ( cls : List[str] ) -> Tuple: try: delete_repo(token=cls._token , repo_id='''test-model-flax''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-model-flax-org''' ) except HTTPError: pass def A ( self : int ) -> Tuple: UpperCAmelCase : List[Any] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) model.push_to_hub('''test-model-flax''' , use_auth_token=self._token ) UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) UpperCAmelCase : List[Any] = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Tuple = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Union[str, Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''test-model-flax''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(__snake_case , repo_id='''test-model-flax''' , push_to_hub=__snake_case , use_auth_token=self._token ) UpperCAmelCase : str = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) UpperCAmelCase : Any = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : str = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Optional[Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) def A ( self : Optional[Any] ) -> Union[str, Any]: UpperCAmelCase : Dict = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Optional[Any] = FlaxBertModel(__snake_case ) model.push_to_hub('''valid_org/test-model-flax-org''' , use_auth_token=self._token ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) UpperCAmelCase : List[Any] = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : int = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Any = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-model-flax-org''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained( __snake_case , repo_id='''valid_org/test-model-flax-org''' , push_to_hub=__snake_case , use_auth_token=self._token ) UpperCAmelCase : str = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) UpperCAmelCase : Any = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Optional[Any] = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : int = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Union[str, Any]: UpperCAmelCase : str = True UpperCAmelCase : int = flatten_dict(modela.params ) UpperCAmelCase : Dict = flatten_dict(modela.params ) for key in flat_params_a.keys(): if np.sum(np.abs(flat_params_a[key] - flat_params_a[key] ) ) > 1e-4: UpperCAmelCase : Dict = False return models_are_equal @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : Tuple ) -> Union[str, Any]: UpperCAmelCase : List[Any] = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) UpperCAmelCase : int = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(__snake_case , __snake_case ) ) with self.assertRaises(__snake_case ): UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : str = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertTrue(check_models_equal(__snake_case , __snake_case ) ) def A ( self : List[str] ) -> Dict: UpperCAmelCase : Dict = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) UpperCAmelCase : Optional[int] = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(__snake_case , __snake_case ) , max_shard_size='''10KB''' ) with self.assertRaises(__snake_case ): UpperCAmelCase : Any = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertTrue(check_models_equal(__snake_case , __snake_case ) ) def A ( self : Optional[int] ) -> str: UpperCAmelCase : Dict = '''bert''' UpperCAmelCase : int = '''hf-internal-testing/tiny-random-bert-subfolder''' with self.assertRaises(__snake_case ): UpperCAmelCase : Optional[Any] = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertIsNotNone(__snake_case ) def A ( self : Dict ) -> List[Any]: UpperCAmelCase : Optional[int] = '''bert''' UpperCAmelCase : int = '''hf-internal-testing/tiny-random-bert-sharded-subfolder''' with self.assertRaises(__snake_case ): UpperCAmelCase : Dict = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertIsNotNone(__snake_case )
23
1
'''simple docstring''' def snake_case_ ( _lowerCAmelCase : int ) -> bool: if num < 0: return False UpperCAmelCase : int = num UpperCAmelCase : int = 0 while num > 0: UpperCAmelCase : Any = rev_num * 10 + (num % 10) num //= 10 return num_copy == rev_num if __name__ == "__main__": import doctest doctest.testmod()
23
'''simple docstring''' import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE: """simple docstring""" def __init__( self : Optional[int] , __snake_case : str , __snake_case : Union[str, Any]=2 , __snake_case : Optional[int]=8 , __snake_case : Any=True , __snake_case : Union[str, Any]=True , __snake_case : Dict=True , __snake_case : int=True , __snake_case : List[Any]=99 , __snake_case : str=16 , __snake_case : Tuple=5 , __snake_case : Tuple=2 , __snake_case : str=36 , __snake_case : Dict="gelu" , __snake_case : str=0.0 , __snake_case : Optional[int]=0.0 , __snake_case : Optional[int]=512 , __snake_case : Optional[Any]=16 , __snake_case : int=2 , __snake_case : int=0.02 , __snake_case : str=3 , __snake_case : Dict=4 , __snake_case : str=None , ) -> Optional[int]: UpperCAmelCase : Optional[int] = parent UpperCAmelCase : Tuple = batch_size UpperCAmelCase : List[str] = seq_length UpperCAmelCase : List[Any] = is_training UpperCAmelCase : int = use_input_mask UpperCAmelCase : Any = use_token_type_ids UpperCAmelCase : str = use_labels UpperCAmelCase : Union[str, Any] = vocab_size UpperCAmelCase : List[str] = hidden_size UpperCAmelCase : Optional[Any] = num_hidden_layers UpperCAmelCase : Union[str, Any] = num_attention_heads UpperCAmelCase : Optional[Any] = intermediate_size UpperCAmelCase : Union[str, Any] = hidden_act UpperCAmelCase : int = hidden_dropout_prob UpperCAmelCase : Optional[int] = attention_probs_dropout_prob UpperCAmelCase : Union[str, Any] = max_position_embeddings UpperCAmelCase : str = type_vocab_size UpperCAmelCase : List[str] = type_sequence_label_size UpperCAmelCase : Tuple = initializer_range UpperCAmelCase : Optional[Any] = num_labels UpperCAmelCase : Optional[int] = num_choices UpperCAmelCase : Any = scope def A ( self : Tuple ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase : Optional[int] = None if self.use_input_mask: UpperCAmelCase : Dict = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase : Dict = None if self.use_token_type_ids: UpperCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase : str = None UpperCAmelCase : Tuple = None UpperCAmelCase : int = None if self.use_labels: UpperCAmelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase : str = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase : List[str] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : int ) -> Tuple: return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__snake_case , initializer_range=self.initializer_range , ) def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Optional[Any] = self.get_config() UpperCAmelCase : int = 300 return config def A ( self : Optional[Any] ) -> Any: ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) : Tuple = self.prepare_config_and_inputs() UpperCAmelCase : Dict = True UpperCAmelCase : Tuple = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def A ( self : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : int , __snake_case : Optional[int] , __snake_case : int , __snake_case : Dict , __snake_case : Tuple , __snake_case : Optional[Any] ) -> List[str]: UpperCAmelCase : int = MraModel(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Tuple = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case ) UpperCAmelCase : Optional[int] = model(__snake_case , token_type_ids=__snake_case ) UpperCAmelCase : Dict = model(__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Tuple , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : List[str] , __snake_case : int , __snake_case : Union[str, Any] , __snake_case : Optional[Any] , __snake_case : Any , __snake_case : List[Any] , __snake_case : Optional[Any] , ) -> Tuple: UpperCAmelCase : str = True UpperCAmelCase : Tuple = MraModel(__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Optional[int] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , encoder_hidden_states=__snake_case , encoder_attention_mask=__snake_case , ) UpperCAmelCase : Optional[Any] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , encoder_hidden_states=__snake_case , ) UpperCAmelCase : str = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Tuple , __snake_case : str , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : int ) -> Any: UpperCAmelCase : Dict = MraForMaskedLM(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Optional[int] = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : Tuple , __snake_case : Tuple , __snake_case : Dict , __snake_case : Dict , __snake_case : Any , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Tuple ) -> Optional[int]: UpperCAmelCase : List[str] = MraForQuestionAnswering(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : List[Any] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , start_positions=__snake_case , end_positions=__snake_case , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : str , __snake_case : Optional[int] , __snake_case : List[str] , __snake_case : str , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Union[str, Any] , __snake_case : List[Any] ) -> int: UpperCAmelCase : int = self.num_labels UpperCAmelCase : Union[str, Any] = MraForSequenceClassification(__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : List[str] = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : str , __snake_case : Dict , __snake_case : Optional[Any] , __snake_case : Dict , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Dict ) -> int: UpperCAmelCase : Tuple = self.num_labels UpperCAmelCase : List[str] = MraForTokenClassification(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : str = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : str , __snake_case : int , __snake_case : Any , __snake_case : Tuple , __snake_case : Optional[int] , __snake_case : List[str] , __snake_case : str , __snake_case : Union[str, Any] ) -> Optional[Any]: UpperCAmelCase : Tuple = self.num_choices UpperCAmelCase : int = MraForMultipleChoice(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : str = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[str] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[str] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : str ) -> Dict: UpperCAmelCase : Any = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) : List[str] = config_and_inputs UpperCAmelCase : Any = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = () def A ( self : int ) -> Union[str, Any]: UpperCAmelCase : List[str] = MraModelTester(self ) UpperCAmelCase : Optional[int] = ConfigTester(self , config_class=__snake_case , hidden_size=37 ) def A ( self : Optional[Any] ) -> str: self.config_tester.run_common_tests() def A ( self : Tuple ) -> Optional[Any]: UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__snake_case ) def A ( self : List[Any] ) -> Optional[Any]: UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase : List[Any] = type self.model_tester.create_and_check_model(*__snake_case ) def A ( self : Tuple ) -> Dict: UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__snake_case ) def A ( self : Tuple ) -> List[str]: UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__snake_case ) def A ( self : int ) -> Dict: UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__snake_case ) def A ( self : Dict ) -> Optional[int]: UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__snake_case ) def A ( self : Any ) -> Optional[int]: UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__snake_case ) @slow def A ( self : Dict ) -> Any: for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase : str = MraModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) @unittest.skip(reason='''MRA does not output attentions''' ) def A ( self : str ) -> Optional[Any]: return @require_torch class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" @slow def A ( self : Tuple ) -> List[Any]: UpperCAmelCase : int = MraModel.from_pretrained('''uw-madison/mra-base-512-4''' ) UpperCAmelCase : Optional[Any] = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : List[Any] = model(__snake_case )[0] UpperCAmelCase : Optional[Any] = torch.Size((1, 256, 768) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Any = torch.tensor( [[[-0.01_40, 0.08_30, -0.03_81], [0.15_46, 0.14_02, 0.02_20], [0.11_62, 0.08_51, 0.01_65]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) ) @slow def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Optional[int] = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-512-4''' ) UpperCAmelCase : Dict = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : List[Any] = model(__snake_case )[0] UpperCAmelCase : int = 50265 UpperCAmelCase : int = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Union[str, Any] = torch.tensor( [[[9.25_95, -3.60_38, 11.88_19], [9.38_69, -3.26_93, 11.09_56], [11.85_24, -3.49_38, 13.12_10]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) ) @slow def A ( self : str ) -> List[Any]: UpperCAmelCase : List[Any] = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-4096-8-d3''' ) UpperCAmelCase : List[Any] = torch.arange(4096 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : Tuple = model(__snake_case )[0] UpperCAmelCase : Optional[int] = 50265 UpperCAmelCase : Tuple = torch.Size((1, 4096, vocab_size) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Optional[int] = torch.tensor( [[[5.47_89, -2.35_64, 7.50_64], [7.90_67, -1.33_69, 9.96_68], [9.07_12, -1.81_06, 7.03_80]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) )
23
1
'''simple docstring''' from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxCrossAttnUpBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, FlaxUpBlockaD, ) @flax.struct.dataclass class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = 42 @flax_register_to_config class SCREAMING_SNAKE_CASE( nn.Module , A__ , A__ ): """simple docstring""" lowerCamelCase__ = 32 lowerCamelCase__ = 4 lowerCamelCase__ = 4 lowerCamelCase__ = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) lowerCamelCase__ = ("UpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D") lowerCamelCase__ = False lowerCamelCase__ = (320, 640, 1_280, 1_280) lowerCamelCase__ = 2 lowerCamelCase__ = 8 lowerCamelCase__ = None lowerCamelCase__ = 1_280 lowerCamelCase__ = 0.0 lowerCamelCase__ = False lowerCamelCase__ = jnp.floataa lowerCamelCase__ = True lowerCamelCase__ = 0 lowerCamelCase__ = False def A ( self : List[Any] , __snake_case : jax.random.KeyArray ) -> FrozenDict: # init input tensors UpperCAmelCase : List[Any] = (1, self.in_channels, self.sample_size, self.sample_size) UpperCAmelCase : int = jnp.zeros(__snake_case , dtype=jnp.floataa ) UpperCAmelCase : List[Any] = jnp.ones((1,) , dtype=jnp.intaa ) UpperCAmelCase : List[Any] = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) UpperCAmelCase , UpperCAmelCase : Optional[Any] = jax.random.split(__snake_case ) UpperCAmelCase : Dict = {'''params''': params_rng, '''dropout''': dropout_rng} return self.init(__snake_case , __snake_case , __snake_case , __snake_case )["params"] def A ( self : Tuple ) -> int: UpperCAmelCase : List[str] = self.block_out_channels UpperCAmelCase : Optional[Any] = block_out_channels[0] * 4 if self.num_attention_heads is not None: raise ValueError( '''At the moment it is not possible to define the number of attention heads via `num_attention_heads` because of a naming issue as described in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131. Passing `num_attention_heads` will only be supported in diffusers v0.19.''' ) # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. UpperCAmelCase : Union[str, Any] = self.num_attention_heads or self.attention_head_dim # input UpperCAmelCase : Union[str, Any] = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time UpperCAmelCase : Optional[Any] = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) UpperCAmelCase : Optional[Any] = FlaxTimestepEmbedding(__snake_case , dtype=self.dtype ) UpperCAmelCase : Tuple = self.only_cross_attention if isinstance(__snake_case , __snake_case ): UpperCAmelCase : Optional[int] = (only_cross_attention,) * len(self.down_block_types ) if isinstance(__snake_case , __snake_case ): UpperCAmelCase : Optional[int] = (num_attention_heads,) * len(self.down_block_types ) # down UpperCAmelCase : Optional[Any] = [] UpperCAmelCase : str = block_out_channels[0] for i, down_block_type in enumerate(self.down_block_types ): UpperCAmelCase : Dict = output_channel UpperCAmelCase : Any = block_out_channels[i] UpperCAmelCase : Optional[int] = i == len(__snake_case ) - 1 if down_block_type == "CrossAttnDownBlock2D": UpperCAmelCase : Tuple = FlaxCrossAttnDownBlockaD( in_channels=__snake_case , out_channels=__snake_case , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: UpperCAmelCase : Any = FlaxDownBlockaD( in_channels=__snake_case , out_channels=__snake_case , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(__snake_case ) UpperCAmelCase : List[Any] = down_blocks # mid UpperCAmelCase : Tuple = FlaxUNetMidBlockaDCrossAttn( in_channels=block_out_channels[-1] , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) # up UpperCAmelCase : Dict = [] UpperCAmelCase : List[Any] = list(reversed(__snake_case ) ) UpperCAmelCase : Optional[int] = list(reversed(__snake_case ) ) UpperCAmelCase : List[Any] = list(reversed(__snake_case ) ) UpperCAmelCase : List[Any] = reversed_block_out_channels[0] for i, up_block_type in enumerate(self.up_block_types ): UpperCAmelCase : int = output_channel UpperCAmelCase : Any = reversed_block_out_channels[i] UpperCAmelCase : List[Any] = reversed_block_out_channels[min(i + 1 , len(__snake_case ) - 1 )] UpperCAmelCase : Tuple = i == len(__snake_case ) - 1 if up_block_type == "CrossAttnUpBlock2D": UpperCAmelCase : List[str] = FlaxCrossAttnUpBlockaD( in_channels=__snake_case , out_channels=__snake_case , prev_output_channel=__snake_case , num_layers=self.layers_per_block + 1 , num_attention_heads=reversed_num_attention_heads[i] , add_upsample=not is_final_block , dropout=self.dropout , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: UpperCAmelCase : List[str] = FlaxUpBlockaD( in_channels=__snake_case , out_channels=__snake_case , prev_output_channel=__snake_case , num_layers=self.layers_per_block + 1 , add_upsample=not is_final_block , dropout=self.dropout , dtype=self.dtype , ) up_blocks.append(__snake_case ) UpperCAmelCase : int = output_channel UpperCAmelCase : str = up_blocks # out UpperCAmelCase : str = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) UpperCAmelCase : Union[str, Any] = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self : Tuple , __snake_case : Tuple , __snake_case : List[str] , __snake_case : str , __snake_case : int=None , __snake_case : Optional[int]=None , __snake_case : bool = True , __snake_case : bool = False , ) -> Union[FlaxUNetaDConditionOutput, Tuple]: # 1. time if not isinstance(__snake_case , jnp.ndarray ): UpperCAmelCase : List[str] = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(__snake_case , jnp.ndarray ) and len(timesteps.shape ) == 0: UpperCAmelCase : Union[str, Any] = timesteps.astype(dtype=jnp.floataa ) UpperCAmelCase : Any = jnp.expand_dims(__snake_case , 0 ) UpperCAmelCase : List[Any] = self.time_proj(__snake_case ) UpperCAmelCase : List[str] = self.time_embedding(__snake_case ) # 2. pre-process UpperCAmelCase : List[str] = jnp.transpose(__snake_case , (0, 2, 3, 1) ) UpperCAmelCase : List[str] = self.conv_in(__snake_case ) # 3. down UpperCAmelCase : List[Any] = (sample,) for down_block in self.down_blocks: if isinstance(__snake_case , __snake_case ): UpperCAmelCase , UpperCAmelCase : List[Any] = down_block(__snake_case , __snake_case , __snake_case , deterministic=not train ) else: UpperCAmelCase , UpperCAmelCase : Optional[Any] = down_block(__snake_case , __snake_case , deterministic=not train ) down_block_res_samples += res_samples if down_block_additional_residuals is not None: UpperCAmelCase : List[str] = () for down_block_res_sample, down_block_additional_residual in zip( __snake_case , __snake_case ): down_block_res_sample += down_block_additional_residual new_down_block_res_samples += (down_block_res_sample,) UpperCAmelCase : str = new_down_block_res_samples # 4. mid UpperCAmelCase : Any = self.mid_block(__snake_case , __snake_case , __snake_case , deterministic=not train ) if mid_block_additional_residual is not None: sample += mid_block_additional_residual # 5. up for up_block in self.up_blocks: UpperCAmelCase : Tuple = down_block_res_samples[-(self.layers_per_block + 1) :] UpperCAmelCase : int = down_block_res_samples[: -(self.layers_per_block + 1)] if isinstance(__snake_case , __snake_case ): UpperCAmelCase : Tuple = up_block( __snake_case , temb=__snake_case , encoder_hidden_states=__snake_case , res_hidden_states_tuple=__snake_case , deterministic=not train , ) else: UpperCAmelCase : List[Any] = up_block(__snake_case , temb=__snake_case , res_hidden_states_tuple=__snake_case , deterministic=not train ) # 6. post-process UpperCAmelCase : Optional[Any] = self.conv_norm_out(__snake_case ) UpperCAmelCase : str = nn.silu(__snake_case ) UpperCAmelCase : Optional[Any] = self.conv_out(__snake_case ) UpperCAmelCase : Dict = jnp.transpose(__snake_case , (0, 3, 1, 2) ) if not return_dict: return (sample,) return FlaxUNetaDConditionOutput(sample=__snake_case )
23
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : Any ) -> str: UpperCAmelCase : Any = { '''task_specific_params''': { '''summarization''': {'''length_penalty''': 1.0, '''max_length''': 128, '''min_length''': 12, '''num_beams''': 4}, '''summarization_cnn''': {'''length_penalty''': 2.0, '''max_length''': 142, '''min_length''': 56, '''num_beams''': 4}, '''summarization_xsum''': {'''length_penalty''': 1.0, '''max_length''': 62, '''min_length''': 11, '''num_beams''': 6}, } } UpperCAmelCase : int = { '''task_specific_params.summarization.length_penalty''': 1.0, '''task_specific_params.summarization.max_length''': 128, '''task_specific_params.summarization.min_length''': 12, '''task_specific_params.summarization.num_beams''': 4, '''task_specific_params.summarization_cnn.length_penalty''': 2.0, '''task_specific_params.summarization_cnn.max_length''': 142, '''task_specific_params.summarization_cnn.min_length''': 56, '''task_specific_params.summarization_cnn.num_beams''': 4, '''task_specific_params.summarization_xsum.length_penalty''': 1.0, '''task_specific_params.summarization_xsum.max_length''': 62, '''task_specific_params.summarization_xsum.min_length''': 11, '''task_specific_params.summarization_xsum.num_beams''': 6, } self.assertEqual(flatten_dict(__snake_case ) , __snake_case ) def A ( self : int ) -> str: UpperCAmelCase : Any = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(__snake_case ) , x.transpose() ) ) UpperCAmelCase : str = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def A ( self : str ) -> Union[str, Any]: UpperCAmelCase : Any = np.random.randn(3 , 4 ) UpperCAmelCase : List[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , transpose(__snake_case ).numpy() ) ) UpperCAmelCase : Tuple = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Any = torch.tensor(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , transpose(__snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_tf def A ( self : List[str] ) -> Optional[Any]: UpperCAmelCase : int = np.random.randn(3 , 4 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , transpose(__snake_case ).numpy() ) ) UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : str = tf.constant(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , transpose(__snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_flax def A ( self : Tuple ) -> Any: UpperCAmelCase : List[Any] = np.random.randn(3 , 4 ) UpperCAmelCase : List[str] = jnp.array(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , np.asarray(transpose(__snake_case ) ) ) ) UpperCAmelCase : Dict = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , np.asarray(transpose(__snake_case , axes=(1, 2, 0) ) ) ) ) def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , np.reshape(__snake_case , (4, 3) ) ) ) UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , np.reshape(__snake_case , (12, 5) ) ) ) @require_torch def A ( self : Union[str, Any] ) -> int: UpperCAmelCase : Dict = np.random.randn(3 , 4 ) UpperCAmelCase : Optional[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , reshape(__snake_case , (4, 3) ).numpy() ) ) UpperCAmelCase : List[str] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : List[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , reshape(__snake_case , (12, 5) ).numpy() ) ) @require_tf def A ( self : int ) -> List[str]: UpperCAmelCase : List[Any] = np.random.randn(3 , 4 ) UpperCAmelCase : List[str] = tf.constant(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , reshape(__snake_case , (4, 3) ).numpy() ) ) UpperCAmelCase : List[Any] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Optional[Any] = tf.constant(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , reshape(__snake_case , (12, 5) ).numpy() ) ) @require_flax def A ( self : Any ) -> Dict: UpperCAmelCase : Tuple = np.random.randn(3 , 4 ) UpperCAmelCase : Union[str, Any] = jnp.array(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , np.asarray(reshape(__snake_case , (4, 3) ) ) ) ) UpperCAmelCase : Any = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Optional[Any] = jnp.array(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , np.asarray(reshape(__snake_case , (12, 5) ) ) ) ) def A ( self : List[Any] ) -> List[Any]: UpperCAmelCase : Union[str, Any] = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(__snake_case ) , np.squeeze(__snake_case ) ) ) UpperCAmelCase : str = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , np.squeeze(__snake_case , axis=2 ) ) ) @require_torch def A ( self : Optional[int] ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : List[str] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , squeeze(__snake_case ).numpy() ) ) UpperCAmelCase : Any = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : str = torch.tensor(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , squeeze(__snake_case , axis=2 ).numpy() ) ) @require_tf def A ( self : Optional[Any] ) -> Dict: UpperCAmelCase : int = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , squeeze(__snake_case ).numpy() ) ) UpperCAmelCase : List[str] = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , squeeze(__snake_case , axis=2 ).numpy() ) ) @require_flax def A ( self : List[Any] ) -> Dict: UpperCAmelCase : Optional[int] = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , np.asarray(squeeze(__snake_case ) ) ) ) UpperCAmelCase : str = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , np.asarray(squeeze(__snake_case , axis=2 ) ) ) ) def A ( self : Optional[Any] ) -> int: UpperCAmelCase : Optional[Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , np.expand_dims(__snake_case , axis=1 ) ) ) @require_torch def A ( self : List[str] ) -> Tuple: UpperCAmelCase : Tuple = np.random.randn(3 , 4 ) UpperCAmelCase : Tuple = torch.tensor(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , expand_dims(__snake_case , axis=1 ).numpy() ) ) @require_tf def A ( self : List[str] ) -> Union[str, Any]: UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 ) UpperCAmelCase : Any = tf.constant(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , expand_dims(__snake_case , axis=1 ).numpy() ) ) @require_flax def A ( self : Any ) -> List[Any]: UpperCAmelCase : List[str] = np.random.randn(3 , 4 ) UpperCAmelCase : str = jnp.array(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , np.asarray(expand_dims(__snake_case , axis=1 ) ) ) )
23
1
'''simple docstring''' from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) UpperCamelCase__: Dict = logging.get_logger(__name__) # pylint: disable=invalid-name UpperCamelCase__: str = "\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-prior\")\n >>> pipe_prior.to(\"cuda\")\n >>> prompt = \"red cat, 4k photo\"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained(\"kandinsky-community/kandinsky-2-2-decoder\")\n >>> pipe.to(\"cuda\")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save(\"cat.png\")\n ```\n" def snake_case_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any]=8 ) -> Any: UpperCAmelCase : Union[str, Any] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 UpperCAmelCase : List[Any] = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" def __init__( self : List[str] , __snake_case : UNetaDConditionModel , __snake_case : DDPMScheduler , __snake_case : VQModel , ) -> Dict: super().__init__() self.register_modules( unet=__snake_case , scheduler=__snake_case , movq=__snake_case , ) UpperCAmelCase : Any = 2 ** (len(self.movq.config.block_out_channels ) - 1) def A ( self : Tuple , __snake_case : Tuple , __snake_case : List[str] , __snake_case : Any , __snake_case : int , __snake_case : Optional[Any] , __snake_case : int ) -> Optional[int]: if latents is None: UpperCAmelCase : Optional[int] = randn_tensor(__snake_case , generator=__snake_case , device=__snake_case , dtype=__snake_case ) else: if latents.shape != shape: raise ValueError(F"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) UpperCAmelCase : List[Any] = latents.to(__snake_case ) UpperCAmelCase : Optional[int] = latents * scheduler.init_noise_sigma return latents def A ( self : Dict , __snake_case : Dict=0 ) -> Any: if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('''Please install accelerate via `pip install accelerate`''' ) UpperCAmelCase : Tuple = torch.device(F"""cuda:{gpu_id}""" ) UpperCAmelCase : Tuple = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(__snake_case , __snake_case ) def A ( self : List[str] , __snake_case : Tuple=0 ) -> List[Any]: if is_accelerate_available() and is_accelerate_version('''>=''' , '''0.17.0.dev0''' ): from accelerate import cpu_offload_with_hook else: raise ImportError('''`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.''' ) UpperCAmelCase : Dict = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('''cpu''' , silence_dtype_warnings=__snake_case ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) UpperCAmelCase : Union[str, Any] = None for cpu_offloaded_model in [self.unet, self.movq]: UpperCAmelCase , UpperCAmelCase : Any = cpu_offload_with_hook(__snake_case , __snake_case , prev_module_hook=__snake_case ) # We'll offload the last model manually. UpperCAmelCase : Tuple = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def A ( self : List[str] ) -> Dict: if not hasattr(self.unet , '''_hf_hook''' ): return self.device for module in self.unet.modules(): if ( hasattr(__snake_case , '''_hf_hook''' ) and hasattr(module._hf_hook , '''execution_device''' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(__snake_case ) def __call__( self : Tuple , __snake_case : Union[torch.FloatTensor, List[torch.FloatTensor]] , __snake_case : Union[torch.FloatTensor, List[torch.FloatTensor]] , __snake_case : int = 512 , __snake_case : int = 512 , __snake_case : int = 100 , __snake_case : float = 4.0 , __snake_case : int = 1 , __snake_case : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __snake_case : Optional[torch.FloatTensor] = None , __snake_case : Optional[str] = "pil" , __snake_case : bool = True , ) -> str: UpperCAmelCase : int = self._execution_device UpperCAmelCase : Optional[int] = guidance_scale > 1.0 if isinstance(__snake_case , __snake_case ): UpperCAmelCase : Union[str, Any] = torch.cat(__snake_case , dim=0 ) UpperCAmelCase : str = image_embeds.shape[0] * num_images_per_prompt if isinstance(__snake_case , __snake_case ): UpperCAmelCase : Tuple = torch.cat(__snake_case , dim=0 ) if do_classifier_free_guidance: UpperCAmelCase : Optional[Any] = image_embeds.repeat_interleave(__snake_case , dim=0 ) UpperCAmelCase : int = negative_image_embeds.repeat_interleave(__snake_case , dim=0 ) UpperCAmelCase : Union[str, Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=__snake_case ) self.scheduler.set_timesteps(__snake_case , device=__snake_case ) UpperCAmelCase : Optional[int] = self.scheduler.timesteps UpperCAmelCase : Dict = self.unet.config.in_channels UpperCAmelCase , UpperCAmelCase : str = downscale_height_and_width(__snake_case , __snake_case , self.movq_scale_factor ) # create initial latent UpperCAmelCase : Optional[Any] = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , __snake_case , __snake_case , __snake_case , self.scheduler , ) for i, t in enumerate(self.progress_bar(__snake_case ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase : Optional[Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase : str = {'''image_embeds''': image_embeds} UpperCAmelCase : Tuple = self.unet( sample=__snake_case , timestep=__snake_case , encoder_hidden_states=__snake_case , added_cond_kwargs=__snake_case , return_dict=__snake_case , )[0] if do_classifier_free_guidance: UpperCAmelCase , UpperCAmelCase : List[Any] = noise_pred.split(latents.shape[1] , dim=1 ) UpperCAmelCase , UpperCAmelCase : str = noise_pred.chunk(2 ) UpperCAmelCase , UpperCAmelCase : int = variance_pred.chunk(2 ) UpperCAmelCase : Any = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) UpperCAmelCase : Optional[Any] = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , '''variance_type''' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): UpperCAmelCase , UpperCAmelCase : Dict = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase : Optional[int] = self.scheduler.step( __snake_case , __snake_case , __snake_case , generator=__snake_case , )[0] # post-processing UpperCAmelCase : Optional[int] = self.movq.decode(__snake_case , force_not_quantize=__snake_case )['''sample'''] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: UpperCAmelCase : List[Any] = image * 0.5 + 0.5 UpperCAmelCase : int = image.clamp(0 , 1 ) UpperCAmelCase : Any = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": UpperCAmelCase : Optional[Any] = self.numpy_to_pil(__snake_case ) if not return_dict: return (image,) return ImagePipelineOutput(images=__snake_case )
23
'''simple docstring''' import argparse import os import re import packaging.version UpperCamelCase__: Union[str, Any] = "examples/" UpperCamelCase__: Optional[Any] = { "examples": (re.compile(r"^check_min_version\(\"[^\"]+\"\)\s*$", re.MULTILINE), "check_min_version(\"VERSION\")\n"), "init": (re.compile(r"^__version__\s+=\s+\"([^\"]+)\"\s*$", re.MULTILINE), "__version__ = \"VERSION\"\n"), "setup": (re.compile(r"^(\s*)version\s*=\s*\"[^\"]+\",", re.MULTILINE), r"\1version=\"VERSION\","), "doc": (re.compile(r"^(\s*)release\s*=\s*\"[^\"]+\"$", re.MULTILINE), "release = \"VERSION\"\n"), } UpperCamelCase__: Optional[int] = { "init": "src/diffusers/__init__.py", "setup": "setup.py", } UpperCamelCase__: List[Any] = "README.md" def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int ) -> Optional[int]: with open(_lowerCAmelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: UpperCAmelCase : Optional[int] = f.read() UpperCAmelCase , UpperCAmelCase : List[Any] = REPLACE_PATTERNS[pattern] UpperCAmelCase : List[Any] = replace.replace('''VERSION''' , _lowerCAmelCase ) UpperCAmelCase : Optional[Any] = re_pattern.sub(_lowerCAmelCase , _lowerCAmelCase ) with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Any ) -> Optional[int]: for folder, directories, fnames in os.walk(_lowerCAmelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase , pattern='''examples''' ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : str=False ) -> List[str]: for pattern, fname in REPLACE_FILES.items(): update_version_in_file(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not patch: update_version_in_examples(_lowerCAmelCase ) def snake_case_ ( ) -> Optional[Any]: UpperCAmelCase : Optional[int] = '''🤗 Transformers currently provides the following architectures''' UpperCAmelCase : Optional[int] = '''1. Want to contribute a new model?''' with open(_lowerCAmelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: UpperCAmelCase : Optional[Any] = f.readlines() # Find the start of the list. UpperCAmelCase : List[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 UpperCAmelCase : Optional[Any] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): UpperCAmelCase : Optional[int] = lines[index].replace( '''https://huggingface.co/docs/diffusers/main/model_doc''' , '''https://huggingface.co/docs/diffusers/model_doc''' , ) index += 1 with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(_lowerCAmelCase ) def snake_case_ ( ) -> Optional[Any]: with open(REPLACE_FILES['''init'''] , '''r''' ) as f: UpperCAmelCase : Union[str, Any] = f.read() UpperCAmelCase : int = REPLACE_PATTERNS['''init'''][0].search(_lowerCAmelCase ).groups()[0] return packaging.version.parse(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : List[str]=False ) -> Any: UpperCAmelCase : Optional[Any] = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: UpperCAmelCase : Optional[int] = default_version.base_version elif patch: UpperCAmelCase : Union[str, Any] = f"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: UpperCAmelCase : Union[str, Any] = f"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. UpperCAmelCase : Dict = input(f"""Which version are you releasing? [{default_version}]""" ) if len(_lowerCAmelCase ) == 0: UpperCAmelCase : Tuple = default_version print(f"""Updating version to {version}.""" ) global_version_update(_lowerCAmelCase , patch=_lowerCAmelCase ) def snake_case_ ( ) -> Any: UpperCAmelCase : List[Any] = get_version() UpperCAmelCase : List[str] = f"""{current_version.major}.{current_version.minor + 1}.0.dev0""" UpperCAmelCase : List[Any] = current_version.base_version # Check with the user we got that right. UpperCAmelCase : Optional[int] = input(f"""Which version are we developing now? [{dev_version}]""" ) if len(_lowerCAmelCase ) == 0: UpperCAmelCase : Dict = dev_version print(f"""Updating version to {version}.""" ) global_version_update(_lowerCAmelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": UpperCamelCase__: Union[str, Any] = argparse.ArgumentParser() parser.add_argument("--post_release", action="store_true", help="Whether this is pre or post release.") parser.add_argument("--patch", action="store_true", help="Whether or not this is a patch release.") UpperCamelCase__: Optional[Any] = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("Nothing to do after a patch :-)") else: post_release_work()
23
1
'''simple docstring''' from __future__ import annotations UpperCamelCase__: List[str] = 10 def snake_case_ ( _lowerCAmelCase : list[int] ) -> list[int]: UpperCAmelCase : Tuple = 1 UpperCAmelCase : List[Any] = max(_lowerCAmelCase ) while placement <= max_digit: # declare and initialize empty buckets UpperCAmelCase : list[list] = [[] for _ in range(_lowerCAmelCase )] # split list_of_ints between the buckets for i in list_of_ints: UpperCAmelCase : List[Any] = int((i / placement) % RADIX ) buckets[tmp].append(_lowerCAmelCase ) # put each buckets' contents into list_of_ints UpperCAmelCase : Dict = 0 for b in range(_lowerCAmelCase ): for i in buckets[b]: UpperCAmelCase : str = i a += 1 # move to next placement *= RADIX return list_of_ints if __name__ == "__main__": import doctest doctest.testmod()
23
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake UpperCamelCase__: Tuple = numpy.array([0, 0]) UpperCamelCase__: Union[str, Any] = numpy.array([0.5, 0.8660254]) UpperCamelCase__: Dict = numpy.array([1, 0]) UpperCamelCase__: int = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] , _lowerCAmelCase : int ) -> list[numpy.ndarray]: UpperCAmelCase : Union[str, Any] = initial_vectors for _ in range(_lowerCAmelCase ): UpperCAmelCase : Union[str, Any] = iteration_step(_lowerCAmelCase ) return vectors def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] ) -> list[numpy.ndarray]: UpperCAmelCase : Tuple = [] for i, start_vector in enumerate(vectors[:-1] ): UpperCAmelCase : List[str] = vectors[i + 1] new_vectors.append(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def snake_case_ ( _lowerCAmelCase : numpy.ndarray , _lowerCAmelCase : float ) -> numpy.ndarray: UpperCAmelCase : List[str] = numpy.radians(_lowerCAmelCase ) UpperCAmelCase , UpperCAmelCase : Tuple = numpy.cos(_lowerCAmelCase ), numpy.sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = numpy.array(((c, -s), (s, c)) ) return numpy.dot(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] ) -> None: UpperCAmelCase : List[Any] = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() UpperCAmelCase , UpperCAmelCase : str = zip(*_lowerCAmelCase ) plt.plot(_lowerCAmelCase , _lowerCAmelCase ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__: List[Any] = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
23
1
'''simple docstring''' from __future__ import annotations from typing import Any def snake_case_ ( _lowerCAmelCase : list[Any] ) -> None: create_state_space_tree(_lowerCAmelCase , [] , 0 ) def snake_case_ ( _lowerCAmelCase : list[Any] , _lowerCAmelCase : list[Any] , _lowerCAmelCase : int ) -> None: if index == len(_lowerCAmelCase ): print(_lowerCAmelCase ) return create_state_space_tree(_lowerCAmelCase , _lowerCAmelCase , index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(_lowerCAmelCase , _lowerCAmelCase , index + 1 ) current_subsequence.pop() if __name__ == "__main__": UpperCamelCase__: list[Any] = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(["A", "B", "C"]) generate_all_subsequences(seq)
23
'''simple docstring''' from manim import * class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" def A ( self : Union[str, Any] ) -> List[str]: UpperCAmelCase : Optional[Any] = Rectangle(height=0.5 , width=0.5 ) UpperCAmelCase : str = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) UpperCAmelCase : Tuple = [mem.copy() for i in range(6 )] UpperCAmelCase : Optional[int] = [mem.copy() for i in range(6 )] UpperCAmelCase : Dict = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Any = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Union[str, Any] = VGroup(__snake_case , __snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Optional[Any] = Text('''CPU''' , font_size=24 ) UpperCAmelCase : Union[str, Any] = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__snake_case ) UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(4 )] UpperCAmelCase : Union[str, Any] = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : List[str] = Text('''GPU''' , font_size=24 ) UpperCAmelCase : Dict = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) gpu.move_to([-1, -1, 0] ) self.add(__snake_case ) UpperCAmelCase : int = [mem.copy() for i in range(6 )] UpperCAmelCase : Union[str, Any] = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : List[str] = Text('''Model''' , font_size=24 ) UpperCAmelCase : Tuple = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) model.move_to([3, -1.0, 0] ) self.add(__snake_case ) UpperCAmelCase : Any = [] for i, rect in enumerate(__snake_case ): rect.set_stroke(__snake_case ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) UpperCAmelCase : Dict = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(__snake_case , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=__snake_case ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=__snake_case , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=__snake_case , buff=0.0 ) self.add(__snake_case ) cpu_targs.append(__snake_case ) UpperCAmelCase : int = [mem.copy() for i in range(6 )] UpperCAmelCase : int = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Any = Text('''Loaded Checkpoint''' , font_size=24 ) UpperCAmelCase : Union[str, Any] = Group(__snake_case , __snake_case ).arrange(__snake_case , aligned_edge=__snake_case , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) UpperCAmelCase : Optional[int] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCAmelCase : str = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(__snake_case , __snake_case ) UpperCAmelCase : Tuple = MarkupText( F"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(__snake_case , DOWN * 2.4 , aligned_edge=key_text.get_left() ) UpperCAmelCase : List[Any] = MarkupText( F"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__snake_case ) , Write(__snake_case ) ) self.play(Write(__snake_case , run_time=1 ) , Create(__snake_case , run_time=1 ) ) UpperCAmelCase : Tuple = [] UpperCAmelCase : int = [] for i, rect in enumerate(__snake_case ): UpperCAmelCase : Any = fill.copy().set_fill(__snake_case , opacity=0.7 ) target.move_to(__snake_case ) first_animations.append(GrowFromCenter(__snake_case , run_time=1 ) ) UpperCAmelCase : List[str] = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(__snake_case , run_time=1.5 ) ) self.play(*__snake_case ) self.play(*__snake_case ) self.wait()
23
1
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE( A__ , A__ , A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = AltDiffusionPipeline lowerCamelCase__ = TEXT_TO_IMAGE_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS def A ( self : Dict ) -> int: torch.manual_seed(0 ) UpperCAmelCase : str = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) UpperCAmelCase : Dict = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) UpperCAmelCase : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) UpperCAmelCase : Dict = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5002 , ) UpperCAmelCase : List[Any] = CLIPTextModel(__snake_case ) UpperCAmelCase : Optional[Any] = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) UpperCAmelCase : Optional[int] = 77 UpperCAmelCase : Optional[int] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def A ( self : Optional[Any] , __snake_case : Dict , __snake_case : List[str]=0 ) -> Union[str, Any]: if str(__snake_case ).startswith('''mps''' ): UpperCAmelCase : str = torch.manual_seed(__snake_case ) else: UpperCAmelCase : Tuple = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) UpperCAmelCase : Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def A ( self : Union[str, Any] ) -> List[str]: super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def A ( self : Tuple ) -> List[str]: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def A ( self : Tuple ) -> Optional[int]: UpperCAmelCase : List[str] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase : Any = self.get_dummy_components() torch.manual_seed(0 ) UpperCAmelCase : Optional[int] = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase : List[str] = RobertaSeriesModelWithTransformation(__snake_case ) UpperCAmelCase : str = text_encoder UpperCAmelCase : Optional[int] = AltDiffusionPipeline(**__snake_case ) UpperCAmelCase : str = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : Optional[int] = self.get_dummy_inputs(__snake_case ) UpperCAmelCase : Optional[int] = '''A photo of an astronaut''' UpperCAmelCase : List[Any] = alt_pipe(**__snake_case ) UpperCAmelCase : Optional[Any] = output.images UpperCAmelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : List[str] = np.array( [0.5_74_81_62, 0.60_44_71_45, 0.48_82_12_17, 0.50_10_06_36, 0.5_43_11_85, 0.45_76_36_83, 0.49_65_76_96, 0.48_13_27_33, 0.47_57_30_93] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A ( self : int ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase : int = self.get_dummy_components() UpperCAmelCase : int = PNDMScheduler(skip_prk_steps=__snake_case ) torch.manual_seed(0 ) UpperCAmelCase : int = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase : Union[str, Any] = RobertaSeriesModelWithTransformation(__snake_case ) UpperCAmelCase : Union[str, Any] = text_encoder UpperCAmelCase : Optional[int] = AltDiffusionPipeline(**__snake_case ) UpperCAmelCase : Dict = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : int = self.get_dummy_inputs(__snake_case ) UpperCAmelCase : Optional[int] = alt_pipe(**__snake_case ) UpperCAmelCase : Optional[int] = output.images UpperCAmelCase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : Optional[int] = np.array( [0.51_60_50_93, 0.5_70_72_41, 0.47_36_55_07, 0.50_57_88_86, 0.5_63_38_77, 0.4_64_25_03, 0.5_18_20_81, 0.48_76_34_84, 0.49_08_42_37] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : str ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : List[Any] ) -> Any: # make sure here that pndm scheduler skips prk UpperCAmelCase : List[Any] = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''' , safety_checker=__snake_case ) UpperCAmelCase : Tuple = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : List[Any] = '''A painting of a squirrel eating a burger''' UpperCAmelCase : Any = torch.manual_seed(0 ) UpperCAmelCase : Optional[int] = alt_pipe([prompt] , generator=__snake_case , guidance_scale=6.0 , num_inference_steps=20 , output_type='''np''' ) UpperCAmelCase : Dict = output.images UpperCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase : List[str] = np.array([0.10_10, 0.08_00, 0.07_94, 0.08_85, 0.08_43, 0.07_62, 0.07_69, 0.07_29, 0.05_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A ( self : Tuple ) -> int: UpperCAmelCase : int = DDIMScheduler.from_pretrained('''BAAI/AltDiffusion''' , subfolder='''scheduler''' ) UpperCAmelCase : Tuple = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''' , scheduler=__snake_case , safety_checker=__snake_case ) UpperCAmelCase : Dict = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : Tuple = '''A painting of a squirrel eating a burger''' UpperCAmelCase : Optional[int] = torch.manual_seed(0 ) UpperCAmelCase : List[Any] = alt_pipe([prompt] , generator=__snake_case , num_inference_steps=2 , output_type='''numpy''' ) UpperCAmelCase : Dict = output.images UpperCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase : Union[str, Any] = np.array([0.40_19, 0.40_52, 0.38_10, 0.41_19, 0.39_16, 0.39_82, 0.46_51, 0.41_95, 0.53_23] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
23
'''simple docstring''' import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( "The converted tokenizer will be the `slow` tokenizer. To use the fast, update your `tokenizers` library and re-run the tokenizer conversion" ) UpperCamelCase__: str = None UpperCamelCase__: int = { "7B": 11008, "13B": 13824, "30B": 17920, "65B": 22016, "70B": 28672, } UpperCamelCase__: List[Any] = { "7B": 1, "7Bf": 1, "13B": 2, "13Bf": 2, "30B": 4, "65B": 8, "70B": 8, "70Bf": 8, } def snake_case_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple=1 , _lowerCAmelCase : List[Any]=256 ) -> Optional[Any]: return multiple_of * ((int(ffn_dim_multiplier * int(8 * n / 3 ) ) + multiple_of - 1) // multiple_of) def snake_case_ ( _lowerCAmelCase : List[str] ) -> str: with open(_lowerCAmelCase , '''r''' ) as f: return json.load(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any ) -> Optional[Any]: with open(_lowerCAmelCase , '''w''' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any=True ) -> List[Any]: os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : int = os.path.join(_lowerCAmelCase , '''tmp''' ) os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : List[str] = read_json(os.path.join(_lowerCAmelCase , '''params.json''' ) ) UpperCAmelCase : str = NUM_SHARDS[model_size] UpperCAmelCase : Any = params['''n_layers'''] UpperCAmelCase : str = params['''n_heads'''] UpperCAmelCase : Any = n_heads // num_shards UpperCAmelCase : List[str] = params['''dim'''] UpperCAmelCase : Optional[Any] = dim // n_heads UpperCAmelCase : str = 1_0_0_0_0.0 UpperCAmelCase : Optional[int] = 1.0 / (base ** (torch.arange(0 , _lowerCAmelCase , 2 ).float() / dims_per_head)) if "n_kv_heads" in params: UpperCAmelCase : Tuple = params['''n_kv_heads'''] # for GQA / MQA UpperCAmelCase : Optional[int] = n_heads_per_shard // num_key_value_heads UpperCAmelCase : Optional[Any] = dim // num_key_value_heads else: # compatibility with other checkpoints UpperCAmelCase : List[str] = n_heads UpperCAmelCase : Optional[int] = n_heads_per_shard UpperCAmelCase : List[str] = dim # permute for sliced rotary def permute(_lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]=n_heads , _lowerCAmelCase : int=dim , _lowerCAmelCase : Dict=dim ): return w.view(_lowerCAmelCase , dima // n_heads // 2 , 2 , _lowerCAmelCase ).transpose(1 , 2 ).reshape(_lowerCAmelCase , _lowerCAmelCase ) print(f"""Fetching all parameters from the checkpoint at {input_base_path}.""" ) # Load weights if model_size == "7B": # Not sharded # (The sharded implementation would also work, but this is simpler.) UpperCAmelCase : int = torch.load(os.path.join(_lowerCAmelCase , '''consolidated.00.pth''' ) , map_location='''cpu''' ) else: # Sharded UpperCAmelCase : Optional[Any] = [ torch.load(os.path.join(_lowerCAmelCase , f"""consolidated.{i:02d}.pth""" ) , map_location='''cpu''' ) for i in range(_lowerCAmelCase ) ] UpperCAmelCase : Any = 0 UpperCAmelCase : str = {'''weight_map''': {}} for layer_i in range(_lowerCAmelCase ): UpperCAmelCase : Optional[Any] = f"""pytorch_model-{layer_i + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase : Optional[int] = { f"""model.layers.{layer_i}.self_attn.q_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wq.weight"""] ), f"""model.layers.{layer_i}.self_attn.k_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wk.weight"""] ), f"""model.layers.{layer_i}.self_attn.v_proj.weight""": loaded[f"""layers.{layer_i}.attention.wv.weight"""], f"""model.layers.{layer_i}.self_attn.o_proj.weight""": loaded[f"""layers.{layer_i}.attention.wo.weight"""], f"""model.layers.{layer_i}.mlp.gate_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w1.weight"""], f"""model.layers.{layer_i}.mlp.down_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w2.weight"""], f"""model.layers.{layer_i}.mlp.up_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w3.weight"""], f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[f"""layers.{layer_i}.attention_norm.weight"""], f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[f"""layers.{layer_i}.ffn_norm.weight"""], } else: # Sharded # Note that attention.w{q,k,v,o}, feed_fordward.w[1,2,3], attention_norm.weight and ffn_norm.weight share # the same storage object, saving attention_norm and ffn_norm will save other weights too, which is # redundant as other weights will be stitched from multiple shards. To avoid that, they are cloned. UpperCAmelCase : List[str] = { f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.attention_norm.weight""" ].clone(), f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.ffn_norm.weight""" ].clone(), } UpperCAmelCase : Union[str, Any] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wq.weight"""].view(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase : Optional[Any] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wk.weight"""].view( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) UpperCAmelCase : str = torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wv.weight"""].view( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = torch.cat( [loaded[i][f"""layers.{layer_i}.attention.wo.weight"""] for i in range(_lowerCAmelCase )] , dim=1 ) UpperCAmelCase : Any = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w1.weight"""] for i in range(_lowerCAmelCase )] , dim=0 ) UpperCAmelCase : str = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w2.weight"""] for i in range(_lowerCAmelCase )] , dim=1 ) UpperCAmelCase : Tuple = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w3.weight"""] for i in range(_lowerCAmelCase )] , dim=0 ) UpperCAmelCase : Any = inv_freq for k, v in state_dict.items(): UpperCAmelCase : List[Any] = filename param_count += v.numel() torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase : Optional[int] = f"""pytorch_model-{n_layers + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase : str = { '''model.embed_tokens.weight''': loaded['''tok_embeddings.weight'''], '''model.norm.weight''': loaded['''norm.weight'''], '''lm_head.weight''': loaded['''output.weight'''], } else: UpperCAmelCase : Any = { '''model.norm.weight''': loaded[0]['''norm.weight'''], '''model.embed_tokens.weight''': torch.cat( [loaded[i]['''tok_embeddings.weight'''] for i in range(_lowerCAmelCase )] , dim=1 ), '''lm_head.weight''': torch.cat([loaded[i]['''output.weight'''] for i in range(_lowerCAmelCase )] , dim=0 ), } for k, v in state_dict.items(): UpperCAmelCase : Optional[int] = filename param_count += v.numel() torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , _lowerCAmelCase ) ) # Write configs UpperCAmelCase : Union[str, Any] = {'''total_size''': param_count * 2} write_json(_lowerCAmelCase , os.path.join(_lowerCAmelCase , '''pytorch_model.bin.index.json''' ) ) UpperCAmelCase : int = params['''ffn_dim_multiplier'''] if '''ffn_dim_multiplier''' in params else 1 UpperCAmelCase : Tuple = params['''multiple_of'''] if '''multiple_of''' in params else 256 UpperCAmelCase : Any = LlamaConfig( hidden_size=_lowerCAmelCase , intermediate_size=compute_intermediate_size(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) , num_attention_heads=params['''n_heads'''] , num_hidden_layers=params['''n_layers'''] , rms_norm_eps=params['''norm_eps'''] , num_key_value_heads=_lowerCAmelCase , ) config.save_pretrained(_lowerCAmelCase ) # Make space so we can load the model properly now. del state_dict del loaded gc.collect() print('''Loading the checkpoint in a Llama model.''' ) UpperCAmelCase : Optional[int] = LlamaForCausalLM.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa , low_cpu_mem_usage=_lowerCAmelCase ) # Avoid saving this as part of the config. del model.config._name_or_path print('''Saving in the Transformers format.''' ) model.save_pretrained(_lowerCAmelCase , safe_serialization=_lowerCAmelCase ) shutil.rmtree(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any] ) -> List[str]: # Initialize the tokenizer based on the `spm` model UpperCAmelCase : Dict = LlamaTokenizer if LlamaTokenizerFast is None else LlamaTokenizerFast print(f"""Saving a {tokenizer_class.__name__} to {tokenizer_path}.""" ) UpperCAmelCase : List[Any] = tokenizer_class(_lowerCAmelCase ) tokenizer.save_pretrained(_lowerCAmelCase ) def snake_case_ ( ) -> List[Any]: UpperCAmelCase : int = argparse.ArgumentParser() parser.add_argument( '''--input_dir''' , help='''Location of LLaMA weights, which contains tokenizer.model and model folders''' , ) parser.add_argument( '''--model_size''' , choices=['''7B''', '''7Bf''', '''13B''', '''13Bf''', '''30B''', '''65B''', '''70B''', '''70Bf''', '''tokenizer_only'''] , ) parser.add_argument( '''--output_dir''' , help='''Location to write HF model and tokenizer''' , ) parser.add_argument('''--safe_serialization''' , type=_lowerCAmelCase , help='''Whether or not to save using `safetensors`.''' ) UpperCAmelCase : List[Any] = parser.parse_args() if args.model_size != "tokenizer_only": write_model( model_path=args.output_dir , input_base_path=os.path.join(args.input_dir , args.model_size ) , model_size=args.model_size , safe_serialization=args.safe_serialization , ) UpperCAmelCase : Optional[int] = os.path.join(args.input_dir , '''tokenizer.model''' ) write_tokenizer(args.output_dir , _lowerCAmelCase ) if __name__ == "__main__": main()
23
1
'''simple docstring''' import random def snake_case_ ( _lowerCAmelCase : list , _lowerCAmelCase : str ) -> tuple: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Tuple = [], [], [] for element in data: if element < pivot: less.append(_lowerCAmelCase ) elif element > pivot: greater.append(_lowerCAmelCase ) else: equal.append(_lowerCAmelCase ) return less, equal, greater def snake_case_ ( _lowerCAmelCase : list , _lowerCAmelCase : int ) -> Dict: # index = len(items) // 2 when trying to find the median # (value of index when items is sorted) # invalid input if index >= len(_lowerCAmelCase ) or index < 0: return None UpperCAmelCase : int = items[random.randint(0 , len(_lowerCAmelCase ) - 1 )] UpperCAmelCase : List[Any] = 0 UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Any = _partition(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : List[str] = len(_lowerCAmelCase ) UpperCAmelCase : str = len(_lowerCAmelCase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(_lowerCAmelCase , _lowerCAmelCase ) # must be in larger else: return quick_select(_lowerCAmelCase , index - (m + count) )
23
'''simple docstring''' from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Optional[int] = tau * frequency / samplerate UpperCAmelCase : List[Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = cos(_lowerCAmelCase ) UpperCAmelCase : int = _sin / (2 * q_factor) UpperCAmelCase : Any = (1 - _cos) / 2 UpperCAmelCase : List[Any] = 1 - _cos UpperCAmelCase : Union[str, Any] = 1 + alpha UpperCAmelCase : Any = -2 * _cos UpperCAmelCase : Dict = 1 - alpha UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Any = tau * frequency / samplerate UpperCAmelCase : Tuple = sin(_lowerCAmelCase ) UpperCAmelCase : Tuple = cos(_lowerCAmelCase ) UpperCAmelCase : Dict = _sin / (2 * q_factor) UpperCAmelCase : int = (1 + _cos) / 2 UpperCAmelCase : List[Any] = -1 - _cos UpperCAmelCase : Tuple = 1 + alpha UpperCAmelCase : List[str] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha UpperCAmelCase : Tuple = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Optional[int] = tau * frequency / samplerate UpperCAmelCase : Optional[int] = sin(_lowerCAmelCase ) UpperCAmelCase : Tuple = cos(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = _sin / (2 * q_factor) UpperCAmelCase : Union[str, Any] = _sin / 2 UpperCAmelCase : Any = 0 UpperCAmelCase : int = -ba UpperCAmelCase : Optional[Any] = 1 + alpha UpperCAmelCase : List[Any] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha UpperCAmelCase : int = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : List[str] = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : str = cos(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = _sin / (2 * q_factor) UpperCAmelCase : List[str] = 1 - alpha UpperCAmelCase : Any = -2 * _cos UpperCAmelCase : Optional[int] = 1 + alpha UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([ba, ba, ba] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : Optional[Any] = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = cos(_lowerCAmelCase ) UpperCAmelCase : Dict = _sin / (2 * q_factor) UpperCAmelCase : str = 10 ** (gain_db / 40) UpperCAmelCase : int = 1 + alpha * big_a UpperCAmelCase : Union[str, Any] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha * big_a UpperCAmelCase : Union[str, Any] = 1 + alpha / big_a UpperCAmelCase : Tuple = -2 * _cos UpperCAmelCase : Any = 1 - alpha / big_a UpperCAmelCase : Optional[Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : Any = tau * frequency / samplerate UpperCAmelCase : Optional[int] = sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = cos(_lowerCAmelCase ) UpperCAmelCase : str = _sin / (2 * q_factor) UpperCAmelCase : List[str] = 10 ** (gain_db / 40) UpperCAmelCase : Optional[int] = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase : int = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase : int = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase : Optional[int] = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase : str = 2 * sqrt(_lowerCAmelCase ) * alpha UpperCAmelCase : Dict = big_a * (pmc + aaa) UpperCAmelCase : Any = 2 * big_a * mpc UpperCAmelCase : Union[str, Any] = big_a * (pmc - aaa) UpperCAmelCase : Optional[int] = ppmc + aaa UpperCAmelCase : Optional[Any] = -2 * pmpc UpperCAmelCase : Optional[Any] = ppmc - aaa UpperCAmelCase : int = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : int = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = cos(_lowerCAmelCase ) UpperCAmelCase : Any = _sin / (2 * q_factor) UpperCAmelCase : int = 10 ** (gain_db / 40) UpperCAmelCase : List[str] = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase : Union[str, Any] = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase : Optional[Any] = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase : Union[str, Any] = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase : List[str] = 2 * sqrt(_lowerCAmelCase ) * alpha UpperCAmelCase : Any = big_a * (ppmc + aaa) UpperCAmelCase : str = -2 * big_a * pmpc UpperCAmelCase : List[Any] = big_a * (ppmc - aaa) UpperCAmelCase : Optional[Any] = pmc + aaa UpperCAmelCase : Any = 2 * mpc UpperCAmelCase : str = pmc - aaa UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt
23
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_camembert import CamembertTokenizer else: UpperCamelCase__: Tuple = None UpperCamelCase__: List[Any] = logging.get_logger(__name__) UpperCamelCase__: List[Any] = {"vocab_file": "sentencepiece.bpe.model", "tokenizer_file": "tokenizer.json"} UpperCamelCase__: int = { "vocab_file": { "camembert-base": "https://huggingface.co/camembert-base/resolve/main/sentencepiece.bpe.model", }, "tokenizer_file": { "camembert-base": "https://huggingface.co/camembert-base/resolve/main/tokenizer.json", }, } UpperCamelCase__: int = { "camembert-base": 512, } UpperCamelCase__: Optional[int] = "▁" class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = ["""input_ids""", """attention_mask"""] lowerCamelCase__ = CamembertTokenizer def __init__( self : Any , __snake_case : Optional[Any]=None , __snake_case : Tuple=None , __snake_case : List[str]="<s>" , __snake_case : str="</s>" , __snake_case : Tuple="</s>" , __snake_case : str="<s>" , __snake_case : Optional[int]="<unk>" , __snake_case : List[str]="<pad>" , __snake_case : int="<mask>" , __snake_case : List[Any]=["<s>NOTUSED", "</s>NOTUSED"] , **__snake_case : str , ) -> Dict: # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase : List[str] = AddedToken(__snake_case , lstrip=__snake_case , rstrip=__snake_case ) if isinstance(__snake_case , __snake_case ) else mask_token super().__init__( __snake_case , tokenizer_file=__snake_case , bos_token=__snake_case , eos_token=__snake_case , sep_token=__snake_case , cls_token=__snake_case , unk_token=__snake_case , pad_token=__snake_case , mask_token=__snake_case , additional_special_tokens=__snake_case , **__snake_case , ) UpperCAmelCase : Union[str, Any] = vocab_file UpperCAmelCase : Union[str, Any] = False if not self.vocab_file else True def A ( self : List[Any] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase : List[Any] = [self.cls_token_id] UpperCAmelCase : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A ( self : Union[str, Any] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: UpperCAmelCase : Optional[int] = [self.sep_token_id] UpperCAmelCase : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A ( self : str , __snake_case : str , __snake_case : Optional[str] = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(__snake_case ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase : List[str] = os.path.join( __snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__snake_case ): copyfile(self.vocab_file , __snake_case ) return (out_vocab_file,)
23
'''simple docstring''' from __future__ import annotations def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : str ) -> bool: UpperCAmelCase : str = get_failure_array(_lowerCAmelCase ) # 2) Step through text searching for pattern UpperCAmelCase , UpperCAmelCase : Optional[Any] = 0, 0 # index into text, pattern while i < len(_lowerCAmelCase ): if pattern[j] == text[i]: if j == (len(_lowerCAmelCase ) - 1): return True j += 1 # if this is a prefix in our pattern # just go back far enough to continue elif j > 0: UpperCAmelCase : Optional[Any] = failure[j - 1] continue i += 1 return False def snake_case_ ( _lowerCAmelCase : str ) -> list[int]: UpperCAmelCase : Optional[Any] = [0] UpperCAmelCase : str = 0 UpperCAmelCase : List[str] = 1 while j < len(_lowerCAmelCase ): if pattern[i] == pattern[j]: i += 1 elif i > 0: UpperCAmelCase : Union[str, Any] = failure[i - 1] continue j += 1 failure.append(_lowerCAmelCase ) return failure if __name__ == "__main__": # Test 1) UpperCamelCase__: str = "abc1abc12" UpperCamelCase__: str = "alskfjaldsabc1abc1abc12k23adsfabcabc" UpperCamelCase__: Any = "alskfjaldsk23adsfabcabc" assert kmp(pattern, texta) and not kmp(pattern, texta) # Test 2) UpperCamelCase__: Tuple = "ABABX" UpperCamelCase__: Union[str, Any] = "ABABZABABYABABX" assert kmp(pattern, text) # Test 3) UpperCamelCase__: Any = "AAAB" UpperCamelCase__: str = "ABAAAAAB" assert kmp(pattern, text) # Test 4) UpperCamelCase__: int = "abcdabcy" UpperCamelCase__: Any = "abcxabcdabxabcdabcdabcy" assert kmp(pattern, text) # Test 5) UpperCamelCase__: List[str] = "aabaabaaa" assert get_failure_array(pattern) == [0, 1, 0, 1, 2, 3, 4, 5, 2]
23
1
'''simple docstring''' import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__: Any = logging.get_logger(__name__) UpperCamelCase__: Optional[int] = { "asapp/sew-d-tiny-100k": "https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json", # See all SEW-D models at https://huggingface.co/models?filter=sew-d } class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = """sew-d""" def __init__( self : Dict , __snake_case : int=32 , __snake_case : Union[str, Any]=768 , __snake_case : int=12 , __snake_case : Optional[Any]=12 , __snake_case : List[Any]=3072 , __snake_case : int=2 , __snake_case : Tuple=512 , __snake_case : Optional[int]=256 , __snake_case : Dict=True , __snake_case : str=True , __snake_case : List[str]=("p2c", "c2p") , __snake_case : Optional[Any]="layer_norm" , __snake_case : Optional[int]="gelu_python" , __snake_case : Dict=0.1 , __snake_case : Optional[int]=0.1 , __snake_case : List[Any]=0.1 , __snake_case : List[Any]=0.0 , __snake_case : str=0.1 , __snake_case : Dict=0.02 , __snake_case : Union[str, Any]=1E-7 , __snake_case : List[Any]=1E-5 , __snake_case : Optional[Any]="group" , __snake_case : List[Any]="gelu" , __snake_case : Tuple=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , __snake_case : Optional[int]=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , __snake_case : Optional[int]=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , __snake_case : Tuple=False , __snake_case : Optional[Any]=128 , __snake_case : List[str]=16 , __snake_case : Tuple=True , __snake_case : Any=0.05 , __snake_case : Optional[Any]=10 , __snake_case : Union[str, Any]=2 , __snake_case : Optional[Any]=0.0 , __snake_case : int=10 , __snake_case : int=0 , __snake_case : Union[str, Any]="mean" , __snake_case : Any=False , __snake_case : int=False , __snake_case : Any=256 , __snake_case : Optional[int]=0 , __snake_case : Optional[Any]=1 , __snake_case : str=2 , **__snake_case : int , ) -> Any: super().__init__(**__snake_case , pad_token_id=__snake_case , bos_token_id=__snake_case , eos_token_id=__snake_case ) UpperCAmelCase : List[Any] = hidden_size UpperCAmelCase : int = feat_extract_norm UpperCAmelCase : List[str] = feat_extract_activation UpperCAmelCase : Tuple = list(__snake_case ) UpperCAmelCase : Any = list(__snake_case ) UpperCAmelCase : Optional[int] = list(__snake_case ) UpperCAmelCase : Any = conv_bias UpperCAmelCase : Optional[Any] = num_conv_pos_embeddings UpperCAmelCase : Optional[Any] = num_conv_pos_embedding_groups UpperCAmelCase : List[str] = len(self.conv_dim ) UpperCAmelCase : int = num_hidden_layers UpperCAmelCase : Any = intermediate_size UpperCAmelCase : Tuple = squeeze_factor UpperCAmelCase : List[Any] = max_position_embeddings UpperCAmelCase : List[str] = position_buckets UpperCAmelCase : List[str] = share_att_key UpperCAmelCase : Union[str, Any] = relative_attention UpperCAmelCase : List[str] = norm_rel_ebd UpperCAmelCase : Optional[int] = list(__snake_case ) UpperCAmelCase : Optional[Any] = hidden_act UpperCAmelCase : int = num_attention_heads UpperCAmelCase : Union[str, Any] = hidden_dropout UpperCAmelCase : Tuple = attention_dropout UpperCAmelCase : List[str] = activation_dropout UpperCAmelCase : Optional[int] = feat_proj_dropout UpperCAmelCase : List[Any] = final_dropout UpperCAmelCase : int = layer_norm_eps UpperCAmelCase : Tuple = feature_layer_norm_eps UpperCAmelCase : List[str] = initializer_range UpperCAmelCase : Tuple = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect.''' '''It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,''' F"""but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)""" F"""= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase : Dict = apply_spec_augment UpperCAmelCase : str = mask_time_prob UpperCAmelCase : Tuple = mask_time_length UpperCAmelCase : Optional[int] = mask_time_min_masks UpperCAmelCase : Optional[int] = mask_feature_prob UpperCAmelCase : Optional[Any] = mask_feature_length UpperCAmelCase : str = mask_feature_min_masks # ctc loss UpperCAmelCase : str = ctc_loss_reduction UpperCAmelCase : Optional[int] = ctc_zero_infinity # sequence classification UpperCAmelCase : Any = use_weighted_layer_sum UpperCAmelCase : int = classifier_proj_size @property def A ( self : Dict ) -> Dict: return functools.reduce(operator.mul , self.conv_stride , 1 )
23
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() UpperCamelCase__: int = logging.get_logger(__name__) UpperCamelCase__: Dict = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "adapter_layer": "encoder.layers.*.adapter_layer", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", "pooling_layer.linear": "projector", "pooling_layer.projection": "classifier", } UpperCamelCase__: Optional[Any] = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", "projector", "classifier", ] def snake_case_ ( _lowerCAmelCase : str ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = {} with open(_lowerCAmelCase , '''r''' ) as file: for line_number, line in enumerate(_lowerCAmelCase ): UpperCAmelCase : List[str] = line.strip() if line: UpperCAmelCase : str = line.split() UpperCAmelCase : Union[str, Any] = line_number UpperCAmelCase : List[Any] = words[0] UpperCAmelCase : Union[str, Any] = value return result def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : str ) -> int: for attribute in key.split('''.''' ): UpperCAmelCase : Dict = getattr(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Dict = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(_lowerCAmelCase ): UpperCAmelCase : Any = PARAM_MAPPING[full_name.split('''.''' )[-1]] UpperCAmelCase : Dict = '''param''' if weight_type is not None and weight_type != "param": UpperCAmelCase : Optional[int] = getattr(_lowerCAmelCase , _lowerCAmelCase ).shape elif weight_type is not None and weight_type == "param": UpperCAmelCase : List[Any] = hf_pointer for attribute in hf_param_name.split('''.''' ): UpperCAmelCase : Optional[Any] = getattr(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : int = shape_pointer.shape # let's reduce dimension UpperCAmelCase : Union[str, Any] = value[0] else: UpperCAmelCase : List[Any] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": UpperCAmelCase : int = value elif weight_type == "weight_g": UpperCAmelCase : str = value elif weight_type == "weight_v": UpperCAmelCase : Dict = value elif weight_type == "bias": UpperCAmelCase : str = value elif weight_type == "param": for attribute in hf_param_name.split('''.''' ): UpperCAmelCase : int = getattr(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = value else: UpperCAmelCase : Tuple = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[Any] ) -> List[Any]: UpperCAmelCase : List[str] = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(_lowerCAmelCase ): UpperCAmelCase : List[str] = PARAM_MAPPING[full_name.split('''.''' )[-1]] UpperCAmelCase : Any = '''param''' if weight_type is not None and weight_type != "param": UpperCAmelCase : Optional[int] = '''.'''.join([key, weight_type] ) elif weight_type is not None and weight_type == "param": UpperCAmelCase : Optional[int] = '''.'''.join([key, hf_param_name] ) else: UpperCAmelCase : List[Any] = key UpperCAmelCase : Tuple = value if '''lm_head''' in full_key else value[0] UpperCamelCase__: Tuple = { "W_a": "linear_1.weight", "W_b": "linear_2.weight", "b_a": "linear_1.bias", "b_b": "linear_2.bias", "ln_W": "norm.weight", "ln_b": "norm.bias", } def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : Any=None , _lowerCAmelCase : Optional[Any]=None ) -> int: UpperCAmelCase : List[Any] = False for key, mapped_key in MAPPING.items(): UpperCAmelCase : int = '''wav2vec2.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: UpperCAmelCase : Optional[Any] = True if "*" in mapped_key: UpperCAmelCase : Tuple = name.split(_lowerCAmelCase )[0].split('''.''' )[-2] UpperCAmelCase : List[Any] = mapped_key.replace('''*''' , _lowerCAmelCase ) if "weight_g" in name: UpperCAmelCase : str = '''weight_g''' elif "weight_v" in name: UpperCAmelCase : int = '''weight_v''' elif "bias" in name: UpperCAmelCase : int = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj UpperCAmelCase : List[str] = '''weight''' else: UpperCAmelCase : Dict = None if hf_dict is not None: rename_dict(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) else: set_recursively(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return is_used return is_used def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Any: UpperCAmelCase : Dict = [] UpperCAmelCase : Dict = fairseq_model.state_dict() UpperCAmelCase : Union[str, Any] = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase : Dict = False if "conv_layers" in name: load_conv_layer( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , hf_model.config.feat_extract_norm == '''group''' , ) UpperCAmelCase : Any = True else: UpperCAmelCase : Optional[Any] = load_wavaveca_layer(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not is_used: unused_weights.append(_lowerCAmelCase ) logger.warning(f"""Unused weights: {unused_weights}""" ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: UpperCAmelCase : Any = full_name.split('''conv_layers.''' )[-1] UpperCAmelCase : Optional[int] = name.split('''.''' ) UpperCAmelCase : Tuple = int(items[0] ) UpperCAmelCase : Tuple = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) UpperCAmelCase : Tuple = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) UpperCAmelCase : Union[str, Any] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.""" ) UpperCAmelCase : Union[str, Any] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.""" ) UpperCAmelCase : List[str] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(_lowerCAmelCase ) @torch.no_grad() def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict=None , _lowerCAmelCase : List[Any]=None , _lowerCAmelCase : int=True , _lowerCAmelCase : Optional[int]=False ) -> Dict: if config_path is not None: UpperCAmelCase : List[str] = WavaVecaConfig.from_pretrained(_lowerCAmelCase ) else: UpperCAmelCase : List[Any] = WavaVecaConfig() if is_seq_class: UpperCAmelCase : Optional[Any] = read_txt_into_dict(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = idalabel UpperCAmelCase : Optional[Any] = WavaVecaForSequenceClassification(_lowerCAmelCase ) UpperCAmelCase : Dict = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase , ) feature_extractor.save_pretrained(_lowerCAmelCase ) elif is_finetuned: if dict_path: UpperCAmelCase : Dict = Dictionary.load(_lowerCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq UpperCAmelCase : Any = target_dict.pad_index UpperCAmelCase : Tuple = target_dict.bos_index UpperCAmelCase : Optional[int] = target_dict.eos_index UpperCAmelCase : Union[str, Any] = len(target_dict.symbols ) UpperCAmelCase : Dict = os.path.join(_lowerCAmelCase , '''vocab.json''' ) if not os.path.isdir(_lowerCAmelCase ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(_lowerCAmelCase ) ) return os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : List[Any] = target_dict.indices # fairseq has the <pad> and <s> switched UpperCAmelCase : List[str] = 0 UpperCAmelCase : List[str] = 1 with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = WavaVecaCTCTokenizer( _lowerCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=_lowerCAmelCase , ) UpperCAmelCase : int = True if config.feat_extract_norm == '''layer''' else False UpperCAmelCase : int = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase , ) UpperCAmelCase : str = WavaVecaProcessor(feature_extractor=_lowerCAmelCase , tokenizer=_lowerCAmelCase ) processor.save_pretrained(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = WavaVecaForCTC(_lowerCAmelCase ) else: UpperCAmelCase : Dict = WavaVecaForPreTraining(_lowerCAmelCase ) if is_finetuned or is_seq_class: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : List[str] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: UpperCAmelCase : Optional[Any] = argparse.Namespace(task='''audio_pretraining''' ) UpperCAmelCase : List[Any] = fairseq.tasks.setup_task(_lowerCAmelCase ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : int = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=_lowerCAmelCase ) UpperCAmelCase : Optional[int] = model[0].eval() recursively_load_weights(_lowerCAmelCase , _lowerCAmelCase , not is_finetuned ) hf_wavavec.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": UpperCamelCase__: Dict = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--not_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) parser.add_argument( "--is_seq_class", action="store_true", help="Whether the model to convert is a fine-tuned sequence classification model or not", ) UpperCamelCase__: Any = parser.parse_args() UpperCamelCase__: int = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
23
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase__: Optional[Any] = { "configuration_clap": [ "CLAP_PRETRAINED_MODEL_ARCHIVE_LIST", "ClapAudioConfig", "ClapConfig", "ClapTextConfig", ], "processing_clap": ["ClapProcessor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: Dict = [ "CLAP_PRETRAINED_MODEL_ARCHIVE_LIST", "ClapModel", "ClapPreTrainedModel", "ClapTextModel", "ClapTextModelWithProjection", "ClapAudioModel", "ClapAudioModelWithProjection", ] UpperCamelCase__: Optional[Any] = ["ClapFeatureExtractor"] if TYPE_CHECKING: from .configuration_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioConfig, ClapConfig, ClapTextConfig, ) from .processing_clap import ClapProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clap import ClapFeatureExtractor from .modeling_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioModel, ClapAudioModelWithProjection, ClapModel, ClapPreTrainedModel, ClapTextModel, ClapTextModelWithProjection, ) else: import sys UpperCamelCase__: Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
'''simple docstring''' import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : str ) -> int: with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights UpperCAmelCase : Optional[Any] = FlaxDiffusionPipeline.from_pretrained( '''hf-internal-testing/tiny-stable-diffusion-pipe''' , safety_checker=__snake_case , cache_dir=__snake_case ) UpperCAmelCase : str = [t[-1] for t in os.walk(os.path.join(__snake_case , os.listdir(__snake_case )[0] , '''snapshots''' ) )] UpperCAmelCase : str = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith('''.bin''' ) for f in files ) @slow @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : List[str] ) -> Dict: UpperCAmelCase , UpperCAmelCase : str = FlaxStableDiffusionPipeline.from_pretrained( '''hf-internal-testing/tiny-stable-diffusion-pipe''' , safety_checker=__snake_case ) UpperCAmelCase : List[Any] = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : List[str] = jax.random.PRNGKey(0 ) UpperCAmelCase : Optional[Any] = 4 UpperCAmelCase : Optional[Any] = jax.device_count() UpperCAmelCase : Tuple = num_samples * [prompt] UpperCAmelCase : int = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Tuple = replicate(__snake_case ) UpperCAmelCase : Any = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Optional[Any] = shard(__snake_case ) UpperCAmelCase : Optional[int] = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1E-3 assert np.abs(np.abs(__snake_case , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5E-1 UpperCAmelCase : Union[str, Any] = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(__snake_case ) == num_samples def A ( self : List[Any] ) -> List[str]: UpperCAmelCase , UpperCAmelCase : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''flax''' , safety_checker=__snake_case ) UpperCAmelCase : Dict = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : Optional[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase : Any = 50 UpperCAmelCase : Union[str, Any] = jax.device_count() UpperCAmelCase : int = num_samples * [prompt] UpperCAmelCase : Union[str, Any] = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Dict = replicate(__snake_case ) UpperCAmelCase : int = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Tuple = shard(__snake_case ) UpperCAmelCase : Tuple = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5E-1 def A ( self : int ) -> Dict: UpperCAmelCase , UpperCAmelCase : List[str] = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__snake_case ) UpperCAmelCase : Dict = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : Union[str, Any] = jax.random.PRNGKey(0 ) UpperCAmelCase : List[str] = 50 UpperCAmelCase : Union[str, Any] = jax.device_count() UpperCAmelCase : List[Any] = num_samples * [prompt] UpperCAmelCase : int = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Tuple = replicate(__snake_case ) UpperCAmelCase : List[Any] = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Optional[int] = shard(__snake_case ) UpperCAmelCase : Optional[Any] = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def A ( self : int ) -> Any: UpperCAmelCase , UpperCAmelCase : Dict = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa ) UpperCAmelCase : List[str] = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : List[str] = jax.random.PRNGKey(0 ) UpperCAmelCase : Union[str, Any] = 50 UpperCAmelCase : Optional[int] = jax.device_count() UpperCAmelCase : List[str] = num_samples * [prompt] UpperCAmelCase : Dict = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Tuple = replicate(__snake_case ) UpperCAmelCase : Any = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : str = shard(__snake_case ) UpperCAmelCase : Optional[int] = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def A ( self : Tuple ) -> Optional[Any]: UpperCAmelCase : int = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , set_alpha_to_one=__snake_case , steps_offset=1 , ) UpperCAmelCase , UpperCAmelCase : Tuple = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , scheduler=__snake_case , safety_checker=__snake_case , ) UpperCAmelCase : Tuple = scheduler.create_state() UpperCAmelCase : Dict = scheduler_state UpperCAmelCase : str = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : int = jax.random.PRNGKey(0 ) UpperCAmelCase : Union[str, Any] = 50 UpperCAmelCase : Optional[Any] = jax.device_count() UpperCAmelCase : Any = num_samples * [prompt] UpperCAmelCase : Dict = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : str = replicate(__snake_case ) UpperCAmelCase : List[str] = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Optional[int] = shard(__snake_case ) UpperCAmelCase : Dict = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5E-1 def A ( self : Any ) -> Tuple: UpperCAmelCase : List[Any] = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : Union[str, Any] = jax.device_count() UpperCAmelCase : List[Any] = num_samples * [prompt] UpperCAmelCase : str = jax.random.split(jax.random.PRNGKey(0 ) , __snake_case ) UpperCAmelCase , UpperCAmelCase : Tuple = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__snake_case , ) UpperCAmelCase : Dict = replicate(__snake_case ) UpperCAmelCase : Optional[Any] = pipeline.prepare_inputs(__snake_case ) UpperCAmelCase : List[str] = shard(__snake_case ) UpperCAmelCase : Any = pipeline(__snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase : Optional[int] = images[2, 0, 256, 10:17, 1] # With memory efficient attention UpperCAmelCase , UpperCAmelCase : Any = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__snake_case , use_memory_efficient_attention=__snake_case , ) UpperCAmelCase : int = replicate(__snake_case ) UpperCAmelCase : int = pipeline.prepare_inputs(__snake_case ) UpperCAmelCase : List[Any] = shard(__snake_case ) UpperCAmelCase : Optional[Any] = pipeline(__snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase : int = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
23
1
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss UpperCamelCase__: Dict = pytest.mark.integration @require_faiss class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" def A ( self : List[str] ) -> int: UpperCAmelCase : Union[str, Any] = Dataset.from_dict({'''filename''': ['''my_name-train''' + '''_''' + str(__snake_case ) for x in np.arange(30 ).tolist()]} ) return dset def A ( self : List[Any] ) -> Optional[int]: import faiss UpperCAmelCase : Dataset = self._create_dummy_dataset() UpperCAmelCase : Tuple = dset.map( lambda __snake_case , __snake_case : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=__snake_case , keep_in_memory=__snake_case ) UpperCAmelCase : Optional[int] = dset.add_faiss_index('''vecs''' , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT ) UpperCAmelCase , UpperCAmelCase : Optional[int] = dset.get_nearest_examples('''vecs''' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['''filename'''][0] , '''my_name-train_29''' ) dset.drop_index('''vecs''' ) def A ( self : Dict ) -> Tuple: import faiss UpperCAmelCase : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='''vecs''' , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT , ) UpperCAmelCase , UpperCAmelCase : Optional[int] = dset.get_nearest_examples('''vecs''' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['''filename'''][0] , '''my_name-train_29''' ) def A ( self : Tuple ) -> Tuple: import faiss UpperCAmelCase : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='''vecs''' , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=__snake_case ) as tmp_file: dset.save_faiss_index('''vecs''' , tmp_file.name ) dset.load_faiss_index('''vecs2''' , tmp_file.name ) os.unlink(tmp_file.name ) UpperCAmelCase , UpperCAmelCase : List[str] = dset.get_nearest_examples('''vecs2''' , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples['''filename'''][0] , '''my_name-train_29''' ) def A ( self : Optional[int] ) -> Dict: UpperCAmelCase : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name='''vecs''' ) dset.drop_index('''vecs''' ) self.assertRaises(__snake_case , partial(dset.get_nearest_examples , '''vecs2''' , np.ones(5 , dtype=np.floataa ) ) ) def A ( self : str ) -> str: from elasticsearch import Elasticsearch UpperCAmelCase : Dataset = self._create_dummy_dataset() with patch('''elasticsearch.Elasticsearch.search''' ) as mocked_search, patch( '''elasticsearch.client.IndicesClient.create''' ) as mocked_index_create, patch('''elasticsearch.helpers.streaming_bulk''' ) as mocked_bulk: UpperCAmelCase : Dict = {'''acknowledged''': True} mocked_bulk.return_value([(True, None)] * 30 ) UpperCAmelCase : Dict = {'''hits''': {'''hits''': [{'''_score''': 1, '''_id''': 29}]}} UpperCAmelCase : List[Any] = Elasticsearch() dset.add_elasticsearch_index('''filename''' , es_client=__snake_case ) UpperCAmelCase , UpperCAmelCase : Any = dset.get_nearest_examples('''filename''' , '''my_name-train_29''' ) self.assertEqual(examples['''filename'''][0] , '''my_name-train_29''' ) @require_faiss class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" def A ( self : Optional[int] ) -> Union[str, Any]: import faiss UpperCAmelCase : Tuple = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 10 ) # single query UpperCAmelCase : Optional[Any] = np.zeros(5 , dtype=np.floataa ) UpperCAmelCase : List[str] = 1 UpperCAmelCase , UpperCAmelCase : Optional[int] = index.search(__snake_case ) self.assertRaises(__snake_case , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries UpperCAmelCase : Optional[int] = np.eye(5 , dtype=np.floataa )[::-1] UpperCAmelCase , UpperCAmelCase : List[str] = index.search_batch(__snake_case ) self.assertRaises(__snake_case , index.search_batch , queries[0] ) UpperCAmelCase : List[Any] = [scores[0] for scores in total_scores] UpperCAmelCase : Optional[Any] = [indices[0] for indices in total_indices] self.assertGreater(np.min(__snake_case ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , __snake_case ) def A ( self : str ) -> Union[str, Any]: import faiss UpperCAmelCase : Optional[Any] = FaissIndex(string_factory='''Flat''' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) UpperCAmelCase : int = FaissIndex(string_factory='''LSH''' ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(__snake_case ): UpperCAmelCase : List[str] = FaissIndex(string_factory='''Flat''' , custom_index=faiss.IndexFlat(5 ) ) def A ( self : int ) -> List[str]: import faiss UpperCAmelCase : Any = faiss.IndexFlat(5 ) UpperCAmelCase : Optional[int] = FaissIndex(custom_index=__snake_case ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def A ( self : str ) -> List[str]: import faiss UpperCAmelCase : int = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=__snake_case ) as tmp_file: index.save(tmp_file.name ) UpperCAmelCase : int = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) UpperCAmelCase : str = np.zeros(5 , dtype=np.floataa ) UpperCAmelCase : List[str] = 1 UpperCAmelCase , UpperCAmelCase : List[Any] = index.search(__snake_case ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def snake_case_ ( _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: import faiss UpperCAmelCase : List[str] = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) UpperCAmelCase : Dict = '''index.faiss''' UpperCAmelCase : Dict = f"""mock://{index_name}""" index.save(_lowerCAmelCase , storage_options=mockfs.storage_options ) UpperCAmelCase : Tuple = FaissIndex.load(_lowerCAmelCase , storage_options=mockfs.storage_options ) UpperCAmelCase : Union[str, Any] = np.zeros(5 , dtype=np.floataa ) UpperCAmelCase : Any = 1 UpperCAmelCase , UpperCAmelCase : Tuple = index.search(_lowerCAmelCase ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" def A ( self : str ) -> Union[str, Any]: from elasticsearch import Elasticsearch with patch('''elasticsearch.Elasticsearch.search''' ) as mocked_search, patch( '''elasticsearch.client.IndicesClient.create''' ) as mocked_index_create, patch('''elasticsearch.helpers.streaming_bulk''' ) as mocked_bulk: UpperCAmelCase : Any = Elasticsearch() UpperCAmelCase : List[Any] = {'''acknowledged''': True} UpperCAmelCase : Optional[Any] = ElasticSearchIndex(es_client=__snake_case ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(['''foo''', '''bar''', '''foobar'''] ) # single query UpperCAmelCase : Optional[Any] = '''foo''' UpperCAmelCase : List[str] = {'''hits''': {'''hits''': [{'''_score''': 1, '''_id''': 0}]}} UpperCAmelCase , UpperCAmelCase : int = index.search(__snake_case ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout UpperCAmelCase : str = '''foo''' UpperCAmelCase : Optional[int] = {'''hits''': {'''hits''': [{'''_score''': 1, '''_id''': 0}]}} UpperCAmelCase , UpperCAmelCase : Tuple = index.search(__snake_case , request_timeout=30 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries UpperCAmelCase : Any = ['''foo''', '''bar''', '''foobar'''] UpperCAmelCase : Optional[Any] = {'''hits''': {'''hits''': [{'''_score''': 1, '''_id''': 1}]}} UpperCAmelCase , UpperCAmelCase : Optional[int] = index.search_batch(__snake_case ) UpperCAmelCase : int = [scores[0] for scores in total_scores] UpperCAmelCase : List[str] = [indices[0] for indices in total_indices] self.assertGreater(np.min(__snake_case ) , 0 ) self.assertListEqual([1, 1, 1] , __snake_case ) # batched queries with timeout UpperCAmelCase : Optional[int] = ['''foo''', '''bar''', '''foobar'''] UpperCAmelCase : int = {'''hits''': {'''hits''': [{'''_score''': 1, '''_id''': 1}]}} UpperCAmelCase , UpperCAmelCase : str = index.search_batch(__snake_case , request_timeout=30 ) UpperCAmelCase : Optional[Any] = [scores[0] for scores in total_scores] UpperCAmelCase : Tuple = [indices[0] for indices in total_indices] self.assertGreater(np.min(__snake_case ) , 0 ) self.assertListEqual([1, 1, 1] , __snake_case )
23
'''simple docstring''' import random from .binary_exp_mod import bin_exp_mod def snake_case_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]=1000 ) -> int: if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd UpperCAmelCase : str = n - 1 UpperCAmelCase : List[Any] = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) UpperCAmelCase : List[str] = 0 while count < prec: UpperCAmelCase : int = random.randint(2 , n - 1 ) UpperCAmelCase : List[str] = bin_exp_mod(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if b != 1: UpperCAmelCase : int = True for _ in range(_lowerCAmelCase ): if b == n - 1: UpperCAmelCase : Dict = False break UpperCAmelCase : str = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": UpperCamelCase__: Optional[int] = abs(int(input("Enter bound : ").strip())) print("Here's the list of primes:") print(", ".join(str(i) for i in range(n + 1) if is_prime_big(i)))
23
1
'''simple docstring''' import numpy as np def snake_case_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[int] ) -> Dict: UpperCAmelCase : Union[str, Any] = int(np.ceil((x_end - xa) / h ) ) UpperCAmelCase : Union[str, Any] = np.zeros((n + 1,) ) UpperCAmelCase : List[Any] = ya UpperCAmelCase : Optional[Any] = xa for k in range(_lowerCAmelCase ): UpperCAmelCase : str = f(_lowerCAmelCase , y[k] ) UpperCAmelCase : Union[str, Any] = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) UpperCAmelCase : List[str] = f(x + 0.5 * h , y[k] + 0.5 * h * ka ) UpperCAmelCase : Tuple = f(x + h , y[k] + h * ka ) UpperCAmelCase : Any = y[k] + (1 / 6) * h * (ka + 2 * ka + 2 * ka + ka) x += h return y if __name__ == "__main__": import doctest doctest.testmod()
23
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__: Optional[int] = logging.get_logger(__name__) def snake_case_ ( _lowerCAmelCase : Optional[int] ) -> Optional[int]: UpperCAmelCase : Tuple = DPTConfig(embedding_type='''hybrid''' ) if "large" in checkpoint_url: UpperCAmelCase : Tuple = 1024 UpperCAmelCase : List[Any] = 4096 UpperCAmelCase : str = 24 UpperCAmelCase : List[Any] = 16 UpperCAmelCase : str = [5, 11, 17, 23] UpperCAmelCase : List[Any] = [256, 512, 1024, 1024] UpperCAmelCase : Tuple = (1, 384, 384) if "nyu" or "midas" in checkpoint_url: UpperCAmelCase : Optional[Any] = 768 UpperCAmelCase : Tuple = [1, 1, 1, 0.5] UpperCAmelCase : int = [256, 512, 768, 768] UpperCAmelCase : Any = 150 UpperCAmelCase : Tuple = 16 UpperCAmelCase : Any = (1, 384, 384) UpperCAmelCase : Optional[Any] = False UpperCAmelCase : Tuple = '''project''' if "ade" in checkpoint_url: UpperCAmelCase : Any = True UpperCAmelCase : str = 768 UpperCAmelCase : Optional[int] = [1, 1, 1, 0.5] UpperCAmelCase : List[Any] = 150 UpperCAmelCase : List[Any] = 16 UpperCAmelCase : str = '''huggingface/label-files''' UpperCAmelCase : Tuple = '''ade20k-id2label.json''' UpperCAmelCase : Any = json.load(open(cached_download(hf_hub_url(_lowerCAmelCase , _lowerCAmelCase , repo_type='''dataset''' ) ) , '''r''' ) ) UpperCAmelCase : Optional[Any] = {int(_lowerCAmelCase ): v for k, v in idalabel.items()} UpperCAmelCase : List[Any] = idalabel UpperCAmelCase : Optional[int] = {v: k for k, v in idalabel.items()} UpperCAmelCase : Union[str, Any] = [1, 150, 480, 480] return config, expected_shape def snake_case_ ( _lowerCAmelCase : Union[str, Any] ) -> int: UpperCAmelCase : List[str] = ['''pretrained.model.head.weight''', '''pretrained.model.head.bias'''] for k in ignore_keys: state_dict.pop(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Tuple ) -> Any: if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): UpperCAmelCase : Tuple = name.replace('''pretrained.model''' , '''dpt.encoder''' ) if "pretrained.model" in name: UpperCAmelCase : Union[str, Any] = name.replace('''pretrained.model''' , '''dpt.embeddings''' ) if "patch_embed" in name: UpperCAmelCase : int = name.replace('''patch_embed''' , '''''' ) if "pos_embed" in name: UpperCAmelCase : Tuple = name.replace('''pos_embed''' , '''position_embeddings''' ) if "attn.proj" in name: UpperCAmelCase : Any = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "proj" in name and "project" not in name: UpperCAmelCase : str = name.replace('''proj''' , '''projection''' ) if "blocks" in name: UpperCAmelCase : Any = name.replace('''blocks''' , '''layer''' ) if "mlp.fc1" in name: UpperCAmelCase : Optional[int] = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: UpperCAmelCase : Optional[Any] = name.replace('''mlp.fc2''' , '''output.dense''' ) if "norm1" in name and "backbone" not in name: UpperCAmelCase : Dict = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name and "backbone" not in name: UpperCAmelCase : Tuple = name.replace('''norm2''' , '''layernorm_after''' ) if "scratch.output_conv" in name: UpperCAmelCase : Tuple = name.replace('''scratch.output_conv''' , '''head''' ) if "scratch" in name: UpperCAmelCase : str = name.replace('''scratch''' , '''neck''' ) if "layer1_rn" in name: UpperCAmelCase : Dict = name.replace('''layer1_rn''' , '''convs.0''' ) if "layer2_rn" in name: UpperCAmelCase : int = name.replace('''layer2_rn''' , '''convs.1''' ) if "layer3_rn" in name: UpperCAmelCase : Tuple = name.replace('''layer3_rn''' , '''convs.2''' ) if "layer4_rn" in name: UpperCAmelCase : int = name.replace('''layer4_rn''' , '''convs.3''' ) if "refinenet" in name: UpperCAmelCase : List[str] = int(name[len('''neck.refinenet''' ) : len('''neck.refinenet''' ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 UpperCAmelCase : str = name.replace(f"""refinenet{layer_idx}""" , f"""fusion_stage.layers.{abs(layer_idx-4 )}""" ) if "out_conv" in name: UpperCAmelCase : List[str] = name.replace('''out_conv''' , '''projection''' ) if "resConfUnit1" in name: UpperCAmelCase : Union[str, Any] = name.replace('''resConfUnit1''' , '''residual_layer1''' ) if "resConfUnit2" in name: UpperCAmelCase : Any = name.replace('''resConfUnit2''' , '''residual_layer2''' ) if "conv1" in name: UpperCAmelCase : Optional[int] = name.replace('''conv1''' , '''convolution1''' ) if "conv2" in name: UpperCAmelCase : Tuple = name.replace('''conv2''' , '''convolution2''' ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: UpperCAmelCase : Dict = name.replace('''pretrained.act_postprocess1.0.project.0''' , '''neck.reassemble_stage.readout_projects.0.0''' ) if "pretrained.act_postprocess2.0.project.0" in name: UpperCAmelCase : int = name.replace('''pretrained.act_postprocess2.0.project.0''' , '''neck.reassemble_stage.readout_projects.1.0''' ) if "pretrained.act_postprocess3.0.project.0" in name: UpperCAmelCase : Any = name.replace('''pretrained.act_postprocess3.0.project.0''' , '''neck.reassemble_stage.readout_projects.2.0''' ) if "pretrained.act_postprocess4.0.project.0" in name: UpperCAmelCase : Optional[Any] = name.replace('''pretrained.act_postprocess4.0.project.0''' , '''neck.reassemble_stage.readout_projects.3.0''' ) # resize blocks if "pretrained.act_postprocess1.3" in name: UpperCAmelCase : List[Any] = name.replace('''pretrained.act_postprocess1.3''' , '''neck.reassemble_stage.layers.0.projection''' ) if "pretrained.act_postprocess1.4" in name: UpperCAmelCase : Any = name.replace('''pretrained.act_postprocess1.4''' , '''neck.reassemble_stage.layers.0.resize''' ) if "pretrained.act_postprocess2.3" in name: UpperCAmelCase : Optional[int] = name.replace('''pretrained.act_postprocess2.3''' , '''neck.reassemble_stage.layers.1.projection''' ) if "pretrained.act_postprocess2.4" in name: UpperCAmelCase : str = name.replace('''pretrained.act_postprocess2.4''' , '''neck.reassemble_stage.layers.1.resize''' ) if "pretrained.act_postprocess3.3" in name: UpperCAmelCase : List[str] = name.replace('''pretrained.act_postprocess3.3''' , '''neck.reassemble_stage.layers.2.projection''' ) if "pretrained.act_postprocess4.3" in name: UpperCAmelCase : Tuple = name.replace('''pretrained.act_postprocess4.3''' , '''neck.reassemble_stage.layers.3.projection''' ) if "pretrained.act_postprocess4.4" in name: UpperCAmelCase : int = name.replace('''pretrained.act_postprocess4.4''' , '''neck.reassemble_stage.layers.3.resize''' ) if "pretrained" in name: UpperCAmelCase : Optional[int] = name.replace('''pretrained''' , '''dpt''' ) if "bn" in name: UpperCAmelCase : Dict = name.replace('''bn''' , '''batch_norm''' ) if "head" in name: UpperCAmelCase : Any = name.replace('''head''' , '''head.head''' ) if "encoder.norm" in name: UpperCAmelCase : Optional[int] = name.replace('''encoder.norm''' , '''layernorm''' ) if "auxlayer" in name: UpperCAmelCase : Union[str, Any] = name.replace('''auxlayer''' , '''auxiliary_head.head''' ) if "backbone" in name: UpperCAmelCase : List[Any] = name.replace('''backbone''' , '''backbone.bit.encoder''' ) if ".." in name: UpperCAmelCase : Optional[int] = name.replace('''..''' , '''.''' ) if "stem.conv" in name: UpperCAmelCase : Optional[Any] = name.replace('''stem.conv''' , '''bit.embedder.convolution''' ) if "blocks" in name: UpperCAmelCase : Optional[int] = name.replace('''blocks''' , '''layers''' ) if "convolution" in name and "backbone" in name: UpperCAmelCase : List[Any] = name.replace('''convolution''' , '''conv''' ) if "layer" in name and "backbone" in name: UpperCAmelCase : List[str] = name.replace('''layer''' , '''layers''' ) if "backbone.bit.encoder.bit" in name: UpperCAmelCase : List[Any] = name.replace('''backbone.bit.encoder.bit''' , '''backbone.bit''' ) if "embedder.conv" in name: UpperCAmelCase : List[Any] = name.replace('''embedder.conv''' , '''embedder.convolution''' ) if "backbone.bit.encoder.stem.norm" in name: UpperCAmelCase : Tuple = name.replace('''backbone.bit.encoder.stem.norm''' , '''backbone.bit.embedder.norm''' ) return name def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : List[Any] ) -> Optional[Any]: for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase : Optional[int] = state_dict.pop(f"""dpt.encoder.layer.{i}.attn.qkv.weight""" ) UpperCAmelCase : Tuple = state_dict.pop(f"""dpt.encoder.layer.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase : Tuple = in_proj_weight[: config.hidden_size, :] UpperCAmelCase : int = in_proj_bias[: config.hidden_size] UpperCAmelCase : List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase : List[str] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase : str = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase : Union[str, Any] = in_proj_bias[-config.hidden_size :] def snake_case_ ( ) -> List[str]: UpperCAmelCase : Optional[int] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' UpperCAmelCase : Optional[int] = Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ) return im @torch.no_grad() def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[str] ) -> Any: UpperCAmelCase , UpperCAmelCase : int = get_dpt_config(_lowerCAmelCase ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") UpperCAmelCase : List[Any] = torch.load(_lowerCAmelCase , map_location='''cpu''' ) # remove certain keys remove_ignore_keys_(_lowerCAmelCase ) # rename keys for key in state_dict.copy().keys(): UpperCAmelCase : Any = state_dict.pop(_lowerCAmelCase ) UpperCAmelCase : List[Any] = val # read in qkv matrices read_in_q_k_v(_lowerCAmelCase , _lowerCAmelCase ) # load HuggingFace model UpperCAmelCase : Optional[Any] = DPTForSemanticSegmentation(_lowerCAmelCase ) if '''ade''' in checkpoint_url else DPTForDepthEstimation(_lowerCAmelCase ) model.load_state_dict(_lowerCAmelCase ) model.eval() # Check outputs on an image UpperCAmelCase : int = 480 if '''ade''' in checkpoint_url else 384 UpperCAmelCase : List[Any] = DPTImageProcessor(size=_lowerCAmelCase ) UpperCAmelCase : Dict = prepare_img() UpperCAmelCase : Optional[int] = image_processor(_lowerCAmelCase , return_tensors='''pt''' ) # forward pass UpperCAmelCase : Any = model(**_lowerCAmelCase ).logits if '''ade''' in checkpoint_url else model(**_lowerCAmelCase ).predicted_depth if show_prediction: UpperCAmelCase : Dict = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode='''bicubic''' , align_corners=_lowerCAmelCase , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 255 ).show() if pytorch_dump_folder_path is not None: Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(_lowerCAmelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_lowerCAmelCase ) if push_to_hub: model.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) image_processor.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) if __name__ == "__main__": UpperCamelCase__: Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint_url", default="https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt", type=str, help="URL of the original DPT checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=False, help="Path to the output PyTorch model directory.", ) parser.add_argument( "--push_to_hub", action="store_true", ) parser.add_argument( "--model_name", default="dpt-large", type=str, help="Name of the model, in case you're pushing to the hub.", ) parser.add_argument( "--show_prediction", action="store_true", ) UpperCamelCase__: Tuple = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
23
1
'''simple docstring''' import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) UpperCamelCase__: List[Any] = pytest.mark.integration @pytest.mark.parametrize('''path''' , ['''paws''', '''csv'''] ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int ) -> str: inspect_dataset(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[Any] = path + '''.py''' assert script_name in os.listdir(_lowerCAmelCase ) assert "__pycache__" not in os.listdir(_lowerCAmelCase ) @pytest.mark.filterwarnings('''ignore:inspect_metric is deprecated:FutureWarning''' ) @pytest.mark.filterwarnings('''ignore:metric_module_factory is deprecated:FutureWarning''' ) @pytest.mark.parametrize('''path''' , ['''accuracy'''] ) def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : List[str] ) -> List[str]: inspect_metric(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Dict = path + '''.py''' assert script_name in os.listdir(_lowerCAmelCase ) assert "__pycache__" not in os.listdir(_lowerCAmelCase ) @pytest.mark.parametrize( '''path, config_name, expected_splits''' , [ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] , ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any] ) -> Any: UpperCAmelCase : Tuple = get_dataset_config_info(_lowerCAmelCase , config_name=_lowerCAmelCase ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' , [ ('''paws''', None, ValueError), ] , ) def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[Any] ) -> Dict: with pytest.raises(_lowerCAmelCase ): get_dataset_config_info(_lowerCAmelCase , config_name=_lowerCAmelCase ) @pytest.mark.parametrize( '''path, expected''' , [ ('''squad''', '''plain_text'''), ('''acronym_identification''', '''default'''), ('''lhoestq/squad''', '''plain_text'''), ('''lhoestq/test''', '''default'''), ('''lhoestq/demo1''', '''lhoestq--demo1'''), ('''dalle-mini/wit''', '''dalle-mini--wit'''), ] , ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple ) -> Tuple: UpperCAmelCase : List[str] = get_dataset_config_names(_lowerCAmelCase ) assert expected in config_names @pytest.mark.parametrize( '''path, expected_configs, expected_splits_in_first_config''' , [ ('''squad''', ['''plain_text'''], ['''train''', '''validation''']), ('''dalle-mini/wit''', ['''dalle-mini--wit'''], ['''train''']), ('''paws''', ['''labeled_final''', '''labeled_swap''', '''unlabeled_final'''], ['''train''', '''test''', '''validation''']), ] , ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] ) -> Dict: UpperCAmelCase : str = get_dataset_infos(_lowerCAmelCase ) assert list(infos.keys() ) == expected_configs UpperCAmelCase : int = expected_configs[0] assert expected_config in infos UpperCAmelCase : str = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( '''path, expected_config, expected_splits''' , [ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] , ) def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : Any , _lowerCAmelCase : Tuple ) -> Tuple: UpperCAmelCase : int = get_dataset_infos(_lowerCAmelCase ) assert expected_config in infos UpperCAmelCase : Dict = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' , [ ('''paws''', None, ValueError), ] , ) def snake_case_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int ) -> int: with pytest.raises(_lowerCAmelCase ): get_dataset_split_names(_lowerCAmelCase , config_name=_lowerCAmelCase )
23
'''simple docstring''' # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers UpperCamelCase__: Optional[int] = "3" print("Python version:", sys.version) print("transformers version:", transformers.__version__) try: import torch print("Torch version:", torch.__version__) print("Cuda available:", torch.cuda.is_available()) print("Cuda version:", torch.version.cuda) print("CuDNN version:", torch.backends.cudnn.version()) print("Number of GPUs available:", torch.cuda.device_count()) print("NCCL version:", torch.cuda.nccl.version()) except ImportError: print("Torch version:", None) try: import deepspeed print("DeepSpeed version:", deepspeed.__version__) except ImportError: print("DeepSpeed version:", None) try: import tensorflow as tf print("TensorFlow version:", tf.__version__) print("TF GPUs available:", bool(tf.config.list_physical_devices("GPU"))) print("Number of TF GPUs available:", len(tf.config.list_physical_devices("GPU"))) except ImportError: print("TensorFlow version:", None)
23
1
'''simple docstring''' import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE: """simple docstring""" def __init__( self : Optional[int] , __snake_case : List[str] , __snake_case : Optional[Any]=13 , __snake_case : int=32 , __snake_case : Optional[int]=3 , __snake_case : int=4 , __snake_case : Optional[int]=[10, 20, 30, 40] , __snake_case : str=[2, 2, 3, 2] , __snake_case : Any=True , __snake_case : Optional[Any]=True , __snake_case : Tuple=37 , __snake_case : Optional[Any]="gelu" , __snake_case : Dict=10 , __snake_case : str=0.02 , __snake_case : Dict=["stage2", "stage3", "stage4"] , __snake_case : Tuple=[2, 3, 4] , __snake_case : List[Any]=None , ) -> Optional[Any]: UpperCAmelCase : Any = parent UpperCAmelCase : Tuple = batch_size UpperCAmelCase : List[str] = image_size UpperCAmelCase : List[Any] = num_channels UpperCAmelCase : Union[str, Any] = num_stages UpperCAmelCase : int = hidden_sizes UpperCAmelCase : Optional[int] = depths UpperCAmelCase : str = is_training UpperCAmelCase : Optional[Any] = use_labels UpperCAmelCase : Optional[int] = intermediate_size UpperCAmelCase : List[str] = hidden_act UpperCAmelCase : int = num_labels UpperCAmelCase : Any = initializer_range UpperCAmelCase : Tuple = out_features UpperCAmelCase : List[Any] = out_indices UpperCAmelCase : Union[str, Any] = scope def A ( self : List[str] ) -> Union[str, Any]: UpperCAmelCase : Union[str, Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase : Optional[int] = None if self.use_labels: UpperCAmelCase : List[str] = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase : Optional[int] = self.get_config() return config, pixel_values, labels def A ( self : Tuple ) -> Tuple: return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=__snake_case , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def A ( self : int , __snake_case : List[str] , __snake_case : Union[str, Any] , __snake_case : List[Any] ) -> Optional[Any]: UpperCAmelCase : List[str] = ConvNextModel(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Any = model(__snake_case ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def A ( self : Tuple , __snake_case : Dict , __snake_case : Any , __snake_case : List[str] ) -> Optional[int]: UpperCAmelCase : Optional[int] = ConvNextForImageClassification(__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Any = model(__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : str , __snake_case : Union[str, Any] , __snake_case : Any , __snake_case : str ) -> Dict: UpperCAmelCase : Union[str, Any] = ConvNextBackbone(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : List[str] = model(__snake_case ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None UpperCAmelCase : Dict = None UpperCAmelCase : int = ConvNextBackbone(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : int = model(__snake_case ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def A ( self : Union[str, Any] ) -> Dict: UpperCAmelCase : Any = self.prepare_config_and_inputs() UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : int = config_and_inputs UpperCAmelCase : Dict = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE( A__ , A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) lowerCamelCase__ = ( {"""feature-extraction""": ConvNextModel, """image-classification""": ConvNextForImageClassification} if is_torch_available() else {} ) lowerCamelCase__ = True lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False def A ( self : List[Any] ) -> Any: UpperCAmelCase : Optional[int] = ConvNextModelTester(self ) UpperCAmelCase : List[Any] = ConfigTester(self , config_class=__snake_case , has_text_modality=__snake_case , hidden_size=37 ) def A ( self : Optional[int] ) -> List[str]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def A ( self : List[str] ) -> Optional[Any]: return @unittest.skip(reason='''ConvNext does not use inputs_embeds''' ) def A ( self : Any ) -> Union[str, Any]: pass @unittest.skip(reason='''ConvNext does not support input and output embeddings''' ) def A ( self : Optional[Any] ) -> Optional[int]: pass @unittest.skip(reason='''ConvNext does not use feedforward chunking''' ) def A ( self : Optional[int] ) -> int: pass def A ( self : Optional[int] ) -> Optional[int]: UpperCAmelCase , UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase : int = model_class(__snake_case ) UpperCAmelCase : str = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase : Optional[Any] = [*signature.parameters.keys()] UpperCAmelCase : Tuple = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __snake_case ) def A ( self : Optional[int] ) -> Union[str, Any]: UpperCAmelCase : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__snake_case ) def A ( self : Union[str, Any] ) -> List[str]: UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*__snake_case ) def A ( self : List[Any] ) -> Optional[Any]: def check_hidden_states_output(__snake_case : List[str] , __snake_case : Tuple , __snake_case : List[str] ): UpperCAmelCase : List[Any] = model_class(__snake_case ) model.to(__snake_case ) model.eval() with torch.no_grad(): UpperCAmelCase : int = model(**self._prepare_for_class(__snake_case , __snake_case ) ) UpperCAmelCase : Dict = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCAmelCase : Union[str, Any] = self.model_tester.num_stages self.assertEqual(len(__snake_case ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) UpperCAmelCase , UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase : List[Any] = True check_hidden_states_output(__snake_case , __snake_case , __snake_case ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase : List[Any] = True check_hidden_states_output(__snake_case , __snake_case , __snake_case ) def A ( self : Optional[Any] ) -> Union[str, Any]: UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__snake_case ) @slow def A ( self : List[Any] ) -> List[str]: for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase : List[str] = ConvNextModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) def snake_case_ ( ) -> int: UpperCAmelCase : List[Any] = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" @cached_property def A ( self : Optional[Any] ) -> Dict: return AutoImageProcessor.from_pretrained('''facebook/convnext-tiny-224''' ) if is_vision_available() else None @slow def A ( self : Union[str, Any] ) -> Dict: UpperCAmelCase : Dict = ConvNextForImageClassification.from_pretrained('''facebook/convnext-tiny-224''' ).to(__snake_case ) UpperCAmelCase : Tuple = self.default_image_processor UpperCAmelCase : Optional[Any] = prepare_img() UpperCAmelCase : str = image_processor(images=__snake_case , return_tensors='''pt''' ).to(__snake_case ) # forward pass with torch.no_grad(): UpperCAmelCase : List[str] = model(**__snake_case ) # verify the logits UpperCAmelCase : Any = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , __snake_case ) UpperCAmelCase : Optional[Any] = torch.tensor([-0.02_60, -0.47_39, 0.19_11] ).to(__snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __snake_case , atol=1E-4 ) ) @require_torch class SCREAMING_SNAKE_CASE( unittest.TestCase , A__ ): """simple docstring""" lowerCamelCase__ = (ConvNextBackbone,) if is_torch_available() else () lowerCamelCase__ = ConvNextConfig lowerCamelCase__ = False def A ( self : str ) -> str: UpperCAmelCase : List[Any] = ConvNextModelTester(self )
23
'''simple docstring''' from __future__ import annotations def snake_case_ ( _lowerCAmelCase : list[int | float] , _lowerCAmelCase : int , _lowerCAmelCase : int ) -> int | float: if len(_lowerCAmelCase ) == 0: raise ValueError('''find_max() arg is an empty sequence''' ) if ( left >= len(_lowerCAmelCase ) or left < -len(_lowerCAmelCase ) or right >= len(_lowerCAmelCase ) or right < -len(_lowerCAmelCase ) ): raise IndexError('''list index out of range''' ) if left == right: return nums[left] UpperCAmelCase : List[Any] = (left + right) >> 1 # the middle UpperCAmelCase : Optional[Any] = find_max(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # find max in range[left, mid] UpperCAmelCase : Dict = find_max(_lowerCAmelCase , mid + 1 , _lowerCAmelCase ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
23
1
'''simple docstring''' from __future__ import annotations def snake_case_ ( _lowerCAmelCase : int ) -> list[int]: UpperCAmelCase : Tuple = [True] * limit UpperCAmelCase : Any = False UpperCAmelCase : Optional[int] = False UpperCAmelCase : Optional[Any] = True for i in range(3 , int(limit**0.5 + 1 ) , 2 ): UpperCAmelCase : Optional[int] = i * 2 while index < limit: UpperCAmelCase : int = False UpperCAmelCase : Optional[Any] = index + i UpperCAmelCase : Optional[int] = [2] for i in range(3 , _lowerCAmelCase , 2 ): if is_prime[i]: primes.append(_lowerCAmelCase ) return primes def snake_case_ ( _lowerCAmelCase : int = 1000000 ) -> int: UpperCAmelCase : Optional[int] = prime_sieve(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = 0 UpperCAmelCase : List[Any] = 0 for i in range(len(_lowerCAmelCase ) ): for j in range(i + length , len(_lowerCAmelCase ) ): UpperCAmelCase : Any = sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: UpperCAmelCase : List[str] = j - i UpperCAmelCase : str = sol return largest if __name__ == "__main__": print(F"{solution() = }")
23
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = 42 lowerCamelCase__ = 42 def __init__( self : Union[str, Any] , __snake_case : UNetaDModel , __snake_case : ScoreSdeVeScheduler ) -> int: super().__init__() self.register_modules(unet=__snake_case , scheduler=__snake_case ) @torch.no_grad() def __call__( self : Optional[int] , __snake_case : int = 1 , __snake_case : int = 2000 , __snake_case : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __snake_case : Optional[str] = "pil" , __snake_case : bool = True , **__snake_case : Optional[int] , ) -> Union[ImagePipelineOutput, Tuple]: UpperCAmelCase : str = self.unet.config.sample_size UpperCAmelCase : Union[str, Any] = (batch_size, 3, img_size, img_size) UpperCAmelCase : int = self.unet UpperCAmelCase : Any = randn_tensor(__snake_case , generator=__snake_case ) * self.scheduler.init_noise_sigma UpperCAmelCase : List[Any] = sample.to(self.device ) self.scheduler.set_timesteps(__snake_case ) self.scheduler.set_sigmas(__snake_case ) for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): UpperCAmelCase : Any = self.scheduler.sigmas[i] * torch.ones(shape[0] , device=self.device ) # correction step for _ in range(self.scheduler.config.correct_steps ): UpperCAmelCase : Union[str, Any] = self.unet(__snake_case , __snake_case ).sample UpperCAmelCase : Optional[Any] = self.scheduler.step_correct(__snake_case , __snake_case , generator=__snake_case ).prev_sample # prediction step UpperCAmelCase : Optional[Any] = model(__snake_case , __snake_case ).sample UpperCAmelCase : List[str] = self.scheduler.step_pred(__snake_case , __snake_case , __snake_case , generator=__snake_case ) UpperCAmelCase , UpperCAmelCase : Optional[Any] = output.prev_sample, output.prev_sample_mean UpperCAmelCase : int = sample_mean.clamp(0 , 1 ) UpperCAmelCase : Union[str, Any] = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCAmelCase : Optional[Any] = self.numpy_to_pil(__snake_case ) if not return_dict: return (sample,) return ImagePipelineOutput(images=__snake_case )
23
1
'''simple docstring''' import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : Union[str, Any] ) -> Any: UpperCAmelCase : int = checkpoint UpperCAmelCase : Any = {} UpperCAmelCase : Optional[Any] = vae_state_dict['''encoder.conv_in.weight'''] UpperCAmelCase : Dict = vae_state_dict['''encoder.conv_in.bias'''] UpperCAmelCase : Tuple = vae_state_dict['''encoder.conv_out.weight'''] UpperCAmelCase : Optional[int] = vae_state_dict['''encoder.conv_out.bias'''] UpperCAmelCase : int = vae_state_dict['''encoder.norm_out.weight'''] UpperCAmelCase : Union[str, Any] = vae_state_dict['''encoder.norm_out.bias'''] UpperCAmelCase : List[Any] = vae_state_dict['''decoder.conv_in.weight'''] UpperCAmelCase : List[Any] = vae_state_dict['''decoder.conv_in.bias'''] UpperCAmelCase : str = vae_state_dict['''decoder.conv_out.weight'''] UpperCAmelCase : Any = vae_state_dict['''decoder.conv_out.bias'''] UpperCAmelCase : List[Any] = vae_state_dict['''decoder.norm_out.weight'''] UpperCAmelCase : Any = vae_state_dict['''decoder.norm_out.bias'''] UpperCAmelCase : int = vae_state_dict['''quant_conv.weight'''] UpperCAmelCase : Union[str, Any] = vae_state_dict['''quant_conv.bias'''] UpperCAmelCase : Any = vae_state_dict['''post_quant_conv.weight'''] UpperCAmelCase : List[Any] = vae_state_dict['''post_quant_conv.bias'''] # Retrieves the keys for the encoder down blocks only UpperCAmelCase : Optional[Any] = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''encoder.down''' in layer} ) UpperCAmelCase : Optional[Any] = { layer_id: [key for key in vae_state_dict if f"""down.{layer_id}""" in key] for layer_id in range(_lowerCAmelCase ) } # Retrieves the keys for the decoder up blocks only UpperCAmelCase : Any = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''decoder.up''' in layer} ) UpperCAmelCase : List[str] = { layer_id: [key for key in vae_state_dict if f"""up.{layer_id}""" in key] for layer_id in range(_lowerCAmelCase ) } for i in range(_lowerCAmelCase ): UpperCAmelCase : Dict = [key for key in down_blocks[i] if f"""down.{i}""" in key and f"""down.{i}.downsample""" not in key] if f"""encoder.down.{i}.downsample.conv.weight""" in vae_state_dict: UpperCAmelCase : Optional[int] = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.weight""" ) UpperCAmelCase : List[Any] = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.bias""" ) UpperCAmelCase : Tuple = renew_vae_resnet_paths(_lowerCAmelCase ) UpperCAmelCase : str = {'''old''': f"""down.{i}.block""", '''new''': f"""down_blocks.{i}.resnets"""} assign_to_checkpoint(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , additional_replacements=[meta_path] , config=_lowerCAmelCase ) UpperCAmelCase : List[Any] = [key for key in vae_state_dict if '''encoder.mid.block''' in key] UpperCAmelCase : Union[str, Any] = 2 for i in range(1 , num_mid_res_blocks + 1 ): UpperCAmelCase : Dict = [key for key in mid_resnets if f"""encoder.mid.block_{i}""" in key] UpperCAmelCase : Tuple = renew_vae_resnet_paths(_lowerCAmelCase ) UpperCAmelCase : Dict = {'''old''': f"""mid.block_{i}""", '''new''': f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , additional_replacements=[meta_path] , config=_lowerCAmelCase ) UpperCAmelCase : Any = [key for key in vae_state_dict if '''encoder.mid.attn''' in key] UpperCAmelCase : List[Any] = renew_vae_attention_paths(_lowerCAmelCase ) UpperCAmelCase : Dict = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , additional_replacements=[meta_path] , config=_lowerCAmelCase ) conv_attn_to_linear(_lowerCAmelCase ) for i in range(_lowerCAmelCase ): UpperCAmelCase : List[Any] = num_up_blocks - 1 - i UpperCAmelCase : Dict = [ key for key in up_blocks[block_id] if f"""up.{block_id}""" in key and f"""up.{block_id}.upsample""" not in key ] if f"""decoder.up.{block_id}.upsample.conv.weight""" in vae_state_dict: UpperCAmelCase : List[Any] = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.weight""" ] UpperCAmelCase : List[Any] = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.bias""" ] UpperCAmelCase : List[Any] = renew_vae_resnet_paths(_lowerCAmelCase ) UpperCAmelCase : int = {'''old''': f"""up.{block_id}.block""", '''new''': f"""up_blocks.{i}.resnets"""} assign_to_checkpoint(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , additional_replacements=[meta_path] , config=_lowerCAmelCase ) UpperCAmelCase : Dict = [key for key in vae_state_dict if '''decoder.mid.block''' in key] UpperCAmelCase : Tuple = 2 for i in range(1 , num_mid_res_blocks + 1 ): UpperCAmelCase : List[str] = [key for key in mid_resnets if f"""decoder.mid.block_{i}""" in key] UpperCAmelCase : List[str] = renew_vae_resnet_paths(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = {'''old''': f"""mid.block_{i}""", '''new''': f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , additional_replacements=[meta_path] , config=_lowerCAmelCase ) UpperCAmelCase : int = [key for key in vae_state_dict if '''decoder.mid.attn''' in key] UpperCAmelCase : Dict = renew_vae_attention_paths(_lowerCAmelCase ) UpperCAmelCase : Tuple = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , additional_replacements=[meta_path] , config=_lowerCAmelCase ) conv_attn_to_linear(_lowerCAmelCase ) return new_checkpoint def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : str , ) -> Tuple: # Only support V1 UpperCAmelCase : str = requests.get( ''' https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml''' ) UpperCAmelCase : Optional[Any] = io.BytesIO(r.content ) UpperCAmelCase : int = OmegaConf.load(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = 512 UpperCAmelCase : List[str] = '''cuda''' if torch.cuda.is_available() else '''cpu''' if checkpoint_path.endswith('''safetensors''' ): from safetensors import safe_open UpperCAmelCase : int = {} with safe_open(_lowerCAmelCase , framework='''pt''' , device='''cpu''' ) as f: for key in f.keys(): UpperCAmelCase : str = f.get_tensor(_lowerCAmelCase ) else: UpperCAmelCase : Dict = torch.load(_lowerCAmelCase , map_location=_lowerCAmelCase )['''state_dict'''] # Convert the VAE model. UpperCAmelCase : List[Any] = create_vae_diffusers_config(_lowerCAmelCase , image_size=_lowerCAmelCase ) UpperCAmelCase : Tuple = custom_convert_ldm_vae_checkpoint(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Tuple = AutoencoderKL(**_lowerCAmelCase ) vae.load_state_dict(_lowerCAmelCase ) vae.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": UpperCamelCase__: Dict = argparse.ArgumentParser() parser.add_argument("--vae_pt_path", default=None, type=str, required=True, help="Path to the VAE.pt to convert.") parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the VAE.pt to convert.") UpperCamelCase__: str = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
23
'''simple docstring''' import warnings from contextlib import contextmanager from ....processing_utils import ProcessorMixin class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = """MCTCTFeatureExtractor""" lowerCamelCase__ = """AutoTokenizer""" def __init__( self : Dict , __snake_case : Optional[int] , __snake_case : List[str] ) -> str: super().__init__(__snake_case , __snake_case ) UpperCAmelCase : List[Any] = self.feature_extractor UpperCAmelCase : Union[str, Any] = False def __call__( self : Any , *__snake_case : List[str] , **__snake_case : Any ) -> List[Any]: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*__snake_case , **__snake_case ) if "raw_speech" in kwargs: warnings.warn('''Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.''' ) UpperCAmelCase : int = kwargs.pop('''raw_speech''' ) else: UpperCAmelCase : Union[str, Any] = kwargs.pop('''audio''' , __snake_case ) UpperCAmelCase : Optional[Any] = kwargs.pop('''sampling_rate''' , __snake_case ) UpperCAmelCase : Dict = kwargs.pop('''text''' , __snake_case ) if len(__snake_case ) > 0: UpperCAmelCase : Any = args[0] UpperCAmelCase : Optional[int] = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if audio is not None: UpperCAmelCase : List[str] = self.feature_extractor(__snake_case , *__snake_case , sampling_rate=__snake_case , **__snake_case ) if text is not None: UpperCAmelCase : int = self.tokenizer(__snake_case , **__snake_case ) if text is None: return inputs elif audio is None: return encodings else: UpperCAmelCase : str = encodings['''input_ids'''] return inputs def A ( self : List[Any] , *__snake_case : List[Any] , **__snake_case : List[Any] ) -> str: return self.tokenizer.batch_decode(*__snake_case , **__snake_case ) def A ( self : List[Any] , *__snake_case : int , **__snake_case : Optional[int] ) -> Any: # For backward compatibility if self._in_target_context_manager: return self.current_processor.pad(*__snake_case , **__snake_case ) UpperCAmelCase : List[Any] = kwargs.pop('''input_features''' , __snake_case ) UpperCAmelCase : Optional[Any] = kwargs.pop('''labels''' , __snake_case ) if len(__snake_case ) > 0: UpperCAmelCase : List[str] = args[0] UpperCAmelCase : List[Any] = args[1:] if input_features is not None: UpperCAmelCase : Tuple = self.feature_extractor.pad(__snake_case , *__snake_case , **__snake_case ) if labels is not None: UpperCAmelCase : Optional[int] = self.tokenizer.pad(__snake_case , **__snake_case ) if labels is None: return input_features elif input_features is None: return labels else: UpperCAmelCase : List[str] = labels['''input_ids'''] return input_features def A ( self : Union[str, Any] , *__snake_case : Optional[Any] , **__snake_case : Optional[int] ) -> Optional[Any]: return self.tokenizer.decode(*__snake_case , **__snake_case ) @contextmanager def A ( self : Any ) -> Optional[int]: warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your audio inputs, or in a separate call.''' ) UpperCAmelCase : Dict = True UpperCAmelCase : List[Any] = self.tokenizer yield UpperCAmelCase : Tuple = self.feature_extractor UpperCAmelCase : List[Any] = False
23
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: UpperCamelCase__: Union[str, Any] = None UpperCamelCase__: List[Any] = logging.get_logger(__name__) UpperCamelCase__: Dict = {"vocab_file": "sentencepiece.bpe.model", "tokenizer_file": "tokenizer.json"} UpperCamelCase__: int = { "vocab_file": { "facebook/mbart-large-en-ro": ( "https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model" ), "facebook/mbart-large-cc25": ( "https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model" ), }, "tokenizer_file": { "facebook/mbart-large-en-ro": "https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json", "facebook/mbart-large-cc25": "https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json", }, } UpperCamelCase__: List[str] = { "facebook/mbart-large-en-ro": 1024, "facebook/mbart-large-cc25": 1024, } # fmt: off UpperCamelCase__: Optional[Any] = ["ar_AR", "cs_CZ", "de_DE", "en_XX", "es_XX", "et_EE", "fi_FI", "fr_XX", "gu_IN", "hi_IN", "it_IT", "ja_XX", "kk_KZ", "ko_KR", "lt_LT", "lv_LV", "my_MM", "ne_NP", "nl_XX", "ro_RO", "ru_RU", "si_LK", "tr_TR", "vi_VN", "zh_CN"] class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = ["""input_ids""", """attention_mask"""] lowerCamelCase__ = MBartTokenizer lowerCamelCase__ = [] lowerCamelCase__ = [] def __init__( self : int , __snake_case : List[Any]=None , __snake_case : Optional[Any]=None , __snake_case : Optional[int]="<s>" , __snake_case : List[Any]="</s>" , __snake_case : Optional[int]="</s>" , __snake_case : List[str]="<s>" , __snake_case : List[str]="<unk>" , __snake_case : Union[str, Any]="<pad>" , __snake_case : Optional[Any]="<mask>" , __snake_case : Tuple=None , __snake_case : List[str]=None , __snake_case : Any=None , **__snake_case : Tuple , ) -> int: # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase : Optional[int] = AddedToken(__snake_case , lstrip=__snake_case , rstrip=__snake_case ) if isinstance(__snake_case , __snake_case ) else mask_token super().__init__( vocab_file=__snake_case , tokenizer_file=__snake_case , bos_token=__snake_case , eos_token=__snake_case , sep_token=__snake_case , cls_token=__snake_case , unk_token=__snake_case , pad_token=__snake_case , mask_token=__snake_case , src_lang=__snake_case , tgt_lang=__snake_case , additional_special_tokens=__snake_case , **__snake_case , ) UpperCAmelCase : Tuple = vocab_file UpperCAmelCase : Union[str, Any] = False if not self.vocab_file else True UpperCAmelCase : Dict = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'''additional_special_tokens''': _additional_special_tokens} ) UpperCAmelCase : int = { lang_code: self.convert_tokens_to_ids(__snake_case ) for lang_code in FAIRSEQ_LANGUAGE_CODES } UpperCAmelCase : Dict = src_lang if src_lang is not None else '''en_XX''' UpperCAmelCase : Optional[Any] = self.convert_tokens_to_ids(self._src_lang ) UpperCAmelCase : Tuple = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def A ( self : Tuple ) -> str: return self._src_lang @src_lang.setter def A ( self : List[Any] , __snake_case : str ) -> None: UpperCAmelCase : Optional[Any] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def A ( self : Union[str, Any] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def A ( self : Optional[int] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: UpperCAmelCase : Union[str, Any] = [self.sep_token_id] UpperCAmelCase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def A ( self : Tuple , __snake_case : Any , __snake_case : str , __snake_case : Optional[str] , __snake_case : Optional[str] , **__snake_case : Any ) -> Tuple: if src_lang is None or tgt_lang is None: raise ValueError('''Translation requires a `src_lang` and a `tgt_lang` for this model''' ) UpperCAmelCase : List[Any] = src_lang UpperCAmelCase : Union[str, Any] = self(__snake_case , add_special_tokens=__snake_case , return_tensors=__snake_case , **__snake_case ) UpperCAmelCase : str = self.convert_tokens_to_ids(__snake_case ) UpperCAmelCase : Union[str, Any] = tgt_lang_id return inputs def A ( self : str , __snake_case : List[str] , __snake_case : str = "en_XX" , __snake_case : Optional[List[str]] = None , __snake_case : str = "ro_RO" , **__snake_case : List[Any] , ) -> BatchEncoding: UpperCAmelCase : Dict = src_lang UpperCAmelCase : List[str] = tgt_lang return super().prepare_seqaseq_batch(__snake_case , __snake_case , **__snake_case ) def A ( self : Any ) -> str: return self.set_src_lang_special_tokens(self.src_lang ) def A ( self : Optional[int] ) -> Optional[int]: return self.set_tgt_lang_special_tokens(self.tgt_lang ) def A ( self : Any , __snake_case : str ) -> None: UpperCAmelCase : Optional[int] = self.convert_tokens_to_ids(__snake_case ) UpperCAmelCase : Dict = [] UpperCAmelCase : Optional[int] = [self.eos_token_id, self.cur_lang_code] UpperCAmelCase : Dict = self.convert_ids_to_tokens(self.prefix_tokens ) UpperCAmelCase : str = self.convert_ids_to_tokens(self.suffix_tokens ) UpperCAmelCase : Optional[int] = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def A ( self : Dict , __snake_case : str ) -> None: UpperCAmelCase : Optional[Any] = self.convert_tokens_to_ids(__snake_case ) UpperCAmelCase : Dict = [] UpperCAmelCase : Optional[Any] = [self.eos_token_id, self.cur_lang_code] UpperCAmelCase : Tuple = self.convert_ids_to_tokens(self.prefix_tokens ) UpperCAmelCase : Any = self.convert_ids_to_tokens(self.suffix_tokens ) UpperCAmelCase : Tuple = processors.TemplateProcessing( single=prefix_tokens_str + ['''$A'''] + suffix_tokens_str , pair=prefix_tokens_str + ['''$A''', '''$B'''] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def A ( self : Any , __snake_case : str , __snake_case : Optional[str] = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(__snake_case ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory.""" ) return UpperCAmelCase : List[str] = os.path.join( __snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__snake_case ): copyfile(self.vocab_file , __snake_case ) return (out_vocab_file,)
23
'''simple docstring''' from math import isclose, sqrt def snake_case_ ( _lowerCAmelCase : float , _lowerCAmelCase : float , _lowerCAmelCase : float ) -> tuple[float, float, float]: UpperCAmelCase : Optional[int] = point_y / 4 / point_x UpperCAmelCase : str = 2 * normal_gradient / (1 + normal_gradient * normal_gradient) UpperCAmelCase : Any = (1 - normal_gradient * normal_gradient) / ( 1 + normal_gradient * normal_gradient ) UpperCAmelCase : Union[str, Any] = (sa - ca * incoming_gradient) / (ca + sa * incoming_gradient) # to find the next point, solve the simultaeneous equations: # y^2 + 4x^2 = 100 # y - b = m * (x - a) # ==> A x^2 + B x + C = 0 UpperCAmelCase : Union[str, Any] = outgoing_gradient**2 + 4 UpperCAmelCase : Dict = 2 * outgoing_gradient * (point_y - outgoing_gradient * point_x) UpperCAmelCase : List[Any] = (point_y - outgoing_gradient * point_x) ** 2 - 100 UpperCAmelCase : List[str] = ( -linear_term - sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) UpperCAmelCase : Optional[int] = ( -linear_term + sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) # two solutions, one of which is our input point UpperCAmelCase : Optional[Any] = x_minus if isclose(_lowerCAmelCase , _lowerCAmelCase ) else x_plus UpperCAmelCase : Union[str, Any] = point_y + outgoing_gradient * (next_x - point_x) return next_x, next_y, outgoing_gradient def snake_case_ ( _lowerCAmelCase : float = 1.4 , _lowerCAmelCase : float = -9.6 ) -> int: UpperCAmelCase : int = 0 UpperCAmelCase : float = first_x_coord UpperCAmelCase : float = first_y_coord UpperCAmelCase : float = (1_0.1 - point_y) / (0.0 - point_x) while not (-0.0_1 <= point_x <= 0.0_1 and point_y > 0): UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : str = next_point(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) num_reflections += 1 return num_reflections if __name__ == "__main__": print(F"{solution() = }")
23
1
'''simple docstring''' from maths.is_square_free import is_square_free from maths.prime_factors import prime_factors def snake_case_ ( _lowerCAmelCase : int ) -> int: UpperCAmelCase : Union[str, Any] = prime_factors(_lowerCAmelCase ) if is_square_free(_lowerCAmelCase ): return -1 if len(_lowerCAmelCase ) % 2 else 1 return 0 if __name__ == "__main__": import doctest doctest.testmod()
23
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase__: str = { "configuration_lxmert": ["LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "LxmertConfig"], "tokenization_lxmert": ["LxmertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: int = ["LxmertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: Union[str, Any] = [ "LxmertEncoder", "LxmertForPreTraining", "LxmertForQuestionAnswering", "LxmertModel", "LxmertPreTrainedModel", "LxmertVisualFeatureEncoder", "LxmertXLayer", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: int = [ "TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFLxmertForPreTraining", "TFLxmertMainLayer", "TFLxmertModel", "TFLxmertPreTrainedModel", "TFLxmertVisualFeatureEncoder", ] if TYPE_CHECKING: from .configuration_lxmert import LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, LxmertConfig from .tokenization_lxmert import LxmertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_lxmert_fast import LxmertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lxmert import ( LxmertEncoder, LxmertForPreTraining, LxmertForQuestionAnswering, LxmertModel, LxmertPreTrainedModel, LxmertVisualFeatureEncoder, LxmertXLayer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_lxmert import ( TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFLxmertForPreTraining, TFLxmertMainLayer, TFLxmertModel, TFLxmertPreTrainedModel, TFLxmertVisualFeatureEncoder, ) else: import sys UpperCamelCase__: Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
1
'''simple docstring''' from math import ceil def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] ) -> Tuple: UpperCAmelCase : List[Any] = list(range(0 , _lowerCAmelCase ) ) UpperCAmelCase : Union[str, Any] = [item for sublist in list(device_map.values() ) for item in sublist] # Duplicate check UpperCAmelCase : Optional[Any] = [] for i in device_map_blocks: if device_map_blocks.count(_lowerCAmelCase ) > 1 and i not in duplicate_blocks: duplicate_blocks.append(_lowerCAmelCase ) # Missing blocks UpperCAmelCase : Optional[int] = [i for i in blocks if i not in device_map_blocks] UpperCAmelCase : int = [i for i in device_map_blocks if i not in blocks] if len(_lowerCAmelCase ) != 0: raise ValueError( '''Duplicate attention blocks specified in device_map. Attention blocks must be specified to one device.''' ''' These attention blocks were specified more than once: ''' + str(_lowerCAmelCase ) ) if len(_lowerCAmelCase ) != 0: raise ValueError( '''There are attention blocks for this model that are not specified in the device_map. Add these attention ''' '''blocks to a device on the device_map: ''' + str(_lowerCAmelCase ) ) if len(_lowerCAmelCase ) != 0: raise ValueError( '''The device_map contains more attention blocks than this model has. Remove these from the device_map:''' + str(_lowerCAmelCase ) ) def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : str ) -> int: UpperCAmelCase : Dict = list(range(_lowerCAmelCase ) ) UpperCAmelCase : Optional[Any] = int(ceil(n_layers / len(_lowerCAmelCase ) ) ) UpperCAmelCase : List[str] = [layers[i : i + n_blocks] for i in range(0 , _lowerCAmelCase , _lowerCAmelCase )] return dict(zip(_lowerCAmelCase , _lowerCAmelCase ) )
23
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE( A__ , A__ , A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = AltDiffusionPipeline lowerCamelCase__ = TEXT_TO_IMAGE_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS def A ( self : Dict ) -> int: torch.manual_seed(0 ) UpperCAmelCase : str = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) UpperCAmelCase : Dict = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) UpperCAmelCase : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) UpperCAmelCase : Dict = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5002 , ) UpperCAmelCase : List[Any] = CLIPTextModel(__snake_case ) UpperCAmelCase : Optional[Any] = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) UpperCAmelCase : Optional[int] = 77 UpperCAmelCase : Optional[int] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def A ( self : Optional[Any] , __snake_case : Dict , __snake_case : List[str]=0 ) -> Union[str, Any]: if str(__snake_case ).startswith('''mps''' ): UpperCAmelCase : str = torch.manual_seed(__snake_case ) else: UpperCAmelCase : Tuple = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) UpperCAmelCase : Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def A ( self : Union[str, Any] ) -> List[str]: super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def A ( self : Tuple ) -> List[str]: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def A ( self : Tuple ) -> Optional[int]: UpperCAmelCase : List[str] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase : Any = self.get_dummy_components() torch.manual_seed(0 ) UpperCAmelCase : Optional[int] = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase : List[str] = RobertaSeriesModelWithTransformation(__snake_case ) UpperCAmelCase : str = text_encoder UpperCAmelCase : Optional[int] = AltDiffusionPipeline(**__snake_case ) UpperCAmelCase : str = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : Optional[int] = self.get_dummy_inputs(__snake_case ) UpperCAmelCase : Optional[int] = '''A photo of an astronaut''' UpperCAmelCase : List[Any] = alt_pipe(**__snake_case ) UpperCAmelCase : Optional[Any] = output.images UpperCAmelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : List[str] = np.array( [0.5_74_81_62, 0.60_44_71_45, 0.48_82_12_17, 0.50_10_06_36, 0.5_43_11_85, 0.45_76_36_83, 0.49_65_76_96, 0.48_13_27_33, 0.47_57_30_93] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A ( self : int ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase : int = self.get_dummy_components() UpperCAmelCase : int = PNDMScheduler(skip_prk_steps=__snake_case ) torch.manual_seed(0 ) UpperCAmelCase : int = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase : Union[str, Any] = RobertaSeriesModelWithTransformation(__snake_case ) UpperCAmelCase : Union[str, Any] = text_encoder UpperCAmelCase : Optional[int] = AltDiffusionPipeline(**__snake_case ) UpperCAmelCase : Dict = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : int = self.get_dummy_inputs(__snake_case ) UpperCAmelCase : Optional[int] = alt_pipe(**__snake_case ) UpperCAmelCase : Optional[int] = output.images UpperCAmelCase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : Optional[int] = np.array( [0.51_60_50_93, 0.5_70_72_41, 0.47_36_55_07, 0.50_57_88_86, 0.5_63_38_77, 0.4_64_25_03, 0.5_18_20_81, 0.48_76_34_84, 0.49_08_42_37] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : str ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : List[Any] ) -> Any: # make sure here that pndm scheduler skips prk UpperCAmelCase : List[Any] = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''' , safety_checker=__snake_case ) UpperCAmelCase : Tuple = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : List[Any] = '''A painting of a squirrel eating a burger''' UpperCAmelCase : Any = torch.manual_seed(0 ) UpperCAmelCase : Optional[int] = alt_pipe([prompt] , generator=__snake_case , guidance_scale=6.0 , num_inference_steps=20 , output_type='''np''' ) UpperCAmelCase : Dict = output.images UpperCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase : List[str] = np.array([0.10_10, 0.08_00, 0.07_94, 0.08_85, 0.08_43, 0.07_62, 0.07_69, 0.07_29, 0.05_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A ( self : Tuple ) -> int: UpperCAmelCase : int = DDIMScheduler.from_pretrained('''BAAI/AltDiffusion''' , subfolder='''scheduler''' ) UpperCAmelCase : Tuple = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''' , scheduler=__snake_case , safety_checker=__snake_case ) UpperCAmelCase : Dict = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : Tuple = '''A painting of a squirrel eating a burger''' UpperCAmelCase : Optional[int] = torch.manual_seed(0 ) UpperCAmelCase : List[Any] = alt_pipe([prompt] , generator=__snake_case , num_inference_steps=2 , output_type='''numpy''' ) UpperCAmelCase : Dict = output.images UpperCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase : Union[str, Any] = np.array([0.40_19, 0.40_52, 0.38_10, 0.41_19, 0.39_16, 0.39_82, 0.46_51, 0.41_95, 0.53_23] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
23
1
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, require_tf, slow if is_tf_available(): import tensorflow as tf if is_tensorflow_text_available(): from transformers.models.bert import TFBertTokenizer UpperCamelCase__: Union[str, Any] = ["bert-base-uncased", "bert-base-cased"] UpperCamelCase__: Optional[Any] = "hf-internal-testing/tiny-bert-tf-only" if is_tf_available(): class SCREAMING_SNAKE_CASE( tf.keras.Model ): """simple docstring""" def __init__( self : Union[str, Any] , __snake_case : Tuple ) -> Dict: super().__init__() UpperCAmelCase : Union[str, Any] = tokenizer UpperCAmelCase : str = AutoConfig.from_pretrained(__snake_case ) UpperCAmelCase : Any = TFAutoModel.from_config(__snake_case ) def A ( self : List[Any] , __snake_case : Optional[Any] ) -> Optional[Any]: UpperCAmelCase : List[Any] = self.tokenizer(__snake_case ) UpperCAmelCase : Optional[int] = self.bert(**__snake_case ) return out["pooler_output"] @require_tf @require_tensorflow_text class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : int ) -> Union[str, Any]: super().setUp() UpperCAmelCase : Dict = [ BertTokenizer.from_pretrained(__snake_case ) for checkpoint in (TOKENIZER_CHECKPOINTS * 2) ] # repeat for when fast_bert_tokenizer=false UpperCAmelCase : str = [TFBertTokenizer.from_pretrained(__snake_case ) for checkpoint in TOKENIZER_CHECKPOINTS] + [ TFBertTokenizer.from_pretrained(__snake_case , use_fast_bert_tokenizer=__snake_case ) for checkpoint in TOKENIZER_CHECKPOINTS ] assert len(self.tokenizers ) == len(self.tf_tokenizers ) UpperCAmelCase : str = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] UpperCAmelCase : Optional[Any] = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def A ( self : Dict ) -> str: for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in (self.test_sentences, self.paired_sentences): UpperCAmelCase : List[str] = tokenizer(__snake_case , return_tensors='''tf''' , padding='''longest''' ) UpperCAmelCase : Dict = tf_tokenizer(__snake_case ) for key in python_outputs.keys(): self.assertTrue(tf.reduce_all(python_outputs[key].shape == tf_outputs[key].shape ) ) self.assertTrue(tf.reduce_all(tf.cast(python_outputs[key] , tf.intaa ) == tf_outputs[key] ) ) @slow def A ( self : Union[str, Any] ) -> Union[str, Any]: for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase : Optional[int] = tf_tokenizer(self.paired_sentences ) UpperCAmelCase : Optional[Any] = tf_tokenizer( text=[sentence[0] for sentence in self.paired_sentences] , text_pair=[sentence[1] for sentence in self.paired_sentences] , ) for key in merged_outputs.keys(): self.assertTrue(tf.reduce_all(tf.cast(merged_outputs[key] , tf.intaa ) == separated_outputs[key] ) ) @slow def A ( self : str ) -> Any: for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase : str = tf.function(__snake_case ) for test_inputs in (self.test_sentences, self.paired_sentences): UpperCAmelCase : int = tf.constant(__snake_case ) UpperCAmelCase : Tuple = compiled_tokenizer(__snake_case ) UpperCAmelCase : List[Any] = tf_tokenizer(__snake_case ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def A ( self : Optional[int] ) -> int: for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase : Tuple = ModelToSave(tokenizer=__snake_case ) UpperCAmelCase : Any = tf.convert_to_tensor(self.test_sentences ) UpperCAmelCase : Optional[Any] = model(__snake_case ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: UpperCAmelCase : Optional[Any] = Path(__snake_case ) / '''saved.model''' model.save(__snake_case ) UpperCAmelCase : Tuple = tf.keras.models.load_model(__snake_case ) UpperCAmelCase : List[Any] = loaded_model(__snake_case ) # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertLessEqual(tf.reduce_max(tf.abs(out - loaded_output ) ) , 1E-5 )
23
'''simple docstring''' import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] ) -> Optional[int]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[int] ) -> Dict: UpperCAmelCase : Dict = tmp_path / '''cache''' UpperCAmelCase : List[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase : str = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict , _lowerCAmelCase : int ) -> Optional[int]: UpperCAmelCase : Any = tmp_path / '''cache''' UpperCAmelCase : Tuple = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Any = features.copy() if features else default_expected_features UpperCAmelCase : List[Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Dict = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Tuple ) -> Tuple: UpperCAmelCase : Optional[Any] = tmp_path / '''cache''' UpperCAmelCase : Optional[int] = {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''} UpperCAmelCase : int = features.copy() if features else default_expected_features UpperCAmelCase : Any = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict ) -> Union[str, Any]: # jsonl_312_path features are {"col_3": "float64", "col_1": "string", "col_2": "int64"} UpperCAmelCase : Tuple = {'''col_2''': '''int64''', '''col_3''': '''float64''', '''col_1''': '''string'''} UpperCAmelCase : List[str] = features.copy() UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = tmp_path / '''cache''' UpperCAmelCase : List[str] = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ) -> Optional[Any]: UpperCAmelCase : Any = tmp_path / '''cache''' UpperCAmelCase : List[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : List[Any] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , split=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('''path_type''' , [str, list] ) def snake_case_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> Dict: if issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase : str = jsonl_path elif issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase : Dict = [jsonl_path] UpperCAmelCase : int = tmp_path / '''cache''' UpperCAmelCase : Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[int] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : Dict=("train",) ) -> Union[str, Any]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) for split in splits: UpperCAmelCase : List[str] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ) -> Any: UpperCAmelCase : Optional[Any] = tmp_path / '''cache''' UpperCAmelCase : List[str] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase : Optional[int] = JsonDatasetReader({'''train''': jsonl_path} , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] ) -> int: UpperCAmelCase : Dict = tmp_path / '''cache''' UpperCAmelCase : Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[int] = features.copy() if features else default_expected_features UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = JsonDatasetReader({'''train''': jsonl_path} , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict ) -> Union[str, Any]: if split: UpperCAmelCase : Optional[int] = {split: jsonl_path} else: UpperCAmelCase : Any = '''train''' UpperCAmelCase : Any = {'''train''': jsonl_path, '''test''': jsonl_path} UpperCAmelCase : Tuple = tmp_path / '''cache''' UpperCAmelCase : int = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[Any] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def snake_case_ ( _lowerCAmelCase : List[str] ) -> str: return json.load(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Dict ) -> str: return [json.loads(_lowerCAmelCase ) for line in buffer] class SCREAMING_SNAKE_CASE: """simple docstring""" @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def A ( self : Union[str, Any] , __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : Optional[int] ) -> Dict: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case ).write() buffer.seek(0 ) UpperCAmelCase : Union[str, Any] = load_json_function(__snake_case ) assert isinstance(__snake_case , __snake_case ) assert isinstance(exported_content[0] , __snake_case ) assert len(__snake_case ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def A ( self : Optional[int] , __snake_case : Optional[Any] , __snake_case : str , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : Optional[Any] ) -> List[Any]: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , orient=__snake_case ).write() buffer.seek(0 ) UpperCAmelCase : Union[str, Any] = load_json(__snake_case ) assert isinstance(__snake_case , __snake_case ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(__snake_case , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(__snake_case ) == 10 @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def A ( self : str , __snake_case : str , __snake_case : str , __snake_case : int ) -> Any: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase : Any = load_json_function(__snake_case ) assert isinstance(__snake_case , __snake_case ) assert isinstance(exported_content[0] , __snake_case ) assert len(__snake_case ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def A ( self : Any , __snake_case : int , __snake_case : Union[str, Any] , __snake_case : Optional[int] , __snake_case : Optional[Any] , __snake_case : List[str] ) -> Any: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , orient=__snake_case , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase : List[str] = load_json(__snake_case ) assert isinstance(__snake_case , __snake_case ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(__snake_case , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(__snake_case ) == 10 def A ( self : List[Any] , __snake_case : str ) -> Dict: with pytest.raises(__snake_case ): with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , num_proc=0 ) @pytest.mark.parametrize('''compression, extension''' , [('''gzip''', '''gz'''), ('''bz2''', '''bz2'''), ('''xz''', '''xz''')] ) def A ( self : Optional[int] , __snake_case : Any , __snake_case : str , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Dict ) -> Union[str, Any]: UpperCAmelCase : List[str] = tmp_path_factory.mktemp('''data''' ) / F"""test.json.{extension}""" UpperCAmelCase : List[Any] = str(shared_datadir / F"""test_file.json.{extension}""" ) JsonDatasetWriter(__snake_case , __snake_case , compression=__snake_case ).write() with fsspec.open(__snake_case , '''rb''' , compression='''infer''' ) as f: UpperCAmelCase : str = f.read() with fsspec.open(__snake_case , '''rb''' , compression='''infer''' ) as f: UpperCAmelCase : Optional[int] = f.read() assert exported_content == original_content
23
1
'''simple docstring''' import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] ) -> Optional[int]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[int] ) -> Dict: UpperCAmelCase : Dict = tmp_path / '''cache''' UpperCAmelCase : List[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase : str = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict , _lowerCAmelCase : int ) -> Optional[int]: UpperCAmelCase : Any = tmp_path / '''cache''' UpperCAmelCase : Tuple = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Any = features.copy() if features else default_expected_features UpperCAmelCase : List[Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Dict = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Tuple ) -> Tuple: UpperCAmelCase : Optional[Any] = tmp_path / '''cache''' UpperCAmelCase : Optional[int] = {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''} UpperCAmelCase : int = features.copy() if features else default_expected_features UpperCAmelCase : Any = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict ) -> Union[str, Any]: # jsonl_312_path features are {"col_3": "float64", "col_1": "string", "col_2": "int64"} UpperCAmelCase : Tuple = {'''col_2''': '''int64''', '''col_3''': '''float64''', '''col_1''': '''string'''} UpperCAmelCase : List[str] = features.copy() UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = tmp_path / '''cache''' UpperCAmelCase : List[str] = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ) -> Optional[Any]: UpperCAmelCase : Any = tmp_path / '''cache''' UpperCAmelCase : List[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : List[Any] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , split=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('''path_type''' , [str, list] ) def snake_case_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> Dict: if issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase : str = jsonl_path elif issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase : Dict = [jsonl_path] UpperCAmelCase : int = tmp_path / '''cache''' UpperCAmelCase : Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[int] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : Dict=("train",) ) -> Union[str, Any]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) for split in splits: UpperCAmelCase : List[str] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ) -> Any: UpperCAmelCase : Optional[Any] = tmp_path / '''cache''' UpperCAmelCase : List[str] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase : Optional[int] = JsonDatasetReader({'''train''': jsonl_path} , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] ) -> int: UpperCAmelCase : Dict = tmp_path / '''cache''' UpperCAmelCase : Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[int] = features.copy() if features else default_expected_features UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = JsonDatasetReader({'''train''': jsonl_path} , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict ) -> Union[str, Any]: if split: UpperCAmelCase : Optional[int] = {split: jsonl_path} else: UpperCAmelCase : Any = '''train''' UpperCAmelCase : Any = {'''train''': jsonl_path, '''test''': jsonl_path} UpperCAmelCase : Tuple = tmp_path / '''cache''' UpperCAmelCase : int = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[Any] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def snake_case_ ( _lowerCAmelCase : List[str] ) -> str: return json.load(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Dict ) -> str: return [json.loads(_lowerCAmelCase ) for line in buffer] class SCREAMING_SNAKE_CASE: """simple docstring""" @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def A ( self : Union[str, Any] , __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : Optional[int] ) -> Dict: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case ).write() buffer.seek(0 ) UpperCAmelCase : Union[str, Any] = load_json_function(__snake_case ) assert isinstance(__snake_case , __snake_case ) assert isinstance(exported_content[0] , __snake_case ) assert len(__snake_case ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def A ( self : Optional[int] , __snake_case : Optional[Any] , __snake_case : str , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : Optional[Any] ) -> List[Any]: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , orient=__snake_case ).write() buffer.seek(0 ) UpperCAmelCase : Union[str, Any] = load_json(__snake_case ) assert isinstance(__snake_case , __snake_case ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(__snake_case , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(__snake_case ) == 10 @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def A ( self : str , __snake_case : str , __snake_case : str , __snake_case : int ) -> Any: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase : Any = load_json_function(__snake_case ) assert isinstance(__snake_case , __snake_case ) assert isinstance(exported_content[0] , __snake_case ) assert len(__snake_case ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def A ( self : Any , __snake_case : int , __snake_case : Union[str, Any] , __snake_case : Optional[int] , __snake_case : Optional[Any] , __snake_case : List[str] ) -> Any: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , orient=__snake_case , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase : List[str] = load_json(__snake_case ) assert isinstance(__snake_case , __snake_case ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(__snake_case , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(__snake_case ) == 10 def A ( self : List[Any] , __snake_case : str ) -> Dict: with pytest.raises(__snake_case ): with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , num_proc=0 ) @pytest.mark.parametrize('''compression, extension''' , [('''gzip''', '''gz'''), ('''bz2''', '''bz2'''), ('''xz''', '''xz''')] ) def A ( self : Optional[int] , __snake_case : Any , __snake_case : str , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Dict ) -> Union[str, Any]: UpperCAmelCase : List[str] = tmp_path_factory.mktemp('''data''' ) / F"""test.json.{extension}""" UpperCAmelCase : List[Any] = str(shared_datadir / F"""test_file.json.{extension}""" ) JsonDatasetWriter(__snake_case , __snake_case , compression=__snake_case ).write() with fsspec.open(__snake_case , '''rb''' , compression='''infer''' ) as f: UpperCAmelCase : str = f.read() with fsspec.open(__snake_case , '''rb''' , compression='''infer''' ) as f: UpperCAmelCase : Optional[int] = f.read() assert exported_content == original_content
23
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCamelCase__: Tuple = logging.get_logger(__name__) UpperCamelCase__: Optional[int] = {"vocab_file": "sentencepiece.bpe.model"} UpperCamelCase__: Optional[int] = { "vocab_file": { "moussaKam/mbarthez": "https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez": "https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez-orangesum-title": ( "https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model" ), }, } UpperCamelCase__: Dict = { "moussaKam/mbarthez": 1024, "moussaKam/barthez": 1024, "moussaKam/barthez-orangesum-title": 1024, } UpperCamelCase__: Tuple = "▁" class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = ["""input_ids""", """attention_mask"""] def __init__( self : List[Any] , __snake_case : List[Any] , __snake_case : Tuple="<s>" , __snake_case : List[Any]="</s>" , __snake_case : int="</s>" , __snake_case : Any="<s>" , __snake_case : Optional[int]="<unk>" , __snake_case : Union[str, Any]="<pad>" , __snake_case : Union[str, Any]="<mask>" , __snake_case : Optional[Dict[str, Any]] = None , **__snake_case : Dict , ) -> None: # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase : int = AddedToken(__snake_case , lstrip=__snake_case , rstrip=__snake_case ) if isinstance(__snake_case , __snake_case ) else mask_token UpperCAmelCase : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__snake_case , eos_token=__snake_case , unk_token=__snake_case , sep_token=__snake_case , cls_token=__snake_case , pad_token=__snake_case , mask_token=__snake_case , sp_model_kwargs=self.sp_model_kwargs , **__snake_case , ) UpperCAmelCase : Optional[int] = vocab_file UpperCAmelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__snake_case ) ) UpperCAmelCase : int = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} UpperCAmelCase : Optional[Any] = len(self.sp_model ) - 1 UpperCAmelCase : Optional[int] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def A ( self : Tuple , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase : Union[str, Any] = [self.cls_token_id] UpperCAmelCase : Optional[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A ( self : int , __snake_case : List[int] , __snake_case : Optional[List[int]] = None , __snake_case : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__snake_case , token_ids_a=__snake_case , already_has_special_tokens=__snake_case ) if token_ids_a is None: return [1] + ([0] * len(__snake_case )) + [1] return [1] + ([0] * len(__snake_case )) + [1, 1] + ([0] * len(__snake_case )) + [1] def A ( self : Optional[int] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: UpperCAmelCase : Tuple = [self.sep_token_id] UpperCAmelCase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def A ( self : Dict ) -> Optional[int]: return len(self.sp_model ) def A ( self : List[str] ) -> Dict: UpperCAmelCase : Optional[Any] = {self.convert_ids_to_tokens(__snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def A ( self : Optional[Any] , __snake_case : str ) -> List[str]: return self.sp_model.encode(__snake_case , out_type=__snake_case ) def A ( self : int , __snake_case : int ) -> int: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCAmelCase : Optional[Any] = self.sp_model.PieceToId(__snake_case ) return spm_id if spm_id else self.unk_token_id def A ( self : int , __snake_case : Any ) -> List[Any]: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(__snake_case ) def A ( self : List[Any] , __snake_case : Union[str, Any] ) -> List[str]: UpperCAmelCase : Optional[Any] = [] UpperCAmelCase : int = '''''' UpperCAmelCase : Union[str, Any] = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__snake_case ) + token UpperCAmelCase : str = True UpperCAmelCase : List[str] = [] else: current_sub_tokens.append(__snake_case ) UpperCAmelCase : Optional[int] = False out_string += self.sp_model.decode(__snake_case ) return out_string.strip() def __getstate__( self : Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase : Optional[Any] = self.__dict__.copy() UpperCAmelCase : Any = None return state def __setstate__( self : Optional[int] , __snake_case : Union[str, Any] ) -> List[Any]: UpperCAmelCase : Dict = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): UpperCAmelCase : Optional[Any] = {} UpperCAmelCase : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def A ( self : Optional[int] , __snake_case : str , __snake_case : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__snake_case ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase : Union[str, Any] = os.path.join( __snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __snake_case ) elif not os.path.isfile(self.vocab_file ): with open(__snake_case , '''wb''' ) as fi: UpperCAmelCase : Any = self.sp_model.serialized_model_proto() fi.write(__snake_case ) return (out_vocab_file,)
23
1
'''simple docstring''' import torch from transformers import CamembertForMaskedLM, CamembertTokenizer def snake_case_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : str=5 ) -> Any: # Adapted from https://github.com/pytorch/fairseq/blob/master/fairseq/models/roberta/hub_interface.py assert masked_input.count('''<mask>''' ) == 1 UpperCAmelCase : Dict = torch.tensor(tokenizer.encode(_lowerCAmelCase , add_special_tokens=_lowerCAmelCase ) ).unsqueeze(0 ) # Batch size 1 UpperCAmelCase : List[str] = model(_lowerCAmelCase )[0] # The last hidden-state is the first element of the output tuple UpperCAmelCase : Tuple = (input_ids.squeeze() == tokenizer.mask_token_id).nonzero().item() UpperCAmelCase : int = logits[0, masked_index, :] UpperCAmelCase : str = logits.softmax(dim=0 ) UpperCAmelCase , UpperCAmelCase : str = prob.topk(k=_lowerCAmelCase , dim=0 ) UpperCAmelCase : str = ''' '''.join( [tokenizer.convert_ids_to_tokens(indices[i].item() ) for i in range(len(_lowerCAmelCase ) )] ) UpperCAmelCase : Any = tokenizer.mask_token UpperCAmelCase : Optional[int] = [] for index, predicted_token_bpe in enumerate(topk_predicted_token_bpe.split(''' ''' ) ): UpperCAmelCase : int = predicted_token_bpe.replace('''\u2581''' , ''' ''' ) if " {0}".format(_lowerCAmelCase ) in masked_input: topk_filled_outputs.append( ( masked_input.replace(''' {0}'''.format(_lowerCAmelCase ) , _lowerCAmelCase ), values[index].item(), predicted_token, ) ) else: topk_filled_outputs.append( ( masked_input.replace(_lowerCAmelCase , _lowerCAmelCase ), values[index].item(), predicted_token, ) ) return topk_filled_outputs UpperCamelCase__: Optional[int] = CamembertTokenizer.from_pretrained("camembert-base") UpperCamelCase__: List[Any] = CamembertForMaskedLM.from_pretrained("camembert-base") model.eval() UpperCamelCase__: int = "Le camembert est <mask> :)" print(fill_mask(masked_input, model, tokenizer, topk=3))
23
'''simple docstring''' import tempfile import unittest import numpy as np from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import BertConfig, is_flax_available from transformers.testing_utils import TOKEN, USER, is_staging_test, require_flax if is_flax_available(): import os from flax.core.frozen_dict import unfreeze from flax.traverse_util import flatten_dict from transformers import FlaxBertModel UpperCamelCase__: Tuple = "0.12" # assumed parallelism: 8 @require_flax @is_staging_test class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" @classmethod def A ( cls : Union[str, Any] ) -> int: UpperCAmelCase : Optional[Any] = TOKEN HfFolder.save_token(__snake_case ) @classmethod def A ( cls : List[str] ) -> Tuple: try: delete_repo(token=cls._token , repo_id='''test-model-flax''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-model-flax-org''' ) except HTTPError: pass def A ( self : int ) -> Tuple: UpperCAmelCase : List[Any] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) model.push_to_hub('''test-model-flax''' , use_auth_token=self._token ) UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) UpperCAmelCase : List[Any] = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Tuple = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Union[str, Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''test-model-flax''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(__snake_case , repo_id='''test-model-flax''' , push_to_hub=__snake_case , use_auth_token=self._token ) UpperCAmelCase : str = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) UpperCAmelCase : Any = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : str = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Optional[Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) def A ( self : Optional[Any] ) -> Union[str, Any]: UpperCAmelCase : Dict = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Optional[Any] = FlaxBertModel(__snake_case ) model.push_to_hub('''valid_org/test-model-flax-org''' , use_auth_token=self._token ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) UpperCAmelCase : List[Any] = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : int = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Any = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-model-flax-org''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained( __snake_case , repo_id='''valid_org/test-model-flax-org''' , push_to_hub=__snake_case , use_auth_token=self._token ) UpperCAmelCase : str = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) UpperCAmelCase : Any = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Optional[Any] = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : int = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Union[str, Any]: UpperCAmelCase : str = True UpperCAmelCase : int = flatten_dict(modela.params ) UpperCAmelCase : Dict = flatten_dict(modela.params ) for key in flat_params_a.keys(): if np.sum(np.abs(flat_params_a[key] - flat_params_a[key] ) ) > 1e-4: UpperCAmelCase : Dict = False return models_are_equal @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : Tuple ) -> Union[str, Any]: UpperCAmelCase : List[Any] = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) UpperCAmelCase : int = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(__snake_case , __snake_case ) ) with self.assertRaises(__snake_case ): UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : str = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertTrue(check_models_equal(__snake_case , __snake_case ) ) def A ( self : List[str] ) -> Dict: UpperCAmelCase : Dict = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) UpperCAmelCase : Optional[int] = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(__snake_case , __snake_case ) , max_shard_size='''10KB''' ) with self.assertRaises(__snake_case ): UpperCAmelCase : Any = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertTrue(check_models_equal(__snake_case , __snake_case ) ) def A ( self : Optional[int] ) -> str: UpperCAmelCase : Dict = '''bert''' UpperCAmelCase : int = '''hf-internal-testing/tiny-random-bert-subfolder''' with self.assertRaises(__snake_case ): UpperCAmelCase : Optional[Any] = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertIsNotNone(__snake_case ) def A ( self : Dict ) -> List[Any]: UpperCAmelCase : Optional[int] = '''bert''' UpperCAmelCase : int = '''hf-internal-testing/tiny-random-bert-sharded-subfolder''' with self.assertRaises(__snake_case ): UpperCAmelCase : Dict = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertIsNotNone(__snake_case )
23
1
'''simple docstring''' from __future__ import annotations def snake_case_ ( _lowerCAmelCase : float , _lowerCAmelCase : float , _lowerCAmelCase : float ) -> dict[str, float]: if (voltage, current, resistance).count(0 ) != 1: raise ValueError('''One and only one argument must be 0''' ) if resistance < 0: raise ValueError('''Resistance cannot be negative''' ) if voltage == 0: return {"voltage": float(current * resistance )} elif current == 0: return {"current": voltage / resistance} elif resistance == 0: return {"resistance": voltage / current} else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
23
'''simple docstring''' import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE: """simple docstring""" def __init__( self : Optional[int] , __snake_case : str , __snake_case : Union[str, Any]=2 , __snake_case : Optional[int]=8 , __snake_case : Any=True , __snake_case : Union[str, Any]=True , __snake_case : Dict=True , __snake_case : int=True , __snake_case : List[Any]=99 , __snake_case : str=16 , __snake_case : Tuple=5 , __snake_case : Tuple=2 , __snake_case : str=36 , __snake_case : Dict="gelu" , __snake_case : str=0.0 , __snake_case : Optional[int]=0.0 , __snake_case : Optional[int]=512 , __snake_case : Optional[Any]=16 , __snake_case : int=2 , __snake_case : int=0.02 , __snake_case : str=3 , __snake_case : Dict=4 , __snake_case : str=None , ) -> Optional[int]: UpperCAmelCase : Optional[int] = parent UpperCAmelCase : Tuple = batch_size UpperCAmelCase : List[str] = seq_length UpperCAmelCase : List[Any] = is_training UpperCAmelCase : int = use_input_mask UpperCAmelCase : Any = use_token_type_ids UpperCAmelCase : str = use_labels UpperCAmelCase : Union[str, Any] = vocab_size UpperCAmelCase : List[str] = hidden_size UpperCAmelCase : Optional[Any] = num_hidden_layers UpperCAmelCase : Union[str, Any] = num_attention_heads UpperCAmelCase : Optional[Any] = intermediate_size UpperCAmelCase : Union[str, Any] = hidden_act UpperCAmelCase : int = hidden_dropout_prob UpperCAmelCase : Optional[int] = attention_probs_dropout_prob UpperCAmelCase : Union[str, Any] = max_position_embeddings UpperCAmelCase : str = type_vocab_size UpperCAmelCase : List[str] = type_sequence_label_size UpperCAmelCase : Tuple = initializer_range UpperCAmelCase : Optional[Any] = num_labels UpperCAmelCase : Optional[int] = num_choices UpperCAmelCase : Any = scope def A ( self : Tuple ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase : Optional[int] = None if self.use_input_mask: UpperCAmelCase : Dict = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase : Dict = None if self.use_token_type_ids: UpperCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase : str = None UpperCAmelCase : Tuple = None UpperCAmelCase : int = None if self.use_labels: UpperCAmelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase : str = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase : List[str] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : int ) -> Tuple: return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__snake_case , initializer_range=self.initializer_range , ) def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Optional[Any] = self.get_config() UpperCAmelCase : int = 300 return config def A ( self : Optional[Any] ) -> Any: ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) : Tuple = self.prepare_config_and_inputs() UpperCAmelCase : Dict = True UpperCAmelCase : Tuple = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def A ( self : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : int , __snake_case : Optional[int] , __snake_case : int , __snake_case : Dict , __snake_case : Tuple , __snake_case : Optional[Any] ) -> List[str]: UpperCAmelCase : int = MraModel(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Tuple = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case ) UpperCAmelCase : Optional[int] = model(__snake_case , token_type_ids=__snake_case ) UpperCAmelCase : Dict = model(__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Tuple , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : List[str] , __snake_case : int , __snake_case : Union[str, Any] , __snake_case : Optional[Any] , __snake_case : Any , __snake_case : List[Any] , __snake_case : Optional[Any] , ) -> Tuple: UpperCAmelCase : str = True UpperCAmelCase : Tuple = MraModel(__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Optional[int] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , encoder_hidden_states=__snake_case , encoder_attention_mask=__snake_case , ) UpperCAmelCase : Optional[Any] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , encoder_hidden_states=__snake_case , ) UpperCAmelCase : str = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Tuple , __snake_case : str , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : int ) -> Any: UpperCAmelCase : Dict = MraForMaskedLM(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Optional[int] = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : Tuple , __snake_case : Tuple , __snake_case : Dict , __snake_case : Dict , __snake_case : Any , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Tuple ) -> Optional[int]: UpperCAmelCase : List[str] = MraForQuestionAnswering(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : List[Any] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , start_positions=__snake_case , end_positions=__snake_case , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : str , __snake_case : Optional[int] , __snake_case : List[str] , __snake_case : str , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Union[str, Any] , __snake_case : List[Any] ) -> int: UpperCAmelCase : int = self.num_labels UpperCAmelCase : Union[str, Any] = MraForSequenceClassification(__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : List[str] = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : str , __snake_case : Dict , __snake_case : Optional[Any] , __snake_case : Dict , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Dict ) -> int: UpperCAmelCase : Tuple = self.num_labels UpperCAmelCase : List[str] = MraForTokenClassification(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : str = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : str , __snake_case : int , __snake_case : Any , __snake_case : Tuple , __snake_case : Optional[int] , __snake_case : List[str] , __snake_case : str , __snake_case : Union[str, Any] ) -> Optional[Any]: UpperCAmelCase : Tuple = self.num_choices UpperCAmelCase : int = MraForMultipleChoice(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : str = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[str] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[str] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : str ) -> Dict: UpperCAmelCase : Any = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) : List[str] = config_and_inputs UpperCAmelCase : Any = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = () def A ( self : int ) -> Union[str, Any]: UpperCAmelCase : List[str] = MraModelTester(self ) UpperCAmelCase : Optional[int] = ConfigTester(self , config_class=__snake_case , hidden_size=37 ) def A ( self : Optional[Any] ) -> str: self.config_tester.run_common_tests() def A ( self : Tuple ) -> Optional[Any]: UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__snake_case ) def A ( self : List[Any] ) -> Optional[Any]: UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase : List[Any] = type self.model_tester.create_and_check_model(*__snake_case ) def A ( self : Tuple ) -> Dict: UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__snake_case ) def A ( self : Tuple ) -> List[str]: UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__snake_case ) def A ( self : int ) -> Dict: UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__snake_case ) def A ( self : Dict ) -> Optional[int]: UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__snake_case ) def A ( self : Any ) -> Optional[int]: UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__snake_case ) @slow def A ( self : Dict ) -> Any: for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase : str = MraModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) @unittest.skip(reason='''MRA does not output attentions''' ) def A ( self : str ) -> Optional[Any]: return @require_torch class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" @slow def A ( self : Tuple ) -> List[Any]: UpperCAmelCase : int = MraModel.from_pretrained('''uw-madison/mra-base-512-4''' ) UpperCAmelCase : Optional[Any] = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : List[Any] = model(__snake_case )[0] UpperCAmelCase : Optional[Any] = torch.Size((1, 256, 768) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Any = torch.tensor( [[[-0.01_40, 0.08_30, -0.03_81], [0.15_46, 0.14_02, 0.02_20], [0.11_62, 0.08_51, 0.01_65]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) ) @slow def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Optional[int] = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-512-4''' ) UpperCAmelCase : Dict = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : List[Any] = model(__snake_case )[0] UpperCAmelCase : int = 50265 UpperCAmelCase : int = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Union[str, Any] = torch.tensor( [[[9.25_95, -3.60_38, 11.88_19], [9.38_69, -3.26_93, 11.09_56], [11.85_24, -3.49_38, 13.12_10]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) ) @slow def A ( self : str ) -> List[Any]: UpperCAmelCase : List[Any] = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-4096-8-d3''' ) UpperCAmelCase : List[Any] = torch.arange(4096 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : Tuple = model(__snake_case )[0] UpperCAmelCase : Optional[int] = 50265 UpperCAmelCase : Tuple = torch.Size((1, 4096, vocab_size) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Optional[int] = torch.tensor( [[[5.47_89, -2.35_64, 7.50_64], [7.90_67, -1.33_69, 9.96_68], [9.07_12, -1.81_06, 7.03_80]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) )
23
1
'''simple docstring''' import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast UpperCamelCase__: int = datasets.utils.logging.get_logger(__name__) @dataclass class SCREAMING_SNAKE_CASE( datasets.BuilderConfig ): """simple docstring""" lowerCamelCase__ = 10_000 lowerCamelCase__ = None lowerCamelCase__ = None class SCREAMING_SNAKE_CASE( datasets.ArrowBasedBuilder ): """simple docstring""" lowerCamelCase__ = ParquetConfig def A ( self : Union[str, Any] ) -> str: return datasets.DatasetInfo(features=self.config.features ) def A ( self : str , __snake_case : Dict ) -> Optional[int]: if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) UpperCAmelCase : Dict = dl_manager.download_and_extract(self.config.data_files ) if isinstance(__snake_case , (str, list, tuple) ): UpperCAmelCase : List[str] = data_files if isinstance(__snake_case , __snake_case ): UpperCAmelCase : Optional[int] = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive UpperCAmelCase : int = [dl_manager.iter_files(__snake_case ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'''files''': files} )] UpperCAmelCase : Union[str, Any] = [] for split_name, files in data_files.items(): if isinstance(__snake_case , __snake_case ): UpperCAmelCase : List[Any] = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive UpperCAmelCase : List[Any] = [dl_manager.iter_files(__snake_case ) for file in files] # Infer features is they are stoed in the arrow schema if self.info.features is None: for file in itertools.chain.from_iterable(__snake_case ): with open(__snake_case , '''rb''' ) as f: UpperCAmelCase : Optional[Any] = datasets.Features.from_arrow_schema(pq.read_schema(__snake_case ) ) break splits.append(datasets.SplitGenerator(name=__snake_case , gen_kwargs={'''files''': files} ) ) return splits def A ( self : List[str] , __snake_case : pa.Table ) -> pa.Table: if self.info.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example UpperCAmelCase : Any = table_cast(__snake_case , self.info.features.arrow_schema ) return pa_table def A ( self : List[Any] , __snake_case : Tuple ) -> Dict: UpperCAmelCase : List[Any] = self.info.features.arrow_schema if self.info.features is not None else None if self.info.features is not None and self.config.columns is not None: if sorted(field.name for field in schema ) != sorted(self.config.columns ): raise ValueError( F"""Tried to load parquet data with columns '{self.config.columns}' with mismatching features '{self.info.features}'""" ) for file_idx, file in enumerate(itertools.chain.from_iterable(__snake_case ) ): with open(__snake_case , '''rb''' ) as f: UpperCAmelCase : Union[str, Any] = pq.ParquetFile(__snake_case ) try: for batch_idx, record_batch in enumerate( parquet_file.iter_batches(batch_size=self.config.batch_size , columns=self.config.columns ) ): UpperCAmelCase : List[Any] = pa.Table.from_batches([record_batch] ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield F"""{file_idx}_{batch_idx}""", self._cast_table(__snake_case ) except ValueError as e: logger.error(F"""Failed to read file '{file}' with error {type(__snake_case )}: {e}""" ) raise
23
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : Any ) -> str: UpperCAmelCase : Any = { '''task_specific_params''': { '''summarization''': {'''length_penalty''': 1.0, '''max_length''': 128, '''min_length''': 12, '''num_beams''': 4}, '''summarization_cnn''': {'''length_penalty''': 2.0, '''max_length''': 142, '''min_length''': 56, '''num_beams''': 4}, '''summarization_xsum''': {'''length_penalty''': 1.0, '''max_length''': 62, '''min_length''': 11, '''num_beams''': 6}, } } UpperCAmelCase : int = { '''task_specific_params.summarization.length_penalty''': 1.0, '''task_specific_params.summarization.max_length''': 128, '''task_specific_params.summarization.min_length''': 12, '''task_specific_params.summarization.num_beams''': 4, '''task_specific_params.summarization_cnn.length_penalty''': 2.0, '''task_specific_params.summarization_cnn.max_length''': 142, '''task_specific_params.summarization_cnn.min_length''': 56, '''task_specific_params.summarization_cnn.num_beams''': 4, '''task_specific_params.summarization_xsum.length_penalty''': 1.0, '''task_specific_params.summarization_xsum.max_length''': 62, '''task_specific_params.summarization_xsum.min_length''': 11, '''task_specific_params.summarization_xsum.num_beams''': 6, } self.assertEqual(flatten_dict(__snake_case ) , __snake_case ) def A ( self : int ) -> str: UpperCAmelCase : Any = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(__snake_case ) , x.transpose() ) ) UpperCAmelCase : str = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def A ( self : str ) -> Union[str, Any]: UpperCAmelCase : Any = np.random.randn(3 , 4 ) UpperCAmelCase : List[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , transpose(__snake_case ).numpy() ) ) UpperCAmelCase : Tuple = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Any = torch.tensor(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , transpose(__snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_tf def A ( self : List[str] ) -> Optional[Any]: UpperCAmelCase : int = np.random.randn(3 , 4 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , transpose(__snake_case ).numpy() ) ) UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : str = tf.constant(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , transpose(__snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_flax def A ( self : Tuple ) -> Any: UpperCAmelCase : List[Any] = np.random.randn(3 , 4 ) UpperCAmelCase : List[str] = jnp.array(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , np.asarray(transpose(__snake_case ) ) ) ) UpperCAmelCase : Dict = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , np.asarray(transpose(__snake_case , axes=(1, 2, 0) ) ) ) ) def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , np.reshape(__snake_case , (4, 3) ) ) ) UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , np.reshape(__snake_case , (12, 5) ) ) ) @require_torch def A ( self : Union[str, Any] ) -> int: UpperCAmelCase : Dict = np.random.randn(3 , 4 ) UpperCAmelCase : Optional[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , reshape(__snake_case , (4, 3) ).numpy() ) ) UpperCAmelCase : List[str] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : List[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , reshape(__snake_case , (12, 5) ).numpy() ) ) @require_tf def A ( self : int ) -> List[str]: UpperCAmelCase : List[Any] = np.random.randn(3 , 4 ) UpperCAmelCase : List[str] = tf.constant(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , reshape(__snake_case , (4, 3) ).numpy() ) ) UpperCAmelCase : List[Any] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Optional[Any] = tf.constant(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , reshape(__snake_case , (12, 5) ).numpy() ) ) @require_flax def A ( self : Any ) -> Dict: UpperCAmelCase : Tuple = np.random.randn(3 , 4 ) UpperCAmelCase : Union[str, Any] = jnp.array(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , np.asarray(reshape(__snake_case , (4, 3) ) ) ) ) UpperCAmelCase : Any = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Optional[Any] = jnp.array(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , np.asarray(reshape(__snake_case , (12, 5) ) ) ) ) def A ( self : List[Any] ) -> List[Any]: UpperCAmelCase : Union[str, Any] = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(__snake_case ) , np.squeeze(__snake_case ) ) ) UpperCAmelCase : str = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , np.squeeze(__snake_case , axis=2 ) ) ) @require_torch def A ( self : Optional[int] ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : List[str] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , squeeze(__snake_case ).numpy() ) ) UpperCAmelCase : Any = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : str = torch.tensor(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , squeeze(__snake_case , axis=2 ).numpy() ) ) @require_tf def A ( self : Optional[Any] ) -> Dict: UpperCAmelCase : int = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , squeeze(__snake_case ).numpy() ) ) UpperCAmelCase : List[str] = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , squeeze(__snake_case , axis=2 ).numpy() ) ) @require_flax def A ( self : List[Any] ) -> Dict: UpperCAmelCase : Optional[int] = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , np.asarray(squeeze(__snake_case ) ) ) ) UpperCAmelCase : str = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , np.asarray(squeeze(__snake_case , axis=2 ) ) ) ) def A ( self : Optional[Any] ) -> int: UpperCAmelCase : Optional[Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , np.expand_dims(__snake_case , axis=1 ) ) ) @require_torch def A ( self : List[str] ) -> Tuple: UpperCAmelCase : Tuple = np.random.randn(3 , 4 ) UpperCAmelCase : Tuple = torch.tensor(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , expand_dims(__snake_case , axis=1 ).numpy() ) ) @require_tf def A ( self : List[str] ) -> Union[str, Any]: UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 ) UpperCAmelCase : Any = tf.constant(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , expand_dims(__snake_case , axis=1 ).numpy() ) ) @require_flax def A ( self : Any ) -> List[Any]: UpperCAmelCase : List[str] = np.random.randn(3 , 4 ) UpperCAmelCase : str = jnp.array(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , np.asarray(expand_dims(__snake_case , axis=1 ) ) ) )
23
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase__: Any = { "configuration_bert": ["BERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "BertConfig", "BertOnnxConfig"], "tokenization_bert": ["BasicTokenizer", "BertTokenizer", "WordpieceTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: List[str] = ["BertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: Optional[Any] = [ "BERT_PRETRAINED_MODEL_ARCHIVE_LIST", "BertForMaskedLM", "BertForMultipleChoice", "BertForNextSentencePrediction", "BertForPreTraining", "BertForQuestionAnswering", "BertForSequenceClassification", "BertForTokenClassification", "BertLayer", "BertLMHeadModel", "BertModel", "BertPreTrainedModel", "load_tf_weights_in_bert", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: List[str] = [ "TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFBertEmbeddings", "TFBertForMaskedLM", "TFBertForMultipleChoice", "TFBertForNextSentencePrediction", "TFBertForPreTraining", "TFBertForQuestionAnswering", "TFBertForSequenceClassification", "TFBertForTokenClassification", "TFBertLMHeadModel", "TFBertMainLayer", "TFBertModel", "TFBertPreTrainedModel", ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: Any = ["TFBertTokenizer"] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: Any = [ "FlaxBertForCausalLM", "FlaxBertForMaskedLM", "FlaxBertForMultipleChoice", "FlaxBertForNextSentencePrediction", "FlaxBertForPreTraining", "FlaxBertForQuestionAnswering", "FlaxBertForSequenceClassification", "FlaxBertForTokenClassification", "FlaxBertModel", "FlaxBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys UpperCamelCase__: Any = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
'''simple docstring''' import argparse import os import re import packaging.version UpperCamelCase__: Union[str, Any] = "examples/" UpperCamelCase__: Optional[Any] = { "examples": (re.compile(r"^check_min_version\(\"[^\"]+\"\)\s*$", re.MULTILINE), "check_min_version(\"VERSION\")\n"), "init": (re.compile(r"^__version__\s+=\s+\"([^\"]+)\"\s*$", re.MULTILINE), "__version__ = \"VERSION\"\n"), "setup": (re.compile(r"^(\s*)version\s*=\s*\"[^\"]+\",", re.MULTILINE), r"\1version=\"VERSION\","), "doc": (re.compile(r"^(\s*)release\s*=\s*\"[^\"]+\"$", re.MULTILINE), "release = \"VERSION\"\n"), } UpperCamelCase__: Optional[int] = { "init": "src/diffusers/__init__.py", "setup": "setup.py", } UpperCamelCase__: List[Any] = "README.md" def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int ) -> Optional[int]: with open(_lowerCAmelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: UpperCAmelCase : Optional[int] = f.read() UpperCAmelCase , UpperCAmelCase : List[Any] = REPLACE_PATTERNS[pattern] UpperCAmelCase : List[Any] = replace.replace('''VERSION''' , _lowerCAmelCase ) UpperCAmelCase : Optional[Any] = re_pattern.sub(_lowerCAmelCase , _lowerCAmelCase ) with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Any ) -> Optional[int]: for folder, directories, fnames in os.walk(_lowerCAmelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase , pattern='''examples''' ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : str=False ) -> List[str]: for pattern, fname in REPLACE_FILES.items(): update_version_in_file(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not patch: update_version_in_examples(_lowerCAmelCase ) def snake_case_ ( ) -> Optional[Any]: UpperCAmelCase : Optional[int] = '''🤗 Transformers currently provides the following architectures''' UpperCAmelCase : Optional[int] = '''1. Want to contribute a new model?''' with open(_lowerCAmelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: UpperCAmelCase : Optional[Any] = f.readlines() # Find the start of the list. UpperCAmelCase : List[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 UpperCAmelCase : Optional[Any] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): UpperCAmelCase : Optional[int] = lines[index].replace( '''https://huggingface.co/docs/diffusers/main/model_doc''' , '''https://huggingface.co/docs/diffusers/model_doc''' , ) index += 1 with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(_lowerCAmelCase ) def snake_case_ ( ) -> Optional[Any]: with open(REPLACE_FILES['''init'''] , '''r''' ) as f: UpperCAmelCase : Union[str, Any] = f.read() UpperCAmelCase : int = REPLACE_PATTERNS['''init'''][0].search(_lowerCAmelCase ).groups()[0] return packaging.version.parse(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : List[str]=False ) -> Any: UpperCAmelCase : Optional[Any] = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: UpperCAmelCase : Optional[int] = default_version.base_version elif patch: UpperCAmelCase : Union[str, Any] = f"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: UpperCAmelCase : Union[str, Any] = f"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. UpperCAmelCase : Dict = input(f"""Which version are you releasing? [{default_version}]""" ) if len(_lowerCAmelCase ) == 0: UpperCAmelCase : Tuple = default_version print(f"""Updating version to {version}.""" ) global_version_update(_lowerCAmelCase , patch=_lowerCAmelCase ) def snake_case_ ( ) -> Any: UpperCAmelCase : List[Any] = get_version() UpperCAmelCase : List[str] = f"""{current_version.major}.{current_version.minor + 1}.0.dev0""" UpperCAmelCase : List[Any] = current_version.base_version # Check with the user we got that right. UpperCAmelCase : Optional[int] = input(f"""Which version are we developing now? [{dev_version}]""" ) if len(_lowerCAmelCase ) == 0: UpperCAmelCase : Dict = dev_version print(f"""Updating version to {version}.""" ) global_version_update(_lowerCAmelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": UpperCamelCase__: Union[str, Any] = argparse.ArgumentParser() parser.add_argument("--post_release", action="store_true", help="Whether this is pre or post release.") parser.add_argument("--patch", action="store_true", help="Whether or not this is a patch release.") UpperCamelCase__: Optional[Any] = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("Nothing to do after a patch :-)") else: post_release_work()
23
1
'''simple docstring''' from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : str , _lowerCAmelCase : Optional[str] = None ) -> str: if version.parse(hfh.__version__ ).release < version.parse('''0.11.0''' ).release: # old versions of hfh don't url-encode the file path UpperCAmelCase : Tuple = quote(_lowerCAmelCase ) return hfh.hf_hub_url(_lowerCAmelCase , _lowerCAmelCase , repo_type='''dataset''' , revision=_lowerCAmelCase )
23
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake UpperCamelCase__: Tuple = numpy.array([0, 0]) UpperCamelCase__: Union[str, Any] = numpy.array([0.5, 0.8660254]) UpperCamelCase__: Dict = numpy.array([1, 0]) UpperCamelCase__: int = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] , _lowerCAmelCase : int ) -> list[numpy.ndarray]: UpperCAmelCase : Union[str, Any] = initial_vectors for _ in range(_lowerCAmelCase ): UpperCAmelCase : Union[str, Any] = iteration_step(_lowerCAmelCase ) return vectors def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] ) -> list[numpy.ndarray]: UpperCAmelCase : Tuple = [] for i, start_vector in enumerate(vectors[:-1] ): UpperCAmelCase : List[str] = vectors[i + 1] new_vectors.append(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def snake_case_ ( _lowerCAmelCase : numpy.ndarray , _lowerCAmelCase : float ) -> numpy.ndarray: UpperCAmelCase : List[str] = numpy.radians(_lowerCAmelCase ) UpperCAmelCase , UpperCAmelCase : Tuple = numpy.cos(_lowerCAmelCase ), numpy.sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = numpy.array(((c, -s), (s, c)) ) return numpy.dot(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] ) -> None: UpperCAmelCase : List[Any] = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() UpperCAmelCase , UpperCAmelCase : str = zip(*_lowerCAmelCase ) plt.plot(_lowerCAmelCase , _lowerCAmelCase ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__: List[Any] = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
23
1
'''simple docstring''' import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" lowerCamelCase__ = inspect.getfile(accelerate.test_utils ) lowerCamelCase__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_cli.py"""] ) lowerCamelCase__ = ["""accelerate""", """launch"""] lowerCamelCase__ = Path.home() / """.cache/huggingface/accelerate""" lowerCamelCase__ = """default_config.yaml""" lowerCamelCase__ = config_folder / config_file lowerCamelCase__ = config_folder / """_default_config.yaml""" lowerCamelCase__ = Path("""tests/test_configs""" ) @classmethod def A ( cls : int ) -> Any: if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def A ( cls : Dict ) -> Optional[int]: if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def A ( self : Union[str, Any] ) -> Tuple: UpperCAmelCase : Tuple = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] , env=os.environ.copy() ) def A ( self : Any ) -> Tuple: for config in sorted(self.test_config_path.glob('''**/*.yaml''' ) ): with self.subTest(config_file=__snake_case ): execute_subprocess_async( self.base_cmd + ['''--config_file''', str(__snake_case ), self.test_file_path] , env=os.environ.copy() ) def A ( self : str ) -> Optional[int]: execute_subprocess_async(['''accelerate''', '''test'''] , env=os.environ.copy() ) class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" lowerCamelCase__ = """test-tpu""" lowerCamelCase__ = """us-central1-a""" lowerCamelCase__ = """ls""" lowerCamelCase__ = ["""accelerate""", """tpu-config"""] lowerCamelCase__ = """cd /usr/share""" lowerCamelCase__ = """tests/test_samples/test_command_file.sh""" lowerCamelCase__ = """Running gcloud compute tpus tpu-vm ssh""" def A ( self : List[Any] ) -> Tuple: UpperCAmelCase : int = run_command( self.cmd + ['''--command''', self.command, '''--tpu_zone''', self.tpu_zone, '''--tpu_name''', self.tpu_name, '''--debug'''] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" , __snake_case , ) def A ( self : Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase : Tuple = run_command( self.cmd + [ '''--config_file''', '''tests/test_configs/0_12_0.yaml''', '''--command''', self.command, '''--tpu_zone''', self.tpu_zone, '''--tpu_name''', self.tpu_name, '''--debug''', ] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" , __snake_case , ) def A ( self : str ) -> Optional[Any]: UpperCAmelCase : List[Any] = run_command( self.cmd + ['''--config_file''', '''tests/test_configs/latest.yaml''', '''--debug'''] , return_stdout=__snake_case ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" , __snake_case , ) def A ( self : str ) -> Optional[int]: UpperCAmelCase : Any = run_command( self.cmd + ['''--config_file''', '''tests/test_configs/latest.yaml''', '''--command''', self.command, '''--debug'''] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" , __snake_case , ) def A ( self : Optional[int] ) -> Tuple: UpperCAmelCase : Any = run_command( self.cmd + [ '''--config_file''', '''tests/test_configs/latest.yaml''', '''--command''', self.command, '''--command''', '''echo "Hello World"''', '''--debug''', ] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all""" , __snake_case , ) def A ( self : Union[str, Any] ) -> Tuple: UpperCAmelCase : List[Any] = run_command( self.cmd + ['''--config_file''', '''tests/test_configs/latest.yaml''', '''--command_file''', self.command_file, '''--debug'''] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" , __snake_case , ) def A ( self : Tuple ) -> Any: UpperCAmelCase : int = run_command( self.cmd + [ '''--config_file''', '''tests/test_configs/0_12_0.yaml''', '''--command_file''', self.command_file, '''--tpu_zone''', self.tpu_zone, '''--tpu_name''', self.tpu_name, '''--debug''', ] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" , __snake_case , ) def A ( self : int ) -> Any: UpperCAmelCase : Tuple = run_command( self.cmd + ['''--config_file''', '''tests/test_configs/latest.yaml''', '''--install_accelerate''', '''--debug'''] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all""" , __snake_case , ) def A ( self : Union[str, Any] ) -> str: UpperCAmelCase : Optional[int] = run_command( self.cmd + [ '''--config_file''', '''tests/test_configs/latest.yaml''', '''--install_accelerate''', '''--accelerate_version''', '''12.0.0''', '''--debug''', ] , return_stdout=__snake_case , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all""" , __snake_case , )
23
'''simple docstring''' from manim import * class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" def A ( self : Union[str, Any] ) -> List[str]: UpperCAmelCase : Optional[Any] = Rectangle(height=0.5 , width=0.5 ) UpperCAmelCase : str = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) UpperCAmelCase : Tuple = [mem.copy() for i in range(6 )] UpperCAmelCase : Optional[int] = [mem.copy() for i in range(6 )] UpperCAmelCase : Dict = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Any = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Union[str, Any] = VGroup(__snake_case , __snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Optional[Any] = Text('''CPU''' , font_size=24 ) UpperCAmelCase : Union[str, Any] = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__snake_case ) UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(4 )] UpperCAmelCase : Union[str, Any] = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : List[str] = Text('''GPU''' , font_size=24 ) UpperCAmelCase : Dict = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) gpu.move_to([-1, -1, 0] ) self.add(__snake_case ) UpperCAmelCase : int = [mem.copy() for i in range(6 )] UpperCAmelCase : Union[str, Any] = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : List[str] = Text('''Model''' , font_size=24 ) UpperCAmelCase : Tuple = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) model.move_to([3, -1.0, 0] ) self.add(__snake_case ) UpperCAmelCase : Any = [] for i, rect in enumerate(__snake_case ): rect.set_stroke(__snake_case ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) UpperCAmelCase : Dict = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(__snake_case , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=__snake_case ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=__snake_case , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=__snake_case , buff=0.0 ) self.add(__snake_case ) cpu_targs.append(__snake_case ) UpperCAmelCase : int = [mem.copy() for i in range(6 )] UpperCAmelCase : int = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Any = Text('''Loaded Checkpoint''' , font_size=24 ) UpperCAmelCase : Union[str, Any] = Group(__snake_case , __snake_case ).arrange(__snake_case , aligned_edge=__snake_case , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) UpperCAmelCase : Optional[int] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCAmelCase : str = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(__snake_case , __snake_case ) UpperCAmelCase : Tuple = MarkupText( F"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(__snake_case , DOWN * 2.4 , aligned_edge=key_text.get_left() ) UpperCAmelCase : List[Any] = MarkupText( F"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__snake_case ) , Write(__snake_case ) ) self.play(Write(__snake_case , run_time=1 ) , Create(__snake_case , run_time=1 ) ) UpperCAmelCase : Tuple = [] UpperCAmelCase : int = [] for i, rect in enumerate(__snake_case ): UpperCAmelCase : Any = fill.copy().set_fill(__snake_case , opacity=0.7 ) target.move_to(__snake_case ) first_animations.append(GrowFromCenter(__snake_case , run_time=1 ) ) UpperCAmelCase : List[str] = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(__snake_case , run_time=1.5 ) ) self.play(*__snake_case ) self.play(*__snake_case ) self.wait()
23
1
'''simple docstring''' import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : Optional[Any] ) -> Optional[Any]: UpperCAmelCase : int = torch.nn.Linear(10 , 10 ) UpperCAmelCase : List[Any] = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase : int = Accelerator() UpperCAmelCase : List[str] = accelerator.prepare(__snake_case ) try: pickle.loads(pickle.dumps(__snake_case ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
23
'''simple docstring''' import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( "The converted tokenizer will be the `slow` tokenizer. To use the fast, update your `tokenizers` library and re-run the tokenizer conversion" ) UpperCamelCase__: str = None UpperCamelCase__: int = { "7B": 11008, "13B": 13824, "30B": 17920, "65B": 22016, "70B": 28672, } UpperCamelCase__: List[Any] = { "7B": 1, "7Bf": 1, "13B": 2, "13Bf": 2, "30B": 4, "65B": 8, "70B": 8, "70Bf": 8, } def snake_case_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple=1 , _lowerCAmelCase : List[Any]=256 ) -> Optional[Any]: return multiple_of * ((int(ffn_dim_multiplier * int(8 * n / 3 ) ) + multiple_of - 1) // multiple_of) def snake_case_ ( _lowerCAmelCase : List[str] ) -> str: with open(_lowerCAmelCase , '''r''' ) as f: return json.load(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any ) -> Optional[Any]: with open(_lowerCAmelCase , '''w''' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any=True ) -> List[Any]: os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : int = os.path.join(_lowerCAmelCase , '''tmp''' ) os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : List[str] = read_json(os.path.join(_lowerCAmelCase , '''params.json''' ) ) UpperCAmelCase : str = NUM_SHARDS[model_size] UpperCAmelCase : Any = params['''n_layers'''] UpperCAmelCase : str = params['''n_heads'''] UpperCAmelCase : Any = n_heads // num_shards UpperCAmelCase : List[str] = params['''dim'''] UpperCAmelCase : Optional[Any] = dim // n_heads UpperCAmelCase : str = 1_0_0_0_0.0 UpperCAmelCase : Optional[int] = 1.0 / (base ** (torch.arange(0 , _lowerCAmelCase , 2 ).float() / dims_per_head)) if "n_kv_heads" in params: UpperCAmelCase : Tuple = params['''n_kv_heads'''] # for GQA / MQA UpperCAmelCase : Optional[int] = n_heads_per_shard // num_key_value_heads UpperCAmelCase : Optional[Any] = dim // num_key_value_heads else: # compatibility with other checkpoints UpperCAmelCase : List[str] = n_heads UpperCAmelCase : Optional[int] = n_heads_per_shard UpperCAmelCase : List[str] = dim # permute for sliced rotary def permute(_lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]=n_heads , _lowerCAmelCase : int=dim , _lowerCAmelCase : Dict=dim ): return w.view(_lowerCAmelCase , dima // n_heads // 2 , 2 , _lowerCAmelCase ).transpose(1 , 2 ).reshape(_lowerCAmelCase , _lowerCAmelCase ) print(f"""Fetching all parameters from the checkpoint at {input_base_path}.""" ) # Load weights if model_size == "7B": # Not sharded # (The sharded implementation would also work, but this is simpler.) UpperCAmelCase : int = torch.load(os.path.join(_lowerCAmelCase , '''consolidated.00.pth''' ) , map_location='''cpu''' ) else: # Sharded UpperCAmelCase : Optional[Any] = [ torch.load(os.path.join(_lowerCAmelCase , f"""consolidated.{i:02d}.pth""" ) , map_location='''cpu''' ) for i in range(_lowerCAmelCase ) ] UpperCAmelCase : Any = 0 UpperCAmelCase : str = {'''weight_map''': {}} for layer_i in range(_lowerCAmelCase ): UpperCAmelCase : Optional[Any] = f"""pytorch_model-{layer_i + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase : Optional[int] = { f"""model.layers.{layer_i}.self_attn.q_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wq.weight"""] ), f"""model.layers.{layer_i}.self_attn.k_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wk.weight"""] ), f"""model.layers.{layer_i}.self_attn.v_proj.weight""": loaded[f"""layers.{layer_i}.attention.wv.weight"""], f"""model.layers.{layer_i}.self_attn.o_proj.weight""": loaded[f"""layers.{layer_i}.attention.wo.weight"""], f"""model.layers.{layer_i}.mlp.gate_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w1.weight"""], f"""model.layers.{layer_i}.mlp.down_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w2.weight"""], f"""model.layers.{layer_i}.mlp.up_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w3.weight"""], f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[f"""layers.{layer_i}.attention_norm.weight"""], f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[f"""layers.{layer_i}.ffn_norm.weight"""], } else: # Sharded # Note that attention.w{q,k,v,o}, feed_fordward.w[1,2,3], attention_norm.weight and ffn_norm.weight share # the same storage object, saving attention_norm and ffn_norm will save other weights too, which is # redundant as other weights will be stitched from multiple shards. To avoid that, they are cloned. UpperCAmelCase : List[str] = { f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.attention_norm.weight""" ].clone(), f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.ffn_norm.weight""" ].clone(), } UpperCAmelCase : Union[str, Any] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wq.weight"""].view(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase : Optional[Any] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wk.weight"""].view( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) UpperCAmelCase : str = torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wv.weight"""].view( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = torch.cat( [loaded[i][f"""layers.{layer_i}.attention.wo.weight"""] for i in range(_lowerCAmelCase )] , dim=1 ) UpperCAmelCase : Any = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w1.weight"""] for i in range(_lowerCAmelCase )] , dim=0 ) UpperCAmelCase : str = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w2.weight"""] for i in range(_lowerCAmelCase )] , dim=1 ) UpperCAmelCase : Tuple = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w3.weight"""] for i in range(_lowerCAmelCase )] , dim=0 ) UpperCAmelCase : Any = inv_freq for k, v in state_dict.items(): UpperCAmelCase : List[Any] = filename param_count += v.numel() torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase : Optional[int] = f"""pytorch_model-{n_layers + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase : str = { '''model.embed_tokens.weight''': loaded['''tok_embeddings.weight'''], '''model.norm.weight''': loaded['''norm.weight'''], '''lm_head.weight''': loaded['''output.weight'''], } else: UpperCAmelCase : Any = { '''model.norm.weight''': loaded[0]['''norm.weight'''], '''model.embed_tokens.weight''': torch.cat( [loaded[i]['''tok_embeddings.weight'''] for i in range(_lowerCAmelCase )] , dim=1 ), '''lm_head.weight''': torch.cat([loaded[i]['''output.weight'''] for i in range(_lowerCAmelCase )] , dim=0 ), } for k, v in state_dict.items(): UpperCAmelCase : Optional[int] = filename param_count += v.numel() torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , _lowerCAmelCase ) ) # Write configs UpperCAmelCase : Union[str, Any] = {'''total_size''': param_count * 2} write_json(_lowerCAmelCase , os.path.join(_lowerCAmelCase , '''pytorch_model.bin.index.json''' ) ) UpperCAmelCase : int = params['''ffn_dim_multiplier'''] if '''ffn_dim_multiplier''' in params else 1 UpperCAmelCase : Tuple = params['''multiple_of'''] if '''multiple_of''' in params else 256 UpperCAmelCase : Any = LlamaConfig( hidden_size=_lowerCAmelCase , intermediate_size=compute_intermediate_size(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) , num_attention_heads=params['''n_heads'''] , num_hidden_layers=params['''n_layers'''] , rms_norm_eps=params['''norm_eps'''] , num_key_value_heads=_lowerCAmelCase , ) config.save_pretrained(_lowerCAmelCase ) # Make space so we can load the model properly now. del state_dict del loaded gc.collect() print('''Loading the checkpoint in a Llama model.''' ) UpperCAmelCase : Optional[int] = LlamaForCausalLM.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa , low_cpu_mem_usage=_lowerCAmelCase ) # Avoid saving this as part of the config. del model.config._name_or_path print('''Saving in the Transformers format.''' ) model.save_pretrained(_lowerCAmelCase , safe_serialization=_lowerCAmelCase ) shutil.rmtree(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any] ) -> List[str]: # Initialize the tokenizer based on the `spm` model UpperCAmelCase : Dict = LlamaTokenizer if LlamaTokenizerFast is None else LlamaTokenizerFast print(f"""Saving a {tokenizer_class.__name__} to {tokenizer_path}.""" ) UpperCAmelCase : List[Any] = tokenizer_class(_lowerCAmelCase ) tokenizer.save_pretrained(_lowerCAmelCase ) def snake_case_ ( ) -> List[Any]: UpperCAmelCase : int = argparse.ArgumentParser() parser.add_argument( '''--input_dir''' , help='''Location of LLaMA weights, which contains tokenizer.model and model folders''' , ) parser.add_argument( '''--model_size''' , choices=['''7B''', '''7Bf''', '''13B''', '''13Bf''', '''30B''', '''65B''', '''70B''', '''70Bf''', '''tokenizer_only'''] , ) parser.add_argument( '''--output_dir''' , help='''Location to write HF model and tokenizer''' , ) parser.add_argument('''--safe_serialization''' , type=_lowerCAmelCase , help='''Whether or not to save using `safetensors`.''' ) UpperCAmelCase : List[Any] = parser.parse_args() if args.model_size != "tokenizer_only": write_model( model_path=args.output_dir , input_base_path=os.path.join(args.input_dir , args.model_size ) , model_size=args.model_size , safe_serialization=args.safe_serialization , ) UpperCAmelCase : Optional[int] = os.path.join(args.input_dir , '''tokenizer.model''' ) write_tokenizer(args.output_dir , _lowerCAmelCase ) if __name__ == "__main__": main()
23
1
'''simple docstring''' import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = (PNDMScheduler,) lowerCamelCase__ = (("""num_inference_steps""", 50),) def A ( self : str , **__snake_case : Tuple ) -> Any: UpperCAmelCase : Optional[int] = { '''num_train_timesteps''': 1000, '''beta_start''': 0.00_01, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', } config.update(**__snake_case ) return config def A ( self : Optional[Any] , __snake_case : Optional[Any]=0 , **__snake_case : Tuple ) -> Union[str, Any]: UpperCAmelCase : List[str] = dict(self.forward_default_kwargs ) UpperCAmelCase : Tuple = kwargs.pop('''num_inference_steps''' , __snake_case ) UpperCAmelCase : Tuple = self.dummy_sample UpperCAmelCase : Union[str, Any] = 0.1 * sample UpperCAmelCase : Any = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase : Union[str, Any] = self.get_scheduler_config(**__snake_case ) UpperCAmelCase : Dict = scheduler_class(**__snake_case ) scheduler.set_timesteps(__snake_case ) # copy over dummy past residuals UpperCAmelCase : str = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(__snake_case ) UpperCAmelCase : List[str] = scheduler_class.from_pretrained(__snake_case ) new_scheduler.set_timesteps(__snake_case ) # copy over dummy past residuals UpperCAmelCase : Optional[Any] = dummy_past_residuals[:] UpperCAmelCase : int = scheduler.step_prk(__snake_case , __snake_case , __snake_case , **__snake_case ).prev_sample UpperCAmelCase : Dict = new_scheduler.step_prk(__snake_case , __snake_case , __snake_case , **__snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" UpperCAmelCase : Optional[Any] = scheduler.step_plms(__snake_case , __snake_case , __snake_case , **__snake_case ).prev_sample UpperCAmelCase : List[str] = new_scheduler.step_plms(__snake_case , __snake_case , __snake_case , **__snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def A ( self : Tuple ) -> Tuple: pass def A ( self : int , __snake_case : List[Any]=0 , **__snake_case : List[Any] ) -> Union[str, Any]: UpperCAmelCase : List[Any] = dict(self.forward_default_kwargs ) UpperCAmelCase : Tuple = kwargs.pop('''num_inference_steps''' , __snake_case ) UpperCAmelCase : Dict = self.dummy_sample UpperCAmelCase : List[Any] = 0.1 * sample UpperCAmelCase : int = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase : Any = self.get_scheduler_config() UpperCAmelCase : Tuple = scheduler_class(**__snake_case ) scheduler.set_timesteps(__snake_case ) # copy over dummy past residuals (must be after setting timesteps) UpperCAmelCase : Union[str, Any] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(__snake_case ) UpperCAmelCase : List[str] = scheduler_class.from_pretrained(__snake_case ) # copy over dummy past residuals new_scheduler.set_timesteps(__snake_case ) # copy over dummy past residual (must be after setting timesteps) UpperCAmelCase : Dict = dummy_past_residuals[:] UpperCAmelCase : Optional[int] = scheduler.step_prk(__snake_case , __snake_case , __snake_case , **__snake_case ).prev_sample UpperCAmelCase : List[Any] = new_scheduler.step_prk(__snake_case , __snake_case , __snake_case , **__snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" UpperCAmelCase : str = scheduler.step_plms(__snake_case , __snake_case , __snake_case , **__snake_case ).prev_sample UpperCAmelCase : Optional[int] = new_scheduler.step_plms(__snake_case , __snake_case , __snake_case , **__snake_case ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def A ( self : Tuple , **__snake_case : Tuple ) -> Any: UpperCAmelCase : Dict = self.scheduler_classes[0] UpperCAmelCase : Tuple = self.get_scheduler_config(**__snake_case ) UpperCAmelCase : Tuple = scheduler_class(**__snake_case ) UpperCAmelCase : Any = 10 UpperCAmelCase : Optional[Any] = self.dummy_model() UpperCAmelCase : int = self.dummy_sample_deter scheduler.set_timesteps(__snake_case ) for i, t in enumerate(scheduler.prk_timesteps ): UpperCAmelCase : Tuple = model(__snake_case , __snake_case ) UpperCAmelCase : int = scheduler.step_prk(__snake_case , __snake_case , __snake_case ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): UpperCAmelCase : Any = model(__snake_case , __snake_case ) UpperCAmelCase : Any = scheduler.step_plms(__snake_case , __snake_case , __snake_case ).prev_sample return sample def A ( self : Optional[int] ) -> Optional[int]: UpperCAmelCase : Union[str, Any] = dict(self.forward_default_kwargs ) UpperCAmelCase : Dict = kwargs.pop('''num_inference_steps''' , __snake_case ) for scheduler_class in self.scheduler_classes: UpperCAmelCase : List[Any] = self.get_scheduler_config() UpperCAmelCase : int = scheduler_class(**__snake_case ) UpperCAmelCase : List[str] = self.dummy_sample UpperCAmelCase : Any = 0.1 * sample if num_inference_steps is not None and hasattr(__snake_case , '''set_timesteps''' ): scheduler.set_timesteps(__snake_case ) elif num_inference_steps is not None and not hasattr(__snake_case , '''set_timesteps''' ): UpperCAmelCase : List[Any] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) UpperCAmelCase : Optional[int] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] UpperCAmelCase : Optional[int] = dummy_past_residuals[:] UpperCAmelCase : List[Any] = scheduler.step_prk(__snake_case , 0 , __snake_case , **__snake_case ).prev_sample UpperCAmelCase : Optional[Any] = scheduler.step_prk(__snake_case , 1 , __snake_case , **__snake_case ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) UpperCAmelCase : str = scheduler.step_plms(__snake_case , 0 , __snake_case , **__snake_case ).prev_sample UpperCAmelCase : List[str] = scheduler.step_plms(__snake_case , 1 , __snake_case , **__snake_case ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def A ( self : List[str] ) -> List[Any]: for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=__snake_case ) def A ( self : Any ) -> Dict: for steps_offset in [0, 1]: self.check_over_configs(steps_offset=__snake_case ) UpperCAmelCase : Union[str, Any] = self.scheduler_classes[0] UpperCAmelCase : str = self.get_scheduler_config(steps_offset=1 ) UpperCAmelCase : Tuple = scheduler_class(**__snake_case ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ) , ) def A ( self : int ) -> Optional[Any]: for beta_start, beta_end in zip([0.00_01, 0.0_01] , [0.0_02, 0.02] ): self.check_over_configs(beta_start=__snake_case , beta_end=__snake_case ) def A ( self : List[Any] ) -> Optional[Any]: for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=__snake_case ) def A ( self : int ) -> Optional[Any]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=__snake_case ) def A ( self : Any ) -> Any: for t in [1, 5, 10]: self.check_over_forward(time_step=__snake_case ) def A ( self : Dict ) -> int: for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=__snake_case ) def A ( self : List[Any] ) -> int: # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 UpperCAmelCase : str = 27 for scheduler_class in self.scheduler_classes: UpperCAmelCase : Tuple = self.dummy_sample UpperCAmelCase : List[Any] = 0.1 * sample UpperCAmelCase : Optional[int] = self.get_scheduler_config() UpperCAmelCase : int = scheduler_class(**__snake_case ) scheduler.set_timesteps(__snake_case ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): UpperCAmelCase : str = scheduler.step_prk(__snake_case , __snake_case , __snake_case ).prev_sample def A ( self : Optional[int] ) -> Optional[int]: with self.assertRaises(__snake_case ): UpperCAmelCase : Optional[int] = self.scheduler_classes[0] UpperCAmelCase : List[str] = self.get_scheduler_config() UpperCAmelCase : List[str] = scheduler_class(**__snake_case ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def A ( self : Tuple ) -> List[str]: UpperCAmelCase : Union[str, Any] = self.full_loop() UpperCAmelCase : str = torch.sum(torch.abs(__snake_case ) ) UpperCAmelCase : List[str] = torch.mean(torch.abs(__snake_case ) ) assert abs(result_sum.item() - 1_98.13_18 ) < 1E-2 assert abs(result_mean.item() - 0.25_80 ) < 1E-3 def A ( self : Union[str, Any] ) -> Dict: UpperCAmelCase : int = self.full_loop(prediction_type='''v_prediction''' ) UpperCAmelCase : List[Any] = torch.sum(torch.abs(__snake_case ) ) UpperCAmelCase : Dict = torch.mean(torch.abs(__snake_case ) ) assert abs(result_sum.item() - 67.39_86 ) < 1E-2 assert abs(result_mean.item() - 0.08_78 ) < 1E-3 def A ( self : Any ) -> List[str]: # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase : int = self.full_loop(set_alpha_to_one=__snake_case , beta_start=0.01 ) UpperCAmelCase : Optional[int] = torch.sum(torch.abs(__snake_case ) ) UpperCAmelCase : str = torch.mean(torch.abs(__snake_case ) ) assert abs(result_sum.item() - 2_30.03_99 ) < 1E-2 assert abs(result_mean.item() - 0.29_95 ) < 1E-3 def A ( self : List[Any] ) -> str: # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase : List[str] = self.full_loop(set_alpha_to_one=__snake_case , beta_start=0.01 ) UpperCAmelCase : Union[str, Any] = torch.sum(torch.abs(__snake_case ) ) UpperCAmelCase : List[str] = torch.mean(torch.abs(__snake_case ) ) assert abs(result_sum.item() - 1_86.94_82 ) < 1E-2 assert abs(result_mean.item() - 0.24_34 ) < 1E-3
23
'''simple docstring''' from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Optional[int] = tau * frequency / samplerate UpperCAmelCase : List[Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = cos(_lowerCAmelCase ) UpperCAmelCase : int = _sin / (2 * q_factor) UpperCAmelCase : Any = (1 - _cos) / 2 UpperCAmelCase : List[Any] = 1 - _cos UpperCAmelCase : Union[str, Any] = 1 + alpha UpperCAmelCase : Any = -2 * _cos UpperCAmelCase : Dict = 1 - alpha UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Any = tau * frequency / samplerate UpperCAmelCase : Tuple = sin(_lowerCAmelCase ) UpperCAmelCase : Tuple = cos(_lowerCAmelCase ) UpperCAmelCase : Dict = _sin / (2 * q_factor) UpperCAmelCase : int = (1 + _cos) / 2 UpperCAmelCase : List[Any] = -1 - _cos UpperCAmelCase : Tuple = 1 + alpha UpperCAmelCase : List[str] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha UpperCAmelCase : Tuple = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Optional[int] = tau * frequency / samplerate UpperCAmelCase : Optional[int] = sin(_lowerCAmelCase ) UpperCAmelCase : Tuple = cos(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = _sin / (2 * q_factor) UpperCAmelCase : Union[str, Any] = _sin / 2 UpperCAmelCase : Any = 0 UpperCAmelCase : int = -ba UpperCAmelCase : Optional[Any] = 1 + alpha UpperCAmelCase : List[Any] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha UpperCAmelCase : int = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : List[str] = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : str = cos(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = _sin / (2 * q_factor) UpperCAmelCase : List[str] = 1 - alpha UpperCAmelCase : Any = -2 * _cos UpperCAmelCase : Optional[int] = 1 + alpha UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([ba, ba, ba] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : Optional[Any] = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = cos(_lowerCAmelCase ) UpperCAmelCase : Dict = _sin / (2 * q_factor) UpperCAmelCase : str = 10 ** (gain_db / 40) UpperCAmelCase : int = 1 + alpha * big_a UpperCAmelCase : Union[str, Any] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha * big_a UpperCAmelCase : Union[str, Any] = 1 + alpha / big_a UpperCAmelCase : Tuple = -2 * _cos UpperCAmelCase : Any = 1 - alpha / big_a UpperCAmelCase : Optional[Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : Any = tau * frequency / samplerate UpperCAmelCase : Optional[int] = sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = cos(_lowerCAmelCase ) UpperCAmelCase : str = _sin / (2 * q_factor) UpperCAmelCase : List[str] = 10 ** (gain_db / 40) UpperCAmelCase : Optional[int] = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase : int = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase : int = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase : Optional[int] = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase : str = 2 * sqrt(_lowerCAmelCase ) * alpha UpperCAmelCase : Dict = big_a * (pmc + aaa) UpperCAmelCase : Any = 2 * big_a * mpc UpperCAmelCase : Union[str, Any] = big_a * (pmc - aaa) UpperCAmelCase : Optional[int] = ppmc + aaa UpperCAmelCase : Optional[Any] = -2 * pmpc UpperCAmelCase : Optional[Any] = ppmc - aaa UpperCAmelCase : int = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : int = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = cos(_lowerCAmelCase ) UpperCAmelCase : Any = _sin / (2 * q_factor) UpperCAmelCase : int = 10 ** (gain_db / 40) UpperCAmelCase : List[str] = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase : Union[str, Any] = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase : Optional[Any] = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase : Union[str, Any] = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase : List[str] = 2 * sqrt(_lowerCAmelCase ) * alpha UpperCAmelCase : Any = big_a * (ppmc + aaa) UpperCAmelCase : str = -2 * big_a * pmpc UpperCAmelCase : List[Any] = big_a * (ppmc - aaa) UpperCAmelCase : Optional[Any] = pmc + aaa UpperCAmelCase : Any = 2 * mpc UpperCAmelCase : str = pmc - aaa UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt
23
1
'''simple docstring''' def snake_case_ ( _lowerCAmelCase : Optional[Any] ) -> List[str]: UpperCAmelCase , UpperCAmelCase : List[str] = [], [] while len(_lowerCAmelCase ) > 1: UpperCAmelCase , UpperCAmelCase : Tuple = min(_lowerCAmelCase ), max(_lowerCAmelCase ) start.append(_lowerCAmelCase ) end.append(_lowerCAmelCase ) collection.remove(_lowerCAmelCase ) collection.remove(_lowerCAmelCase ) end.reverse() return start + collection + end if __name__ == "__main__": UpperCamelCase__: List[Any] = input("Enter numbers separated by a comma:\n").strip() UpperCamelCase__: str = [int(item) for item in user_input.split(",")] print(*merge_sort(unsorted), sep=",")
23
'''simple docstring''' from __future__ import annotations def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : str ) -> bool: UpperCAmelCase : str = get_failure_array(_lowerCAmelCase ) # 2) Step through text searching for pattern UpperCAmelCase , UpperCAmelCase : Optional[Any] = 0, 0 # index into text, pattern while i < len(_lowerCAmelCase ): if pattern[j] == text[i]: if j == (len(_lowerCAmelCase ) - 1): return True j += 1 # if this is a prefix in our pattern # just go back far enough to continue elif j > 0: UpperCAmelCase : Optional[Any] = failure[j - 1] continue i += 1 return False def snake_case_ ( _lowerCAmelCase : str ) -> list[int]: UpperCAmelCase : Optional[Any] = [0] UpperCAmelCase : str = 0 UpperCAmelCase : List[str] = 1 while j < len(_lowerCAmelCase ): if pattern[i] == pattern[j]: i += 1 elif i > 0: UpperCAmelCase : Union[str, Any] = failure[i - 1] continue j += 1 failure.append(_lowerCAmelCase ) return failure if __name__ == "__main__": # Test 1) UpperCamelCase__: str = "abc1abc12" UpperCamelCase__: str = "alskfjaldsabc1abc1abc12k23adsfabcabc" UpperCamelCase__: Any = "alskfjaldsk23adsfabcabc" assert kmp(pattern, texta) and not kmp(pattern, texta) # Test 2) UpperCamelCase__: Tuple = "ABABX" UpperCamelCase__: Union[str, Any] = "ABABZABABYABABX" assert kmp(pattern, text) # Test 3) UpperCamelCase__: Any = "AAAB" UpperCamelCase__: str = "ABAAAAAB" assert kmp(pattern, text) # Test 4) UpperCamelCase__: int = "abcdabcy" UpperCamelCase__: Any = "abcxabcdabxabcdabcdabcy" assert kmp(pattern, text) # Test 5) UpperCamelCase__: List[str] = "aabaabaaa" assert get_failure_array(pattern) == [0, 1, 0, 1, 2, 3, 4, 5, 2]
23
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__: Optional[int] = logging.get_logger(__name__) UpperCamelCase__: str = {"ctrl": "https://huggingface.co/ctrl/resolve/main/config.json"} class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = """ctrl""" lowerCamelCase__ = ["""past_key_values"""] lowerCamelCase__ = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self : str , __snake_case : Optional[int]=246534 , __snake_case : Any=256 , __snake_case : Optional[Any]=1280 , __snake_case : List[Any]=8192 , __snake_case : Tuple=48 , __snake_case : Union[str, Any]=16 , __snake_case : Optional[int]=0.1 , __snake_case : List[Any]=0.1 , __snake_case : Optional[int]=1E-6 , __snake_case : Any=0.02 , __snake_case : Optional[Any]=True , **__snake_case : Dict , ) -> Optional[int]: UpperCAmelCase : Dict = vocab_size UpperCAmelCase : Optional[int] = n_positions UpperCAmelCase : str = n_embd UpperCAmelCase : Any = n_layer UpperCAmelCase : Tuple = n_head UpperCAmelCase : int = dff UpperCAmelCase : Any = resid_pdrop UpperCAmelCase : Any = embd_pdrop UpperCAmelCase : Tuple = layer_norm_epsilon UpperCAmelCase : int = initializer_range UpperCAmelCase : Any = use_cache super().__init__(**__snake_case )
23
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() UpperCamelCase__: int = logging.get_logger(__name__) UpperCamelCase__: Dict = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "adapter_layer": "encoder.layers.*.adapter_layer", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", "pooling_layer.linear": "projector", "pooling_layer.projection": "classifier", } UpperCamelCase__: Optional[Any] = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", "projector", "classifier", ] def snake_case_ ( _lowerCAmelCase : str ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = {} with open(_lowerCAmelCase , '''r''' ) as file: for line_number, line in enumerate(_lowerCAmelCase ): UpperCAmelCase : List[str] = line.strip() if line: UpperCAmelCase : str = line.split() UpperCAmelCase : Union[str, Any] = line_number UpperCAmelCase : List[Any] = words[0] UpperCAmelCase : Union[str, Any] = value return result def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : str ) -> int: for attribute in key.split('''.''' ): UpperCAmelCase : Dict = getattr(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Dict = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(_lowerCAmelCase ): UpperCAmelCase : Any = PARAM_MAPPING[full_name.split('''.''' )[-1]] UpperCAmelCase : Dict = '''param''' if weight_type is not None and weight_type != "param": UpperCAmelCase : Optional[int] = getattr(_lowerCAmelCase , _lowerCAmelCase ).shape elif weight_type is not None and weight_type == "param": UpperCAmelCase : List[Any] = hf_pointer for attribute in hf_param_name.split('''.''' ): UpperCAmelCase : Optional[Any] = getattr(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : int = shape_pointer.shape # let's reduce dimension UpperCAmelCase : Union[str, Any] = value[0] else: UpperCAmelCase : List[Any] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": UpperCAmelCase : int = value elif weight_type == "weight_g": UpperCAmelCase : str = value elif weight_type == "weight_v": UpperCAmelCase : Dict = value elif weight_type == "bias": UpperCAmelCase : str = value elif weight_type == "param": for attribute in hf_param_name.split('''.''' ): UpperCAmelCase : int = getattr(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = value else: UpperCAmelCase : Tuple = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[Any] ) -> List[Any]: UpperCAmelCase : List[str] = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(_lowerCAmelCase ): UpperCAmelCase : List[str] = PARAM_MAPPING[full_name.split('''.''' )[-1]] UpperCAmelCase : Any = '''param''' if weight_type is not None and weight_type != "param": UpperCAmelCase : Optional[int] = '''.'''.join([key, weight_type] ) elif weight_type is not None and weight_type == "param": UpperCAmelCase : Optional[int] = '''.'''.join([key, hf_param_name] ) else: UpperCAmelCase : List[Any] = key UpperCAmelCase : Tuple = value if '''lm_head''' in full_key else value[0] UpperCamelCase__: Tuple = { "W_a": "linear_1.weight", "W_b": "linear_2.weight", "b_a": "linear_1.bias", "b_b": "linear_2.bias", "ln_W": "norm.weight", "ln_b": "norm.bias", } def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : Any=None , _lowerCAmelCase : Optional[Any]=None ) -> int: UpperCAmelCase : List[Any] = False for key, mapped_key in MAPPING.items(): UpperCAmelCase : int = '''wav2vec2.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: UpperCAmelCase : Optional[Any] = True if "*" in mapped_key: UpperCAmelCase : Tuple = name.split(_lowerCAmelCase )[0].split('''.''' )[-2] UpperCAmelCase : List[Any] = mapped_key.replace('''*''' , _lowerCAmelCase ) if "weight_g" in name: UpperCAmelCase : str = '''weight_g''' elif "weight_v" in name: UpperCAmelCase : int = '''weight_v''' elif "bias" in name: UpperCAmelCase : int = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj UpperCAmelCase : List[str] = '''weight''' else: UpperCAmelCase : Dict = None if hf_dict is not None: rename_dict(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) else: set_recursively(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return is_used return is_used def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Any: UpperCAmelCase : Dict = [] UpperCAmelCase : Dict = fairseq_model.state_dict() UpperCAmelCase : Union[str, Any] = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase : Dict = False if "conv_layers" in name: load_conv_layer( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , hf_model.config.feat_extract_norm == '''group''' , ) UpperCAmelCase : Any = True else: UpperCAmelCase : Optional[Any] = load_wavaveca_layer(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not is_used: unused_weights.append(_lowerCAmelCase ) logger.warning(f"""Unused weights: {unused_weights}""" ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: UpperCAmelCase : Any = full_name.split('''conv_layers.''' )[-1] UpperCAmelCase : Optional[int] = name.split('''.''' ) UpperCAmelCase : Tuple = int(items[0] ) UpperCAmelCase : Tuple = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) UpperCAmelCase : Tuple = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) UpperCAmelCase : Union[str, Any] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.""" ) UpperCAmelCase : Union[str, Any] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.""" ) UpperCAmelCase : List[str] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(_lowerCAmelCase ) @torch.no_grad() def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict=None , _lowerCAmelCase : List[Any]=None , _lowerCAmelCase : int=True , _lowerCAmelCase : Optional[int]=False ) -> Dict: if config_path is not None: UpperCAmelCase : List[str] = WavaVecaConfig.from_pretrained(_lowerCAmelCase ) else: UpperCAmelCase : List[Any] = WavaVecaConfig() if is_seq_class: UpperCAmelCase : Optional[Any] = read_txt_into_dict(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = idalabel UpperCAmelCase : Optional[Any] = WavaVecaForSequenceClassification(_lowerCAmelCase ) UpperCAmelCase : Dict = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase , ) feature_extractor.save_pretrained(_lowerCAmelCase ) elif is_finetuned: if dict_path: UpperCAmelCase : Dict = Dictionary.load(_lowerCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq UpperCAmelCase : Any = target_dict.pad_index UpperCAmelCase : Tuple = target_dict.bos_index UpperCAmelCase : Optional[int] = target_dict.eos_index UpperCAmelCase : Union[str, Any] = len(target_dict.symbols ) UpperCAmelCase : Dict = os.path.join(_lowerCAmelCase , '''vocab.json''' ) if not os.path.isdir(_lowerCAmelCase ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(_lowerCAmelCase ) ) return os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : List[Any] = target_dict.indices # fairseq has the <pad> and <s> switched UpperCAmelCase : List[str] = 0 UpperCAmelCase : List[str] = 1 with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = WavaVecaCTCTokenizer( _lowerCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=_lowerCAmelCase , ) UpperCAmelCase : int = True if config.feat_extract_norm == '''layer''' else False UpperCAmelCase : int = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase , ) UpperCAmelCase : str = WavaVecaProcessor(feature_extractor=_lowerCAmelCase , tokenizer=_lowerCAmelCase ) processor.save_pretrained(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = WavaVecaForCTC(_lowerCAmelCase ) else: UpperCAmelCase : Dict = WavaVecaForPreTraining(_lowerCAmelCase ) if is_finetuned or is_seq_class: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : List[str] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: UpperCAmelCase : Optional[Any] = argparse.Namespace(task='''audio_pretraining''' ) UpperCAmelCase : List[Any] = fairseq.tasks.setup_task(_lowerCAmelCase ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : int = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=_lowerCAmelCase ) UpperCAmelCase : Optional[int] = model[0].eval() recursively_load_weights(_lowerCAmelCase , _lowerCAmelCase , not is_finetuned ) hf_wavavec.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": UpperCamelCase__: Dict = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--not_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) parser.add_argument( "--is_seq_class", action="store_true", help="Whether the model to convert is a fine-tuned sequence classification model or not", ) UpperCamelCase__: Any = parser.parse_args() UpperCamelCase__: int = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
23
1
'''simple docstring''' import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer, GPTaConfig, GPTaLMHeadModel, GPTaTokenizer, RobertaConfig, RobertaForMaskedLM, RobertaTokenizer, ) from utils import git_log, init_gpu_params, logger, set_seed UpperCamelCase__: Any = { "distilbert": (DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer), "roberta": (RobertaConfig, RobertaForMaskedLM, RobertaTokenizer), "bert": (BertConfig, BertForMaskedLM, BertTokenizer), "gpt2": (GPTaConfig, GPTaLMHeadModel, GPTaTokenizer), } def snake_case_ ( _lowerCAmelCase : Optional[Any] ) -> Tuple: assert (args.mlm and args.alpha_mlm > 0.0) or (not args.mlm and args.alpha_mlm == 0.0) assert (args.alpha_mlm > 0.0 and args.alpha_clm == 0.0) or (args.alpha_mlm == 0.0 and args.alpha_clm > 0.0) if args.mlm: assert os.path.isfile(args.token_counts ) assert (args.student_type in ["roberta", "distilbert"]) and (args.teacher_type in ["roberta", "bert"]) else: assert (args.student_type in ["gpt2"]) and (args.teacher_type in ["gpt2"]) assert args.teacher_type == args.student_type or ( args.student_type == "distilbert" and args.teacher_type == "bert" ) assert os.path.isfile(args.student_config ) if args.student_pretrained_weights is not None: assert os.path.isfile(args.student_pretrained_weights ) if args.freeze_token_type_embds: assert args.student_type in ["roberta"] assert args.alpha_ce >= 0.0 assert args.alpha_mlm >= 0.0 assert args.alpha_clm >= 0.0 assert args.alpha_mse >= 0.0 assert args.alpha_cos >= 0.0 assert args.alpha_ce + args.alpha_mlm + args.alpha_clm + args.alpha_mse + args.alpha_cos > 0.0 def snake_case_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[Any] ) -> str: if args.student_type == "roberta": UpperCAmelCase : Any = False elif args.student_type == "gpt2": UpperCAmelCase : List[Any] = False def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] ) -> str: if args.student_type == "roberta": UpperCAmelCase : List[Any] = False def snake_case_ ( ) -> Union[str, Any]: UpperCAmelCase : str = argparse.ArgumentParser(description='''Training''' ) parser.add_argument('''--force''' , action='''store_true''' , help='''Overwrite dump_path if it already exists.''' ) parser.add_argument( '''--dump_path''' , type=_lowerCAmelCase , required=_lowerCAmelCase , help='''The output directory (log, checkpoints, parameters, etc.)''' ) parser.add_argument( '''--data_file''' , type=_lowerCAmelCase , required=_lowerCAmelCase , help='''The binarized file (tokenized + tokens_to_ids) and grouped by sequence.''' , ) parser.add_argument( '''--student_type''' , type=_lowerCAmelCase , choices=['''distilbert''', '''roberta''', '''gpt2'''] , required=_lowerCAmelCase , help='''The student type (DistilBERT, RoBERTa).''' , ) parser.add_argument('''--student_config''' , type=_lowerCAmelCase , required=_lowerCAmelCase , help='''Path to the student configuration.''' ) parser.add_argument( '''--student_pretrained_weights''' , default=_lowerCAmelCase , type=_lowerCAmelCase , help='''Load student initialization checkpoint.''' ) parser.add_argument( '''--teacher_type''' , choices=['''bert''', '''roberta''', '''gpt2'''] , required=_lowerCAmelCase , help='''Teacher type (BERT, RoBERTa).''' ) parser.add_argument('''--teacher_name''' , type=_lowerCAmelCase , required=_lowerCAmelCase , help='''The teacher model.''' ) parser.add_argument('''--temperature''' , default=2.0 , type=_lowerCAmelCase , help='''Temperature for the softmax temperature.''' ) parser.add_argument( '''--alpha_ce''' , default=0.5 , type=_lowerCAmelCase , help='''Linear weight for the distillation loss. Must be >=0.''' ) parser.add_argument( '''--alpha_mlm''' , default=0.0 , type=_lowerCAmelCase , help='''Linear weight for the MLM loss. Must be >=0. Should be used in conjunction with `mlm` flag.''' , ) parser.add_argument('''--alpha_clm''' , default=0.5 , type=_lowerCAmelCase , help='''Linear weight for the CLM loss. Must be >=0.''' ) parser.add_argument('''--alpha_mse''' , default=0.0 , type=_lowerCAmelCase , help='''Linear weight of the MSE loss. Must be >=0.''' ) parser.add_argument( '''--alpha_cos''' , default=0.0 , type=_lowerCAmelCase , help='''Linear weight of the cosine embedding loss. Must be >=0.''' ) parser.add_argument( '''--mlm''' , action='''store_true''' , help='''The LM step: MLM or CLM. If `mlm` is True, the MLM is used over CLM.''' ) parser.add_argument( '''--mlm_mask_prop''' , default=0.1_5 , type=_lowerCAmelCase , help='''Proportion of tokens for which we need to make a prediction.''' , ) parser.add_argument('''--word_mask''' , default=0.8 , type=_lowerCAmelCase , help='''Proportion of tokens to mask out.''' ) parser.add_argument('''--word_keep''' , default=0.1 , type=_lowerCAmelCase , help='''Proportion of tokens to keep.''' ) parser.add_argument('''--word_rand''' , default=0.1 , type=_lowerCAmelCase , help='''Proportion of tokens to randomly replace.''' ) parser.add_argument( '''--mlm_smoothing''' , default=0.7 , type=_lowerCAmelCase , help='''Smoothing parameter to emphasize more rare tokens (see XLM, similar to word2vec).''' , ) parser.add_argument('''--token_counts''' , type=_lowerCAmelCase , help='''The token counts in the data_file for MLM.''' ) parser.add_argument( '''--restrict_ce_to_mask''' , action='''store_true''' , help='''If true, compute the distillation loss only the [MLM] prediction distribution.''' , ) parser.add_argument( '''--freeze_pos_embs''' , action='''store_true''' , help='''Freeze positional embeddings during distillation. For student_type in [\'roberta\', \'gpt2\'] only.''' , ) parser.add_argument( '''--freeze_token_type_embds''' , action='''store_true''' , help='''Freeze token type embeddings during distillation if existent. For student_type in [\'roberta\'] only.''' , ) parser.add_argument('''--n_epoch''' , type=_lowerCAmelCase , default=3 , help='''Number of pass on the whole dataset.''' ) parser.add_argument('''--batch_size''' , type=_lowerCAmelCase , default=5 , help='''Batch size (for each process).''' ) parser.add_argument( '''--group_by_size''' , action='''store_false''' , help='''If true, group sequences that have similar length into the same batch. Default is true.''' , ) parser.add_argument( '''--gradient_accumulation_steps''' , type=_lowerCAmelCase , default=50 , help='''Gradient accumulation for larger training batches.''' , ) parser.add_argument('''--warmup_prop''' , default=0.0_5 , type=_lowerCAmelCase , help='''Linear warmup proportion.''' ) parser.add_argument('''--weight_decay''' , default=0.0 , type=_lowerCAmelCase , help='''Weight decay if we apply some.''' ) parser.add_argument('''--learning_rate''' , default=5e-4 , type=_lowerCAmelCase , help='''The initial learning rate for Adam.''' ) parser.add_argument('''--adam_epsilon''' , default=1e-6 , type=_lowerCAmelCase , help='''Epsilon for Adam optimizer.''' ) parser.add_argument('''--max_grad_norm''' , default=5.0 , type=_lowerCAmelCase , help='''Max gradient norm.''' ) parser.add_argument('''--initializer_range''' , default=0.0_2 , type=_lowerCAmelCase , help='''Random initialization range.''' ) parser.add_argument( '''--fp16''' , action='''store_true''' , help='''Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit''' , ) parser.add_argument( '''--fp16_opt_level''' , type=_lowerCAmelCase , default='''O1''' , help=( '''For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\'].''' '''See details at https://nvidia.github.io/apex/amp.html''' ) , ) parser.add_argument('''--n_gpu''' , type=_lowerCAmelCase , default=1 , help='''Number of GPUs in the node.''' ) parser.add_argument('''--local_rank''' , type=_lowerCAmelCase , default=-1 , help='''Distributed training - Local rank''' ) parser.add_argument('''--seed''' , type=_lowerCAmelCase , default=56 , help='''Random seed''' ) parser.add_argument('''--log_interval''' , type=_lowerCAmelCase , default=500 , help='''Tensorboard logging interval.''' ) parser.add_argument('''--checkpoint_interval''' , type=_lowerCAmelCase , default=4000 , help='''Checkpoint interval.''' ) UpperCAmelCase : Optional[int] = parser.parse_args() sanity_checks(_lowerCAmelCase ) # ARGS # init_gpu_params(_lowerCAmelCase ) set_seed(_lowerCAmelCase ) if args.is_master: if os.path.exists(args.dump_path ): if not args.force: raise ValueError( f"""Serialization dir {args.dump_path} already exists, but you have not precised wheter to overwrite""" ''' itUse `--force` if you want to overwrite it''' ) else: shutil.rmtree(args.dump_path ) if not os.path.exists(args.dump_path ): os.makedirs(args.dump_path ) logger.info(f"""Experiment will be dumped and logged in {args.dump_path}""" ) # SAVE PARAMS # logger.info(f"""Param: {args}""" ) with open(os.path.join(args.dump_path , '''parameters.json''' ) , '''w''' ) as f: json.dump(vars(_lowerCAmelCase ) , _lowerCAmelCase , indent=4 ) git_log(args.dump_path ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : Optional[Any] = MODEL_CLASSES[args.student_type] UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : List[str] = MODEL_CLASSES[args.teacher_type] # TOKENIZER # UpperCAmelCase : List[Any] = teacher_tokenizer_class.from_pretrained(args.teacher_name ) UpperCAmelCase : str = {} for tok_name, tok_symbol in tokenizer.special_tokens_map.items(): UpperCAmelCase : int = tokenizer.all_special_tokens.index(_lowerCAmelCase ) UpperCAmelCase : List[Any] = tokenizer.all_special_ids[idx] logger.info(f"""Special tokens {special_tok_ids}""" ) UpperCAmelCase : Any = special_tok_ids UpperCAmelCase : List[Any] = tokenizer.max_model_input_sizes[args.teacher_name] # DATA LOADER # logger.info(f"""Loading data from {args.data_file}""" ) with open(args.data_file , '''rb''' ) as fp: UpperCAmelCase : str = pickle.load(_lowerCAmelCase ) if args.mlm: logger.info(f"""Loading token counts from {args.token_counts} (already pre-computed)""" ) with open(args.token_counts , '''rb''' ) as fp: UpperCAmelCase : List[str] = pickle.load(_lowerCAmelCase ) UpperCAmelCase : str = np.maximum(_lowerCAmelCase , 1 ) ** -args.mlm_smoothing for idx in special_tok_ids.values(): UpperCAmelCase : Optional[Any] = 0.0 # do not predict special tokens UpperCAmelCase : str = torch.from_numpy(_lowerCAmelCase ) else: UpperCAmelCase : Optional[int] = None UpperCAmelCase : List[str] = LmSeqsDataset(params=_lowerCAmelCase , data=_lowerCAmelCase ) logger.info('''Data loader created.''' ) # STUDENT # logger.info(f"""Loading student config from {args.student_config}""" ) UpperCAmelCase : Dict = student_config_class.from_pretrained(args.student_config ) UpperCAmelCase : List[str] = True if args.student_pretrained_weights is not None: logger.info(f"""Loading pretrained weights from {args.student_pretrained_weights}""" ) UpperCAmelCase : Optional[int] = student_model_class.from_pretrained(args.student_pretrained_weights , config=_lowerCAmelCase ) else: UpperCAmelCase : Optional[Any] = student_model_class(_lowerCAmelCase ) if args.n_gpu > 0: student.to(f"""cuda:{args.local_rank}""" ) logger.info('''Student loaded.''' ) # TEACHER # UpperCAmelCase : List[Any] = teacher_model_class.from_pretrained(args.teacher_name , output_hidden_states=_lowerCAmelCase ) if args.n_gpu > 0: teacher.to(f"""cuda:{args.local_rank}""" ) logger.info(f"""Teacher loaded from {args.teacher_name}.""" ) # FREEZING # if args.freeze_pos_embs: freeze_pos_embeddings(_lowerCAmelCase , _lowerCAmelCase ) if args.freeze_token_type_embds: freeze_token_type_embeddings(_lowerCAmelCase , _lowerCAmelCase ) # SANITY CHECKS # assert student.config.vocab_size == teacher.config.vocab_size assert student.config.hidden_size == teacher.config.hidden_size assert student.config.max_position_embeddings == teacher.config.max_position_embeddings if args.mlm: assert token_probs.size(0 ) == stu_architecture_config.vocab_size # DISTILLER # torch.cuda.empty_cache() UpperCAmelCase : Union[str, Any] = Distiller( params=_lowerCAmelCase , dataset=_lowerCAmelCase , token_probs=_lowerCAmelCase , student=_lowerCAmelCase , teacher=_lowerCAmelCase ) distiller.train() logger.info('''Let\'s go get some drinks.''' ) if __name__ == "__main__": main()
23
'''simple docstring''' import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : str ) -> int: with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights UpperCAmelCase : Optional[Any] = FlaxDiffusionPipeline.from_pretrained( '''hf-internal-testing/tiny-stable-diffusion-pipe''' , safety_checker=__snake_case , cache_dir=__snake_case ) UpperCAmelCase : str = [t[-1] for t in os.walk(os.path.join(__snake_case , os.listdir(__snake_case )[0] , '''snapshots''' ) )] UpperCAmelCase : str = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith('''.bin''' ) for f in files ) @slow @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : List[str] ) -> Dict: UpperCAmelCase , UpperCAmelCase : str = FlaxStableDiffusionPipeline.from_pretrained( '''hf-internal-testing/tiny-stable-diffusion-pipe''' , safety_checker=__snake_case ) UpperCAmelCase : List[Any] = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : List[str] = jax.random.PRNGKey(0 ) UpperCAmelCase : Optional[Any] = 4 UpperCAmelCase : Optional[Any] = jax.device_count() UpperCAmelCase : Tuple = num_samples * [prompt] UpperCAmelCase : int = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Tuple = replicate(__snake_case ) UpperCAmelCase : Any = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Optional[Any] = shard(__snake_case ) UpperCAmelCase : Optional[int] = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1E-3 assert np.abs(np.abs(__snake_case , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5E-1 UpperCAmelCase : Union[str, Any] = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(__snake_case ) == num_samples def A ( self : List[Any] ) -> List[str]: UpperCAmelCase , UpperCAmelCase : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''flax''' , safety_checker=__snake_case ) UpperCAmelCase : Dict = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : Optional[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase : Any = 50 UpperCAmelCase : Union[str, Any] = jax.device_count() UpperCAmelCase : int = num_samples * [prompt] UpperCAmelCase : Union[str, Any] = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Dict = replicate(__snake_case ) UpperCAmelCase : int = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Tuple = shard(__snake_case ) UpperCAmelCase : Tuple = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5E-1 def A ( self : int ) -> Dict: UpperCAmelCase , UpperCAmelCase : List[str] = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__snake_case ) UpperCAmelCase : Dict = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : Union[str, Any] = jax.random.PRNGKey(0 ) UpperCAmelCase : List[str] = 50 UpperCAmelCase : Union[str, Any] = jax.device_count() UpperCAmelCase : List[Any] = num_samples * [prompt] UpperCAmelCase : int = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Tuple = replicate(__snake_case ) UpperCAmelCase : List[Any] = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Optional[int] = shard(__snake_case ) UpperCAmelCase : Optional[Any] = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def A ( self : int ) -> Any: UpperCAmelCase , UpperCAmelCase : Dict = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa ) UpperCAmelCase : List[str] = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : List[str] = jax.random.PRNGKey(0 ) UpperCAmelCase : Union[str, Any] = 50 UpperCAmelCase : Optional[int] = jax.device_count() UpperCAmelCase : List[str] = num_samples * [prompt] UpperCAmelCase : Dict = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Tuple = replicate(__snake_case ) UpperCAmelCase : Any = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : str = shard(__snake_case ) UpperCAmelCase : Optional[int] = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def A ( self : Tuple ) -> Optional[Any]: UpperCAmelCase : int = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , set_alpha_to_one=__snake_case , steps_offset=1 , ) UpperCAmelCase , UpperCAmelCase : Tuple = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , scheduler=__snake_case , safety_checker=__snake_case , ) UpperCAmelCase : Tuple = scheduler.create_state() UpperCAmelCase : Dict = scheduler_state UpperCAmelCase : str = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : int = jax.random.PRNGKey(0 ) UpperCAmelCase : Union[str, Any] = 50 UpperCAmelCase : Optional[Any] = jax.device_count() UpperCAmelCase : Any = num_samples * [prompt] UpperCAmelCase : Dict = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : str = replicate(__snake_case ) UpperCAmelCase : List[str] = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Optional[int] = shard(__snake_case ) UpperCAmelCase : Dict = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5E-1 def A ( self : Any ) -> Tuple: UpperCAmelCase : List[Any] = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : Union[str, Any] = jax.device_count() UpperCAmelCase : List[Any] = num_samples * [prompt] UpperCAmelCase : str = jax.random.split(jax.random.PRNGKey(0 ) , __snake_case ) UpperCAmelCase , UpperCAmelCase : Tuple = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__snake_case , ) UpperCAmelCase : Dict = replicate(__snake_case ) UpperCAmelCase : Optional[Any] = pipeline.prepare_inputs(__snake_case ) UpperCAmelCase : List[str] = shard(__snake_case ) UpperCAmelCase : Any = pipeline(__snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase : Optional[int] = images[2, 0, 256, 10:17, 1] # With memory efficient attention UpperCAmelCase , UpperCAmelCase : Any = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__snake_case , use_memory_efficient_attention=__snake_case , ) UpperCAmelCase : int = replicate(__snake_case ) UpperCAmelCase : int = pipeline.prepare_inputs(__snake_case ) UpperCAmelCase : List[Any] = shard(__snake_case ) UpperCAmelCase : Optional[Any] = pipeline(__snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase : int = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
23
1
'''simple docstring''' import argparse from tax import checkpoints from transformers import AutoConfig, FlaxAutoModelForSeqaSeqLM def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ) -> List[Any]: UpperCAmelCase : List[str] = AutoConfig.from_pretrained(_lowerCAmelCase ) UpperCAmelCase : Tuple = FlaxAutoModelForSeqaSeqLM.from_config(config=_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = checkpoints.load_tax_checkpoint(_lowerCAmelCase ) UpperCAmelCase : Dict = '''wi_0''' in tax_model['''target''']['''encoder''']['''layers_0''']['''mlp'''] if config.model_type == "t5": UpperCAmelCase : str = '''SelfAttention''' if config.model_type == "longt5" and config.encoder_attention_type == "local": UpperCAmelCase : List[Any] = '''LocalSelfAttention''' elif config.model_type == "longt5" and config.encoder_attention_type == "transient-global": UpperCAmelCase : Optional[Any] = '''TransientGlobalSelfAttention''' else: raise ValueError( '''Given config is expected to have `model_type=\'t5\'`, or `model_type=\'longt5` with `encoder_attention_type`''' ''' attribute with a value from [\'local\', \'transient-global].''' ) # Encoder for layer_index in range(config.num_layers ): UpperCAmelCase : Dict = f"""layers_{str(_lowerCAmelCase )}""" # Self-Attention UpperCAmelCase : Any = tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''key''']['''kernel'''] UpperCAmelCase : Optional[Any] = tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''out''']['''kernel'''] UpperCAmelCase : List[str] = tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''query''']['''kernel'''] UpperCAmelCase : Dict = tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''value''']['''kernel'''] # Global input layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": UpperCAmelCase : List[Any] = tax_model['''target''']['''encoder'''][layer_name]['''attention''']['''T5LayerNorm_0''']['''scale'''] # Layer Normalization UpperCAmelCase : Optional[int] = tax_model['''target''']['''encoder'''][layer_name]['''pre_attention_layer_norm''']['''scale'''] if split_mlp_wi: UpperCAmelCase : Optional[Any] = tax_model['''target''']['''encoder'''][layer_name]['''mlp''']['''wi_0''']['''kernel'''] UpperCAmelCase : int = tax_model['''target''']['''encoder'''][layer_name]['''mlp''']['''wi_1''']['''kernel'''] else: UpperCAmelCase : List[Any] = tax_model['''target''']['''encoder'''][layer_name]['''mlp''']['''wi''']['''kernel'''] UpperCAmelCase : List[str] = tax_model['''target''']['''encoder'''][layer_name]['''mlp''']['''wo''']['''kernel'''] # Layer Normalization UpperCAmelCase : Tuple = tax_model['''target''']['''encoder'''][layer_name]['''pre_mlp_layer_norm''']['''scale'''] # Assigning UpperCAmelCase : List[Any] = flax_model.params['''encoder''']['''block'''][str(_lowerCAmelCase )]['''layer'''] UpperCAmelCase : Union[str, Any] = tax_attention_key UpperCAmelCase : int = tax_attention_out UpperCAmelCase : Union[str, Any] = tax_attention_query UpperCAmelCase : Any = tax_attention_value UpperCAmelCase : str = tax_attention_layer_norm # Global input layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": UpperCAmelCase : List[Any] = tax_global_layer_norm if split_mlp_wi: UpperCAmelCase : str = tax_mlp_wi_a UpperCAmelCase : int = tax_mlp_wi_a else: UpperCAmelCase : List[str] = tax_mlp_wi UpperCAmelCase : Any = tax_mlp_wo UpperCAmelCase : int = tax_mlp_layer_norm UpperCAmelCase : List[Any] = flax_model_encoder_layer_block # Only for layer 0: UpperCAmelCase : Tuple = tax_model['''target''']['''encoder''']['''relpos_bias''']['''rel_embedding'''].T UpperCAmelCase : Any = tax_encoder_rel_embedding # Side/global relative position_bias + layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": UpperCAmelCase : Tuple = tax_model['''target''']['''encoder''']['''side_relpos_bias''']['''rel_embedding'''].T UpperCAmelCase : int = tax_encoder_global_rel_embedding # Assigning UpperCAmelCase : int = tax_model['''target''']['''encoder''']['''encoder_norm''']['''scale'''] UpperCAmelCase : List[Any] = tax_encoder_norm # Decoder for layer_index in range(config.num_layers ): UpperCAmelCase : str = f"""layers_{str(_lowerCAmelCase )}""" # Self-Attention UpperCAmelCase : List[Any] = tax_model['''target''']['''decoder'''][layer_name]['''self_attention''']['''key''']['''kernel'''] UpperCAmelCase : int = tax_model['''target''']['''decoder'''][layer_name]['''self_attention''']['''out''']['''kernel'''] UpperCAmelCase : Dict = tax_model['''target''']['''decoder'''][layer_name]['''self_attention''']['''query''']['''kernel'''] UpperCAmelCase : Dict = tax_model['''target''']['''decoder'''][layer_name]['''self_attention''']['''value''']['''kernel'''] # Layer Normalization UpperCAmelCase : Dict = tax_model['''target''']['''decoder'''][layer_name]['''pre_self_attention_layer_norm'''][ '''scale''' ] # Encoder-Decoder-Attention UpperCAmelCase : Optional[int] = tax_model['''target''']['''decoder'''][layer_name]['''encoder_decoder_attention'''] UpperCAmelCase : List[Any] = tax_enc_dec_attention_module['''key''']['''kernel'''] UpperCAmelCase : Any = tax_enc_dec_attention_module['''out''']['''kernel'''] UpperCAmelCase : Optional[int] = tax_enc_dec_attention_module['''query''']['''kernel'''] UpperCAmelCase : str = tax_enc_dec_attention_module['''value''']['''kernel'''] # Layer Normalization UpperCAmelCase : Dict = tax_model['''target''']['''decoder'''][layer_name]['''pre_cross_attention_layer_norm''']['''scale'''] # MLP if split_mlp_wi: UpperCAmelCase : Dict = tax_model['''target''']['''decoder'''][layer_name]['''mlp''']['''wi_0''']['''kernel'''] UpperCAmelCase : Optional[int] = tax_model['''target''']['''decoder'''][layer_name]['''mlp''']['''wi_1''']['''kernel'''] else: UpperCAmelCase : Tuple = tax_model['''target''']['''decoder'''][layer_name]['''mlp''']['''wi''']['''kernel'''] UpperCAmelCase : List[str] = tax_model['''target''']['''decoder'''][layer_name]['''mlp''']['''wo''']['''kernel'''] # Layer Normalization UpperCAmelCase : List[Any] = tax_model['''target''']['''decoder'''][layer_name]['''pre_mlp_layer_norm''']['''scale'''] # Assigning UpperCAmelCase : List[Any] = flax_model.params['''decoder''']['''block'''][str(_lowerCAmelCase )]['''layer'''] UpperCAmelCase : Optional[int] = tax_attention_key UpperCAmelCase : Optional[int] = tax_attention_out UpperCAmelCase : Union[str, Any] = tax_attention_query UpperCAmelCase : List[str] = tax_attention_value UpperCAmelCase : List[str] = tax_pre_attention_layer_norm UpperCAmelCase : List[str] = tax_enc_dec_attention_key UpperCAmelCase : int = tax_enc_dec_attention_out UpperCAmelCase : Dict = tax_enc_dec_attention_query UpperCAmelCase : List[str] = tax_enc_dec_attention_value UpperCAmelCase : str = tax_cross_layer_norm if split_mlp_wi: UpperCAmelCase : Tuple = tax_mlp_wi_a UpperCAmelCase : int = tax_mlp_wi_a else: UpperCAmelCase : List[str] = tax_mlp_wi UpperCAmelCase : Union[str, Any] = tax_mlp_wo UpperCAmelCase : str = txa_mlp_layer_norm UpperCAmelCase : str = flax_model_decoder_layer_block # Decoder Normalization UpperCAmelCase : List[Any] = tax_model['''target''']['''decoder''']['''decoder_norm''']['''scale'''] UpperCAmelCase : Any = txa_decoder_norm # Only for layer 0: UpperCAmelCase : str = tax_model['''target''']['''decoder''']['''relpos_bias''']['''rel_embedding'''].T UpperCAmelCase : Optional[int] = tax_decoder_rel_embedding # Token Embeddings UpperCAmelCase : Union[str, Any] = tax_model['''target''']['''token_embedder''']['''embedding'''] UpperCAmelCase : str = txa_token_embeddings # LM Head (only in v1.1 and LongT5 checkpoints) if "logits_dense" in tax_model["target"]["decoder"]: UpperCAmelCase : Optional[Any] = tax_model['''target''']['''decoder''']['''logits_dense''']['''kernel'''] flax_model.save_pretrained(_lowerCAmelCase ) print('''T5X Model was sucessfully converted!''' ) if __name__ == "__main__": UpperCamelCase__: List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--t5x_checkpoint_path", default=None, type=str, required=True, help="Path the T5X checkpoint." ) parser.add_argument("--config_name", default=None, type=str, required=True, help="Config name of LongT5/T5 model.") parser.add_argument( "--flax_dump_folder_path", default=None, type=str, required=True, help="Path to the output FLAX model." ) UpperCamelCase__: List[Any] = parser.parse_args() convert_tax_checkpoint_to_flax(args.tax_checkpoint_path, args.config_name, args.flax_dump_folder_path)
23
'''simple docstring''' import random from .binary_exp_mod import bin_exp_mod def snake_case_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]=1000 ) -> int: if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd UpperCAmelCase : str = n - 1 UpperCAmelCase : List[Any] = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) UpperCAmelCase : List[str] = 0 while count < prec: UpperCAmelCase : int = random.randint(2 , n - 1 ) UpperCAmelCase : List[str] = bin_exp_mod(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if b != 1: UpperCAmelCase : int = True for _ in range(_lowerCAmelCase ): if b == n - 1: UpperCAmelCase : Dict = False break UpperCAmelCase : str = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": UpperCamelCase__: Optional[int] = abs(int(input("Enter bound : ").strip())) print("Here's the list of primes:") print(", ".join(str(i) for i in range(n + 1) if is_prime_big(i)))
23
1
'''simple docstring''' import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( "The converted tokenizer will be the `slow` tokenizer. To use the fast, update your `tokenizers` library and re-run the tokenizer conversion" ) UpperCamelCase__: str = None UpperCamelCase__: int = { "7B": 11008, "13B": 13824, "30B": 17920, "65B": 22016, "70B": 28672, } UpperCamelCase__: List[Any] = { "7B": 1, "7Bf": 1, "13B": 2, "13Bf": 2, "30B": 4, "65B": 8, "70B": 8, "70Bf": 8, } def snake_case_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple=1 , _lowerCAmelCase : List[Any]=256 ) -> Optional[Any]: return multiple_of * ((int(ffn_dim_multiplier * int(8 * n / 3 ) ) + multiple_of - 1) // multiple_of) def snake_case_ ( _lowerCAmelCase : List[str] ) -> str: with open(_lowerCAmelCase , '''r''' ) as f: return json.load(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any ) -> Optional[Any]: with open(_lowerCAmelCase , '''w''' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any=True ) -> List[Any]: os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : int = os.path.join(_lowerCAmelCase , '''tmp''' ) os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : List[str] = read_json(os.path.join(_lowerCAmelCase , '''params.json''' ) ) UpperCAmelCase : str = NUM_SHARDS[model_size] UpperCAmelCase : Any = params['''n_layers'''] UpperCAmelCase : str = params['''n_heads'''] UpperCAmelCase : Any = n_heads // num_shards UpperCAmelCase : List[str] = params['''dim'''] UpperCAmelCase : Optional[Any] = dim // n_heads UpperCAmelCase : str = 1_0_0_0_0.0 UpperCAmelCase : Optional[int] = 1.0 / (base ** (torch.arange(0 , _lowerCAmelCase , 2 ).float() / dims_per_head)) if "n_kv_heads" in params: UpperCAmelCase : Tuple = params['''n_kv_heads'''] # for GQA / MQA UpperCAmelCase : Optional[int] = n_heads_per_shard // num_key_value_heads UpperCAmelCase : Optional[Any] = dim // num_key_value_heads else: # compatibility with other checkpoints UpperCAmelCase : List[str] = n_heads UpperCAmelCase : Optional[int] = n_heads_per_shard UpperCAmelCase : List[str] = dim # permute for sliced rotary def permute(_lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]=n_heads , _lowerCAmelCase : int=dim , _lowerCAmelCase : Dict=dim ): return w.view(_lowerCAmelCase , dima // n_heads // 2 , 2 , _lowerCAmelCase ).transpose(1 , 2 ).reshape(_lowerCAmelCase , _lowerCAmelCase ) print(f"""Fetching all parameters from the checkpoint at {input_base_path}.""" ) # Load weights if model_size == "7B": # Not sharded # (The sharded implementation would also work, but this is simpler.) UpperCAmelCase : int = torch.load(os.path.join(_lowerCAmelCase , '''consolidated.00.pth''' ) , map_location='''cpu''' ) else: # Sharded UpperCAmelCase : Optional[Any] = [ torch.load(os.path.join(_lowerCAmelCase , f"""consolidated.{i:02d}.pth""" ) , map_location='''cpu''' ) for i in range(_lowerCAmelCase ) ] UpperCAmelCase : Any = 0 UpperCAmelCase : str = {'''weight_map''': {}} for layer_i in range(_lowerCAmelCase ): UpperCAmelCase : Optional[Any] = f"""pytorch_model-{layer_i + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase : Optional[int] = { f"""model.layers.{layer_i}.self_attn.q_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wq.weight"""] ), f"""model.layers.{layer_i}.self_attn.k_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wk.weight"""] ), f"""model.layers.{layer_i}.self_attn.v_proj.weight""": loaded[f"""layers.{layer_i}.attention.wv.weight"""], f"""model.layers.{layer_i}.self_attn.o_proj.weight""": loaded[f"""layers.{layer_i}.attention.wo.weight"""], f"""model.layers.{layer_i}.mlp.gate_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w1.weight"""], f"""model.layers.{layer_i}.mlp.down_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w2.weight"""], f"""model.layers.{layer_i}.mlp.up_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w3.weight"""], f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[f"""layers.{layer_i}.attention_norm.weight"""], f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[f"""layers.{layer_i}.ffn_norm.weight"""], } else: # Sharded # Note that attention.w{q,k,v,o}, feed_fordward.w[1,2,3], attention_norm.weight and ffn_norm.weight share # the same storage object, saving attention_norm and ffn_norm will save other weights too, which is # redundant as other weights will be stitched from multiple shards. To avoid that, they are cloned. UpperCAmelCase : List[str] = { f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.attention_norm.weight""" ].clone(), f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.ffn_norm.weight""" ].clone(), } UpperCAmelCase : Union[str, Any] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wq.weight"""].view(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase : Optional[Any] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wk.weight"""].view( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) UpperCAmelCase : str = torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wv.weight"""].view( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = torch.cat( [loaded[i][f"""layers.{layer_i}.attention.wo.weight"""] for i in range(_lowerCAmelCase )] , dim=1 ) UpperCAmelCase : Any = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w1.weight"""] for i in range(_lowerCAmelCase )] , dim=0 ) UpperCAmelCase : str = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w2.weight"""] for i in range(_lowerCAmelCase )] , dim=1 ) UpperCAmelCase : Tuple = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w3.weight"""] for i in range(_lowerCAmelCase )] , dim=0 ) UpperCAmelCase : Any = inv_freq for k, v in state_dict.items(): UpperCAmelCase : List[Any] = filename param_count += v.numel() torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase : Optional[int] = f"""pytorch_model-{n_layers + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase : str = { '''model.embed_tokens.weight''': loaded['''tok_embeddings.weight'''], '''model.norm.weight''': loaded['''norm.weight'''], '''lm_head.weight''': loaded['''output.weight'''], } else: UpperCAmelCase : Any = { '''model.norm.weight''': loaded[0]['''norm.weight'''], '''model.embed_tokens.weight''': torch.cat( [loaded[i]['''tok_embeddings.weight'''] for i in range(_lowerCAmelCase )] , dim=1 ), '''lm_head.weight''': torch.cat([loaded[i]['''output.weight'''] for i in range(_lowerCAmelCase )] , dim=0 ), } for k, v in state_dict.items(): UpperCAmelCase : Optional[int] = filename param_count += v.numel() torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , _lowerCAmelCase ) ) # Write configs UpperCAmelCase : Union[str, Any] = {'''total_size''': param_count * 2} write_json(_lowerCAmelCase , os.path.join(_lowerCAmelCase , '''pytorch_model.bin.index.json''' ) ) UpperCAmelCase : int = params['''ffn_dim_multiplier'''] if '''ffn_dim_multiplier''' in params else 1 UpperCAmelCase : Tuple = params['''multiple_of'''] if '''multiple_of''' in params else 256 UpperCAmelCase : Any = LlamaConfig( hidden_size=_lowerCAmelCase , intermediate_size=compute_intermediate_size(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) , num_attention_heads=params['''n_heads'''] , num_hidden_layers=params['''n_layers'''] , rms_norm_eps=params['''norm_eps'''] , num_key_value_heads=_lowerCAmelCase , ) config.save_pretrained(_lowerCAmelCase ) # Make space so we can load the model properly now. del state_dict del loaded gc.collect() print('''Loading the checkpoint in a Llama model.''' ) UpperCAmelCase : Optional[int] = LlamaForCausalLM.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa , low_cpu_mem_usage=_lowerCAmelCase ) # Avoid saving this as part of the config. del model.config._name_or_path print('''Saving in the Transformers format.''' ) model.save_pretrained(_lowerCAmelCase , safe_serialization=_lowerCAmelCase ) shutil.rmtree(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any] ) -> List[str]: # Initialize the tokenizer based on the `spm` model UpperCAmelCase : Dict = LlamaTokenizer if LlamaTokenizerFast is None else LlamaTokenizerFast print(f"""Saving a {tokenizer_class.__name__} to {tokenizer_path}.""" ) UpperCAmelCase : List[Any] = tokenizer_class(_lowerCAmelCase ) tokenizer.save_pretrained(_lowerCAmelCase ) def snake_case_ ( ) -> List[Any]: UpperCAmelCase : int = argparse.ArgumentParser() parser.add_argument( '''--input_dir''' , help='''Location of LLaMA weights, which contains tokenizer.model and model folders''' , ) parser.add_argument( '''--model_size''' , choices=['''7B''', '''7Bf''', '''13B''', '''13Bf''', '''30B''', '''65B''', '''70B''', '''70Bf''', '''tokenizer_only'''] , ) parser.add_argument( '''--output_dir''' , help='''Location to write HF model and tokenizer''' , ) parser.add_argument('''--safe_serialization''' , type=_lowerCAmelCase , help='''Whether or not to save using `safetensors`.''' ) UpperCAmelCase : List[Any] = parser.parse_args() if args.model_size != "tokenizer_only": write_model( model_path=args.output_dir , input_base_path=os.path.join(args.input_dir , args.model_size ) , model_size=args.model_size , safe_serialization=args.safe_serialization , ) UpperCAmelCase : Optional[int] = os.path.join(args.input_dir , '''tokenizer.model''' ) write_tokenizer(args.output_dir , _lowerCAmelCase ) if __name__ == "__main__": main()
23
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__: Optional[int] = logging.get_logger(__name__) def snake_case_ ( _lowerCAmelCase : Optional[int] ) -> Optional[int]: UpperCAmelCase : Tuple = DPTConfig(embedding_type='''hybrid''' ) if "large" in checkpoint_url: UpperCAmelCase : Tuple = 1024 UpperCAmelCase : List[Any] = 4096 UpperCAmelCase : str = 24 UpperCAmelCase : List[Any] = 16 UpperCAmelCase : str = [5, 11, 17, 23] UpperCAmelCase : List[Any] = [256, 512, 1024, 1024] UpperCAmelCase : Tuple = (1, 384, 384) if "nyu" or "midas" in checkpoint_url: UpperCAmelCase : Optional[Any] = 768 UpperCAmelCase : Tuple = [1, 1, 1, 0.5] UpperCAmelCase : int = [256, 512, 768, 768] UpperCAmelCase : Any = 150 UpperCAmelCase : Tuple = 16 UpperCAmelCase : Any = (1, 384, 384) UpperCAmelCase : Optional[Any] = False UpperCAmelCase : Tuple = '''project''' if "ade" in checkpoint_url: UpperCAmelCase : Any = True UpperCAmelCase : str = 768 UpperCAmelCase : Optional[int] = [1, 1, 1, 0.5] UpperCAmelCase : List[Any] = 150 UpperCAmelCase : List[Any] = 16 UpperCAmelCase : str = '''huggingface/label-files''' UpperCAmelCase : Tuple = '''ade20k-id2label.json''' UpperCAmelCase : Any = json.load(open(cached_download(hf_hub_url(_lowerCAmelCase , _lowerCAmelCase , repo_type='''dataset''' ) ) , '''r''' ) ) UpperCAmelCase : Optional[Any] = {int(_lowerCAmelCase ): v for k, v in idalabel.items()} UpperCAmelCase : List[Any] = idalabel UpperCAmelCase : Optional[int] = {v: k for k, v in idalabel.items()} UpperCAmelCase : Union[str, Any] = [1, 150, 480, 480] return config, expected_shape def snake_case_ ( _lowerCAmelCase : Union[str, Any] ) -> int: UpperCAmelCase : List[str] = ['''pretrained.model.head.weight''', '''pretrained.model.head.bias'''] for k in ignore_keys: state_dict.pop(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Tuple ) -> Any: if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): UpperCAmelCase : Tuple = name.replace('''pretrained.model''' , '''dpt.encoder''' ) if "pretrained.model" in name: UpperCAmelCase : Union[str, Any] = name.replace('''pretrained.model''' , '''dpt.embeddings''' ) if "patch_embed" in name: UpperCAmelCase : int = name.replace('''patch_embed''' , '''''' ) if "pos_embed" in name: UpperCAmelCase : Tuple = name.replace('''pos_embed''' , '''position_embeddings''' ) if "attn.proj" in name: UpperCAmelCase : Any = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "proj" in name and "project" not in name: UpperCAmelCase : str = name.replace('''proj''' , '''projection''' ) if "blocks" in name: UpperCAmelCase : Any = name.replace('''blocks''' , '''layer''' ) if "mlp.fc1" in name: UpperCAmelCase : Optional[int] = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: UpperCAmelCase : Optional[Any] = name.replace('''mlp.fc2''' , '''output.dense''' ) if "norm1" in name and "backbone" not in name: UpperCAmelCase : Dict = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name and "backbone" not in name: UpperCAmelCase : Tuple = name.replace('''norm2''' , '''layernorm_after''' ) if "scratch.output_conv" in name: UpperCAmelCase : Tuple = name.replace('''scratch.output_conv''' , '''head''' ) if "scratch" in name: UpperCAmelCase : str = name.replace('''scratch''' , '''neck''' ) if "layer1_rn" in name: UpperCAmelCase : Dict = name.replace('''layer1_rn''' , '''convs.0''' ) if "layer2_rn" in name: UpperCAmelCase : int = name.replace('''layer2_rn''' , '''convs.1''' ) if "layer3_rn" in name: UpperCAmelCase : Tuple = name.replace('''layer3_rn''' , '''convs.2''' ) if "layer4_rn" in name: UpperCAmelCase : int = name.replace('''layer4_rn''' , '''convs.3''' ) if "refinenet" in name: UpperCAmelCase : List[str] = int(name[len('''neck.refinenet''' ) : len('''neck.refinenet''' ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 UpperCAmelCase : str = name.replace(f"""refinenet{layer_idx}""" , f"""fusion_stage.layers.{abs(layer_idx-4 )}""" ) if "out_conv" in name: UpperCAmelCase : List[str] = name.replace('''out_conv''' , '''projection''' ) if "resConfUnit1" in name: UpperCAmelCase : Union[str, Any] = name.replace('''resConfUnit1''' , '''residual_layer1''' ) if "resConfUnit2" in name: UpperCAmelCase : Any = name.replace('''resConfUnit2''' , '''residual_layer2''' ) if "conv1" in name: UpperCAmelCase : Optional[int] = name.replace('''conv1''' , '''convolution1''' ) if "conv2" in name: UpperCAmelCase : Tuple = name.replace('''conv2''' , '''convolution2''' ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: UpperCAmelCase : Dict = name.replace('''pretrained.act_postprocess1.0.project.0''' , '''neck.reassemble_stage.readout_projects.0.0''' ) if "pretrained.act_postprocess2.0.project.0" in name: UpperCAmelCase : int = name.replace('''pretrained.act_postprocess2.0.project.0''' , '''neck.reassemble_stage.readout_projects.1.0''' ) if "pretrained.act_postprocess3.0.project.0" in name: UpperCAmelCase : Any = name.replace('''pretrained.act_postprocess3.0.project.0''' , '''neck.reassemble_stage.readout_projects.2.0''' ) if "pretrained.act_postprocess4.0.project.0" in name: UpperCAmelCase : Optional[Any] = name.replace('''pretrained.act_postprocess4.0.project.0''' , '''neck.reassemble_stage.readout_projects.3.0''' ) # resize blocks if "pretrained.act_postprocess1.3" in name: UpperCAmelCase : List[Any] = name.replace('''pretrained.act_postprocess1.3''' , '''neck.reassemble_stage.layers.0.projection''' ) if "pretrained.act_postprocess1.4" in name: UpperCAmelCase : Any = name.replace('''pretrained.act_postprocess1.4''' , '''neck.reassemble_stage.layers.0.resize''' ) if "pretrained.act_postprocess2.3" in name: UpperCAmelCase : Optional[int] = name.replace('''pretrained.act_postprocess2.3''' , '''neck.reassemble_stage.layers.1.projection''' ) if "pretrained.act_postprocess2.4" in name: UpperCAmelCase : str = name.replace('''pretrained.act_postprocess2.4''' , '''neck.reassemble_stage.layers.1.resize''' ) if "pretrained.act_postprocess3.3" in name: UpperCAmelCase : List[str] = name.replace('''pretrained.act_postprocess3.3''' , '''neck.reassemble_stage.layers.2.projection''' ) if "pretrained.act_postprocess4.3" in name: UpperCAmelCase : Tuple = name.replace('''pretrained.act_postprocess4.3''' , '''neck.reassemble_stage.layers.3.projection''' ) if "pretrained.act_postprocess4.4" in name: UpperCAmelCase : int = name.replace('''pretrained.act_postprocess4.4''' , '''neck.reassemble_stage.layers.3.resize''' ) if "pretrained" in name: UpperCAmelCase : Optional[int] = name.replace('''pretrained''' , '''dpt''' ) if "bn" in name: UpperCAmelCase : Dict = name.replace('''bn''' , '''batch_norm''' ) if "head" in name: UpperCAmelCase : Any = name.replace('''head''' , '''head.head''' ) if "encoder.norm" in name: UpperCAmelCase : Optional[int] = name.replace('''encoder.norm''' , '''layernorm''' ) if "auxlayer" in name: UpperCAmelCase : Union[str, Any] = name.replace('''auxlayer''' , '''auxiliary_head.head''' ) if "backbone" in name: UpperCAmelCase : List[Any] = name.replace('''backbone''' , '''backbone.bit.encoder''' ) if ".." in name: UpperCAmelCase : Optional[int] = name.replace('''..''' , '''.''' ) if "stem.conv" in name: UpperCAmelCase : Optional[Any] = name.replace('''stem.conv''' , '''bit.embedder.convolution''' ) if "blocks" in name: UpperCAmelCase : Optional[int] = name.replace('''blocks''' , '''layers''' ) if "convolution" in name and "backbone" in name: UpperCAmelCase : List[Any] = name.replace('''convolution''' , '''conv''' ) if "layer" in name and "backbone" in name: UpperCAmelCase : List[str] = name.replace('''layer''' , '''layers''' ) if "backbone.bit.encoder.bit" in name: UpperCAmelCase : List[Any] = name.replace('''backbone.bit.encoder.bit''' , '''backbone.bit''' ) if "embedder.conv" in name: UpperCAmelCase : List[Any] = name.replace('''embedder.conv''' , '''embedder.convolution''' ) if "backbone.bit.encoder.stem.norm" in name: UpperCAmelCase : Tuple = name.replace('''backbone.bit.encoder.stem.norm''' , '''backbone.bit.embedder.norm''' ) return name def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : List[Any] ) -> Optional[Any]: for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase : Optional[int] = state_dict.pop(f"""dpt.encoder.layer.{i}.attn.qkv.weight""" ) UpperCAmelCase : Tuple = state_dict.pop(f"""dpt.encoder.layer.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase : Tuple = in_proj_weight[: config.hidden_size, :] UpperCAmelCase : int = in_proj_bias[: config.hidden_size] UpperCAmelCase : List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase : List[str] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase : str = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase : Union[str, Any] = in_proj_bias[-config.hidden_size :] def snake_case_ ( ) -> List[str]: UpperCAmelCase : Optional[int] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' UpperCAmelCase : Optional[int] = Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ) return im @torch.no_grad() def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[str] ) -> Any: UpperCAmelCase , UpperCAmelCase : int = get_dpt_config(_lowerCAmelCase ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") UpperCAmelCase : List[Any] = torch.load(_lowerCAmelCase , map_location='''cpu''' ) # remove certain keys remove_ignore_keys_(_lowerCAmelCase ) # rename keys for key in state_dict.copy().keys(): UpperCAmelCase : Any = state_dict.pop(_lowerCAmelCase ) UpperCAmelCase : List[Any] = val # read in qkv matrices read_in_q_k_v(_lowerCAmelCase , _lowerCAmelCase ) # load HuggingFace model UpperCAmelCase : Optional[Any] = DPTForSemanticSegmentation(_lowerCAmelCase ) if '''ade''' in checkpoint_url else DPTForDepthEstimation(_lowerCAmelCase ) model.load_state_dict(_lowerCAmelCase ) model.eval() # Check outputs on an image UpperCAmelCase : int = 480 if '''ade''' in checkpoint_url else 384 UpperCAmelCase : List[Any] = DPTImageProcessor(size=_lowerCAmelCase ) UpperCAmelCase : Dict = prepare_img() UpperCAmelCase : Optional[int] = image_processor(_lowerCAmelCase , return_tensors='''pt''' ) # forward pass UpperCAmelCase : Any = model(**_lowerCAmelCase ).logits if '''ade''' in checkpoint_url else model(**_lowerCAmelCase ).predicted_depth if show_prediction: UpperCAmelCase : Dict = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode='''bicubic''' , align_corners=_lowerCAmelCase , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 255 ).show() if pytorch_dump_folder_path is not None: Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(_lowerCAmelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_lowerCAmelCase ) if push_to_hub: model.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) image_processor.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) if __name__ == "__main__": UpperCamelCase__: Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint_url", default="https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt", type=str, help="URL of the original DPT checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=False, help="Path to the output PyTorch model directory.", ) parser.add_argument( "--push_to_hub", action="store_true", ) parser.add_argument( "--model_name", default="dpt-large", type=str, help="Name of the model, in case you're pushing to the hub.", ) parser.add_argument( "--show_prediction", action="store_true", ) UpperCamelCase__: Tuple = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
23
1
'''simple docstring''' import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->UnCLIP class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = 42 lowerCamelCase__ = None def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any=0.9_9_9 , _lowerCAmelCase : Optional[Any]="cosine" , ) -> List[Any]: if alpha_transform_type == "cosine": def alpha_bar_fn(_lowerCAmelCase : List[Any] ): return math.cos((t + 0.0_0_8) / 1.0_0_8 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(_lowerCAmelCase : Any ): return math.exp(t * -1_2.0 ) else: raise ValueError(f"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) UpperCAmelCase : Any = [] for i in range(_lowerCAmelCase ): UpperCAmelCase : Dict = i / num_diffusion_timesteps UpperCAmelCase : List[Any] = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(_lowerCAmelCase ) / alpha_bar_fn(_lowerCAmelCase ) , _lowerCAmelCase ) ) return torch.tensor(_lowerCAmelCase , dtype=torch.floataa ) class SCREAMING_SNAKE_CASE( A__ , A__ ): """simple docstring""" @register_to_config def __init__( self : Union[str, Any] , __snake_case : int = 1000 , __snake_case : str = "fixed_small_log" , __snake_case : bool = True , __snake_case : Optional[float] = 1.0 , __snake_case : str = "epsilon" , __snake_case : str = "squaredcos_cap_v2" , ) -> str: if beta_schedule != "squaredcos_cap_v2": raise ValueError('''UnCLIPScheduler only supports `beta_schedule`: \'squaredcos_cap_v2\'''' ) UpperCAmelCase : Union[str, Any] = betas_for_alpha_bar(__snake_case ) UpperCAmelCase : List[Any] = 1.0 - self.betas UpperCAmelCase : List[Any] = torch.cumprod(self.alphas , dim=0 ) UpperCAmelCase : List[Any] = torch.tensor(1.0 ) # standard deviation of the initial noise distribution UpperCAmelCase : int = 1.0 # setable values UpperCAmelCase : str = None UpperCAmelCase : Union[str, Any] = torch.from_numpy(np.arange(0 , __snake_case )[::-1].copy() ) UpperCAmelCase : Optional[Any] = variance_type def A ( self : Optional[Any] , __snake_case : torch.FloatTensor , __snake_case : Optional[int] = None ) -> torch.FloatTensor: return sample def A ( self : Dict , __snake_case : int , __snake_case : Union[str, torch.device] = None ) -> Optional[Any]: UpperCAmelCase : List[str] = num_inference_steps UpperCAmelCase : Optional[int] = (self.config.num_train_timesteps - 1) / (self.num_inference_steps - 1) UpperCAmelCase : str = (np.arange(0 , __snake_case ) * step_ratio).round()[::-1].copy().astype(np.intaa ) UpperCAmelCase : Optional[int] = torch.from_numpy(__snake_case ).to(__snake_case ) def A ( self : Any , __snake_case : str , __snake_case : List[str]=None , __snake_case : str=None , __snake_case : List[str]=None ) -> int: if prev_timestep is None: UpperCAmelCase : Optional[int] = t - 1 UpperCAmelCase : Any = self.alphas_cumprod[t] UpperCAmelCase : Any = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase : Any = 1 - alpha_prod_t UpperCAmelCase : int = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase : Optional[int] = self.betas[t] else: UpperCAmelCase : List[str] = 1 - alpha_prod_t / alpha_prod_t_prev # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample UpperCAmelCase : int = beta_prod_t_prev / beta_prod_t * beta if variance_type is None: UpperCAmelCase : Any = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small_log": UpperCAmelCase : Optional[Any] = torch.log(torch.clamp(__snake_case , min=1E-20 ) ) UpperCAmelCase : int = torch.exp(0.5 * variance ) elif variance_type == "learned_range": # NOTE difference with DDPM scheduler UpperCAmelCase : Tuple = variance.log() UpperCAmelCase : List[Any] = beta.log() UpperCAmelCase : List[Any] = (predicted_variance + 1) / 2 UpperCAmelCase : List[Any] = frac * max_log + (1 - frac) * min_log return variance def A ( self : Union[str, Any] , __snake_case : torch.FloatTensor , __snake_case : int , __snake_case : torch.FloatTensor , __snake_case : Optional[int] = None , __snake_case : int=None , __snake_case : bool = True , ) -> Union[UnCLIPSchedulerOutput, Tuple]: UpperCAmelCase : Optional[int] = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type == "learned_range": UpperCAmelCase , UpperCAmelCase : Tuple = torch.split(__snake_case , sample.shape[1] , dim=1 ) else: UpperCAmelCase : int = None # 1. compute alphas, betas if prev_timestep is None: UpperCAmelCase : Optional[Any] = t - 1 UpperCAmelCase : str = self.alphas_cumprod[t] UpperCAmelCase : Dict = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase : Union[str, Any] = 1 - alpha_prod_t UpperCAmelCase : Dict = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase : Tuple = self.betas[t] UpperCAmelCase : Optional[Any] = self.alphas[t] else: UpperCAmelCase : List[str] = 1 - alpha_prod_t / alpha_prod_t_prev UpperCAmelCase : Union[str, Any] = 1 - beta # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": UpperCAmelCase : Dict = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": UpperCAmelCase : Union[str, Any] = model_output else: raise ValueError( F"""prediction_type given as {self.config.prediction_type} must be one of `epsilon` or `sample`""" ''' for the UnCLIPScheduler.''' ) # 3. Clip "predicted x_0" if self.config.clip_sample: UpperCAmelCase : int = torch.clamp( __snake_case , -self.config.clip_sample_range , self.config.clip_sample_range ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase : Union[str, Any] = (alpha_prod_t_prev ** 0.5 * beta) / beta_prod_t UpperCAmelCase : Optional[int] = alpha ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase : Union[str, Any] = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise UpperCAmelCase : int = 0 if t > 0: UpperCAmelCase : Union[str, Any] = randn_tensor( model_output.shape , dtype=model_output.dtype , generator=__snake_case , device=model_output.device ) UpperCAmelCase : Optional[Any] = self._get_variance( __snake_case , predicted_variance=__snake_case , prev_timestep=__snake_case , ) if self.variance_type == "fixed_small_log": UpperCAmelCase : Tuple = variance elif self.variance_type == "learned_range": UpperCAmelCase : List[Any] = (0.5 * variance).exp() else: raise ValueError( F"""variance_type given as {self.variance_type} must be one of `fixed_small_log` or `learned_range`""" ''' for the UnCLIPScheduler.''' ) UpperCAmelCase : Dict = variance * variance_noise UpperCAmelCase : Tuple = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return UnCLIPSchedulerOutput(prev_sample=__snake_case , pred_original_sample=__snake_case ) def A ( self : Optional[Any] , __snake_case : torch.FloatTensor , __snake_case : torch.FloatTensor , __snake_case : torch.IntTensor , ) -> torch.FloatTensor: # Make sure alphas_cumprod and timestep have same device and dtype as original_samples UpperCAmelCase : Dict = self.alphas_cumprod.to(device=original_samples.device , dtype=original_samples.dtype ) UpperCAmelCase : Tuple = timesteps.to(original_samples.device ) UpperCAmelCase : int = alphas_cumprod[timesteps] ** 0.5 UpperCAmelCase : Optional[int] = sqrt_alpha_prod.flatten() while len(sqrt_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase : Optional[int] = sqrt_alpha_prod.unsqueeze(-1 ) UpperCAmelCase : Any = (1 - alphas_cumprod[timesteps]) ** 0.5 UpperCAmelCase : Any = sqrt_one_minus_alpha_prod.flatten() while len(sqrt_one_minus_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase : int = sqrt_one_minus_alpha_prod.unsqueeze(-1 ) UpperCAmelCase : Any = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples
23
'''simple docstring''' # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers UpperCamelCase__: Optional[int] = "3" print("Python version:", sys.version) print("transformers version:", transformers.__version__) try: import torch print("Torch version:", torch.__version__) print("Cuda available:", torch.cuda.is_available()) print("Cuda version:", torch.version.cuda) print("CuDNN version:", torch.backends.cudnn.version()) print("Number of GPUs available:", torch.cuda.device_count()) print("NCCL version:", torch.cuda.nccl.version()) except ImportError: print("Torch version:", None) try: import deepspeed print("DeepSpeed version:", deepspeed.__version__) except ImportError: print("DeepSpeed version:", None) try: import tensorflow as tf print("TensorFlow version:", tf.__version__) print("TF GPUs available:", bool(tf.config.list_physical_devices("GPU"))) print("Number of TF GPUs available:", len(tf.config.list_physical_devices("GPU"))) except ImportError: print("TensorFlow version:", None)
23
1
'''simple docstring''' from math import factorial UpperCamelCase__: Union[str, Any] = {str(d): factorial(d) for d in range(10)} def snake_case_ ( _lowerCAmelCase : int ) -> int: return sum(DIGIT_FACTORIAL[d] for d in str(_lowerCAmelCase ) ) def snake_case_ ( ) -> int: UpperCAmelCase : Tuple = 7 * factorial(9 ) + 1 return sum(i for i in range(3 , _lowerCAmelCase ) if sum_of_digit_factorial(_lowerCAmelCase ) == i ) if __name__ == "__main__": print(F"{solution() = }")
23
'''simple docstring''' from __future__ import annotations def snake_case_ ( _lowerCAmelCase : list[int | float] , _lowerCAmelCase : int , _lowerCAmelCase : int ) -> int | float: if len(_lowerCAmelCase ) == 0: raise ValueError('''find_max() arg is an empty sequence''' ) if ( left >= len(_lowerCAmelCase ) or left < -len(_lowerCAmelCase ) or right >= len(_lowerCAmelCase ) or right < -len(_lowerCAmelCase ) ): raise IndexError('''list index out of range''' ) if left == right: return nums[left] UpperCAmelCase : List[Any] = (left + right) >> 1 # the middle UpperCAmelCase : Optional[Any] = find_max(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # find max in range[left, mid] UpperCAmelCase : Dict = find_max(_lowerCAmelCase , mid + 1 , _lowerCAmelCase ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
23
1
import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel UpperCAmelCase__ = HfApi() UpperCAmelCase__ = {} # fmt: off UpperCAmelCase__ = torch.tensor([ -0.7515, -1.6883, 0.2420, 0.0300, 0.6347, 1.3433, -1.1743, -3.7467, 1.2342, -2.2485, 0.4636, 0.8076, -0.7991, 0.3969, 0.8498, 0.9189, -1.8887, -3.3522, 0.7639, 0.2040, 0.6271, -2.7148, -1.6316, 3.0839, 0.3186, 0.2721, -0.9759, -1.2461, 2.6257, 1.3557 ]) UpperCAmelCase__ = torch.tensor([ -2.3639, -2.5344, 0.0054, -0.6674, 1.5990, 1.0158, 0.3124, -2.1436, 1.8795, -2.5429, -0.1566, -0.3973, 1.2490, 2.6447, 1.2283, -0.5208, -2.8154, -3.5119, 2.3838, 1.2033, 1.7201, -2.1256, -1.4576, 2.7948, 2.4204, -0.9752, -1.2546, 0.8027, 3.2758, 3.1365 ]) UpperCAmelCase__ = torch.tensor([ -0.6531, -0.6891, -0.3172, -0.5375, -0.9140, -0.5367, -0.1175, -0.7869, -0.3808, -0.4513, -0.2098, -0.0083, 0.3183, 0.5140, 0.2247, -0.1304, -0.1302, -0.2802, -0.2084, -0.2025, -0.4967, -0.4873, -0.0861, 0.6925, 0.0250, 0.1290, -0.1543, 0.6316, 1.0460, 1.4943 ]) UpperCAmelCase__ = torch.tensor([ 0.0911, 0.1107, 0.0182, 0.0435, -0.0805, -0.0608, 0.0381, 0.2172, -0.0280, 0.1327, -0.0299, -0.0255, -0.0050, -0.1170, -0.1046, 0.0309, 0.1367, 0.1728, -0.0533, -0.0748, -0.0534, 0.1624, 0.0384, -0.1805, -0.0707, 0.0642, 0.0220, -0.0134, -0.1333, -0.1505 ]) UpperCAmelCase__ = torch.tensor([ 0.1321, 0.1337, 0.0440, 0.0622, -0.0591, -0.0370, 0.0503, 0.2133, -0.0177, 0.1415, -0.0116, -0.0112, 0.0044, -0.0980, -0.0789, 0.0395, 0.1502, 0.1785, -0.0488, -0.0514, -0.0404, 0.1539, 0.0454, -0.1559, -0.0665, 0.0659, 0.0383, -0.0005, -0.1266, -0.1386 ]) UpperCAmelCase__ = torch.tensor([ 0.1154, 0.1218, 0.0307, 0.0526, -0.0711, -0.0541, 0.0366, 0.2078, -0.0267, 0.1317, -0.0226, -0.0193, -0.0014, -0.1055, -0.0902, 0.0330, 0.1391, 0.1709, -0.0562, -0.0693, -0.0560, 0.1482, 0.0381, -0.1683, -0.0681, 0.0661, 0.0331, -0.0046, -0.1268, -0.1431 ]) UpperCAmelCase__ = torch.tensor([ 0.1192, 0.1240, 0.0414, 0.0606, -0.0557, -0.0412, 0.0430, 0.2042, -0.0200, 0.1385, -0.0115, -0.0132, 0.0017, -0.0965, -0.0802, 0.0398, 0.1433, 0.1747, -0.0458, -0.0533, -0.0407, 0.1545, 0.0419, -0.1574, -0.0645, 0.0626, 0.0341, -0.0010, -0.1199, -0.1390 ]) UpperCAmelCase__ = torch.tensor([ 0.1075, 0.1074, 0.0205, 0.0431, -0.0774, -0.0607, 0.0298, 0.2042, -0.0320, 0.1267, -0.0281, -0.0250, -0.0064, -0.1091, -0.0946, 0.0290, 0.1328, 0.1650, -0.0580, -0.0738, -0.0586, 0.1440, 0.0337, -0.1746, -0.0712, 0.0605, 0.0250, -0.0099, -0.1316, -0.1473 ]) UpperCAmelCase__ = torch.tensor([ -1.4572, -2.0481, -0.0414, -0.6005, 1.4136, 0.5848, 0.4028, -2.7330, 1.2212, -2.1228, 0.2155, 0.4039, 0.7662, 2.0535, 0.7477, -0.3243, -2.1758, -2.7648, 1.6947, 0.7026, 1.2338, -1.6078, -0.8682, 2.2810, 1.8574, -0.5718, -0.5586, -0.0186, 2.3415, 2.1251]) UpperCAmelCase__ = torch.tensor([ -1.3690, -1.9720, -0.4090, -0.6966, 1.4660, 0.9938, -0.1385, -2.7324, 0.7736, -1.8917, 0.2923, 0.4293, 0.1693, 1.4112, 1.1887, -0.3181, -2.2160, -2.6381, 1.3170, 0.8163, 0.9240, -1.6544, -0.6099, 2.5259, 1.6430, -0.9090, -0.9392, -0.0126, 2.4268, 2.3266 ]) UpperCAmelCase__ = torch.tensor([ -1.3525, -1.9628, -0.3956, -0.6860, 1.4664, 1.0014, -0.1259, -2.7212, 0.7772, -1.8811, 0.2996, 0.4388, 0.1704, 1.4029, 1.1701, -0.3027, -2.2053, -2.6287, 1.3350, 0.8131, 0.9274, -1.6292, -0.6098, 2.5131, 1.6505, -0.8958, -0.9298, -0.0151, 2.4257, 2.3355 ]) UpperCAmelCase__ = torch.tensor([ -2.0585, -2.7897, -0.2850, -0.8940, 1.9052, 0.5702, 0.6345, -3.8959, 1.5932, -3.2319, 0.1974, 0.0287, 1.7566, 2.6543, 0.8387, -0.5351, -3.2736, -4.3375, 2.9029, 1.6390, 1.4640, -2.1701, -1.9013, 2.9341, 3.4981, -0.6255, -1.1644, -0.1591, 3.7097, 3.2066 ]) UpperCAmelCase__ = torch.tensor([ -2.3139, -2.5594, -0.0197, -0.6785, 1.7001, 1.1606, 0.3075, -2.1740, 1.8071, -2.5630, -0.0926, -0.3811, 1.2116, 2.6246, 1.2731, -0.5398, -2.8153, -3.6140, 2.3893, 1.3262, 1.6258, -2.1856, -1.3267, 2.8395, 2.3779, -1.0623, -1.2468, 0.8959, 3.3367, 3.2243 ]) UpperCAmelCase__ = torch.tensor([ -2.0628, -2.7667, -0.2089, -0.8263, 2.0539, 0.5992, 0.6495, -3.8336, 1.6025, -3.2817, 0.1721, -0.0633, 1.7516, 2.7039, 0.8100, -0.5908, -3.2113, -4.4343, 2.9257, 1.3632, 1.5562, -2.1489, -1.9894, 3.0560, 3.3396, -0.7328, -1.0417, 0.0383, 3.7093, 3.2343 ]) UpperCAmelCase__ = torch.tensor([ -1.4574, -2.0569, -0.0473, -0.6117, 1.4018, 0.5769, 0.4129, -2.7344, 1.2241, -2.1397, 0.2000, 0.3937, 0.7616, 2.0453, 0.7324, -0.3391, -2.1746, -2.7744, 1.6963, 0.6921, 1.2187, -1.6172, -0.8877, 2.2439, 1.8471, -0.5839, -0.5605, -0.0464, 2.3250, 2.1219 ]) # fmt: on UpperCAmelCase__ = api.list_models(filter="diffusers") for mod in models: if "google" in mod.author or mod.modelId == "CompVis/ldm-celebahq-256": UpperCAmelCase__ = "/home/patrick/google_checkpoints/" + mod.modelId.split("/")[-1] print(f"""Started running {mod.modelId}!!!""") if mod.modelId.startswith("CompVis"): UpperCAmelCase__ = UNetaDModel.from_pretrained(local_checkpoint, subfolder="unet") else: UpperCAmelCase__ = UNetaDModel.from_pretrained(local_checkpoint) torch.manual_seed(0) random.seed(0) UpperCAmelCase__ = torch.randn(1, model.config.in_channels, model.config.sample_size, model.config.sample_size) UpperCAmelCase__ = torch.tensor([10] * noise.shape[0]) with torch.no_grad(): UpperCAmelCase__ = model(noise, time_step).sample assert torch.allclose( logits[0, 0, 0, :30], results["_".join("_".join(mod.modelId.split("/")).split("-"))], atol=1E-3 ) print(f"""{mod.modelId} has passed successfully!!!""")
0
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = 42 lowerCamelCase__ = 42 def __init__( self : Union[str, Any] , __snake_case : UNetaDModel , __snake_case : ScoreSdeVeScheduler ) -> int: super().__init__() self.register_modules(unet=__snake_case , scheduler=__snake_case ) @torch.no_grad() def __call__( self : Optional[int] , __snake_case : int = 1 , __snake_case : int = 2000 , __snake_case : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __snake_case : Optional[str] = "pil" , __snake_case : bool = True , **__snake_case : Optional[int] , ) -> Union[ImagePipelineOutput, Tuple]: UpperCAmelCase : str = self.unet.config.sample_size UpperCAmelCase : Union[str, Any] = (batch_size, 3, img_size, img_size) UpperCAmelCase : int = self.unet UpperCAmelCase : Any = randn_tensor(__snake_case , generator=__snake_case ) * self.scheduler.init_noise_sigma UpperCAmelCase : List[Any] = sample.to(self.device ) self.scheduler.set_timesteps(__snake_case ) self.scheduler.set_sigmas(__snake_case ) for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): UpperCAmelCase : Any = self.scheduler.sigmas[i] * torch.ones(shape[0] , device=self.device ) # correction step for _ in range(self.scheduler.config.correct_steps ): UpperCAmelCase : Union[str, Any] = self.unet(__snake_case , __snake_case ).sample UpperCAmelCase : Optional[Any] = self.scheduler.step_correct(__snake_case , __snake_case , generator=__snake_case ).prev_sample # prediction step UpperCAmelCase : Optional[Any] = model(__snake_case , __snake_case ).sample UpperCAmelCase : List[str] = self.scheduler.step_pred(__snake_case , __snake_case , __snake_case , generator=__snake_case ) UpperCAmelCase , UpperCAmelCase : Optional[Any] = output.prev_sample, output.prev_sample_mean UpperCAmelCase : int = sample_mean.clamp(0 , 1 ) UpperCAmelCase : Union[str, Any] = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCAmelCase : Optional[Any] = self.numpy_to_pil(__snake_case ) if not return_dict: return (sample,) return ImagePipelineOutput(images=__snake_case )
23
0
'''simple docstring''' import json import logging import os import sys from pathlib import Path import finetune_rag from transformers.file_utils import is_apex_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, require_ray, require_torch_gpu, require_torch_multi_gpu, ) logging.basicConfig(level=logging.DEBUG) SCREAMING_SNAKE_CASE_: Optional[int] =logging.getLogger() SCREAMING_SNAKE_CASE_: Union[str, Any] =logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class __A ( UpperCamelCase__ ): def _lowercase (self : Optional[Any] , __a : str ): os.makedirs(__a , exist_ok=__a ) UpperCAmelCase_ = {"source": "What is love ?", "target": "life"} UpperCAmelCase_ = {"train": 12, "val": 2, "test": 2} for split in ["train", "test", "val"]: for field in ["source", "target"]: UpperCAmelCase_ = "\n".join([contents[field]] * n_lines[split] ) with open(os.path.join(__a , f"""{split}.{field}""" ) , "w" ) as f: f.write(__a ) def _lowercase (self : Optional[int] , __a : int , __a : str = "pytorch" ): UpperCAmelCase_ = self.get_auto_remove_tmp_dir() UpperCAmelCase_ = os.path.join(__a , "output" ) UpperCAmelCase_ = os.path.join(__a , "data" ) self._create_dummy_data(data_dir=__a ) UpperCAmelCase_ = f""" --data_dir {data_dir} \ --output_dir {output_dir} \ --model_name_or_path facebook/rag-sequence-base \ --model_type rag_sequence \ --do_train \ --do_predict \ --n_val -1 \ --val_check_interval 1.0 \ --train_batch_size 2 \ --eval_batch_size 1 \ --max_source_length 25 \ --max_target_length 25 \ --val_max_target_length 25 \ --test_max_target_length 25 \ --label_smoothing 0.1 \ --dropout 0.1 \ --attention_dropout 0.1 \ --weight_decay 0.001 \ --adam_epsilon 1e-08 \ --max_grad_norm 0.1 \ --lr_scheduler polynomial \ --learning_rate 3e-04 \ --num_train_epochs 1 \ --warmup_steps 4 \ --gradient_accumulation_steps 1 \ --distributed-port 8787 \ --use_dummy_dataset 1 \ --distributed_retriever {distributed_retriever} \ """.split() if gpus > 0: testargs.append(f"""--gpus={gpus}""" ) if is_apex_available(): testargs.append("--fp16" ) else: testargs.append("--gpus=0" ) testargs.append("--distributed_backend=ddp_cpu" ) testargs.append("--num_processes=2" ) UpperCAmelCase_ = [sys.executable, str(Path(finetune_rag.__file__ ).resolve() )] + testargs execute_subprocess_async(__a , env=self.get_env() ) UpperCAmelCase_ = os.path.join(__a , "metrics.json" ) with open(__a ) as f: UpperCAmelCase_ = json.load(__a ) return result @require_torch_gpu def _lowercase (self : Optional[int] ): UpperCAmelCase_ = self._run_finetune(gpus=1 ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 ) @require_torch_multi_gpu def _lowercase (self : Dict ): UpperCAmelCase_ = self._run_finetune(gpus=2 ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 ) @require_torch_gpu @require_ray def _lowercase (self : Optional[int] ): UpperCAmelCase_ = self._run_finetune(gpus=1 , distributed_retriever="ray" ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 ) @require_torch_multi_gpu @require_ray def _lowercase (self : Any ): UpperCAmelCase_ = self._run_finetune(gpus=1 , distributed_retriever="ray" ) self.assertGreaterEqual(result["test"][0]["test_avg_em"] , 0.2 )
1
'''simple docstring''' import warnings from contextlib import contextmanager from ....processing_utils import ProcessorMixin class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = """MCTCTFeatureExtractor""" lowerCamelCase__ = """AutoTokenizer""" def __init__( self : Dict , __snake_case : Optional[int] , __snake_case : List[str] ) -> str: super().__init__(__snake_case , __snake_case ) UpperCAmelCase : List[Any] = self.feature_extractor UpperCAmelCase : Union[str, Any] = False def __call__( self : Any , *__snake_case : List[str] , **__snake_case : Any ) -> List[Any]: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*__snake_case , **__snake_case ) if "raw_speech" in kwargs: warnings.warn('''Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.''' ) UpperCAmelCase : int = kwargs.pop('''raw_speech''' ) else: UpperCAmelCase : Union[str, Any] = kwargs.pop('''audio''' , __snake_case ) UpperCAmelCase : Optional[Any] = kwargs.pop('''sampling_rate''' , __snake_case ) UpperCAmelCase : Dict = kwargs.pop('''text''' , __snake_case ) if len(__snake_case ) > 0: UpperCAmelCase : Any = args[0] UpperCAmelCase : Optional[int] = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if audio is not None: UpperCAmelCase : List[str] = self.feature_extractor(__snake_case , *__snake_case , sampling_rate=__snake_case , **__snake_case ) if text is not None: UpperCAmelCase : int = self.tokenizer(__snake_case , **__snake_case ) if text is None: return inputs elif audio is None: return encodings else: UpperCAmelCase : str = encodings['''input_ids'''] return inputs def A ( self : List[Any] , *__snake_case : List[Any] , **__snake_case : List[Any] ) -> str: return self.tokenizer.batch_decode(*__snake_case , **__snake_case ) def A ( self : List[Any] , *__snake_case : int , **__snake_case : Optional[int] ) -> Any: # For backward compatibility if self._in_target_context_manager: return self.current_processor.pad(*__snake_case , **__snake_case ) UpperCAmelCase : List[Any] = kwargs.pop('''input_features''' , __snake_case ) UpperCAmelCase : Optional[Any] = kwargs.pop('''labels''' , __snake_case ) if len(__snake_case ) > 0: UpperCAmelCase : List[str] = args[0] UpperCAmelCase : List[Any] = args[1:] if input_features is not None: UpperCAmelCase : Tuple = self.feature_extractor.pad(__snake_case , *__snake_case , **__snake_case ) if labels is not None: UpperCAmelCase : Optional[int] = self.tokenizer.pad(__snake_case , **__snake_case ) if labels is None: return input_features elif input_features is None: return labels else: UpperCAmelCase : List[str] = labels['''input_ids'''] return input_features def A ( self : Union[str, Any] , *__snake_case : Optional[Any] , **__snake_case : Optional[int] ) -> Optional[Any]: return self.tokenizer.decode(*__snake_case , **__snake_case ) @contextmanager def A ( self : Any ) -> Optional[int]: warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your audio inputs, or in a separate call.''' ) UpperCAmelCase : Dict = True UpperCAmelCase : List[Any] = self.tokenizer yield UpperCAmelCase : Tuple = self.feature_extractor UpperCAmelCase : List[Any] = False
23
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCamelCase : List[Any] = { 'configuration_pegasus_x': ['PEGASUS_X_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PegasusXConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase : Union[str, Any] = [ 'PEGASUS_X_PRETRAINED_MODEL_ARCHIVE_LIST', 'PegasusXForConditionalGeneration', 'PegasusXModel', 'PegasusXPreTrainedModel', ] if TYPE_CHECKING: from .configuration_pegasus_x import PEGASUS_X_PRETRAINED_CONFIG_ARCHIVE_MAP, PegasusXConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pegasus_x import ( PEGASUS_X_PRETRAINED_MODEL_ARCHIVE_LIST, PegasusXForConditionalGeneration, PegasusXModel, PegasusXPreTrainedModel, ) else: import sys lowerCamelCase : List[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
2
'''simple docstring''' from math import isclose, sqrt def snake_case_ ( _lowerCAmelCase : float , _lowerCAmelCase : float , _lowerCAmelCase : float ) -> tuple[float, float, float]: UpperCAmelCase : Optional[int] = point_y / 4 / point_x UpperCAmelCase : str = 2 * normal_gradient / (1 + normal_gradient * normal_gradient) UpperCAmelCase : Any = (1 - normal_gradient * normal_gradient) / ( 1 + normal_gradient * normal_gradient ) UpperCAmelCase : Union[str, Any] = (sa - ca * incoming_gradient) / (ca + sa * incoming_gradient) # to find the next point, solve the simultaeneous equations: # y^2 + 4x^2 = 100 # y - b = m * (x - a) # ==> A x^2 + B x + C = 0 UpperCAmelCase : Union[str, Any] = outgoing_gradient**2 + 4 UpperCAmelCase : Dict = 2 * outgoing_gradient * (point_y - outgoing_gradient * point_x) UpperCAmelCase : List[Any] = (point_y - outgoing_gradient * point_x) ** 2 - 100 UpperCAmelCase : List[str] = ( -linear_term - sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) UpperCAmelCase : Optional[int] = ( -linear_term + sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) # two solutions, one of which is our input point UpperCAmelCase : Optional[Any] = x_minus if isclose(_lowerCAmelCase , _lowerCAmelCase ) else x_plus UpperCAmelCase : Union[str, Any] = point_y + outgoing_gradient * (next_x - point_x) return next_x, next_y, outgoing_gradient def snake_case_ ( _lowerCAmelCase : float = 1.4 , _lowerCAmelCase : float = -9.6 ) -> int: UpperCAmelCase : int = 0 UpperCAmelCase : float = first_x_coord UpperCAmelCase : float = first_y_coord UpperCAmelCase : float = (1_0.1 - point_y) / (0.0 - point_x) while not (-0.0_1 <= point_x <= 0.0_1 and point_y > 0): UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : str = next_point(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) num_reflections += 1 return num_reflections if __name__ == "__main__": print(F"{solution() = }")
23
0
'''simple docstring''' import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / 'utils')) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class A ( unittest.TestCase ): def __lowerCAmelCase ( self ) -> Union[str, Any]: """simple docstring""" A : Dict = mock.Mock() A : str = 500 A : List[str] = {} A : List[Any] = HTTPError A : str = {} # Download this model to make sure it's in the cache. A : Tuple = BertTokenizer.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=SCREAMING_SNAKE_CASE ) as mock_head: A : Dict = BertTokenizer.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def __lowerCAmelCase ( self ) -> Dict: """simple docstring""" A : str = mock.Mock() A : Optional[Any] = 500 A : str = {} A : Tuple = HTTPError A : str = {} # Download this model to make sure it's in the cache. A : str = GPTaTokenizerFast.from_pretrained('''gpt2''' ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=SCREAMING_SNAKE_CASE ) as mock_head: A : Union[str, Any] = GPTaTokenizerFast.from_pretrained('''gpt2''' ) # This check we did call the fake head request mock_head.assert_called() def __lowerCAmelCase ( self ) -> Any: """simple docstring""" try: A : Dict = tempfile.mktemp() with open(SCREAMING_SNAKE_CASE , '''wb''' ) as f: http_get('''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''' , SCREAMING_SNAKE_CASE ) A : List[str] = AlbertTokenizer.from_pretrained(SCREAMING_SNAKE_CASE ) finally: os.remove(SCREAMING_SNAKE_CASE ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile('''tokenizer.json''' ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open('''tokenizer.json''' , '''wb''' ) as f: http_get('''https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json''' , SCREAMING_SNAKE_CASE ) A : List[Any] = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 1000 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove('''tokenizer.json''' ) def __lowerCAmelCase ( self ) -> Any: """simple docstring""" A : int = AlbertTokenizer.from_pretrained('''https://huggingface.co/albert-base-v1/resolve/main/spiece.model''' ) @is_staging_test class A ( unittest.TestCase ): __magic_name__ = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''bla''', '''blou'''] @classmethod def __lowerCAmelCase ( cls ) -> Optional[int]: """simple docstring""" A : List[str] = TOKEN HfFolder.save_token(SCREAMING_SNAKE_CASE ) @classmethod def __lowerCAmelCase ( cls ) -> List[str]: """simple docstring""" try: delete_repo(token=cls._token , repo_id='''test-tokenizer''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-tokenizer-org''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''test-dynamic-tokenizer''' ) except HTTPError: pass def __lowerCAmelCase ( self ) -> List[str]: """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: A : str = os.path.join(SCREAMING_SNAKE_CASE , '''vocab.txt''' ) with open(SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) A : List[Any] = BertTokenizer(SCREAMING_SNAKE_CASE ) tokenizer.push_to_hub('''test-tokenizer''' , use_auth_token=self._token ) A : Optional[Any] = BertTokenizer.from_pretrained(F'{USER}/test-tokenizer' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='''test-tokenizer''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(SCREAMING_SNAKE_CASE , repo_id='''test-tokenizer''' , push_to_hub=SCREAMING_SNAKE_CASE , use_auth_token=self._token ) A : int = BertTokenizer.from_pretrained(F'{USER}/test-tokenizer' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def __lowerCAmelCase ( self ) -> Any: """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: A : Union[str, Any] = os.path.join(SCREAMING_SNAKE_CASE , '''vocab.txt''' ) with open(SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) A : Optional[int] = BertTokenizer(SCREAMING_SNAKE_CASE ) tokenizer.push_to_hub('''valid_org/test-tokenizer-org''' , use_auth_token=self._token ) A : List[Any] = BertTokenizer.from_pretrained('''valid_org/test-tokenizer-org''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-tokenizer-org''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( SCREAMING_SNAKE_CASE , repo_id='''valid_org/test-tokenizer-org''' , push_to_hub=SCREAMING_SNAKE_CASE , use_auth_token=self._token ) A : str = BertTokenizer.from_pretrained('''valid_org/test-tokenizer-org''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def __lowerCAmelCase ( self ) -> str: """simple docstring""" CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: A : List[Any] = os.path.join(SCREAMING_SNAKE_CASE , '''vocab.txt''' ) with open(SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) A : Optional[int] = CustomTokenizer(SCREAMING_SNAKE_CASE ) # No fast custom tokenizer tokenizer.push_to_hub('''test-dynamic-tokenizer''' , use_auth_token=self._token ) A : Dict = AutoTokenizer.from_pretrained(F'{USER}/test-dynamic-tokenizer' , trust_remote_code=SCREAMING_SNAKE_CASE ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , '''CustomTokenizer''' ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: A : Optional[Any] = os.path.join(SCREAMING_SNAKE_CASE , '''vocab.txt''' ) with open(SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) A : str = BertTokenizerFast.from_pretrained(SCREAMING_SNAKE_CASE ) bert_tokenizer.save_pretrained(SCREAMING_SNAKE_CASE ) A : Any = CustomTokenizerFast.from_pretrained(SCREAMING_SNAKE_CASE ) tokenizer.push_to_hub('''test-dynamic-tokenizer''' , use_auth_token=self._token ) A : Union[str, Any] = AutoTokenizer.from_pretrained(F'{USER}/test-dynamic-tokenizer' , trust_remote_code=SCREAMING_SNAKE_CASE ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , '''CustomTokenizerFast''' ) A : str = AutoTokenizer.from_pretrained( F'{USER}/test-dynamic-tokenizer' , use_fast=SCREAMING_SNAKE_CASE , trust_remote_code=SCREAMING_SNAKE_CASE ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , '''CustomTokenizer''' ) class A ( unittest.TestCase ): def __lowerCAmelCase ( self ) -> Dict: """simple docstring""" A : Optional[Any] = Trie() trie.add('''Hello 友達''' ) self.assertEqual(trie.data , {'''H''': {'''e''': {'''l''': {'''l''': {'''o''': {''' ''': {'''友''': {'''達''': {'''''': 1}}}}}}}}} ) trie.add('''Hello''' ) trie.data self.assertEqual(trie.data , {'''H''': {'''e''': {'''l''': {'''l''': {'''o''': {'''''': 1, ''' ''': {'''友''': {'''達''': {'''''': 1}}}}}}}}} ) def __lowerCAmelCase ( self ) -> List[str]: """simple docstring""" A : List[str] = Trie() self.assertEqual(trie.split('''[CLS] This is a extra_id_100''' ) , ['''[CLS] This is a extra_id_100'''] ) trie.add('''[CLS]''' ) trie.add('''extra_id_1''' ) trie.add('''extra_id_100''' ) self.assertEqual(trie.split('''[CLS] This is a extra_id_100''' ) , ['''[CLS]''', ''' This is a ''', '''extra_id_100'''] ) def __lowerCAmelCase ( self ) -> Tuple: """simple docstring""" A : Dict = Trie() trie.add('''A''' ) self.assertEqual(trie.split('''ABC''' ) , ['''A''', '''BC'''] ) self.assertEqual(trie.split('''BCA''' ) , ['''BC''', '''A'''] ) def __lowerCAmelCase ( self ) -> Dict: """simple docstring""" A : List[Any] = Trie() trie.add('''TOKEN]''' ) trie.add('''[SPECIAL_TOKEN]''' ) self.assertEqual(trie.split('''This is something [SPECIAL_TOKEN]''' ) , ['''This is something ''', '''[SPECIAL_TOKEN]'''] ) def __lowerCAmelCase ( self ) -> List[str]: """simple docstring""" A : str = Trie() trie.add('''A''' ) trie.add('''P''' ) trie.add('''[SPECIAL_TOKEN]''' ) self.assertEqual(trie.split('''This is something [SPECIAL_TOKEN]''' ) , ['''This is something ''', '''[SPECIAL_TOKEN]'''] ) def __lowerCAmelCase ( self ) -> Union[str, Any]: """simple docstring""" A : Any = Trie() trie.add('''AB''' ) trie.add('''B''' ) trie.add('''C''' ) self.assertEqual(trie.split('''ABC''' ) , ['''AB''', '''C'''] ) def __lowerCAmelCase ( self ) -> Any: """simple docstring""" A : List[str] = Trie() trie.add('''ABC''' ) trie.add('''B''' ) trie.add('''CD''' ) self.assertEqual(trie.split('''ABCD''' ) , ['''ABC''', '''D'''] ) def __lowerCAmelCase ( self ) -> Optional[int]: """simple docstring""" A : Tuple = Trie() A : Optional[int] = trie.cut_text('''ABC''' , [0, 0, 2, 1, 2, 3] ) self.assertEqual(SCREAMING_SNAKE_CASE , ['''AB''', '''C'''] )
3
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase__: str = { "configuration_lxmert": ["LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "LxmertConfig"], "tokenization_lxmert": ["LxmertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: int = ["LxmertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: Union[str, Any] = [ "LxmertEncoder", "LxmertForPreTraining", "LxmertForQuestionAnswering", "LxmertModel", "LxmertPreTrainedModel", "LxmertVisualFeatureEncoder", "LxmertXLayer", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: int = [ "TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFLxmertForPreTraining", "TFLxmertMainLayer", "TFLxmertModel", "TFLxmertPreTrainedModel", "TFLxmertVisualFeatureEncoder", ] if TYPE_CHECKING: from .configuration_lxmert import LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, LxmertConfig from .tokenization_lxmert import LxmertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_lxmert_fast import LxmertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lxmert import ( LxmertEncoder, LxmertForPreTraining, LxmertForQuestionAnswering, LxmertModel, LxmertPreTrainedModel, LxmertVisualFeatureEncoder, LxmertXLayer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_lxmert import ( TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFLxmertForPreTraining, TFLxmertMainLayer, TFLxmertModel, TFLxmertPreTrainedModel, TFLxmertVisualFeatureEncoder, ) else: import sys UpperCamelCase__: Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
0
'''simple docstring''' import enum import warnings from ..tokenization_utils import TruncationStrategy from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING __snake_case =logging.get_logger(__name__) class UpperCAmelCase_ ( enum.Enum ): lowerCamelCase : Optional[int] = 0 lowerCamelCase : str = 1 @add_end_docstrings(__lowercase ) class UpperCAmelCase_ ( __lowercase ): lowerCamelCase : Tuple = '''generated''' def __init__( self : Optional[int] , *UpperCAmelCase__ : Optional[Any] , **UpperCAmelCase__ : Optional[int] ) -> Tuple: super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ ) self.check_model_type( TF_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING if self.framework == 'tf' else MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING ) def __UpperCAmelCase ( self : Dict , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : Dict=None , UpperCAmelCase__ : List[str]=None , UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : Any=None , UpperCAmelCase__ : Dict=None , **UpperCAmelCase__ : List[str] , ) -> Union[str, Any]: lowerCAmelCase = {} if truncation is not None: lowerCAmelCase = truncation lowerCAmelCase = generate_kwargs lowerCAmelCase = {} if return_tensors is not None and return_type is None: lowerCAmelCase = ReturnType.TENSORS if return_tensors else ReturnType.TEXT if return_type is not None: lowerCAmelCase = return_type if clean_up_tokenization_spaces is not None: lowerCAmelCase = clean_up_tokenization_spaces if stop_sequence is not None: lowerCAmelCase = self.tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) if len(UpperCAmelCase__ ) > 1: warnings.warn( 'Stopping on a multiple token sequence is not yet supported on transformers. The first token of' ' the stop sequence will be used as the stop sequence string in the interim.' ) lowerCAmelCase = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def __UpperCAmelCase ( self : List[Any] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int ) -> List[Any]: return True def __UpperCAmelCase ( self : Optional[int] , *UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : List[Any] ) -> Optional[int]: lowerCAmelCase = self.model.config.prefix if self.model.config.prefix is not None else '' if isinstance(args[0] , UpperCAmelCase__ ): if self.tokenizer.pad_token_id is None: raise ValueError('Please make sure that the tokenizer has a pad_token_id when using a batch input' ) lowerCAmelCase = ([prefix + arg for arg in args[0]],) lowerCAmelCase = True elif isinstance(args[0] , UpperCAmelCase__ ): lowerCAmelCase = (prefix + args[0],) lowerCAmelCase = False else: raise ValueError( F''' `args[0]`: {args[0]} have the wrong format. The should be either of type `str` or type `list`''' ) lowerCAmelCase = self.tokenizer(*UpperCAmelCase__ , padding=UpperCAmelCase__ , truncation=UpperCAmelCase__ , return_tensors=self.framework ) # This is produced by tokenizers but is an invalid generate kwargs if "token_type_ids" in inputs: del inputs["token_type_ids"] return inputs def __call__( self : Optional[Any] , *UpperCAmelCase__ : List[Any] , **UpperCAmelCase__ : str ) -> List[Any]: lowerCAmelCase = super().__call__(*UpperCAmelCase__ , **UpperCAmelCase__ ) if ( isinstance(args[0] , UpperCAmelCase__ ) and all(isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) for el in args[0] ) and all(len(UpperCAmelCase__ ) == 1 for res in result ) ): return [res[0] for res in result] return result def __UpperCAmelCase ( self : Optional[Any] , UpperCAmelCase__ : Any , UpperCAmelCase__ : Optional[Any]=TruncationStrategy.DO_NOT_TRUNCATE , **UpperCAmelCase__ : int ) -> int: lowerCAmelCase = self._parse_and_tokenize(UpperCAmelCase__ , truncation=UpperCAmelCase__ , **UpperCAmelCase__ ) return inputs def __UpperCAmelCase ( self : int , UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : str ) -> Tuple: if self.framework == "pt": lowerCAmelCase , lowerCAmelCase = model_inputs['input_ids'].shape elif self.framework == "tf": lowerCAmelCase , lowerCAmelCase = tf.shape(model_inputs['input_ids'] ).numpy() lowerCAmelCase = generate_kwargs.get('min_length' , self.model.config.min_length ) lowerCAmelCase = generate_kwargs.get('max_length' , self.model.config.max_length ) self.check_inputs(UpperCAmelCase__ , generate_kwargs['min_length'] , generate_kwargs['max_length'] ) lowerCAmelCase = self.model.generate(**UpperCAmelCase__ , **UpperCAmelCase__ ) lowerCAmelCase = output_ids.shape[0] if self.framework == "pt": lowerCAmelCase = output_ids.reshape(UpperCAmelCase__ , out_b // in_b , *output_ids.shape[1:] ) elif self.framework == "tf": lowerCAmelCase = tf.reshape(UpperCAmelCase__ , (in_b, out_b // in_b, *output_ids.shape[1:]) ) return {"output_ids": output_ids} def __UpperCAmelCase ( self : List[str] , UpperCAmelCase__ : int , UpperCAmelCase__ : List[Any]=ReturnType.TEXT , UpperCAmelCase__ : str=False ) -> int: lowerCAmelCase = [] for output_ids in model_outputs["output_ids"][0]: if return_type == ReturnType.TENSORS: lowerCAmelCase = {F'''{self.return_name}_token_ids''': output_ids} elif return_type == ReturnType.TEXT: lowerCAmelCase = { F'''{self.return_name}_text''': self.tokenizer.decode( UpperCAmelCase__ , skip_special_tokens=UpperCAmelCase__ , clean_up_tokenization_spaces=UpperCAmelCase__ , ) } records.append(UpperCAmelCase__ ) return records @add_end_docstrings(__lowercase ) class UpperCAmelCase_ ( __lowercase ): lowerCamelCase : Union[str, Any] = '''summary''' def __call__( self : Any , *UpperCAmelCase__ : int , **UpperCAmelCase__ : str ) -> Union[str, Any]: return super().__call__(*UpperCAmelCase__ , **UpperCAmelCase__ ) def __UpperCAmelCase ( self : Optional[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int ) -> bool: if max_length < min_length: logger.warning(F'''Your min_length={min_length} must be inferior than your max_length={max_length}.''' ) if input_length < max_length: logger.warning( F'''Your max_length is set to {max_length}, but your input_length is only {input_length}. Since this is ''' 'a summarization task, where outputs shorter than the input are typically wanted, you might ' F'''consider decreasing max_length manually, e.g. summarizer(\'...\', max_length={input_length//2})''' ) @add_end_docstrings(__lowercase ) class UpperCAmelCase_ ( __lowercase ): lowerCamelCase : Union[str, Any] = '''translation''' def __UpperCAmelCase ( self : str , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int ) -> Union[str, Any]: if input_length > 0.9 * max_length: logger.warning( F'''Your input_length: {input_length} is bigger than 0.9 * max_length: {max_length}. You might consider ''' 'increasing your max_length manually, e.g. translator(\'...\', max_length=400)' ) return True def __UpperCAmelCase ( self : Union[str, Any] , *UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Optional[int]=TruncationStrategy.DO_NOT_TRUNCATE , UpperCAmelCase__ : str=None , UpperCAmelCase__ : Optional[int]=None ) -> str: if getattr(self.tokenizer , '_build_translation_inputs' , UpperCAmelCase__ ): return self.tokenizer._build_translation_inputs( *UpperCAmelCase__ , return_tensors=self.framework , truncation=UpperCAmelCase__ , src_lang=UpperCAmelCase__ , tgt_lang=UpperCAmelCase__ ) else: return super()._parse_and_tokenize(*UpperCAmelCase__ , truncation=UpperCAmelCase__ ) def __UpperCAmelCase ( self : Optional[int] , UpperCAmelCase__ : List[Any]=None , UpperCAmelCase__ : Union[str, Any]=None , **UpperCAmelCase__ : int ) -> Tuple: lowerCAmelCase , lowerCAmelCase , lowerCAmelCase = super()._sanitize_parameters(**UpperCAmelCase__ ) if src_lang is not None: lowerCAmelCase = src_lang if tgt_lang is not None: lowerCAmelCase = tgt_lang if src_lang is None and tgt_lang is None: # Backward compatibility, direct arguments use is preferred. lowerCAmelCase = kwargs.get('task' , self.task ) lowerCAmelCase = task.split('_' ) if task and len(UpperCAmelCase__ ) == 4: # translation, XX, to YY lowerCAmelCase = items[1] lowerCAmelCase = items[3] return preprocess_params, forward_params, postprocess_params def __call__( self : Optional[int] , *UpperCAmelCase__ : int , **UpperCAmelCase__ : Any ) -> List[Any]: return super().__call__(*UpperCAmelCase__ , **UpperCAmelCase__ )
4
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE( A__ , A__ , A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = AltDiffusionPipeline lowerCamelCase__ = TEXT_TO_IMAGE_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS def A ( self : Dict ) -> int: torch.manual_seed(0 ) UpperCAmelCase : str = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) UpperCAmelCase : Dict = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) UpperCAmelCase : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) UpperCAmelCase : Dict = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5002 , ) UpperCAmelCase : List[Any] = CLIPTextModel(__snake_case ) UpperCAmelCase : Optional[Any] = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) UpperCAmelCase : Optional[int] = 77 UpperCAmelCase : Optional[int] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def A ( self : Optional[Any] , __snake_case : Dict , __snake_case : List[str]=0 ) -> Union[str, Any]: if str(__snake_case ).startswith('''mps''' ): UpperCAmelCase : str = torch.manual_seed(__snake_case ) else: UpperCAmelCase : Tuple = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) UpperCAmelCase : Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def A ( self : Union[str, Any] ) -> List[str]: super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def A ( self : Tuple ) -> List[str]: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def A ( self : Tuple ) -> Optional[int]: UpperCAmelCase : List[str] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase : Any = self.get_dummy_components() torch.manual_seed(0 ) UpperCAmelCase : Optional[int] = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase : List[str] = RobertaSeriesModelWithTransformation(__snake_case ) UpperCAmelCase : str = text_encoder UpperCAmelCase : Optional[int] = AltDiffusionPipeline(**__snake_case ) UpperCAmelCase : str = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : Optional[int] = self.get_dummy_inputs(__snake_case ) UpperCAmelCase : Optional[int] = '''A photo of an astronaut''' UpperCAmelCase : List[Any] = alt_pipe(**__snake_case ) UpperCAmelCase : Optional[Any] = output.images UpperCAmelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : List[str] = np.array( [0.5_74_81_62, 0.60_44_71_45, 0.48_82_12_17, 0.50_10_06_36, 0.5_43_11_85, 0.45_76_36_83, 0.49_65_76_96, 0.48_13_27_33, 0.47_57_30_93] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A ( self : int ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase : int = self.get_dummy_components() UpperCAmelCase : int = PNDMScheduler(skip_prk_steps=__snake_case ) torch.manual_seed(0 ) UpperCAmelCase : int = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase : Union[str, Any] = RobertaSeriesModelWithTransformation(__snake_case ) UpperCAmelCase : Union[str, Any] = text_encoder UpperCAmelCase : Optional[int] = AltDiffusionPipeline(**__snake_case ) UpperCAmelCase : Dict = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : int = self.get_dummy_inputs(__snake_case ) UpperCAmelCase : Optional[int] = alt_pipe(**__snake_case ) UpperCAmelCase : Optional[int] = output.images UpperCAmelCase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : Optional[int] = np.array( [0.51_60_50_93, 0.5_70_72_41, 0.47_36_55_07, 0.50_57_88_86, 0.5_63_38_77, 0.4_64_25_03, 0.5_18_20_81, 0.48_76_34_84, 0.49_08_42_37] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : str ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : List[Any] ) -> Any: # make sure here that pndm scheduler skips prk UpperCAmelCase : List[Any] = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''' , safety_checker=__snake_case ) UpperCAmelCase : Tuple = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : List[Any] = '''A painting of a squirrel eating a burger''' UpperCAmelCase : Any = torch.manual_seed(0 ) UpperCAmelCase : Optional[int] = alt_pipe([prompt] , generator=__snake_case , guidance_scale=6.0 , num_inference_steps=20 , output_type='''np''' ) UpperCAmelCase : Dict = output.images UpperCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase : List[str] = np.array([0.10_10, 0.08_00, 0.07_94, 0.08_85, 0.08_43, 0.07_62, 0.07_69, 0.07_29, 0.05_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A ( self : Tuple ) -> int: UpperCAmelCase : int = DDIMScheduler.from_pretrained('''BAAI/AltDiffusion''' , subfolder='''scheduler''' ) UpperCAmelCase : Tuple = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''' , scheduler=__snake_case , safety_checker=__snake_case ) UpperCAmelCase : Dict = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : Tuple = '''A painting of a squirrel eating a burger''' UpperCAmelCase : Optional[int] = torch.manual_seed(0 ) UpperCAmelCase : List[Any] = alt_pipe([prompt] , generator=__snake_case , num_inference_steps=2 , output_type='''numpy''' ) UpperCAmelCase : Dict = output.images UpperCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase : Union[str, Any] = np.array([0.40_19, 0.40_52, 0.38_10, 0.41_19, 0.39_16, 0.39_82, 0.46_51, 0.41_95, 0.53_23] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
23
0
import json import multiprocessing import os import re from collections import defaultdict import torch from accelerate import Accelerator from accelerate.utils import set_seed from arguments import HumanEvalArguments from datasets import load_dataset, load_metric from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from tqdm import tqdm import transformers from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, StoppingCriteria, StoppingCriteriaList UpperCAmelCase__ = ['''\nclass''', '''\ndef''', '''\n#''', '''\n@''', '''\nprint''', '''\nif'''] class lowerCamelCase__ ( lowerCAmelCase): def __init__(self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase=None , UpperCAmelCase=1 ) -> Dict: _lowercase =tokenizer _lowercase =dataset _lowercase =len(UpperCAmelCase ) if n_tasks is None else n_tasks _lowercase =n_copies def __iter__(self ) -> Optional[Any]: _lowercase =[] for task in range(self.n_tasks ): # without strip, the model generate commented codes ... prompts.append(self.tokenizer.eos_token + self.dataset[task]['''prompt'''].strip() ) _lowercase =self.tokenizer(UpperCAmelCase , padding=UpperCAmelCase , return_tensors='''pt''' ) for task in range(self.n_tasks ): for _ in range(self.n_copies ): yield { "ids": outputs.input_ids[task], "task_id": task, "input_len": outputs.attention_mask[task].sum(), } class lowerCamelCase__ ( lowerCAmelCase): def __init__(self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Optional[Any]: _lowercase =start_length _lowercase =eof_strings _lowercase =tokenizer def __call__(self , UpperCAmelCase , UpperCAmelCase , **UpperCAmelCase ) -> Dict: _lowercase =self.tokenizer.batch_decode(input_ids[:, self.start_length :] ) _lowercase =[] for decoded_generation in decoded_generations: done.append(any(stop_string in decoded_generation for stop_string in self.eof_strings ) ) return all(UpperCAmelCase ) def UpperCAmelCase_ ( __snake_case ) -> Optional[Any]: """simple docstring""" _lowercase =re.split('''(%s)''' % '''|'''.join(__snake_case ) , __snake_case ) # last string should be "" return "".join(string_list[:-2] ) def UpperCAmelCase_ ( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case , __snake_case=20 , **__snake_case ) -> Tuple: """simple docstring""" _lowercase =defaultdict(__snake_case ) # dict of list of generated tokens for step, batch in tqdm(enumerate(__snake_case ) ): with torch.no_grad(): _lowercase =batch['''ids'''].shape[-1] _lowercase =accelerator.unwrap_model(__snake_case ).generate( input_ids=batch['''ids'''][:, : batch['''input_len''']] , num_return_sequences=__snake_case , **__snake_case ) # each task is generated batch_size times _lowercase =batch['''task_id'''].repeat(__snake_case ) _lowercase =accelerator.pad_across_processes( __snake_case , dim=1 , pad_index=tokenizer.pad_token_id ) _lowercase , _lowercase =accelerator.gather((generated_tokens, generated_tasks) ) _lowercase =generated_tokens.cpu().numpy() _lowercase =generated_tasks.cpu().numpy() for task, generated_tokens in zip(__snake_case , __snake_case ): gen_token_dict[task].append(__snake_case ) _lowercase =[[] for _ in range(__snake_case )] for task, generated_tokens in gen_token_dict.items(): for s in generated_tokens: _lowercase =tokenizer.decode(__snake_case , skip_special_tokens=__snake_case , clean_up_tokenization_spaces=__snake_case ) code_gens[task].append(remove_last_block(__snake_case ) ) return code_gens def UpperCAmelCase_ ( ) -> str: """simple docstring""" _lowercase =HfArgumentParser(__snake_case ) _lowercase =parser.parse_args() transformers.logging.set_verbosity_error() # enables code execution in code_eval metric _lowercase =args.HF_ALLOW_CODE_EVAL # make sure tokenizer plays nice with multiprocessing _lowercase ='''false''' if args.num_workers is None: _lowercase =multiprocessing.cpu_count() # Use dataset load to feed to accelerate _lowercase =Accelerator() set_seed(args.seed , device_specific=__snake_case ) # Load model and tokenizer _lowercase =AutoTokenizer.from_pretrained(args.model_ckpt ) _lowercase =tokenizer.eos_token _lowercase =AutoModelForCausalLM.from_pretrained(args.model_ckpt ) # Generation settings _lowercase ={ '''do_sample''': args.do_sample, '''temperature''': args.temperature, '''max_new_tokens''': args.max_new_tokens, '''top_p''': args.top_p, '''top_k''': args.top_k, '''stopping_criteria''': StoppingCriteriaList([EndOfFunctionCriteria(0 , __snake_case , __snake_case )] ), } # Load evaluation dataset and metric _lowercase =load_dataset('''openai_humaneval''' ) _lowercase =load_metric('''code_eval''' ) _lowercase =args.num_tasks if args.num_tasks is not None else len(human_eval['''test'''] ) _lowercase =args.n_samples // args.batch_size _lowercase =TokenizedDataset(__snake_case , human_eval['''test'''] , n_copies=__snake_case , n_tasks=__snake_case ) # do not confuse args.batch_size, which is actually the num_return_sequences _lowercase =DataLoader(__snake_case , batch_size=1 ) # Run a quick test to see if code evaluation is enabled try: _lowercase =code_eval_metric.compute(references=[''''''] , predictions=[['''''']] ) except ValueError as exception: print( '''Code evaluation not enabled. Read the warning below carefully and then use `--HF_ALLOW_CODE_EVAL="1"`''' ''' flag to enable code evaluation.''' ) raise exception _lowercase , _lowercase =accelerator.prepare(__snake_case , __snake_case ) _lowercase =complete_code( __snake_case , __snake_case , __snake_case , __snake_case , n_tasks=__snake_case , batch_size=args.batch_size , **__snake_case , ) if accelerator.is_main_process: _lowercase =[] for task in tqdm(range(__snake_case ) ): _lowercase =human_eval['''test'''][task]['''test'''] _lowercase =F"check({human_eval['test'][task]['entry_point']})" references.append('''\n''' + test_func + '''\n''' + entry_point ) # Evaluate completions with "code_eval" metric _lowercase , _lowercase =code_eval_metric.compute( references=__snake_case , predictions=__snake_case , num_workers=args.num_workers ) print(F"Results: {pass_at_k}" ) # Save results to json file with open(args.output_file , '''w''' ) as fp: json.dump(__snake_case , __snake_case ) # For some reason the folliwng seems to be necessary sometimes for code_eval to work nice with multiprocessing # https://stackoverflow.com/questions/60804599/python-multiprocessing-keeps-spawning-the-whole-script if __name__ == "__main__": main()
5
'''simple docstring''' import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] ) -> Optional[int]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[int] ) -> Dict: UpperCAmelCase : Dict = tmp_path / '''cache''' UpperCAmelCase : List[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase : str = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict , _lowerCAmelCase : int ) -> Optional[int]: UpperCAmelCase : Any = tmp_path / '''cache''' UpperCAmelCase : Tuple = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Any = features.copy() if features else default_expected_features UpperCAmelCase : List[Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Dict = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Tuple ) -> Tuple: UpperCAmelCase : Optional[Any] = tmp_path / '''cache''' UpperCAmelCase : Optional[int] = {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''} UpperCAmelCase : int = features.copy() if features else default_expected_features UpperCAmelCase : Any = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict ) -> Union[str, Any]: # jsonl_312_path features are {"col_3": "float64", "col_1": "string", "col_2": "int64"} UpperCAmelCase : Tuple = {'''col_2''': '''int64''', '''col_3''': '''float64''', '''col_1''': '''string'''} UpperCAmelCase : List[str] = features.copy() UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = tmp_path / '''cache''' UpperCAmelCase : List[str] = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ) -> Optional[Any]: UpperCAmelCase : Any = tmp_path / '''cache''' UpperCAmelCase : List[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : List[Any] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , split=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('''path_type''' , [str, list] ) def snake_case_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> Dict: if issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase : str = jsonl_path elif issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase : Dict = [jsonl_path] UpperCAmelCase : int = tmp_path / '''cache''' UpperCAmelCase : Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[int] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : Dict=("train",) ) -> Union[str, Any]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) for split in splits: UpperCAmelCase : List[str] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ) -> Any: UpperCAmelCase : Optional[Any] = tmp_path / '''cache''' UpperCAmelCase : List[str] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase : Optional[int] = JsonDatasetReader({'''train''': jsonl_path} , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] ) -> int: UpperCAmelCase : Dict = tmp_path / '''cache''' UpperCAmelCase : Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[int] = features.copy() if features else default_expected_features UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = JsonDatasetReader({'''train''': jsonl_path} , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict ) -> Union[str, Any]: if split: UpperCAmelCase : Optional[int] = {split: jsonl_path} else: UpperCAmelCase : Any = '''train''' UpperCAmelCase : Any = {'''train''': jsonl_path, '''test''': jsonl_path} UpperCAmelCase : Tuple = tmp_path / '''cache''' UpperCAmelCase : int = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[Any] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def snake_case_ ( _lowerCAmelCase : List[str] ) -> str: return json.load(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Dict ) -> str: return [json.loads(_lowerCAmelCase ) for line in buffer] class SCREAMING_SNAKE_CASE: """simple docstring""" @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def A ( self : Union[str, Any] , __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : Optional[int] ) -> Dict: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case ).write() buffer.seek(0 ) UpperCAmelCase : Union[str, Any] = load_json_function(__snake_case ) assert isinstance(__snake_case , __snake_case ) assert isinstance(exported_content[0] , __snake_case ) assert len(__snake_case ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def A ( self : Optional[int] , __snake_case : Optional[Any] , __snake_case : str , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : Optional[Any] ) -> List[Any]: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , orient=__snake_case ).write() buffer.seek(0 ) UpperCAmelCase : Union[str, Any] = load_json(__snake_case ) assert isinstance(__snake_case , __snake_case ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(__snake_case , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(__snake_case ) == 10 @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def A ( self : str , __snake_case : str , __snake_case : str , __snake_case : int ) -> Any: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase : Any = load_json_function(__snake_case ) assert isinstance(__snake_case , __snake_case ) assert isinstance(exported_content[0] , __snake_case ) assert len(__snake_case ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def A ( self : Any , __snake_case : int , __snake_case : Union[str, Any] , __snake_case : Optional[int] , __snake_case : Optional[Any] , __snake_case : List[str] ) -> Any: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , orient=__snake_case , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase : List[str] = load_json(__snake_case ) assert isinstance(__snake_case , __snake_case ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(__snake_case , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(__snake_case ) == 10 def A ( self : List[Any] , __snake_case : str ) -> Dict: with pytest.raises(__snake_case ): with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , num_proc=0 ) @pytest.mark.parametrize('''compression, extension''' , [('''gzip''', '''gz'''), ('''bz2''', '''bz2'''), ('''xz''', '''xz''')] ) def A ( self : Optional[int] , __snake_case : Any , __snake_case : str , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Dict ) -> Union[str, Any]: UpperCAmelCase : List[str] = tmp_path_factory.mktemp('''data''' ) / F"""test.json.{extension}""" UpperCAmelCase : List[Any] = str(shared_datadir / F"""test_file.json.{extension}""" ) JsonDatasetWriter(__snake_case , __snake_case , compression=__snake_case ).write() with fsspec.open(__snake_case , '''rb''' , compression='''infer''' ) as f: UpperCAmelCase : str = f.read() with fsspec.open(__snake_case , '''rb''' , compression='''infer''' ) as f: UpperCAmelCase : Optional[int] = f.read() assert exported_content == original_content
23
0
from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __A( a ): snake_case_ = ['''image_processor''', '''tokenizer'''] snake_case_ = '''BlipImageProcessor''' snake_case_ = ('''BertTokenizer''', '''BertTokenizerFast''') def __init__( self , _snake_case , _snake_case ) -> Dict: '''simple docstring''' __a = False super().__init__(_snake_case , _snake_case ) __a = self.image_processor def __call__( self , _snake_case = None , _snake_case = None , _snake_case = True , _snake_case = False , _snake_case = None , _snake_case = None , _snake_case = 0 , _snake_case = None , _snake_case = None , _snake_case = False , _snake_case = False , _snake_case = False , _snake_case = False , _snake_case = False , _snake_case = True , _snake_case = None , **_snake_case , ) -> BatchEncoding: '''simple docstring''' if images is None and text is None: raise ValueError('''You have to specify either images or text.''' ) # Get only text if images is None: __a = self.tokenizer __a = self.tokenizer( text=_snake_case , add_special_tokens=_snake_case , padding=_snake_case , truncation=_snake_case , max_length=_snake_case , stride=_snake_case , pad_to_multiple_of=_snake_case , return_attention_mask=_snake_case , return_overflowing_tokens=_snake_case , return_special_tokens_mask=_snake_case , return_offsets_mapping=_snake_case , return_token_type_ids=_snake_case , return_length=_snake_case , verbose=_snake_case , return_tensors=_snake_case , **_snake_case , ) return text_encoding # add pixel_values __a = self.image_processor(_snake_case , return_tensors=_snake_case ) if text is not None: __a = self.tokenizer( text=_snake_case , add_special_tokens=_snake_case , padding=_snake_case , truncation=_snake_case , max_length=_snake_case , stride=_snake_case , pad_to_multiple_of=_snake_case , return_attention_mask=_snake_case , return_overflowing_tokens=_snake_case , return_special_tokens_mask=_snake_case , return_offsets_mapping=_snake_case , return_token_type_ids=_snake_case , return_length=_snake_case , verbose=_snake_case , return_tensors=_snake_case , **_snake_case , ) else: __a = None if text_encoding is not None: encoding_image_processor.update(_snake_case ) return encoding_image_processor def SCREAMING_SNAKE_CASE_ ( self , *_snake_case , **_snake_case ) -> int: '''simple docstring''' return self.tokenizer.batch_decode(*_snake_case , **_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , *_snake_case , **_snake_case ) -> Optional[int]: '''simple docstring''' return self.tokenizer.decode(*_snake_case , **_snake_case ) @property def SCREAMING_SNAKE_CASE_ ( self ) -> List[Any]: '''simple docstring''' __a = self.tokenizer.model_input_names __a = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
6
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCamelCase__: Tuple = logging.get_logger(__name__) UpperCamelCase__: Optional[int] = {"vocab_file": "sentencepiece.bpe.model"} UpperCamelCase__: Optional[int] = { "vocab_file": { "moussaKam/mbarthez": "https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez": "https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez-orangesum-title": ( "https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model" ), }, } UpperCamelCase__: Dict = { "moussaKam/mbarthez": 1024, "moussaKam/barthez": 1024, "moussaKam/barthez-orangesum-title": 1024, } UpperCamelCase__: Tuple = "▁" class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = ["""input_ids""", """attention_mask"""] def __init__( self : List[Any] , __snake_case : List[Any] , __snake_case : Tuple="<s>" , __snake_case : List[Any]="</s>" , __snake_case : int="</s>" , __snake_case : Any="<s>" , __snake_case : Optional[int]="<unk>" , __snake_case : Union[str, Any]="<pad>" , __snake_case : Union[str, Any]="<mask>" , __snake_case : Optional[Dict[str, Any]] = None , **__snake_case : Dict , ) -> None: # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase : int = AddedToken(__snake_case , lstrip=__snake_case , rstrip=__snake_case ) if isinstance(__snake_case , __snake_case ) else mask_token UpperCAmelCase : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__snake_case , eos_token=__snake_case , unk_token=__snake_case , sep_token=__snake_case , cls_token=__snake_case , pad_token=__snake_case , mask_token=__snake_case , sp_model_kwargs=self.sp_model_kwargs , **__snake_case , ) UpperCAmelCase : Optional[int] = vocab_file UpperCAmelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__snake_case ) ) UpperCAmelCase : int = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} UpperCAmelCase : Optional[Any] = len(self.sp_model ) - 1 UpperCAmelCase : Optional[int] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def A ( self : Tuple , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase : Union[str, Any] = [self.cls_token_id] UpperCAmelCase : Optional[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A ( self : int , __snake_case : List[int] , __snake_case : Optional[List[int]] = None , __snake_case : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__snake_case , token_ids_a=__snake_case , already_has_special_tokens=__snake_case ) if token_ids_a is None: return [1] + ([0] * len(__snake_case )) + [1] return [1] + ([0] * len(__snake_case )) + [1, 1] + ([0] * len(__snake_case )) + [1] def A ( self : Optional[int] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: UpperCAmelCase : Tuple = [self.sep_token_id] UpperCAmelCase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def A ( self : Dict ) -> Optional[int]: return len(self.sp_model ) def A ( self : List[str] ) -> Dict: UpperCAmelCase : Optional[Any] = {self.convert_ids_to_tokens(__snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def A ( self : Optional[Any] , __snake_case : str ) -> List[str]: return self.sp_model.encode(__snake_case , out_type=__snake_case ) def A ( self : int , __snake_case : int ) -> int: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCAmelCase : Optional[Any] = self.sp_model.PieceToId(__snake_case ) return spm_id if spm_id else self.unk_token_id def A ( self : int , __snake_case : Any ) -> List[Any]: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(__snake_case ) def A ( self : List[Any] , __snake_case : Union[str, Any] ) -> List[str]: UpperCAmelCase : Optional[Any] = [] UpperCAmelCase : int = '''''' UpperCAmelCase : Union[str, Any] = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__snake_case ) + token UpperCAmelCase : str = True UpperCAmelCase : List[str] = [] else: current_sub_tokens.append(__snake_case ) UpperCAmelCase : Optional[int] = False out_string += self.sp_model.decode(__snake_case ) return out_string.strip() def __getstate__( self : Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase : Optional[Any] = self.__dict__.copy() UpperCAmelCase : Any = None return state def __setstate__( self : Optional[int] , __snake_case : Union[str, Any] ) -> List[Any]: UpperCAmelCase : Dict = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): UpperCAmelCase : Optional[Any] = {} UpperCAmelCase : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def A ( self : Optional[int] , __snake_case : str , __snake_case : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__snake_case ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase : Union[str, Any] = os.path.join( __snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __snake_case ) elif not os.path.isfile(self.vocab_file ): with open(__snake_case , '''wb''' ) as fi: UpperCAmelCase : Any = self.sp_model.serialized_model_proto() fi.write(__snake_case ) return (out_vocab_file,)
23
0
from maths.prime_factors import prime_factors def _snake_case( SCREAMING_SNAKE_CASE__ : int ) -> int: '''simple docstring''' if not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): A__ = f'Input value of [number={number}] must be an integer' raise TypeError(SCREAMING_SNAKE_CASE__ ) if number < 1: raise ValueError('Input must be a positive integer' ) return -1 if len(prime_factors(SCREAMING_SNAKE_CASE__ ) ) % 2 else 1 if __name__ == "__main__": import doctest doctest.testmod()
7
'''simple docstring''' import tempfile import unittest import numpy as np from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import BertConfig, is_flax_available from transformers.testing_utils import TOKEN, USER, is_staging_test, require_flax if is_flax_available(): import os from flax.core.frozen_dict import unfreeze from flax.traverse_util import flatten_dict from transformers import FlaxBertModel UpperCamelCase__: Tuple = "0.12" # assumed parallelism: 8 @require_flax @is_staging_test class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" @classmethod def A ( cls : Union[str, Any] ) -> int: UpperCAmelCase : Optional[Any] = TOKEN HfFolder.save_token(__snake_case ) @classmethod def A ( cls : List[str] ) -> Tuple: try: delete_repo(token=cls._token , repo_id='''test-model-flax''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-model-flax-org''' ) except HTTPError: pass def A ( self : int ) -> Tuple: UpperCAmelCase : List[Any] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) model.push_to_hub('''test-model-flax''' , use_auth_token=self._token ) UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) UpperCAmelCase : List[Any] = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Tuple = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Union[str, Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''test-model-flax''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(__snake_case , repo_id='''test-model-flax''' , push_to_hub=__snake_case , use_auth_token=self._token ) UpperCAmelCase : str = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) UpperCAmelCase : Any = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : str = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Optional[Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) def A ( self : Optional[Any] ) -> Union[str, Any]: UpperCAmelCase : Dict = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Optional[Any] = FlaxBertModel(__snake_case ) model.push_to_hub('''valid_org/test-model-flax-org''' , use_auth_token=self._token ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) UpperCAmelCase : List[Any] = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : int = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Any = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-model-flax-org''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained( __snake_case , repo_id='''valid_org/test-model-flax-org''' , push_to_hub=__snake_case , use_auth_token=self._token ) UpperCAmelCase : str = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) UpperCAmelCase : Any = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Optional[Any] = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : int = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Union[str, Any]: UpperCAmelCase : str = True UpperCAmelCase : int = flatten_dict(modela.params ) UpperCAmelCase : Dict = flatten_dict(modela.params ) for key in flat_params_a.keys(): if np.sum(np.abs(flat_params_a[key] - flat_params_a[key] ) ) > 1e-4: UpperCAmelCase : Dict = False return models_are_equal @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : Tuple ) -> Union[str, Any]: UpperCAmelCase : List[Any] = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) UpperCAmelCase : int = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(__snake_case , __snake_case ) ) with self.assertRaises(__snake_case ): UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : str = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertTrue(check_models_equal(__snake_case , __snake_case ) ) def A ( self : List[str] ) -> Dict: UpperCAmelCase : Dict = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) UpperCAmelCase : Optional[int] = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(__snake_case , __snake_case ) , max_shard_size='''10KB''' ) with self.assertRaises(__snake_case ): UpperCAmelCase : Any = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertTrue(check_models_equal(__snake_case , __snake_case ) ) def A ( self : Optional[int] ) -> str: UpperCAmelCase : Dict = '''bert''' UpperCAmelCase : int = '''hf-internal-testing/tiny-random-bert-subfolder''' with self.assertRaises(__snake_case ): UpperCAmelCase : Optional[Any] = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertIsNotNone(__snake_case ) def A ( self : Dict ) -> List[Any]: UpperCAmelCase : Optional[int] = '''bert''' UpperCAmelCase : int = '''hf-internal-testing/tiny-random-bert-sharded-subfolder''' with self.assertRaises(__snake_case ): UpperCAmelCase : Dict = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertIsNotNone(__snake_case )
23
0
import re def __SCREAMING_SNAKE_CASE (SCREAMING_SNAKE_CASE__ ): snake_case_ = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) ) if __name__ == "__main__": lowerCAmelCase_ = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
8
'''simple docstring''' import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE: """simple docstring""" def __init__( self : Optional[int] , __snake_case : str , __snake_case : Union[str, Any]=2 , __snake_case : Optional[int]=8 , __snake_case : Any=True , __snake_case : Union[str, Any]=True , __snake_case : Dict=True , __snake_case : int=True , __snake_case : List[Any]=99 , __snake_case : str=16 , __snake_case : Tuple=5 , __snake_case : Tuple=2 , __snake_case : str=36 , __snake_case : Dict="gelu" , __snake_case : str=0.0 , __snake_case : Optional[int]=0.0 , __snake_case : Optional[int]=512 , __snake_case : Optional[Any]=16 , __snake_case : int=2 , __snake_case : int=0.02 , __snake_case : str=3 , __snake_case : Dict=4 , __snake_case : str=None , ) -> Optional[int]: UpperCAmelCase : Optional[int] = parent UpperCAmelCase : Tuple = batch_size UpperCAmelCase : List[str] = seq_length UpperCAmelCase : List[Any] = is_training UpperCAmelCase : int = use_input_mask UpperCAmelCase : Any = use_token_type_ids UpperCAmelCase : str = use_labels UpperCAmelCase : Union[str, Any] = vocab_size UpperCAmelCase : List[str] = hidden_size UpperCAmelCase : Optional[Any] = num_hidden_layers UpperCAmelCase : Union[str, Any] = num_attention_heads UpperCAmelCase : Optional[Any] = intermediate_size UpperCAmelCase : Union[str, Any] = hidden_act UpperCAmelCase : int = hidden_dropout_prob UpperCAmelCase : Optional[int] = attention_probs_dropout_prob UpperCAmelCase : Union[str, Any] = max_position_embeddings UpperCAmelCase : str = type_vocab_size UpperCAmelCase : List[str] = type_sequence_label_size UpperCAmelCase : Tuple = initializer_range UpperCAmelCase : Optional[Any] = num_labels UpperCAmelCase : Optional[int] = num_choices UpperCAmelCase : Any = scope def A ( self : Tuple ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase : Optional[int] = None if self.use_input_mask: UpperCAmelCase : Dict = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase : Dict = None if self.use_token_type_ids: UpperCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase : str = None UpperCAmelCase : Tuple = None UpperCAmelCase : int = None if self.use_labels: UpperCAmelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase : str = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase : List[str] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : int ) -> Tuple: return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__snake_case , initializer_range=self.initializer_range , ) def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Optional[Any] = self.get_config() UpperCAmelCase : int = 300 return config def A ( self : Optional[Any] ) -> Any: ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) : Tuple = self.prepare_config_and_inputs() UpperCAmelCase : Dict = True UpperCAmelCase : Tuple = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def A ( self : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : int , __snake_case : Optional[int] , __snake_case : int , __snake_case : Dict , __snake_case : Tuple , __snake_case : Optional[Any] ) -> List[str]: UpperCAmelCase : int = MraModel(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Tuple = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case ) UpperCAmelCase : Optional[int] = model(__snake_case , token_type_ids=__snake_case ) UpperCAmelCase : Dict = model(__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Tuple , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : List[str] , __snake_case : int , __snake_case : Union[str, Any] , __snake_case : Optional[Any] , __snake_case : Any , __snake_case : List[Any] , __snake_case : Optional[Any] , ) -> Tuple: UpperCAmelCase : str = True UpperCAmelCase : Tuple = MraModel(__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Optional[int] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , encoder_hidden_states=__snake_case , encoder_attention_mask=__snake_case , ) UpperCAmelCase : Optional[Any] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , encoder_hidden_states=__snake_case , ) UpperCAmelCase : str = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Tuple , __snake_case : str , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : int ) -> Any: UpperCAmelCase : Dict = MraForMaskedLM(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Optional[int] = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : Tuple , __snake_case : Tuple , __snake_case : Dict , __snake_case : Dict , __snake_case : Any , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Tuple ) -> Optional[int]: UpperCAmelCase : List[str] = MraForQuestionAnswering(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : List[Any] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , start_positions=__snake_case , end_positions=__snake_case , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : str , __snake_case : Optional[int] , __snake_case : List[str] , __snake_case : str , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Union[str, Any] , __snake_case : List[Any] ) -> int: UpperCAmelCase : int = self.num_labels UpperCAmelCase : Union[str, Any] = MraForSequenceClassification(__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : List[str] = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : str , __snake_case : Dict , __snake_case : Optional[Any] , __snake_case : Dict , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Dict ) -> int: UpperCAmelCase : Tuple = self.num_labels UpperCAmelCase : List[str] = MraForTokenClassification(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : str = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : str , __snake_case : int , __snake_case : Any , __snake_case : Tuple , __snake_case : Optional[int] , __snake_case : List[str] , __snake_case : str , __snake_case : Union[str, Any] ) -> Optional[Any]: UpperCAmelCase : Tuple = self.num_choices UpperCAmelCase : int = MraForMultipleChoice(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : str = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[str] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[str] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : str ) -> Dict: UpperCAmelCase : Any = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) : List[str] = config_and_inputs UpperCAmelCase : Any = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = () def A ( self : int ) -> Union[str, Any]: UpperCAmelCase : List[str] = MraModelTester(self ) UpperCAmelCase : Optional[int] = ConfigTester(self , config_class=__snake_case , hidden_size=37 ) def A ( self : Optional[Any] ) -> str: self.config_tester.run_common_tests() def A ( self : Tuple ) -> Optional[Any]: UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__snake_case ) def A ( self : List[Any] ) -> Optional[Any]: UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase : List[Any] = type self.model_tester.create_and_check_model(*__snake_case ) def A ( self : Tuple ) -> Dict: UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__snake_case ) def A ( self : Tuple ) -> List[str]: UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__snake_case ) def A ( self : int ) -> Dict: UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__snake_case ) def A ( self : Dict ) -> Optional[int]: UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__snake_case ) def A ( self : Any ) -> Optional[int]: UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__snake_case ) @slow def A ( self : Dict ) -> Any: for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase : str = MraModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) @unittest.skip(reason='''MRA does not output attentions''' ) def A ( self : str ) -> Optional[Any]: return @require_torch class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" @slow def A ( self : Tuple ) -> List[Any]: UpperCAmelCase : int = MraModel.from_pretrained('''uw-madison/mra-base-512-4''' ) UpperCAmelCase : Optional[Any] = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : List[Any] = model(__snake_case )[0] UpperCAmelCase : Optional[Any] = torch.Size((1, 256, 768) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Any = torch.tensor( [[[-0.01_40, 0.08_30, -0.03_81], [0.15_46, 0.14_02, 0.02_20], [0.11_62, 0.08_51, 0.01_65]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) ) @slow def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Optional[int] = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-512-4''' ) UpperCAmelCase : Dict = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : List[Any] = model(__snake_case )[0] UpperCAmelCase : int = 50265 UpperCAmelCase : int = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Union[str, Any] = torch.tensor( [[[9.25_95, -3.60_38, 11.88_19], [9.38_69, -3.26_93, 11.09_56], [11.85_24, -3.49_38, 13.12_10]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) ) @slow def A ( self : str ) -> List[Any]: UpperCAmelCase : List[Any] = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-4096-8-d3''' ) UpperCAmelCase : List[Any] = torch.arange(4096 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : Tuple = model(__snake_case )[0] UpperCAmelCase : Optional[int] = 50265 UpperCAmelCase : Tuple = torch.Size((1, 4096, vocab_size) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Optional[int] = torch.tensor( [[[5.47_89, -2.35_64, 7.50_64], [7.90_67, -1.33_69, 9.96_68], [9.07_12, -1.81_06, 7.03_80]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) )
23
0
from binascii import hexlify from hashlib import shaaaa from os import urandom # RFC 3526 - More Modular Exponential (MODP) Diffie-Hellman groups for # Internet Key Exchange (IKE) https://tools.ietf.org/html/rfc3526 __lowerCAmelCase : Tuple ={ # 1536-bit 5: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA237327FFFFFFFFFFFFFFFF', base=1_6, ), 'generator': 2, }, # 2048-bit 1_4: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AACAA68FFFFFFFFFFFFFFFF', base=1_6, ), 'generator': 2, }, # 3072-bit 1_5: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A93AD2CAFFFFFFFFFFFFFFFF', base=1_6, ), 'generator': 2, }, # 4096-bit 1_6: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934063199' + 'FFFFFFFFFFFFFFFF', base=1_6, ), 'generator': 2, }, # 6144-bit 1_7: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD129024E08' + '8A67CC74020BBEA63B139B22514A08798E3404DDEF9519B3CD3A431B' + '302B0A6DF25F14374FE1356D6D51C245E485B576625E7EC6F44C42E9' + 'A637ED6B0BFF5CB6F406B7EDEE386BFB5A899FA5AE9F24117C4B1FE6' + '49286651ECE45B3DC2007CB8A163BF0598DA48361C55D39A69163FA8' + 'FD24CF5F83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3BE39E772C' + '180E86039B2783A2EC07A28FB5C55DF06F4C52C9DE2BCBF695581718' + '3995497CEA956AE515D2261898FA051015728E5A8AAAC42DAD33170D' + '04507A33A85521ABDF1CBA64ECFB850458DBEF0A8AEA71575D060C7D' + 'B3970F85A6E1E4C7ABF5AE8CDB0933D71E8C94E04A25619DCEE3D226' + '1AD2EE6BF12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB3143DB5BFC' + 'E0FD108E4B82D120A92108011A723C12A787E6D788719A10BDBA5B26' + '99C327186AF4E23C1A946834B6150BDA2583E9CA2AD44CE8DBBBC2DB' + '04DE8EF92E8EFC141FBECAA6287C59474E6BC05D99B2964FA090C3A2' + '233BA186515BE7ED1F612970CEE2D7AFB81BDD762170481CD0069127' + 'D5B05AA993B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BDF8FF9406' + 'AD9E530EE5DB382F413001AEB06A53ED9027D831179727B0865A8918' + 'DA3EDBEBCF9B14ED44CE6CBACED4BB1BDB7F1447E6CC254B33205151' + '2BD7AF426FB8F401378CD2BF5983CA01C64B92ECF032EA15D1721D03' + 'F482D7CE6E74FEF6D55E702F46980C82B5A84031900B1C9E59E7C97F' + 'BEC7E8F323A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE32806A1D58B' + 'B7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55CDA56C9EC2EF29632' + '387FE8D76E3C0468043E8F663F4860EE12BF2D5B0B7474D6E694F91E' + '6DCC4024FFFFFFFFFFFFFFFF', base=1_6, ), 'generator': 2, }, # 8192-bit 1_8: { 'prime': int( 'FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD1' + '29024E088A67CC74020BBEA63B139B22514A08798E3404DD' + 'EF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245' + 'E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7ED' + 'EE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3D' + 'C2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F' + '83655D23DCA3AD961C62F356208552BB9ED529077096966D' + '670C354E4ABC9804F1746C08CA18217C32905E462E36CE3B' + 'E39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9' + 'DE2BCBF6955817183995497CEA956AE515D2261898FA0510' + '15728E5A8AAAC42DAD33170D04507A33A85521ABDF1CBA64' + 'ECFB850458DBEF0A8AEA71575D060C7DB3970F85A6E1E4C7' + 'ABF5AE8CDB0933D71E8C94E04A25619DCEE3D2261AD2EE6B' + 'F12FFA06D98A0864D87602733EC86A64521F2B18177B200C' + 'BBE117577A615D6C770988C0BAD946E208E24FA074E5AB31' + '43DB5BFCE0FD108E4B82D120A92108011A723C12A787E6D7' + '88719A10BDBA5B2699C327186AF4E23C1A946834B6150BDA' + '2583E9CA2AD44CE8DBBBC2DB04DE8EF92E8EFC141FBECAA6' + '287C59474E6BC05D99B2964FA090C3A2233BA186515BE7ED' + '1F612970CEE2D7AFB81BDD762170481CD0069127D5B05AA9' + '93B4EA988D8FDDC186FFB7DC90A6C08F4DF435C934028492' + '36C3FAB4D27C7026C1D4DCB2602646DEC9751E763DBA37BD' + 'F8FF9406AD9E530EE5DB382F413001AEB06A53ED9027D831' + '179727B0865A8918DA3EDBEBCF9B14ED44CE6CBACED4BB1B' + 'DB7F1447E6CC254B332051512BD7AF426FB8F401378CD2BF' + '5983CA01C64B92ECF032EA15D1721D03F482D7CE6E74FEF6' + 'D55E702F46980C82B5A84031900B1C9E59E7C97FBEC7E8F3' + '23A97A7E36CC88BE0F1D45B7FF585AC54BD407B22B4154AA' + 'CC8F6D7EBF48E1D814CC5ED20F8037E0A79715EEF29BE328' + '06A1D58BB7C5DA76F550AA3D8A1FBFF0EB19CCB1A313D55C' + 'DA56C9EC2EF29632387FE8D76E3C0468043E8F663F4860EE' + '12BF2D5B0B7474D6E694F91E6DBE115974A3926F12FEE5E4' + '38777CB6A932DF8CD8BEC4D073B931BA3BC832B68D9DD300' + '741FA7BF8AFC47ED2576F6936BA424663AAB639C5AE4F568' + '3423B4742BF1C978238F16CBE39D652DE3FDB8BEFC848AD9' + '22222E04A4037C0713EB57A81A23F0C73473FC646CEA306B' + '4BCBC8862F8385DDFA9D4B7FA2C087E879683303ED5BDD3A' + '062B3CF5B3A278A66D2A13F83F44F82DDF310EE074AB6A36' + '4597E899A0255DC164F31CC50846851DF9AB48195DED7EA1' + 'B1D510BD7EE74D73FAF36BC31ECFA268359046F4EB879F92' + '4009438B481C6CD7889A002ED5EE382BC9190DA6FC026E47' + '9558E4475677E9AA9E3050E2765694DFC81F56E880B96E71' + '60C980DD98EDD3DFFFFFFFFFFFFFFFFF', base=1_6, ), 'generator': 2, }, } class _lowercase : '''simple docstring''' def __init__( self :List[Any] , lowerCAmelCase__ :int = 14 ) -> None: if group not in primes: raise ValueError('''Unsupported Group''' ) __SCREAMING_SNAKE_CASE : int = primes[group]['''prime'''] __SCREAMING_SNAKE_CASE : Tuple = primes[group]['''generator'''] __SCREAMING_SNAKE_CASE : Tuple = int(hexlify(urandom(32 ) ) , base=16 ) def __magic_name__( self :Union[str, Any] ) -> str: return hex(self.__private_key )[2:] def __magic_name__( self :Union[str, Any] ) -> str: __SCREAMING_SNAKE_CASE : Dict = pow(self.generator , self.__private_key , self.prime ) return hex(lowerCAmelCase__ )[2:] def __magic_name__( self :List[str] , lowerCAmelCase__ :int ) -> bool: # check if the other public key is valid based on NIST SP800-56 return ( 2 <= key <= self.prime - 2 and pow(lowerCAmelCase__ , (self.prime - 1) // 2 , self.prime ) == 1 ) def __magic_name__( self :List[str] , lowerCAmelCase__ :str ) -> str: __SCREAMING_SNAKE_CASE : Dict = int(lowerCAmelCase__ , base=16 ) if not self.is_valid_public_key(lowerCAmelCase__ ): raise ValueError('''Invalid public key''' ) __SCREAMING_SNAKE_CASE : Union[str, Any] = pow(lowerCAmelCase__ , self.__private_key , self.prime ) return shaaaa(str(lowerCAmelCase__ ).encode() ).hexdigest() @staticmethod def __magic_name__( lowerCAmelCase__ :int , lowerCAmelCase__ :int ) -> bool: # check if the other public key is valid based on NIST SP800-56 return ( 2 <= remote_public_key_str <= prime - 2 and pow(lowerCAmelCase__ , (prime - 1) // 2 , lowerCAmelCase__ ) == 1 ) @staticmethod def __magic_name__( lowerCAmelCase__ :str , lowerCAmelCase__ :str , lowerCAmelCase__ :int = 14 ) -> str: __SCREAMING_SNAKE_CASE : List[Any] = int(lowerCAmelCase__ , base=16 ) __SCREAMING_SNAKE_CASE : Optional[int] = int(lowerCAmelCase__ , base=16 ) __SCREAMING_SNAKE_CASE : str = primes[group]['''prime'''] if not DiffieHellman.is_valid_public_key_static(lowerCAmelCase__ , lowerCAmelCase__ ): raise ValueError('''Invalid public key''' ) __SCREAMING_SNAKE_CASE : Tuple = pow(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) return shaaaa(str(lowerCAmelCase__ ).encode() ).hexdigest() if __name__ == "__main__": import doctest doctest.testmod()
9
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : Any ) -> str: UpperCAmelCase : Any = { '''task_specific_params''': { '''summarization''': {'''length_penalty''': 1.0, '''max_length''': 128, '''min_length''': 12, '''num_beams''': 4}, '''summarization_cnn''': {'''length_penalty''': 2.0, '''max_length''': 142, '''min_length''': 56, '''num_beams''': 4}, '''summarization_xsum''': {'''length_penalty''': 1.0, '''max_length''': 62, '''min_length''': 11, '''num_beams''': 6}, } } UpperCAmelCase : int = { '''task_specific_params.summarization.length_penalty''': 1.0, '''task_specific_params.summarization.max_length''': 128, '''task_specific_params.summarization.min_length''': 12, '''task_specific_params.summarization.num_beams''': 4, '''task_specific_params.summarization_cnn.length_penalty''': 2.0, '''task_specific_params.summarization_cnn.max_length''': 142, '''task_specific_params.summarization_cnn.min_length''': 56, '''task_specific_params.summarization_cnn.num_beams''': 4, '''task_specific_params.summarization_xsum.length_penalty''': 1.0, '''task_specific_params.summarization_xsum.max_length''': 62, '''task_specific_params.summarization_xsum.min_length''': 11, '''task_specific_params.summarization_xsum.num_beams''': 6, } self.assertEqual(flatten_dict(__snake_case ) , __snake_case ) def A ( self : int ) -> str: UpperCAmelCase : Any = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(__snake_case ) , x.transpose() ) ) UpperCAmelCase : str = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def A ( self : str ) -> Union[str, Any]: UpperCAmelCase : Any = np.random.randn(3 , 4 ) UpperCAmelCase : List[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , transpose(__snake_case ).numpy() ) ) UpperCAmelCase : Tuple = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Any = torch.tensor(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , transpose(__snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_tf def A ( self : List[str] ) -> Optional[Any]: UpperCAmelCase : int = np.random.randn(3 , 4 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , transpose(__snake_case ).numpy() ) ) UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : str = tf.constant(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , transpose(__snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_flax def A ( self : Tuple ) -> Any: UpperCAmelCase : List[Any] = np.random.randn(3 , 4 ) UpperCAmelCase : List[str] = jnp.array(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , np.asarray(transpose(__snake_case ) ) ) ) UpperCAmelCase : Dict = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , np.asarray(transpose(__snake_case , axes=(1, 2, 0) ) ) ) ) def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , np.reshape(__snake_case , (4, 3) ) ) ) UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , np.reshape(__snake_case , (12, 5) ) ) ) @require_torch def A ( self : Union[str, Any] ) -> int: UpperCAmelCase : Dict = np.random.randn(3 , 4 ) UpperCAmelCase : Optional[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , reshape(__snake_case , (4, 3) ).numpy() ) ) UpperCAmelCase : List[str] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : List[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , reshape(__snake_case , (12, 5) ).numpy() ) ) @require_tf def A ( self : int ) -> List[str]: UpperCAmelCase : List[Any] = np.random.randn(3 , 4 ) UpperCAmelCase : List[str] = tf.constant(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , reshape(__snake_case , (4, 3) ).numpy() ) ) UpperCAmelCase : List[Any] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Optional[Any] = tf.constant(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , reshape(__snake_case , (12, 5) ).numpy() ) ) @require_flax def A ( self : Any ) -> Dict: UpperCAmelCase : Tuple = np.random.randn(3 , 4 ) UpperCAmelCase : Union[str, Any] = jnp.array(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , np.asarray(reshape(__snake_case , (4, 3) ) ) ) ) UpperCAmelCase : Any = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Optional[Any] = jnp.array(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , np.asarray(reshape(__snake_case , (12, 5) ) ) ) ) def A ( self : List[Any] ) -> List[Any]: UpperCAmelCase : Union[str, Any] = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(__snake_case ) , np.squeeze(__snake_case ) ) ) UpperCAmelCase : str = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , np.squeeze(__snake_case , axis=2 ) ) ) @require_torch def A ( self : Optional[int] ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : List[str] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , squeeze(__snake_case ).numpy() ) ) UpperCAmelCase : Any = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : str = torch.tensor(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , squeeze(__snake_case , axis=2 ).numpy() ) ) @require_tf def A ( self : Optional[Any] ) -> Dict: UpperCAmelCase : int = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , squeeze(__snake_case ).numpy() ) ) UpperCAmelCase : List[str] = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , squeeze(__snake_case , axis=2 ).numpy() ) ) @require_flax def A ( self : List[Any] ) -> Dict: UpperCAmelCase : Optional[int] = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , np.asarray(squeeze(__snake_case ) ) ) ) UpperCAmelCase : str = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , np.asarray(squeeze(__snake_case , axis=2 ) ) ) ) def A ( self : Optional[Any] ) -> int: UpperCAmelCase : Optional[Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , np.expand_dims(__snake_case , axis=1 ) ) ) @require_torch def A ( self : List[str] ) -> Tuple: UpperCAmelCase : Tuple = np.random.randn(3 , 4 ) UpperCAmelCase : Tuple = torch.tensor(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , expand_dims(__snake_case , axis=1 ).numpy() ) ) @require_tf def A ( self : List[str] ) -> Union[str, Any]: UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 ) UpperCAmelCase : Any = tf.constant(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , expand_dims(__snake_case , axis=1 ).numpy() ) ) @require_flax def A ( self : Any ) -> List[Any]: UpperCAmelCase : List[str] = np.random.randn(3 , 4 ) UpperCAmelCase : str = jnp.array(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , np.asarray(expand_dims(__snake_case , axis=1 ) ) ) )
23
0
class _SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__(self : Any , UpperCAmelCase_ : str = "" , UpperCAmelCase_ : bool = False) ->None: '''simple docstring''' lowerCamelCase__: dict[str, RadixNode] ={} # A node will be a leaf if the tree contains its word lowerCamelCase__: Any =is_leaf lowerCamelCase__: List[str] =prefix def SCREAMING_SNAKE_CASE_ (self : List[str] , UpperCAmelCase_ : str) ->tuple[str, str, str]: '''simple docstring''' lowerCamelCase__: Any =0 for q, w in zip(self.prefix , UpperCAmelCase_): if q != w: break x += 1 return self.prefix[:x], self.prefix[x:], word[x:] def SCREAMING_SNAKE_CASE_ (self : Optional[Any] , UpperCAmelCase_ : list[str]) ->None: '''simple docstring''' for word in words: self.insert(UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : List[str] , UpperCAmelCase_ : str) ->None: '''simple docstring''' if self.prefix == word: lowerCamelCase__: Optional[int] =True # Case 2: The node has no edges that have a prefix to the word # Solution: We create an edge from the current node to a new one # containing the word elif word[0] not in self.nodes: lowerCamelCase__: Any =RadixNode(prefix=UpperCAmelCase_ , is_leaf=UpperCAmelCase_) else: lowerCamelCase__: Union[str, Any] =self.nodes[word[0]] lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__: Tuple =incoming_node.match( UpperCAmelCase_) # Case 3: The node prefix is equal to the matching # Solution: We insert remaining word on the next node if remaining_prefix == "": self.nodes[matching_string[0]].insert(UpperCAmelCase_) # Case 4: The word is greater equal to the matching # Solution: Create a node in between both nodes, change # prefixes and add the new node for the remaining word else: lowerCamelCase__: Union[str, Any] =remaining_prefix lowerCamelCase__: Optional[int] =self.nodes[matching_string[0]] lowerCamelCase__: Dict =RadixNode(UpperCAmelCase_ , UpperCAmelCase_) lowerCamelCase__: Tuple =aux_node if remaining_word == "": lowerCamelCase__: Dict =True else: self.nodes[matching_string[0]].insert(UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : int , UpperCAmelCase_ : str) ->bool: '''simple docstring''' lowerCamelCase__: Dict =self.nodes.get(word[0] , UpperCAmelCase_) if not incoming_node: return False else: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__: str =incoming_node.match( UpperCAmelCase_) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # This applies when the word and the prefix are equal elif remaining_word == "": return incoming_node.is_leaf # We have word remaining so we check the next node else: return incoming_node.find(UpperCAmelCase_) def SCREAMING_SNAKE_CASE_ (self : Any , UpperCAmelCase_ : str) ->bool: '''simple docstring''' lowerCamelCase__: Union[str, Any] =self.nodes.get(word[0] , UpperCAmelCase_) if not incoming_node: return False else: lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__: List[Any] =incoming_node.match( UpperCAmelCase_) # If there is remaining prefix, the word can't be on the tree if remaining_prefix != "": return False # We have word remaining so we check the next node elif remaining_word != "": return incoming_node.delete(UpperCAmelCase_) else: # If it is not a leaf, we don't have to delete if not incoming_node.is_leaf: return False else: # We delete the nodes if no edges go from it if len(incoming_node.nodes) == 0: del self.nodes[word[0]] # We merge the current node with its only child if len(self.nodes) == 1 and not self.is_leaf: lowerCamelCase__: int =list(self.nodes.values())[0] lowerCamelCase__: Any =merging_node.is_leaf self.prefix += merging_node.prefix lowerCamelCase__: Tuple =merging_node.nodes # If there is more than 1 edge, we just mark it as non-leaf elif len(incoming_node.nodes) > 1: lowerCamelCase__: Dict =False # If there is 1 edge, we merge it with its child else: lowerCamelCase__: str =list(incoming_node.nodes.values())[0] lowerCamelCase__: Any =merging_node.is_leaf incoming_node.prefix += merging_node.prefix lowerCamelCase__: Optional[Any] =merging_node.nodes return True def SCREAMING_SNAKE_CASE_ (self : List[Any] , UpperCAmelCase_ : int = 0) ->None: '''simple docstring''' if self.prefix != "": print("-" * height , self.prefix , " (leaf)" if self.is_leaf else "") for value in self.nodes.values(): value.print_tree(height + 1) def lowerCAmelCase_ ( ) -> bool: """simple docstring""" lowerCamelCase__: str ="banana bananas bandana band apple all beast".split() lowerCamelCase__: List[Any] =RadixNode() root.insert_many(__a ) assert all(root.find(__a ) for word in words ) assert not root.find("bandanas" ) assert not root.find("apps" ) root.delete("all" ) assert not root.find("all" ) root.delete("banana" ) assert not root.find("banana" ) assert root.find("bananas" ) return True def lowerCAmelCase_ ( ) -> None: """simple docstring""" assert test_trie() def lowerCAmelCase_ ( ) -> None: """simple docstring""" lowerCamelCase__: Optional[int] =RadixNode() lowerCamelCase__: Optional[int] ="banana bananas bandanas bandana band apple all beast".split() root.insert_many(__a ) print("Words:" , __a ) print("Tree:" ) root.print_tree() if __name__ == "__main__": main()
10
'''simple docstring''' import argparse import os import re import packaging.version UpperCamelCase__: Union[str, Any] = "examples/" UpperCamelCase__: Optional[Any] = { "examples": (re.compile(r"^check_min_version\(\"[^\"]+\"\)\s*$", re.MULTILINE), "check_min_version(\"VERSION\")\n"), "init": (re.compile(r"^__version__\s+=\s+\"([^\"]+)\"\s*$", re.MULTILINE), "__version__ = \"VERSION\"\n"), "setup": (re.compile(r"^(\s*)version\s*=\s*\"[^\"]+\",", re.MULTILINE), r"\1version=\"VERSION\","), "doc": (re.compile(r"^(\s*)release\s*=\s*\"[^\"]+\"$", re.MULTILINE), "release = \"VERSION\"\n"), } UpperCamelCase__: Optional[int] = { "init": "src/diffusers/__init__.py", "setup": "setup.py", } UpperCamelCase__: List[Any] = "README.md" def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int ) -> Optional[int]: with open(_lowerCAmelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: UpperCAmelCase : Optional[int] = f.read() UpperCAmelCase , UpperCAmelCase : List[Any] = REPLACE_PATTERNS[pattern] UpperCAmelCase : List[Any] = replace.replace('''VERSION''' , _lowerCAmelCase ) UpperCAmelCase : Optional[Any] = re_pattern.sub(_lowerCAmelCase , _lowerCAmelCase ) with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Any ) -> Optional[int]: for folder, directories, fnames in os.walk(_lowerCAmelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase , pattern='''examples''' ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : str=False ) -> List[str]: for pattern, fname in REPLACE_FILES.items(): update_version_in_file(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not patch: update_version_in_examples(_lowerCAmelCase ) def snake_case_ ( ) -> Optional[Any]: UpperCAmelCase : Optional[int] = '''🤗 Transformers currently provides the following architectures''' UpperCAmelCase : Optional[int] = '''1. Want to contribute a new model?''' with open(_lowerCAmelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: UpperCAmelCase : Optional[Any] = f.readlines() # Find the start of the list. UpperCAmelCase : List[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 UpperCAmelCase : Optional[Any] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): UpperCAmelCase : Optional[int] = lines[index].replace( '''https://huggingface.co/docs/diffusers/main/model_doc''' , '''https://huggingface.co/docs/diffusers/model_doc''' , ) index += 1 with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(_lowerCAmelCase ) def snake_case_ ( ) -> Optional[Any]: with open(REPLACE_FILES['''init'''] , '''r''' ) as f: UpperCAmelCase : Union[str, Any] = f.read() UpperCAmelCase : int = REPLACE_PATTERNS['''init'''][0].search(_lowerCAmelCase ).groups()[0] return packaging.version.parse(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : List[str]=False ) -> Any: UpperCAmelCase : Optional[Any] = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: UpperCAmelCase : Optional[int] = default_version.base_version elif patch: UpperCAmelCase : Union[str, Any] = f"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: UpperCAmelCase : Union[str, Any] = f"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. UpperCAmelCase : Dict = input(f"""Which version are you releasing? [{default_version}]""" ) if len(_lowerCAmelCase ) == 0: UpperCAmelCase : Tuple = default_version print(f"""Updating version to {version}.""" ) global_version_update(_lowerCAmelCase , patch=_lowerCAmelCase ) def snake_case_ ( ) -> Any: UpperCAmelCase : List[Any] = get_version() UpperCAmelCase : List[str] = f"""{current_version.major}.{current_version.minor + 1}.0.dev0""" UpperCAmelCase : List[Any] = current_version.base_version # Check with the user we got that right. UpperCAmelCase : Optional[int] = input(f"""Which version are we developing now? [{dev_version}]""" ) if len(_lowerCAmelCase ) == 0: UpperCAmelCase : Dict = dev_version print(f"""Updating version to {version}.""" ) global_version_update(_lowerCAmelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": UpperCamelCase__: Union[str, Any] = argparse.ArgumentParser() parser.add_argument("--post_release", action="store_true", help="Whether this is pre or post release.") parser.add_argument("--patch", action="store_true", help="Whether or not this is a patch release.") UpperCamelCase__: Optional[Any] = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("Nothing to do after a patch :-)") else: post_release_work()
23
0
from math import factorial def _UpperCAmelCase (UpperCamelCase__ : int , UpperCamelCase__ : int , UpperCamelCase__ : float ): if successes > trials: raise ValueError("successes must be lower or equal to trials" ) if trials < 0 or successes < 0: raise ValueError("the function is defined for non-negative integers" ) if not isinstance(UpperCamelCase__ , UpperCamelCase__ ) or not isinstance(UpperCamelCase__ , UpperCamelCase__ ): raise ValueError("the function is defined for non-negative integers" ) if not 0 < prob < 1: raise ValueError("prob has to be in range of 1 - 0" ) _A : Optional[Any] = (prob**successes) * ((1 - prob) ** (trials - successes)) # Calculate the binomial coefficient: n! / k!(n-k)! _A : Dict = float(factorial(UpperCamelCase__ ) ) coefficient /= factorial(UpperCamelCase__ ) * factorial(trials - successes ) return probability * coefficient if __name__ == "__main__": from doctest import testmod testmod() print('Probability of 2 successes out of 4 trails') print('with probability of 0.75 is:', end=' ') print(binomial_distribution(2, 4, 0.75))
11
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake UpperCamelCase__: Tuple = numpy.array([0, 0]) UpperCamelCase__: Union[str, Any] = numpy.array([0.5, 0.8660254]) UpperCamelCase__: Dict = numpy.array([1, 0]) UpperCamelCase__: int = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] , _lowerCAmelCase : int ) -> list[numpy.ndarray]: UpperCAmelCase : Union[str, Any] = initial_vectors for _ in range(_lowerCAmelCase ): UpperCAmelCase : Union[str, Any] = iteration_step(_lowerCAmelCase ) return vectors def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] ) -> list[numpy.ndarray]: UpperCAmelCase : Tuple = [] for i, start_vector in enumerate(vectors[:-1] ): UpperCAmelCase : List[str] = vectors[i + 1] new_vectors.append(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def snake_case_ ( _lowerCAmelCase : numpy.ndarray , _lowerCAmelCase : float ) -> numpy.ndarray: UpperCAmelCase : List[str] = numpy.radians(_lowerCAmelCase ) UpperCAmelCase , UpperCAmelCase : Tuple = numpy.cos(_lowerCAmelCase ), numpy.sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = numpy.array(((c, -s), (s, c)) ) return numpy.dot(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] ) -> None: UpperCAmelCase : List[Any] = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() UpperCAmelCase , UpperCAmelCase : str = zip(*_lowerCAmelCase ) plt.plot(_lowerCAmelCase , _lowerCAmelCase ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__: List[Any] = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
23
0
import os import re import warnings from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer if TYPE_CHECKING: from ...tokenization_utils_base import TextInput from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = {'vocab_file': 'spiece.model'} UpperCAmelCase_ = { 'vocab_file': { 't5-small': 'https://huggingface.co/t5-small/resolve/main/spiece.model', 't5-base': 'https://huggingface.co/t5-base/resolve/main/spiece.model', 't5-large': 'https://huggingface.co/t5-large/resolve/main/spiece.model', 't5-3b': 'https://huggingface.co/t5-3b/resolve/main/spiece.model', 't5-11b': 'https://huggingface.co/t5-11b/resolve/main/spiece.model', } } # TODO(PVP) - this should be removed in Transformers v5 UpperCAmelCase_ = { 't5-small': 512, 't5-base': 512, 't5-large': 512, 't5-3b': 512, 't5-11b': 512, } UpperCAmelCase_ = '▁' class lowerCamelCase__( __lowerCamelCase): UpperCAmelCase__ : List[str] = VOCAB_FILES_NAMES UpperCAmelCase__ : int = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ : Any = ['input_ids', 'attention_mask'] def __init__( self: Dict , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Union[str, Any]="</s>" , UpperCamelCase_: Optional[Any]="<unk>" , UpperCamelCase_: Any="<pad>" , UpperCamelCase_: Optional[Any]=1_00 , UpperCamelCase_: Optional[int]=None , UpperCamelCase_: Optional[Dict[str, Any]] = None , UpperCamelCase_: Optional[int]=True , **UpperCamelCase_: str , ): # Add extra_ids to the special token list if extra_ids > 0 and additional_special_tokens is None: __lowerCamelCase = [F'<extra_id_{i}>' for i in range(UpperCamelCase_ )] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens __lowerCamelCase = len(set(filter(lambda UpperCamelCase_ : bool("""extra_id""" in str(UpperCamelCase_ ) ) , UpperCamelCase_ ) ) ) if extra_tokens != extra_ids: raise ValueError( F'Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are' """ provided to T5Tokenizer. In this case the additional_special_tokens must include the extra_ids""" """ tokens""" ) if legacy: logger.warning_once( F'You are using the legacy behaviour of the {self.__class__}. This means that tokens that come after special tokens will not be properly handled. We recommend you to' """ read the related pull request available at https://github.com/huggingface/transformers/pull/24565""" ) __lowerCamelCase = legacy __lowerCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , extra_ids=UpperCamelCase_ , additional_special_tokens=UpperCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , legacy=UpperCamelCase_ , **UpperCamelCase_ , ) __lowerCamelCase = vocab_file __lowerCamelCase = extra_ids __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(UpperCamelCase_ ) @staticmethod def lowerCAmelCase__ ( UpperCamelCase_: Optional[int] , UpperCamelCase_: str , UpperCamelCase_: Optional[int] ): if pretrained_model_name_or_path in TaTokenizer.max_model_input_sizes: __lowerCamelCase = TaTokenizer.max_model_input_sizes[pretrained_model_name_or_path] if init_max_model_length is not None and init_max_model_length != max_model_length: return init_max_model_length elif init_max_model_length is None: warnings.warn( """This tokenizer was incorrectly instantiated with a model max length of""" F' {deprecated_max_model_length} which will be corrected in Transformers v5.\nFor now, this' """ behavior is kept to avoid breaking backwards compatibility when padding/encoding with""" """ `truncation is True`.\n- Be aware that you SHOULD NOT rely on""" F' {pretrained_model_name_or_path} automatically truncating your input to' F' {deprecated_max_model_length} when padding/encoding.\n- If you want to encode/pad to sequences' F' longer than {deprecated_max_model_length} you can either instantiate this tokenizer with' """ `model_max_length` or pass `max_length` when encoding/padding.\n- To avoid this warning, please""" """ instantiate this tokenizer with `model_max_length` set to your preferred value.""" , UpperCamelCase_ , ) return max_model_length @property def lowerCAmelCase__ ( self: List[Any] ): return self.sp_model.get_piece_size() + self._extra_ids def lowerCAmelCase__ ( self: List[str] ): __lowerCamelCase = {self.convert_ids_to_tokens(UpperCamelCase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: List[int] , UpperCamelCase_: Optional[List[int]] = None , UpperCamelCase_: bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=UpperCamelCase_ , token_ids_a=UpperCamelCase_ , already_has_special_tokens=UpperCamelCase_ ) # normal case: some special tokens if token_ids_a is None: return ([0] * len(UpperCamelCase_ )) + [1] return ([0] * len(UpperCamelCase_ )) + [1] + ([0] * len(UpperCamelCase_ )) + [1] def lowerCAmelCase__ ( self: List[str] ): return list( set(filter(lambda UpperCamelCase_ : bool(re.search(r"""<extra_id_\d+>""" , UpperCamelCase_ ) ) is not None , self.additional_special_tokens ) ) ) def lowerCAmelCase__ ( self: Union[str, Any] ): return [self._convert_token_to_id(UpperCamelCase_ ) for token in self.get_sentinel_tokens()] def lowerCAmelCase__ ( self: int , UpperCamelCase_: List[int] ): if len(UpperCamelCase_ ) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( F'This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated' """ eos tokens being added.""" ) return token_ids else: return token_ids + [self.eos_token_id] def lowerCAmelCase__ ( self: Optional[Any] , UpperCamelCase_: List[int] , UpperCamelCase_: Optional[List[int]] = None ): __lowerCamelCase = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos ) * [0] return len(token_ids_a + eos + token_ids_a + eos ) * [0] def lowerCAmelCase__ ( self: Optional[int] , UpperCamelCase_: List[int] , UpperCamelCase_: Optional[List[int]] = None ): __lowerCamelCase = self._add_eos_if_not_present(UpperCamelCase_ ) if token_ids_a is None: return token_ids_a else: __lowerCamelCase = self._add_eos_if_not_present(UpperCamelCase_ ) return token_ids_a + token_ids_a def __getstate__( self: Dict ): __lowerCamelCase = self.__dict__.copy() __lowerCamelCase = None return state def __setstate__( self: int , UpperCamelCase_: Dict ): __lowerCamelCase = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): __lowerCamelCase = {} __lowerCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: "TextInput" , **UpperCamelCase_: Dict ): # Replace the SPIECE_UNDERLINE with a space to make sure SPIECE_UNDERLINE is only used at # the beginning of the text if not self.legacy: __lowerCamelCase = SPIECE_UNDERLINE + text.replace(UpperCamelCase_ , """ """ ) return super().tokenize(UpperCamelCase_ , **UpperCamelCase_ ) def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Any , **UpperCamelCase_: List[Any] ): if not self.legacy: __lowerCamelCase = text.startswith(UpperCamelCase_ ) if is_first: __lowerCamelCase = text[1:] __lowerCamelCase = self.sp_model.encode(UpperCamelCase_ , out_type=UpperCamelCase_ ) if not self.legacy and not is_first and not text.startswith(""" """ ) and tokens[0].startswith(UpperCamelCase_ ): __lowerCamelCase = ([tokens[0][1:]] if len(tokens[0] ) > 1 else []) + tokens[1:] return tokens def lowerCAmelCase__ ( self: Tuple , UpperCamelCase_: Tuple ): if token.startswith("""<extra_id_""" ): __lowerCamelCase = re.match(r"""<extra_id_(\d+)>""" , UpperCamelCase_ ) __lowerCamelCase = int(match.group(1 ) ) return self.vocab_size - num - 1 return self.sp_model.piece_to_id(UpperCamelCase_ ) def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: Optional[int] ): if index < self.sp_model.get_piece_size(): __lowerCamelCase = self.sp_model.IdToPiece(UpperCamelCase_ ) else: __lowerCamelCase = F'<extra_id_{self.vocab_size - 1 - index}>' return token def lowerCAmelCase__ ( self: Dict , UpperCamelCase_: Dict ): __lowerCamelCase = [] __lowerCamelCase = """""" __lowerCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(UpperCamelCase_ ) + token __lowerCamelCase = True __lowerCamelCase = [] else: current_sub_tokens.append(UpperCamelCase_ ) __lowerCamelCase = False out_string += self.sp_model.decode(UpperCamelCase_ ) return out_string.strip() def lowerCAmelCase__ ( self: List[str] , UpperCamelCase_: str , UpperCamelCase_: Optional[str] = None ): if not os.path.isdir(UpperCamelCase_ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return __lowerCamelCase = os.path.join( UpperCamelCase_ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCamelCase_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , UpperCamelCase_ ) elif not os.path.isfile(self.vocab_file ): with open(UpperCamelCase_ , """wb""" ) as fi: __lowerCamelCase = self.sp_model.serialized_model_proto() fi.write(UpperCamelCase_ ) return (out_vocab_file,)
12
'''simple docstring''' from manim import * class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" def A ( self : Union[str, Any] ) -> List[str]: UpperCAmelCase : Optional[Any] = Rectangle(height=0.5 , width=0.5 ) UpperCAmelCase : str = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) UpperCAmelCase : Tuple = [mem.copy() for i in range(6 )] UpperCAmelCase : Optional[int] = [mem.copy() for i in range(6 )] UpperCAmelCase : Dict = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Any = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Union[str, Any] = VGroup(__snake_case , __snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Optional[Any] = Text('''CPU''' , font_size=24 ) UpperCAmelCase : Union[str, Any] = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__snake_case ) UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(4 )] UpperCAmelCase : Union[str, Any] = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : List[str] = Text('''GPU''' , font_size=24 ) UpperCAmelCase : Dict = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) gpu.move_to([-1, -1, 0] ) self.add(__snake_case ) UpperCAmelCase : int = [mem.copy() for i in range(6 )] UpperCAmelCase : Union[str, Any] = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : List[str] = Text('''Model''' , font_size=24 ) UpperCAmelCase : Tuple = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) model.move_to([3, -1.0, 0] ) self.add(__snake_case ) UpperCAmelCase : Any = [] for i, rect in enumerate(__snake_case ): rect.set_stroke(__snake_case ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) UpperCAmelCase : Dict = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(__snake_case , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=__snake_case ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=__snake_case , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=__snake_case , buff=0.0 ) self.add(__snake_case ) cpu_targs.append(__snake_case ) UpperCAmelCase : int = [mem.copy() for i in range(6 )] UpperCAmelCase : int = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Any = Text('''Loaded Checkpoint''' , font_size=24 ) UpperCAmelCase : Union[str, Any] = Group(__snake_case , __snake_case ).arrange(__snake_case , aligned_edge=__snake_case , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) UpperCAmelCase : Optional[int] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCAmelCase : str = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(__snake_case , __snake_case ) UpperCAmelCase : Tuple = MarkupText( F"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(__snake_case , DOWN * 2.4 , aligned_edge=key_text.get_left() ) UpperCAmelCase : List[Any] = MarkupText( F"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__snake_case ) , Write(__snake_case ) ) self.play(Write(__snake_case , run_time=1 ) , Create(__snake_case , run_time=1 ) ) UpperCAmelCase : Tuple = [] UpperCAmelCase : int = [] for i, rect in enumerate(__snake_case ): UpperCAmelCase : Any = fill.copy().set_fill(__snake_case , opacity=0.7 ) target.move_to(__snake_case ) first_animations.append(GrowFromCenter(__snake_case , run_time=1 ) ) UpperCAmelCase : List[str] = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(__snake_case , run_time=1.5 ) ) self.play(*__snake_case ) self.play(*__snake_case ) self.wait()
23
0
import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def A_ ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): # Initialise PyTorch model SCREAMING_SNAKE_CASE_: Tuple = FunnelConfig.from_json_file(_UpperCAmelCase ) print(f"Building PyTorch model from configuration: {config}" ) SCREAMING_SNAKE_CASE_: Optional[Any] = FunnelBaseModel(_UpperCAmelCase ) if base_model else FunnelModel(_UpperCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_funnel(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # Save pytorch-model print(f"Save PyTorch model to {pytorch_dump_path}" ) torch.save(model.state_dict() , _UpperCAmelCase ) if __name__ == "__main__": lowerCAmelCase : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--config_file""", default=None, type=str, required=True, help="""The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.""", ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--base_model""", action="""store_true""", help="""Whether you want just the base model (no decoder) or not.""" ) lowerCAmelCase : Optional[int] = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path, args.base_model )
13
'''simple docstring''' import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( "The converted tokenizer will be the `slow` tokenizer. To use the fast, update your `tokenizers` library and re-run the tokenizer conversion" ) UpperCamelCase__: str = None UpperCamelCase__: int = { "7B": 11008, "13B": 13824, "30B": 17920, "65B": 22016, "70B": 28672, } UpperCamelCase__: List[Any] = { "7B": 1, "7Bf": 1, "13B": 2, "13Bf": 2, "30B": 4, "65B": 8, "70B": 8, "70Bf": 8, } def snake_case_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple=1 , _lowerCAmelCase : List[Any]=256 ) -> Optional[Any]: return multiple_of * ((int(ffn_dim_multiplier * int(8 * n / 3 ) ) + multiple_of - 1) // multiple_of) def snake_case_ ( _lowerCAmelCase : List[str] ) -> str: with open(_lowerCAmelCase , '''r''' ) as f: return json.load(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any ) -> Optional[Any]: with open(_lowerCAmelCase , '''w''' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any=True ) -> List[Any]: os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : int = os.path.join(_lowerCAmelCase , '''tmp''' ) os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : List[str] = read_json(os.path.join(_lowerCAmelCase , '''params.json''' ) ) UpperCAmelCase : str = NUM_SHARDS[model_size] UpperCAmelCase : Any = params['''n_layers'''] UpperCAmelCase : str = params['''n_heads'''] UpperCAmelCase : Any = n_heads // num_shards UpperCAmelCase : List[str] = params['''dim'''] UpperCAmelCase : Optional[Any] = dim // n_heads UpperCAmelCase : str = 1_0_0_0_0.0 UpperCAmelCase : Optional[int] = 1.0 / (base ** (torch.arange(0 , _lowerCAmelCase , 2 ).float() / dims_per_head)) if "n_kv_heads" in params: UpperCAmelCase : Tuple = params['''n_kv_heads'''] # for GQA / MQA UpperCAmelCase : Optional[int] = n_heads_per_shard // num_key_value_heads UpperCAmelCase : Optional[Any] = dim // num_key_value_heads else: # compatibility with other checkpoints UpperCAmelCase : List[str] = n_heads UpperCAmelCase : Optional[int] = n_heads_per_shard UpperCAmelCase : List[str] = dim # permute for sliced rotary def permute(_lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]=n_heads , _lowerCAmelCase : int=dim , _lowerCAmelCase : Dict=dim ): return w.view(_lowerCAmelCase , dima // n_heads // 2 , 2 , _lowerCAmelCase ).transpose(1 , 2 ).reshape(_lowerCAmelCase , _lowerCAmelCase ) print(f"""Fetching all parameters from the checkpoint at {input_base_path}.""" ) # Load weights if model_size == "7B": # Not sharded # (The sharded implementation would also work, but this is simpler.) UpperCAmelCase : int = torch.load(os.path.join(_lowerCAmelCase , '''consolidated.00.pth''' ) , map_location='''cpu''' ) else: # Sharded UpperCAmelCase : Optional[Any] = [ torch.load(os.path.join(_lowerCAmelCase , f"""consolidated.{i:02d}.pth""" ) , map_location='''cpu''' ) for i in range(_lowerCAmelCase ) ] UpperCAmelCase : Any = 0 UpperCAmelCase : str = {'''weight_map''': {}} for layer_i in range(_lowerCAmelCase ): UpperCAmelCase : Optional[Any] = f"""pytorch_model-{layer_i + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase : Optional[int] = { f"""model.layers.{layer_i}.self_attn.q_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wq.weight"""] ), f"""model.layers.{layer_i}.self_attn.k_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wk.weight"""] ), f"""model.layers.{layer_i}.self_attn.v_proj.weight""": loaded[f"""layers.{layer_i}.attention.wv.weight"""], f"""model.layers.{layer_i}.self_attn.o_proj.weight""": loaded[f"""layers.{layer_i}.attention.wo.weight"""], f"""model.layers.{layer_i}.mlp.gate_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w1.weight"""], f"""model.layers.{layer_i}.mlp.down_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w2.weight"""], f"""model.layers.{layer_i}.mlp.up_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w3.weight"""], f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[f"""layers.{layer_i}.attention_norm.weight"""], f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[f"""layers.{layer_i}.ffn_norm.weight"""], } else: # Sharded # Note that attention.w{q,k,v,o}, feed_fordward.w[1,2,3], attention_norm.weight and ffn_norm.weight share # the same storage object, saving attention_norm and ffn_norm will save other weights too, which is # redundant as other weights will be stitched from multiple shards. To avoid that, they are cloned. UpperCAmelCase : List[str] = { f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.attention_norm.weight""" ].clone(), f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.ffn_norm.weight""" ].clone(), } UpperCAmelCase : Union[str, Any] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wq.weight"""].view(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase : Optional[Any] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wk.weight"""].view( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) UpperCAmelCase : str = torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wv.weight"""].view( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = torch.cat( [loaded[i][f"""layers.{layer_i}.attention.wo.weight"""] for i in range(_lowerCAmelCase )] , dim=1 ) UpperCAmelCase : Any = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w1.weight"""] for i in range(_lowerCAmelCase )] , dim=0 ) UpperCAmelCase : str = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w2.weight"""] for i in range(_lowerCAmelCase )] , dim=1 ) UpperCAmelCase : Tuple = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w3.weight"""] for i in range(_lowerCAmelCase )] , dim=0 ) UpperCAmelCase : Any = inv_freq for k, v in state_dict.items(): UpperCAmelCase : List[Any] = filename param_count += v.numel() torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase : Optional[int] = f"""pytorch_model-{n_layers + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase : str = { '''model.embed_tokens.weight''': loaded['''tok_embeddings.weight'''], '''model.norm.weight''': loaded['''norm.weight'''], '''lm_head.weight''': loaded['''output.weight'''], } else: UpperCAmelCase : Any = { '''model.norm.weight''': loaded[0]['''norm.weight'''], '''model.embed_tokens.weight''': torch.cat( [loaded[i]['''tok_embeddings.weight'''] for i in range(_lowerCAmelCase )] , dim=1 ), '''lm_head.weight''': torch.cat([loaded[i]['''output.weight'''] for i in range(_lowerCAmelCase )] , dim=0 ), } for k, v in state_dict.items(): UpperCAmelCase : Optional[int] = filename param_count += v.numel() torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , _lowerCAmelCase ) ) # Write configs UpperCAmelCase : Union[str, Any] = {'''total_size''': param_count * 2} write_json(_lowerCAmelCase , os.path.join(_lowerCAmelCase , '''pytorch_model.bin.index.json''' ) ) UpperCAmelCase : int = params['''ffn_dim_multiplier'''] if '''ffn_dim_multiplier''' in params else 1 UpperCAmelCase : Tuple = params['''multiple_of'''] if '''multiple_of''' in params else 256 UpperCAmelCase : Any = LlamaConfig( hidden_size=_lowerCAmelCase , intermediate_size=compute_intermediate_size(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) , num_attention_heads=params['''n_heads'''] , num_hidden_layers=params['''n_layers'''] , rms_norm_eps=params['''norm_eps'''] , num_key_value_heads=_lowerCAmelCase , ) config.save_pretrained(_lowerCAmelCase ) # Make space so we can load the model properly now. del state_dict del loaded gc.collect() print('''Loading the checkpoint in a Llama model.''' ) UpperCAmelCase : Optional[int] = LlamaForCausalLM.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa , low_cpu_mem_usage=_lowerCAmelCase ) # Avoid saving this as part of the config. del model.config._name_or_path print('''Saving in the Transformers format.''' ) model.save_pretrained(_lowerCAmelCase , safe_serialization=_lowerCAmelCase ) shutil.rmtree(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any] ) -> List[str]: # Initialize the tokenizer based on the `spm` model UpperCAmelCase : Dict = LlamaTokenizer if LlamaTokenizerFast is None else LlamaTokenizerFast print(f"""Saving a {tokenizer_class.__name__} to {tokenizer_path}.""" ) UpperCAmelCase : List[Any] = tokenizer_class(_lowerCAmelCase ) tokenizer.save_pretrained(_lowerCAmelCase ) def snake_case_ ( ) -> List[Any]: UpperCAmelCase : int = argparse.ArgumentParser() parser.add_argument( '''--input_dir''' , help='''Location of LLaMA weights, which contains tokenizer.model and model folders''' , ) parser.add_argument( '''--model_size''' , choices=['''7B''', '''7Bf''', '''13B''', '''13Bf''', '''30B''', '''65B''', '''70B''', '''70Bf''', '''tokenizer_only'''] , ) parser.add_argument( '''--output_dir''' , help='''Location to write HF model and tokenizer''' , ) parser.add_argument('''--safe_serialization''' , type=_lowerCAmelCase , help='''Whether or not to save using `safetensors`.''' ) UpperCAmelCase : List[Any] = parser.parse_args() if args.model_size != "tokenizer_only": write_model( model_path=args.output_dir , input_base_path=os.path.join(args.input_dir , args.model_size ) , model_size=args.model_size , safe_serialization=args.safe_serialization , ) UpperCAmelCase : Optional[int] = os.path.join(args.input_dir , '''tokenizer.model''' ) write_tokenizer(args.output_dir , _lowerCAmelCase ) if __name__ == "__main__": main()
23
0
from __future__ import annotations import queue class UpperCamelCase_ : '''simple docstring''' def __init__( self : str , UpperCAmelCase__ : str) ->Tuple: '''simple docstring''' A__ = data A__ = None A__ = None def SCREAMING_SNAKE_CASE ( ) -> TreeNode: """simple docstring""" print('''\n********Press N to stop entering at any point of time********\n''' ) A__ = input('''Enter the value of the root node: ''' ).strip().lower() A__ = queue.Queue() A__ = TreeNode(int(lowercase_ ) ) q.put(lowercase_ ) while not q.empty(): A__ = q.get() A__ = f"""Enter the left node of {node_found.data}: """ A__ = input(lowercase_ ).strip().lower() or '''n''' if check == "n": return tree_node A__ = TreeNode(int(lowercase_ ) ) A__ = left_node q.put(lowercase_ ) A__ = f"""Enter the right node of {node_found.data}: """ A__ = input(lowercase_ ).strip().lower() or '''n''' if check == "n": return tree_node A__ = TreeNode(int(lowercase_ ) ) A__ = right_node q.put(lowercase_ ) raise def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return print(node.data , end=''',''' ) pre_order(node.left ) pre_order(node.right ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return in_order(node.left ) print(node.data , end=''',''' ) in_order(node.right ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return post_order(node.left ) post_order(node.right ) print(node.data , end=''',''' ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ = queue.Queue() q.put(lowercase_ ) while not q.empty(): A__ = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: q.put(node_dequeued.left ) if node_dequeued.right: q.put(node_dequeued.right ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ = queue.Queue() q.put(lowercase_ ) while not q.empty(): A__ = [] while not q.empty(): A__ = q.get() print(node_dequeued.data , end=''',''' ) if node_dequeued.left: list_.append(node_dequeued.left ) if node_dequeued.right: list_.append(node_dequeued.right ) print() for node in list_: q.put(lowercase_ ) def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ = [] A__ = node while n or stack: while n: # start from root node, find its left child print(n.data , end=''',''' ) stack.append(lowercase_ ) A__ = n.left # end of while means current node doesn't have left child A__ = stack.pop() # start to traverse its right child A__ = n.right def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ = [] A__ = node while n or stack: while n: stack.append(lowercase_ ) A__ = n.left A__ = stack.pop() print(n.data , end=''',''' ) A__ = n.right def SCREAMING_SNAKE_CASE ( lowercase_ ) -> None: """simple docstring""" if not isinstance(lowercase_ , lowercase_ ) or not node: return A__ , A__ = [], [] A__ = node stacka.append(lowercase_ ) while stacka: # to find the reversed order of post order, store it in stack2 A__ = stacka.pop() if n.left: stacka.append(n.left ) if n.right: stacka.append(n.right ) stacka.append(lowercase_ ) while stacka: # pop up from stack2 will be the post order print(stacka.pop().data , end=''',''' ) def SCREAMING_SNAKE_CASE ( lowercase_ = "" , lowercase_=50 , lowercase_="*" ) -> str: """simple docstring""" if not s: return "\n" + width * char A__ , A__ = divmod(width - len(lowercase_ ) - 2 , 2 ) return f"""{left * char} {s} {(left + extra) * char}""" if __name__ == "__main__": import doctest doctest.testmod() print(prompt("""Binary Tree Traversals""")) _lowerCamelCase : TreeNode = build_tree() print(prompt("""Pre Order Traversal""")) pre_order(node) print(prompt() + """\n""") print(prompt("""In Order Traversal""")) in_order(node) print(prompt() + """\n""") print(prompt("""Post Order Traversal""")) post_order(node) print(prompt() + """\n""") print(prompt("""Level Order Traversal""")) level_order(node) print(prompt() + """\n""") print(prompt("""Actual Level Order Traversal""")) level_order_actual(node) print("""*""" * 50 + """\n""") print(prompt("""Pre Order Traversal - Iteration Version""")) pre_order_iter(node) print(prompt() + """\n""") print(prompt("""In Order Traversal - Iteration Version""")) in_order_iter(node) print(prompt() + """\n""") print(prompt("""Post Order Traversal - Iteration Version""")) post_order_iter(node) print(prompt())
14
'''simple docstring''' from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Optional[int] = tau * frequency / samplerate UpperCAmelCase : List[Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = cos(_lowerCAmelCase ) UpperCAmelCase : int = _sin / (2 * q_factor) UpperCAmelCase : Any = (1 - _cos) / 2 UpperCAmelCase : List[Any] = 1 - _cos UpperCAmelCase : Union[str, Any] = 1 + alpha UpperCAmelCase : Any = -2 * _cos UpperCAmelCase : Dict = 1 - alpha UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Any = tau * frequency / samplerate UpperCAmelCase : Tuple = sin(_lowerCAmelCase ) UpperCAmelCase : Tuple = cos(_lowerCAmelCase ) UpperCAmelCase : Dict = _sin / (2 * q_factor) UpperCAmelCase : int = (1 + _cos) / 2 UpperCAmelCase : List[Any] = -1 - _cos UpperCAmelCase : Tuple = 1 + alpha UpperCAmelCase : List[str] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha UpperCAmelCase : Tuple = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Optional[int] = tau * frequency / samplerate UpperCAmelCase : Optional[int] = sin(_lowerCAmelCase ) UpperCAmelCase : Tuple = cos(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = _sin / (2 * q_factor) UpperCAmelCase : Union[str, Any] = _sin / 2 UpperCAmelCase : Any = 0 UpperCAmelCase : int = -ba UpperCAmelCase : Optional[Any] = 1 + alpha UpperCAmelCase : List[Any] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha UpperCAmelCase : int = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : List[str] = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : str = cos(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = _sin / (2 * q_factor) UpperCAmelCase : List[str] = 1 - alpha UpperCAmelCase : Any = -2 * _cos UpperCAmelCase : Optional[int] = 1 + alpha UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([ba, ba, ba] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : Optional[Any] = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = cos(_lowerCAmelCase ) UpperCAmelCase : Dict = _sin / (2 * q_factor) UpperCAmelCase : str = 10 ** (gain_db / 40) UpperCAmelCase : int = 1 + alpha * big_a UpperCAmelCase : Union[str, Any] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha * big_a UpperCAmelCase : Union[str, Any] = 1 + alpha / big_a UpperCAmelCase : Tuple = -2 * _cos UpperCAmelCase : Any = 1 - alpha / big_a UpperCAmelCase : Optional[Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : Any = tau * frequency / samplerate UpperCAmelCase : Optional[int] = sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = cos(_lowerCAmelCase ) UpperCAmelCase : str = _sin / (2 * q_factor) UpperCAmelCase : List[str] = 10 ** (gain_db / 40) UpperCAmelCase : Optional[int] = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase : int = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase : int = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase : Optional[int] = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase : str = 2 * sqrt(_lowerCAmelCase ) * alpha UpperCAmelCase : Dict = big_a * (pmc + aaa) UpperCAmelCase : Any = 2 * big_a * mpc UpperCAmelCase : Union[str, Any] = big_a * (pmc - aaa) UpperCAmelCase : Optional[int] = ppmc + aaa UpperCAmelCase : Optional[Any] = -2 * pmpc UpperCAmelCase : Optional[Any] = ppmc - aaa UpperCAmelCase : int = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : int = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = cos(_lowerCAmelCase ) UpperCAmelCase : Any = _sin / (2 * q_factor) UpperCAmelCase : int = 10 ** (gain_db / 40) UpperCAmelCase : List[str] = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase : Union[str, Any] = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase : Optional[Any] = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase : Union[str, Any] = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase : List[str] = 2 * sqrt(_lowerCAmelCase ) * alpha UpperCAmelCase : Any = big_a * (ppmc + aaa) UpperCAmelCase : str = -2 * big_a * pmpc UpperCAmelCase : List[Any] = big_a * (ppmc - aaa) UpperCAmelCase : Optional[Any] = pmc + aaa UpperCAmelCase : Any = 2 * mpc UpperCAmelCase : str = pmc - aaa UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt
23
0
from typing import Optional, Tuple import jax import jax.numpy as jnp from flax import linen as nn from flax.core.frozen_dict import FrozenDict from transformers import CLIPConfig, FlaxPreTrainedModel from transformers.models.clip.modeling_flax_clip import FlaxCLIPVisionModule def UpperCAmelCase ( a_ , a_ , a_=1E-12 ) -> List[str]: """simple docstring""" __A = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(a_ , axis=1 ) , a_min=a_ ) ).T __A = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(a_ , axis=1 ) , a_min=a_ ) ).T return jnp.matmul(a_ , norm_emb_a.T ) class UpperCAmelCase ( nn.Module ): '''simple docstring''' snake_case_ = 42 snake_case_ = jnp.floataa def UpperCamelCase_ ( self : List[str] ): __A = FlaxCLIPVisionModule(self.config.vision_config ) __A = nn.Dense(self.config.projection_dim ,use_bias=A ,dtype=self.dtype ) __A = self.param("concept_embeds" ,jax.nn.initializers.ones ,(17, self.config.projection_dim) ) __A = self.param( "special_care_embeds" ,jax.nn.initializers.ones ,(3, self.config.projection_dim) ) __A = self.param("concept_embeds_weights" ,jax.nn.initializers.ones ,(17,) ) __A = self.param("special_care_embeds_weights" ,jax.nn.initializers.ones ,(3,) ) def __call__( self : Tuple ,A : Any ): __A = self.vision_model(A )[1] __A = self.visual_projection(A ) __A = jax_cosine_distance(A ,self.special_care_embeds ) __A = jax_cosine_distance(A ,self.concept_embeds ) # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign image inputs __A = 0.0 __A = special_cos_dist - self.special_care_embeds_weights[None, :] + adjustment __A = jnp.round(A ,3 ) __A = jnp.any(special_scores > 0 ,axis=1 ,keepdims=A ) # Use a lower threshold if an image has any special care concept __A = is_special_care * 0.01 __A = cos_dist - self.concept_embeds_weights[None, :] + special_adjustment __A = jnp.round(A ,3 ) __A = jnp.any(concept_scores > 0 ,axis=1 ) return has_nsfw_concepts class UpperCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' snake_case_ = CLIPConfig snake_case_ = "clip_input" snake_case_ = FlaxStableDiffusionSafetyCheckerModule def __init__( self : int ,A : CLIPConfig ,A : Optional[Tuple] = None ,A : int = 0 ,A : jnp.dtype = jnp.floataa ,A : bool = True ,**A : Tuple ,): if input_shape is None: __A = (1, 2_24, 2_24, 3) __A = self.module_class(config=A ,dtype=A ,**A ) super().__init__(A ,A ,input_shape=A ,seed=A ,dtype=A ,_do_init=_do_init ) def UpperCamelCase_ ( self : int ,A : jax.random.KeyArray ,A : Tuple ,A : FrozenDict = None ): # init input tensor __A = jax.random.normal(A ,A ) __A , __A = jax.random.split(A ) __A = {"params": params_rng, "dropout": dropout_rng} __A = self.module.init(A ,A )["params"] return random_params def __call__( self : Tuple ,A : Dict ,A : dict = None ,): __A = jnp.transpose(A ,(0, 2, 3, 1) ) return self.module.apply( {"params": params or self.params} ,jnp.array(A ,dtype=jnp.floataa ) ,rngs={} ,)
15
'''simple docstring''' from __future__ import annotations def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : str ) -> bool: UpperCAmelCase : str = get_failure_array(_lowerCAmelCase ) # 2) Step through text searching for pattern UpperCAmelCase , UpperCAmelCase : Optional[Any] = 0, 0 # index into text, pattern while i < len(_lowerCAmelCase ): if pattern[j] == text[i]: if j == (len(_lowerCAmelCase ) - 1): return True j += 1 # if this is a prefix in our pattern # just go back far enough to continue elif j > 0: UpperCAmelCase : Optional[Any] = failure[j - 1] continue i += 1 return False def snake_case_ ( _lowerCAmelCase : str ) -> list[int]: UpperCAmelCase : Optional[Any] = [0] UpperCAmelCase : str = 0 UpperCAmelCase : List[str] = 1 while j < len(_lowerCAmelCase ): if pattern[i] == pattern[j]: i += 1 elif i > 0: UpperCAmelCase : Union[str, Any] = failure[i - 1] continue j += 1 failure.append(_lowerCAmelCase ) return failure if __name__ == "__main__": # Test 1) UpperCamelCase__: str = "abc1abc12" UpperCamelCase__: str = "alskfjaldsabc1abc1abc12k23adsfabcabc" UpperCamelCase__: Any = "alskfjaldsk23adsfabcabc" assert kmp(pattern, texta) and not kmp(pattern, texta) # Test 2) UpperCamelCase__: Tuple = "ABABX" UpperCamelCase__: Union[str, Any] = "ABABZABABYABABX" assert kmp(pattern, text) # Test 3) UpperCamelCase__: Any = "AAAB" UpperCamelCase__: str = "ABAAAAAB" assert kmp(pattern, text) # Test 4) UpperCamelCase__: int = "abcdabcy" UpperCamelCase__: Any = "abcxabcdabxabcdabcdabcy" assert kmp(pattern, text) # Test 5) UpperCamelCase__: List[str] = "aabaabaaa" assert get_failure_array(pattern) == [0, 1, 0, 1, 2, 3, 4, 5, 2]
23
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available, is_torch_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, is_pt_tf_cross_test, slow if is_tf_available(): from transformers import ( AutoConfig, BertConfig, GPTaConfig, TaConfig, TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST if is_torch_available(): from transformers import ( AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelWithLMHead, BertForMaskedLM, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertModel, GPTaLMHeadModel, RobertaForMaskedLM, TaForConditionalGeneration, ) @is_pt_tf_cross_test class __A ( unittest.TestCase ): '''simple docstring''' @slow def UpperCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" for model_name in ["bert-base-uncased"]: lowercase__ : Tuple = AutoConfig.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : Dict = TFAutoModel.from_pretrained(_snake_case ,from_pt=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : List[str] = AutoModel.from_pretrained(_snake_case ,from_tf=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) @slow def UpperCAmelCase ( self : Union[str, Any] ) -> Tuple: """simple docstring""" for model_name in ["bert-base-uncased"]: lowercase__ : Dict = AutoConfig.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : str = TFAutoModelForPreTraining.from_pretrained(_snake_case ,from_pt=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : Optional[Any] = AutoModelForPreTraining.from_pretrained(_snake_case ,from_tf=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) @slow def UpperCAmelCase ( self : Tuple ) -> Dict: """simple docstring""" for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ : Any = AutoConfig.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : List[str] = TFAutoModelForCausalLM.from_pretrained(_snake_case ,from_pt=_snake_case ) lowercase__ , lowercase__ : Optional[Any] = TFAutoModelForCausalLM.from_pretrained( _snake_case ,output_loading_info=_snake_case ,from_pt=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : Union[str, Any] = AutoModelForCausalLM.from_pretrained(_snake_case ,from_tf=_snake_case ) lowercase__ , lowercase__ : Optional[Any] = AutoModelForCausalLM.from_pretrained( _snake_case ,output_loading_info=_snake_case ,from_tf=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) @slow def UpperCAmelCase ( self : Any ) -> Tuple: """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ : Any = AutoConfig.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : Optional[Any] = TFAutoModelWithLMHead.from_pretrained(_snake_case ,from_pt=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : Any = AutoModelWithLMHead.from_pretrained(_snake_case ,from_tf=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) @slow def UpperCAmelCase ( self : List[str] ) -> Optional[Any]: """simple docstring""" for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ : str = AutoConfig.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : Union[str, Any] = TFAutoModelForMaskedLM.from_pretrained(_snake_case ,from_pt=_snake_case ) lowercase__ , lowercase__ : str = TFAutoModelForMaskedLM.from_pretrained( _snake_case ,output_loading_info=_snake_case ,from_pt=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : List[str] = AutoModelForMaskedLM.from_pretrained(_snake_case ,from_tf=_snake_case ) lowercase__ , lowercase__ : Any = AutoModelForMaskedLM.from_pretrained( _snake_case ,output_loading_info=_snake_case ,from_tf=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) @slow def UpperCAmelCase ( self : List[Any] ) -> Dict: """simple docstring""" for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ : Union[str, Any] = AutoConfig.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : List[str] = TFAutoModelForSeqaSeqLM.from_pretrained(_snake_case ,from_pt=_snake_case ) lowercase__ , lowercase__ : List[str] = TFAutoModelForSeqaSeqLM.from_pretrained( _snake_case ,output_loading_info=_snake_case ,from_pt=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : Any = AutoModelForSeqaSeqLM.from_pretrained(_snake_case ,from_tf=_snake_case ) lowercase__ , lowercase__ : Optional[int] = AutoModelForSeqaSeqLM.from_pretrained( _snake_case ,output_loading_info=_snake_case ,from_tf=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) @slow def UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" for model_name in ["bert-base-uncased"]: lowercase__ : Tuple = AutoConfig.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : Any = TFAutoModelForSequenceClassification.from_pretrained(_snake_case ,from_pt=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : Union[str, Any] = AutoModelForSequenceClassification.from_pretrained(_snake_case ,from_tf=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) @slow def UpperCAmelCase ( self : Tuple ) -> Optional[Any]: """simple docstring""" for model_name in ["bert-base-uncased"]: lowercase__ : List[Any] = AutoConfig.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : str = TFAutoModelForQuestionAnswering.from_pretrained(_snake_case ,from_pt=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) lowercase__ : Any = AutoModelForQuestionAnswering.from_pretrained(_snake_case ,from_tf=_snake_case ) self.assertIsNotNone(_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) def UpperCAmelCase ( self : Dict ) -> Any: """simple docstring""" lowercase__ : Optional[Any] = TFAutoModelWithLMHead.from_pretrained(_snake_case ,from_pt=_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) self.assertEqual(model.num_parameters() ,14_410 ) self.assertEqual(model.num_parameters(only_trainable=_snake_case ) ,14_410 ) lowercase__ : Union[str, Any] = AutoModelWithLMHead.from_pretrained(_snake_case ,from_tf=_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) self.assertEqual(model.num_parameters() ,14_410 ) self.assertEqual(model.num_parameters(only_trainable=_snake_case ) ,14_410 ) def UpperCAmelCase ( self : int ) -> List[Any]: """simple docstring""" lowercase__ : List[Any] = TFAutoModelWithLMHead.from_pretrained(_snake_case ,from_pt=_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) self.assertEqual(model.num_parameters() ,14_410 ) self.assertEqual(model.num_parameters(only_trainable=_snake_case ) ,14_410 ) lowercase__ : int = AutoModelWithLMHead.from_pretrained(_snake_case ,from_tf=_snake_case ) self.assertIsInstance(_snake_case ,_snake_case ) self.assertEqual(model.num_parameters() ,14_410 ) self.assertEqual(model.num_parameters(only_trainable=_snake_case ) ,14_410 )
16
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() UpperCamelCase__: int = logging.get_logger(__name__) UpperCamelCase__: Dict = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "adapter_layer": "encoder.layers.*.adapter_layer", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", "pooling_layer.linear": "projector", "pooling_layer.projection": "classifier", } UpperCamelCase__: Optional[Any] = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", "projector", "classifier", ] def snake_case_ ( _lowerCAmelCase : str ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = {} with open(_lowerCAmelCase , '''r''' ) as file: for line_number, line in enumerate(_lowerCAmelCase ): UpperCAmelCase : List[str] = line.strip() if line: UpperCAmelCase : str = line.split() UpperCAmelCase : Union[str, Any] = line_number UpperCAmelCase : List[Any] = words[0] UpperCAmelCase : Union[str, Any] = value return result def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : str ) -> int: for attribute in key.split('''.''' ): UpperCAmelCase : Dict = getattr(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Dict = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(_lowerCAmelCase ): UpperCAmelCase : Any = PARAM_MAPPING[full_name.split('''.''' )[-1]] UpperCAmelCase : Dict = '''param''' if weight_type is not None and weight_type != "param": UpperCAmelCase : Optional[int] = getattr(_lowerCAmelCase , _lowerCAmelCase ).shape elif weight_type is not None and weight_type == "param": UpperCAmelCase : List[Any] = hf_pointer for attribute in hf_param_name.split('''.''' ): UpperCAmelCase : Optional[Any] = getattr(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : int = shape_pointer.shape # let's reduce dimension UpperCAmelCase : Union[str, Any] = value[0] else: UpperCAmelCase : List[Any] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": UpperCAmelCase : int = value elif weight_type == "weight_g": UpperCAmelCase : str = value elif weight_type == "weight_v": UpperCAmelCase : Dict = value elif weight_type == "bias": UpperCAmelCase : str = value elif weight_type == "param": for attribute in hf_param_name.split('''.''' ): UpperCAmelCase : int = getattr(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = value else: UpperCAmelCase : Tuple = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[Any] ) -> List[Any]: UpperCAmelCase : List[str] = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(_lowerCAmelCase ): UpperCAmelCase : List[str] = PARAM_MAPPING[full_name.split('''.''' )[-1]] UpperCAmelCase : Any = '''param''' if weight_type is not None and weight_type != "param": UpperCAmelCase : Optional[int] = '''.'''.join([key, weight_type] ) elif weight_type is not None and weight_type == "param": UpperCAmelCase : Optional[int] = '''.'''.join([key, hf_param_name] ) else: UpperCAmelCase : List[Any] = key UpperCAmelCase : Tuple = value if '''lm_head''' in full_key else value[0] UpperCamelCase__: Tuple = { "W_a": "linear_1.weight", "W_b": "linear_2.weight", "b_a": "linear_1.bias", "b_b": "linear_2.bias", "ln_W": "norm.weight", "ln_b": "norm.bias", } def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : Any=None , _lowerCAmelCase : Optional[Any]=None ) -> int: UpperCAmelCase : List[Any] = False for key, mapped_key in MAPPING.items(): UpperCAmelCase : int = '''wav2vec2.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: UpperCAmelCase : Optional[Any] = True if "*" in mapped_key: UpperCAmelCase : Tuple = name.split(_lowerCAmelCase )[0].split('''.''' )[-2] UpperCAmelCase : List[Any] = mapped_key.replace('''*''' , _lowerCAmelCase ) if "weight_g" in name: UpperCAmelCase : str = '''weight_g''' elif "weight_v" in name: UpperCAmelCase : int = '''weight_v''' elif "bias" in name: UpperCAmelCase : int = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj UpperCAmelCase : List[str] = '''weight''' else: UpperCAmelCase : Dict = None if hf_dict is not None: rename_dict(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) else: set_recursively(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return is_used return is_used def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Any: UpperCAmelCase : Dict = [] UpperCAmelCase : Dict = fairseq_model.state_dict() UpperCAmelCase : Union[str, Any] = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase : Dict = False if "conv_layers" in name: load_conv_layer( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , hf_model.config.feat_extract_norm == '''group''' , ) UpperCAmelCase : Any = True else: UpperCAmelCase : Optional[Any] = load_wavaveca_layer(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not is_used: unused_weights.append(_lowerCAmelCase ) logger.warning(f"""Unused weights: {unused_weights}""" ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: UpperCAmelCase : Any = full_name.split('''conv_layers.''' )[-1] UpperCAmelCase : Optional[int] = name.split('''.''' ) UpperCAmelCase : Tuple = int(items[0] ) UpperCAmelCase : Tuple = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) UpperCAmelCase : Tuple = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) UpperCAmelCase : Union[str, Any] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.""" ) UpperCAmelCase : Union[str, Any] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.""" ) UpperCAmelCase : List[str] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(_lowerCAmelCase ) @torch.no_grad() def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict=None , _lowerCAmelCase : List[Any]=None , _lowerCAmelCase : int=True , _lowerCAmelCase : Optional[int]=False ) -> Dict: if config_path is not None: UpperCAmelCase : List[str] = WavaVecaConfig.from_pretrained(_lowerCAmelCase ) else: UpperCAmelCase : List[Any] = WavaVecaConfig() if is_seq_class: UpperCAmelCase : Optional[Any] = read_txt_into_dict(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = idalabel UpperCAmelCase : Optional[Any] = WavaVecaForSequenceClassification(_lowerCAmelCase ) UpperCAmelCase : Dict = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase , ) feature_extractor.save_pretrained(_lowerCAmelCase ) elif is_finetuned: if dict_path: UpperCAmelCase : Dict = Dictionary.load(_lowerCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq UpperCAmelCase : Any = target_dict.pad_index UpperCAmelCase : Tuple = target_dict.bos_index UpperCAmelCase : Optional[int] = target_dict.eos_index UpperCAmelCase : Union[str, Any] = len(target_dict.symbols ) UpperCAmelCase : Dict = os.path.join(_lowerCAmelCase , '''vocab.json''' ) if not os.path.isdir(_lowerCAmelCase ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(_lowerCAmelCase ) ) return os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : List[Any] = target_dict.indices # fairseq has the <pad> and <s> switched UpperCAmelCase : List[str] = 0 UpperCAmelCase : List[str] = 1 with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = WavaVecaCTCTokenizer( _lowerCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=_lowerCAmelCase , ) UpperCAmelCase : int = True if config.feat_extract_norm == '''layer''' else False UpperCAmelCase : int = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase , ) UpperCAmelCase : str = WavaVecaProcessor(feature_extractor=_lowerCAmelCase , tokenizer=_lowerCAmelCase ) processor.save_pretrained(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = WavaVecaForCTC(_lowerCAmelCase ) else: UpperCAmelCase : Dict = WavaVecaForPreTraining(_lowerCAmelCase ) if is_finetuned or is_seq_class: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : List[str] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: UpperCAmelCase : Optional[Any] = argparse.Namespace(task='''audio_pretraining''' ) UpperCAmelCase : List[Any] = fairseq.tasks.setup_task(_lowerCAmelCase ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : int = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=_lowerCAmelCase ) UpperCAmelCase : Optional[int] = model[0].eval() recursively_load_weights(_lowerCAmelCase , _lowerCAmelCase , not is_finetuned ) hf_wavavec.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": UpperCamelCase__: Dict = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--not_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) parser.add_argument( "--is_seq_class", action="store_true", help="Whether the model to convert is a fine-tuned sequence classification model or not", ) UpperCamelCase__: Any = parser.parse_args() UpperCamelCase__: int = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
23
0
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def _lowercase ( self : str ): __lowercase = FlaxMTaForConditionalGeneration.from_pretrained("google/mt5-small" ) __lowercase = AutoTokenizer.from_pretrained("google/mt5-small" ) __lowercase = tokenizer("Hello there", return_tensors="np" ).input_ids __lowercase = tokenizer("Hi I am", return_tensors="np" ).input_ids __lowercase = shift_tokens_right(UpperCAmelCase__, model.config.pad_token_id, model.config.decoder_start_token_id ) __lowercase = model(UpperCAmelCase__, decoder_input_ids=UpperCAmelCase__ ).logits __lowercase = optax.softmax_cross_entropy(UpperCAmelCase__, onehot(UpperCAmelCase__, logits.shape[-1] ) ).mean() __lowercase = -(labels.shape[-1] * loss.item()) __lowercase = -84.9_127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
17
'''simple docstring''' import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : str ) -> int: with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights UpperCAmelCase : Optional[Any] = FlaxDiffusionPipeline.from_pretrained( '''hf-internal-testing/tiny-stable-diffusion-pipe''' , safety_checker=__snake_case , cache_dir=__snake_case ) UpperCAmelCase : str = [t[-1] for t in os.walk(os.path.join(__snake_case , os.listdir(__snake_case )[0] , '''snapshots''' ) )] UpperCAmelCase : str = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith('''.bin''' ) for f in files ) @slow @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : List[str] ) -> Dict: UpperCAmelCase , UpperCAmelCase : str = FlaxStableDiffusionPipeline.from_pretrained( '''hf-internal-testing/tiny-stable-diffusion-pipe''' , safety_checker=__snake_case ) UpperCAmelCase : List[Any] = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : List[str] = jax.random.PRNGKey(0 ) UpperCAmelCase : Optional[Any] = 4 UpperCAmelCase : Optional[Any] = jax.device_count() UpperCAmelCase : Tuple = num_samples * [prompt] UpperCAmelCase : int = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Tuple = replicate(__snake_case ) UpperCAmelCase : Any = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Optional[Any] = shard(__snake_case ) UpperCAmelCase : Optional[int] = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1E-3 assert np.abs(np.abs(__snake_case , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5E-1 UpperCAmelCase : Union[str, Any] = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(__snake_case ) == num_samples def A ( self : List[Any] ) -> List[str]: UpperCAmelCase , UpperCAmelCase : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''flax''' , safety_checker=__snake_case ) UpperCAmelCase : Dict = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : Optional[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase : Any = 50 UpperCAmelCase : Union[str, Any] = jax.device_count() UpperCAmelCase : int = num_samples * [prompt] UpperCAmelCase : Union[str, Any] = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Dict = replicate(__snake_case ) UpperCAmelCase : int = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Tuple = shard(__snake_case ) UpperCAmelCase : Tuple = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5E-1 def A ( self : int ) -> Dict: UpperCAmelCase , UpperCAmelCase : List[str] = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__snake_case ) UpperCAmelCase : Dict = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : Union[str, Any] = jax.random.PRNGKey(0 ) UpperCAmelCase : List[str] = 50 UpperCAmelCase : Union[str, Any] = jax.device_count() UpperCAmelCase : List[Any] = num_samples * [prompt] UpperCAmelCase : int = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Tuple = replicate(__snake_case ) UpperCAmelCase : List[Any] = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Optional[int] = shard(__snake_case ) UpperCAmelCase : Optional[Any] = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def A ( self : int ) -> Any: UpperCAmelCase , UpperCAmelCase : Dict = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa ) UpperCAmelCase : List[str] = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : List[str] = jax.random.PRNGKey(0 ) UpperCAmelCase : Union[str, Any] = 50 UpperCAmelCase : Optional[int] = jax.device_count() UpperCAmelCase : List[str] = num_samples * [prompt] UpperCAmelCase : Dict = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Tuple = replicate(__snake_case ) UpperCAmelCase : Any = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : str = shard(__snake_case ) UpperCAmelCase : Optional[int] = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def A ( self : Tuple ) -> Optional[Any]: UpperCAmelCase : int = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , set_alpha_to_one=__snake_case , steps_offset=1 , ) UpperCAmelCase , UpperCAmelCase : Tuple = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , scheduler=__snake_case , safety_checker=__snake_case , ) UpperCAmelCase : Tuple = scheduler.create_state() UpperCAmelCase : Dict = scheduler_state UpperCAmelCase : str = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : int = jax.random.PRNGKey(0 ) UpperCAmelCase : Union[str, Any] = 50 UpperCAmelCase : Optional[Any] = jax.device_count() UpperCAmelCase : Any = num_samples * [prompt] UpperCAmelCase : Dict = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : str = replicate(__snake_case ) UpperCAmelCase : List[str] = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Optional[int] = shard(__snake_case ) UpperCAmelCase : Dict = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5E-1 def A ( self : Any ) -> Tuple: UpperCAmelCase : List[Any] = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : Union[str, Any] = jax.device_count() UpperCAmelCase : List[Any] = num_samples * [prompt] UpperCAmelCase : str = jax.random.split(jax.random.PRNGKey(0 ) , __snake_case ) UpperCAmelCase , UpperCAmelCase : Tuple = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__snake_case , ) UpperCAmelCase : Dict = replicate(__snake_case ) UpperCAmelCase : Optional[Any] = pipeline.prepare_inputs(__snake_case ) UpperCAmelCase : List[str] = shard(__snake_case ) UpperCAmelCase : Any = pipeline(__snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase : Optional[int] = images[2, 0, 256, 10:17, 1] # With memory efficient attention UpperCAmelCase , UpperCAmelCase : Any = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__snake_case , use_memory_efficient_attention=__snake_case , ) UpperCAmelCase : int = replicate(__snake_case ) UpperCAmelCase : int = pipeline.prepare_inputs(__snake_case ) UpperCAmelCase : List[Any] = shard(__snake_case ) UpperCAmelCase : Optional[Any] = pipeline(__snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase : int = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
23
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCamelCase : Dict = { '''configuration_biogpt''': ['''BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BioGptConfig'''], '''tokenization_biogpt''': ['''BioGptTokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : Tuple = [ '''BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BioGptForCausalLM''', '''BioGptForTokenClassification''', '''BioGptForSequenceClassification''', '''BioGptModel''', '''BioGptPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys __lowerCamelCase : Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
18
'''simple docstring''' import random from .binary_exp_mod import bin_exp_mod def snake_case_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]=1000 ) -> int: if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd UpperCAmelCase : str = n - 1 UpperCAmelCase : List[Any] = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) UpperCAmelCase : List[str] = 0 while count < prec: UpperCAmelCase : int = random.randint(2 , n - 1 ) UpperCAmelCase : List[str] = bin_exp_mod(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if b != 1: UpperCAmelCase : int = True for _ in range(_lowerCAmelCase ): if b == n - 1: UpperCAmelCase : Dict = False break UpperCAmelCase : str = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": UpperCamelCase__: Optional[int] = abs(int(input("Enter bound : ").strip())) print("Here's the list of primes:") print(", ".join(str(i) for i in range(n + 1) if is_prime_big(i)))
23
0
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging __A =logging.get_logger(__name__) __A ={'''vocab_file''': '''spiece.model'''} __A ={ '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', } } __A ={ '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } # Segments (not really needed) __A =0 __A =1 __A =2 __A =3 __A =4 class _SCREAMING_SNAKE_CASE ( snake_case_ ): lowerCAmelCase__ = VOCAB_FILES_NAMES lowerCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase__ = 'left' def __init__( self , lowercase , lowercase=False , lowercase=True , lowercase=False , lowercase="<s>" , lowercase="</s>" , lowercase="<unk>" , lowercase="<sep>" , lowercase="<pad>" , lowercase="<cls>" , lowercase="<mask>" , lowercase=["<eop>", "<eod>"] , lowercase = None , **lowercase , ) -> None: # Mask token behave like a normal word, i.e. include the space before it lowerCamelCase_ = AddedToken(lowercase , lstrip=lowercase , rstrip=lowercase ) if isinstance(lowercase , lowercase ) else mask_token lowerCamelCase_ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=lowercase , remove_space=lowercase , keep_accents=lowercase , bos_token=lowercase , eos_token=lowercase , unk_token=lowercase , sep_token=lowercase , pad_token=lowercase , cls_token=lowercase , mask_token=lowercase , additional_special_tokens=lowercase , sp_model_kwargs=self.sp_model_kwargs , **lowercase , ) lowerCamelCase_ = 3 lowerCamelCase_ = do_lower_case lowerCamelCase_ = remove_space lowerCamelCase_ = keep_accents lowerCamelCase_ = vocab_file lowerCamelCase_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(lowercase ) @property def SCREAMING_SNAKE_CASE_( self ) -> int: return len(self.sp_model ) def SCREAMING_SNAKE_CASE_( self ) -> Optional[int]: lowerCamelCase_ = {self.convert_ids_to_tokens(lowercase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) -> str: lowerCamelCase_ = self.__dict__.copy() lowerCamelCase_ = None return state def __setstate__( self , lowercase ) -> Optional[Any]: lowerCamelCase_ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): lowerCamelCase_ = {} lowerCamelCase_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE_( self , lowercase ) -> Any: if self.remove_space: lowerCamelCase_ = " ".join(inputs.strip().split() ) else: lowerCamelCase_ = inputs lowerCamelCase_ = outputs.replace("``" , "\"" ).replace("''" , "\"" ) if not self.keep_accents: lowerCamelCase_ = unicodedata.normalize("NFKD" , lowercase ) lowerCamelCase_ = "".join([c for c in outputs if not unicodedata.combining(lowercase )] ) if self.do_lower_case: lowerCamelCase_ = outputs.lower() return outputs def SCREAMING_SNAKE_CASE_( self , lowercase ) -> List[str]: lowerCamelCase_ = self.preprocess_text(lowercase ) lowerCamelCase_ = self.sp_model.encode(lowercase , out_type=lowercase ) lowerCamelCase_ = [] for piece in pieces: if len(lowercase ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): lowerCamelCase_ = self.sp_model.EncodeAsPieces(piece[:-1].replace(lowercase , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowerCamelCase_ = cur_pieces[1:] else: lowerCamelCase_ = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(lowercase ) else: new_pieces.append(lowercase ) return new_pieces def SCREAMING_SNAKE_CASE_( self , lowercase ) -> Union[str, Any]: return self.sp_model.PieceToId(lowercase ) def SCREAMING_SNAKE_CASE_( self , lowercase ) -> List[Any]: return self.sp_model.IdToPiece(lowercase ) def SCREAMING_SNAKE_CASE_( self , lowercase ) -> Tuple: lowerCamelCase_ = "".join(lowercase ).replace(lowercase , " " ).strip() return out_string def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = False , lowercase = None , lowercase = True , **lowercase , ) -> str: lowerCamelCase_ = kwargs.pop("use_source_tokenizer" , lowercase ) lowerCamelCase_ = self.convert_ids_to_tokens(lowercase , skip_special_tokens=lowercase ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 lowerCamelCase_ = [] lowerCamelCase_ = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(lowercase ) ) lowerCamelCase_ = [] sub_texts.append(lowercase ) else: current_sub_text.append(lowercase ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(lowercase ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens lowerCamelCase_ = "".join(lowercase ) lowerCamelCase_ = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: lowerCamelCase_ = self.clean_up_tokenization(lowercase ) return clean_text else: return text def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None ) -> List[int]: lowerCamelCase_ = [self.sep_token_id] lowerCamelCase_ = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None , lowercase = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase , token_ids_a=lowercase , already_has_special_tokens=lowercase ) if token_ids_a is not None: return ([0] * len(lowercase )) + [1] + ([0] * len(lowercase )) + [1, 1] return ([0] * len(lowercase )) + [1, 1] def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None ) -> List[int]: lowerCamelCase_ = [self.sep_token_id] lowerCamelCase_ = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase = None ) -> Tuple[str]: if not os.path.isdir(lowercase ): logger.error(f'Vocabulary path ({save_directory}) should be a directory' ) return lowerCamelCase_ = os.path.join( lowercase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , lowercase ) elif not os.path.isfile(self.vocab_file ): with open(lowercase , "wb" ) as fi: lowerCamelCase_ = self.sp_model.serialized_model_proto() fi.write(lowercase ) return (out_vocab_file,)
19
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__: Optional[int] = logging.get_logger(__name__) def snake_case_ ( _lowerCAmelCase : Optional[int] ) -> Optional[int]: UpperCAmelCase : Tuple = DPTConfig(embedding_type='''hybrid''' ) if "large" in checkpoint_url: UpperCAmelCase : Tuple = 1024 UpperCAmelCase : List[Any] = 4096 UpperCAmelCase : str = 24 UpperCAmelCase : List[Any] = 16 UpperCAmelCase : str = [5, 11, 17, 23] UpperCAmelCase : List[Any] = [256, 512, 1024, 1024] UpperCAmelCase : Tuple = (1, 384, 384) if "nyu" or "midas" in checkpoint_url: UpperCAmelCase : Optional[Any] = 768 UpperCAmelCase : Tuple = [1, 1, 1, 0.5] UpperCAmelCase : int = [256, 512, 768, 768] UpperCAmelCase : Any = 150 UpperCAmelCase : Tuple = 16 UpperCAmelCase : Any = (1, 384, 384) UpperCAmelCase : Optional[Any] = False UpperCAmelCase : Tuple = '''project''' if "ade" in checkpoint_url: UpperCAmelCase : Any = True UpperCAmelCase : str = 768 UpperCAmelCase : Optional[int] = [1, 1, 1, 0.5] UpperCAmelCase : List[Any] = 150 UpperCAmelCase : List[Any] = 16 UpperCAmelCase : str = '''huggingface/label-files''' UpperCAmelCase : Tuple = '''ade20k-id2label.json''' UpperCAmelCase : Any = json.load(open(cached_download(hf_hub_url(_lowerCAmelCase , _lowerCAmelCase , repo_type='''dataset''' ) ) , '''r''' ) ) UpperCAmelCase : Optional[Any] = {int(_lowerCAmelCase ): v for k, v in idalabel.items()} UpperCAmelCase : List[Any] = idalabel UpperCAmelCase : Optional[int] = {v: k for k, v in idalabel.items()} UpperCAmelCase : Union[str, Any] = [1, 150, 480, 480] return config, expected_shape def snake_case_ ( _lowerCAmelCase : Union[str, Any] ) -> int: UpperCAmelCase : List[str] = ['''pretrained.model.head.weight''', '''pretrained.model.head.bias'''] for k in ignore_keys: state_dict.pop(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Tuple ) -> Any: if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): UpperCAmelCase : Tuple = name.replace('''pretrained.model''' , '''dpt.encoder''' ) if "pretrained.model" in name: UpperCAmelCase : Union[str, Any] = name.replace('''pretrained.model''' , '''dpt.embeddings''' ) if "patch_embed" in name: UpperCAmelCase : int = name.replace('''patch_embed''' , '''''' ) if "pos_embed" in name: UpperCAmelCase : Tuple = name.replace('''pos_embed''' , '''position_embeddings''' ) if "attn.proj" in name: UpperCAmelCase : Any = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "proj" in name and "project" not in name: UpperCAmelCase : str = name.replace('''proj''' , '''projection''' ) if "blocks" in name: UpperCAmelCase : Any = name.replace('''blocks''' , '''layer''' ) if "mlp.fc1" in name: UpperCAmelCase : Optional[int] = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: UpperCAmelCase : Optional[Any] = name.replace('''mlp.fc2''' , '''output.dense''' ) if "norm1" in name and "backbone" not in name: UpperCAmelCase : Dict = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name and "backbone" not in name: UpperCAmelCase : Tuple = name.replace('''norm2''' , '''layernorm_after''' ) if "scratch.output_conv" in name: UpperCAmelCase : Tuple = name.replace('''scratch.output_conv''' , '''head''' ) if "scratch" in name: UpperCAmelCase : str = name.replace('''scratch''' , '''neck''' ) if "layer1_rn" in name: UpperCAmelCase : Dict = name.replace('''layer1_rn''' , '''convs.0''' ) if "layer2_rn" in name: UpperCAmelCase : int = name.replace('''layer2_rn''' , '''convs.1''' ) if "layer3_rn" in name: UpperCAmelCase : Tuple = name.replace('''layer3_rn''' , '''convs.2''' ) if "layer4_rn" in name: UpperCAmelCase : int = name.replace('''layer4_rn''' , '''convs.3''' ) if "refinenet" in name: UpperCAmelCase : List[str] = int(name[len('''neck.refinenet''' ) : len('''neck.refinenet''' ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 UpperCAmelCase : str = name.replace(f"""refinenet{layer_idx}""" , f"""fusion_stage.layers.{abs(layer_idx-4 )}""" ) if "out_conv" in name: UpperCAmelCase : List[str] = name.replace('''out_conv''' , '''projection''' ) if "resConfUnit1" in name: UpperCAmelCase : Union[str, Any] = name.replace('''resConfUnit1''' , '''residual_layer1''' ) if "resConfUnit2" in name: UpperCAmelCase : Any = name.replace('''resConfUnit2''' , '''residual_layer2''' ) if "conv1" in name: UpperCAmelCase : Optional[int] = name.replace('''conv1''' , '''convolution1''' ) if "conv2" in name: UpperCAmelCase : Tuple = name.replace('''conv2''' , '''convolution2''' ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: UpperCAmelCase : Dict = name.replace('''pretrained.act_postprocess1.0.project.0''' , '''neck.reassemble_stage.readout_projects.0.0''' ) if "pretrained.act_postprocess2.0.project.0" in name: UpperCAmelCase : int = name.replace('''pretrained.act_postprocess2.0.project.0''' , '''neck.reassemble_stage.readout_projects.1.0''' ) if "pretrained.act_postprocess3.0.project.0" in name: UpperCAmelCase : Any = name.replace('''pretrained.act_postprocess3.0.project.0''' , '''neck.reassemble_stage.readout_projects.2.0''' ) if "pretrained.act_postprocess4.0.project.0" in name: UpperCAmelCase : Optional[Any] = name.replace('''pretrained.act_postprocess4.0.project.0''' , '''neck.reassemble_stage.readout_projects.3.0''' ) # resize blocks if "pretrained.act_postprocess1.3" in name: UpperCAmelCase : List[Any] = name.replace('''pretrained.act_postprocess1.3''' , '''neck.reassemble_stage.layers.0.projection''' ) if "pretrained.act_postprocess1.4" in name: UpperCAmelCase : Any = name.replace('''pretrained.act_postprocess1.4''' , '''neck.reassemble_stage.layers.0.resize''' ) if "pretrained.act_postprocess2.3" in name: UpperCAmelCase : Optional[int] = name.replace('''pretrained.act_postprocess2.3''' , '''neck.reassemble_stage.layers.1.projection''' ) if "pretrained.act_postprocess2.4" in name: UpperCAmelCase : str = name.replace('''pretrained.act_postprocess2.4''' , '''neck.reassemble_stage.layers.1.resize''' ) if "pretrained.act_postprocess3.3" in name: UpperCAmelCase : List[str] = name.replace('''pretrained.act_postprocess3.3''' , '''neck.reassemble_stage.layers.2.projection''' ) if "pretrained.act_postprocess4.3" in name: UpperCAmelCase : Tuple = name.replace('''pretrained.act_postprocess4.3''' , '''neck.reassemble_stage.layers.3.projection''' ) if "pretrained.act_postprocess4.4" in name: UpperCAmelCase : int = name.replace('''pretrained.act_postprocess4.4''' , '''neck.reassemble_stage.layers.3.resize''' ) if "pretrained" in name: UpperCAmelCase : Optional[int] = name.replace('''pretrained''' , '''dpt''' ) if "bn" in name: UpperCAmelCase : Dict = name.replace('''bn''' , '''batch_norm''' ) if "head" in name: UpperCAmelCase : Any = name.replace('''head''' , '''head.head''' ) if "encoder.norm" in name: UpperCAmelCase : Optional[int] = name.replace('''encoder.norm''' , '''layernorm''' ) if "auxlayer" in name: UpperCAmelCase : Union[str, Any] = name.replace('''auxlayer''' , '''auxiliary_head.head''' ) if "backbone" in name: UpperCAmelCase : List[Any] = name.replace('''backbone''' , '''backbone.bit.encoder''' ) if ".." in name: UpperCAmelCase : Optional[int] = name.replace('''..''' , '''.''' ) if "stem.conv" in name: UpperCAmelCase : Optional[Any] = name.replace('''stem.conv''' , '''bit.embedder.convolution''' ) if "blocks" in name: UpperCAmelCase : Optional[int] = name.replace('''blocks''' , '''layers''' ) if "convolution" in name and "backbone" in name: UpperCAmelCase : List[Any] = name.replace('''convolution''' , '''conv''' ) if "layer" in name and "backbone" in name: UpperCAmelCase : List[str] = name.replace('''layer''' , '''layers''' ) if "backbone.bit.encoder.bit" in name: UpperCAmelCase : List[Any] = name.replace('''backbone.bit.encoder.bit''' , '''backbone.bit''' ) if "embedder.conv" in name: UpperCAmelCase : List[Any] = name.replace('''embedder.conv''' , '''embedder.convolution''' ) if "backbone.bit.encoder.stem.norm" in name: UpperCAmelCase : Tuple = name.replace('''backbone.bit.encoder.stem.norm''' , '''backbone.bit.embedder.norm''' ) return name def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : List[Any] ) -> Optional[Any]: for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase : Optional[int] = state_dict.pop(f"""dpt.encoder.layer.{i}.attn.qkv.weight""" ) UpperCAmelCase : Tuple = state_dict.pop(f"""dpt.encoder.layer.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase : Tuple = in_proj_weight[: config.hidden_size, :] UpperCAmelCase : int = in_proj_bias[: config.hidden_size] UpperCAmelCase : List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase : List[str] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase : str = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase : Union[str, Any] = in_proj_bias[-config.hidden_size :] def snake_case_ ( ) -> List[str]: UpperCAmelCase : Optional[int] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' UpperCAmelCase : Optional[int] = Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ) return im @torch.no_grad() def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[str] ) -> Any: UpperCAmelCase , UpperCAmelCase : int = get_dpt_config(_lowerCAmelCase ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") UpperCAmelCase : List[Any] = torch.load(_lowerCAmelCase , map_location='''cpu''' ) # remove certain keys remove_ignore_keys_(_lowerCAmelCase ) # rename keys for key in state_dict.copy().keys(): UpperCAmelCase : Any = state_dict.pop(_lowerCAmelCase ) UpperCAmelCase : List[Any] = val # read in qkv matrices read_in_q_k_v(_lowerCAmelCase , _lowerCAmelCase ) # load HuggingFace model UpperCAmelCase : Optional[Any] = DPTForSemanticSegmentation(_lowerCAmelCase ) if '''ade''' in checkpoint_url else DPTForDepthEstimation(_lowerCAmelCase ) model.load_state_dict(_lowerCAmelCase ) model.eval() # Check outputs on an image UpperCAmelCase : int = 480 if '''ade''' in checkpoint_url else 384 UpperCAmelCase : List[Any] = DPTImageProcessor(size=_lowerCAmelCase ) UpperCAmelCase : Dict = prepare_img() UpperCAmelCase : Optional[int] = image_processor(_lowerCAmelCase , return_tensors='''pt''' ) # forward pass UpperCAmelCase : Any = model(**_lowerCAmelCase ).logits if '''ade''' in checkpoint_url else model(**_lowerCAmelCase ).predicted_depth if show_prediction: UpperCAmelCase : Dict = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode='''bicubic''' , align_corners=_lowerCAmelCase , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 255 ).show() if pytorch_dump_folder_path is not None: Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(_lowerCAmelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_lowerCAmelCase ) if push_to_hub: model.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) image_processor.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) if __name__ == "__main__": UpperCamelCase__: Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint_url", default="https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt", type=str, help="URL of the original DPT checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=False, help="Path to the output PyTorch model directory.", ) parser.add_argument( "--push_to_hub", action="store_true", ) parser.add_argument( "--model_name", default="dpt-large", type=str, help="Name of the model, in case you're pushing to the hub.", ) parser.add_argument( "--show_prediction", action="store_true", ) UpperCamelCase__: Tuple = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
23
0
from math import factorial class __snake_case : def __init__( self ,snake_case ,snake_case ): '''simple docstring''' lowercase : List[str] = real if isinstance(snake_case ,snake_case ): lowercase : Any = [1] * rank else: lowercase : Optional[Any] = rank def __repr__( self ): '''simple docstring''' return ( f"{self.real}+" f"{'+'.join(str(snake_case )+'E'+str(n+1 )for n,dual in enumerate(self.duals ) )}" ) def _SCREAMING_SNAKE_CASE ( self ): '''simple docstring''' lowercase : Dict = self.duals.copy() while cur[-1] == 0: cur.pop(-1 ) return Dual(self.real ,snake_case ) def __add__( self ,snake_case ): '''simple docstring''' if not isinstance(snake_case ,snake_case ): return Dual(self.real + other ,self.duals ) lowercase : Optional[int] = self.duals.copy() lowercase : Optional[Any] = other.duals.copy() if len(snake_case ) > len(snake_case ): o_dual.extend([1] * (len(snake_case ) - len(snake_case )) ) elif len(snake_case ) < len(snake_case ): s_dual.extend([1] * (len(snake_case ) - len(snake_case )) ) lowercase : Dict = [] for i in range(len(snake_case ) ): new_duals.append(s_dual[i] + o_dual[i] ) return Dual(self.real + other.real ,snake_case ) _a : Union[str, Any]= __add__ def __sub__( self ,snake_case ): '''simple docstring''' return self + other * -1 def __mul__( self ,snake_case ): '''simple docstring''' if not isinstance(snake_case ,snake_case ): lowercase : str = [] for i in self.duals: new_duals.append(i * other ) return Dual(self.real * other ,snake_case ) lowercase : Union[str, Any] = [0] * (len(self.duals ) + len(other.duals ) + 1) for i, item in enumerate(self.duals ): for j, jtem in enumerate(other.duals ): new_duals[i + j + 1] += item * jtem for k in range(len(self.duals ) ): new_duals[k] += self.duals[k] * other.real for index in range(len(other.duals ) ): new_duals[index] += other.duals[index] * self.real return Dual(self.real * other.real ,snake_case ) _a : int= __mul__ def __truediv__( self ,snake_case ): '''simple docstring''' if not isinstance(snake_case ,snake_case ): lowercase : Union[str, Any] = [] for i in self.duals: new_duals.append(i / other ) return Dual(self.real / other ,snake_case ) raise ValueError def __floordiv__( self ,snake_case ): '''simple docstring''' if not isinstance(snake_case ,snake_case ): lowercase : List[Any] = [] for i in self.duals: new_duals.append(i // other ) return Dual(self.real // other ,snake_case ) raise ValueError def __pow__( self ,snake_case ): '''simple docstring''' if n < 0 or isinstance(snake_case ,snake_case ): raise ValueError("""power must be a positive integer""" ) if n == 0: return 1 if n == 1: return self lowercase : str = self for _ in range(n - 1 ): x *= self return x def _snake_case( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> Dict: if not callable(SCREAMING_SNAKE_CASE__ ): raise ValueError("""differentiate() requires a function as input for func""" ) if not isinstance(SCREAMING_SNAKE_CASE__ , (float, int) ): raise ValueError("""differentiate() requires a float as input for position""" ) if not isinstance(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ): raise ValueError("""differentiate() requires an int as input for order""" ) lowercase : List[Any] = Dual(SCREAMING_SNAKE_CASE__ , 1 ) lowercase : List[Any] = func(SCREAMING_SNAKE_CASE__ ) if order == 0: return result.real return result.duals[order - 1] * factorial(SCREAMING_SNAKE_CASE__ ) if __name__ == "__main__": import doctest doctest.testmod() def _snake_case( SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: return y**2 * y**4 print(differentiate(f, 9, 2))
20
'''simple docstring''' # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers UpperCamelCase__: Optional[int] = "3" print("Python version:", sys.version) print("transformers version:", transformers.__version__) try: import torch print("Torch version:", torch.__version__) print("Cuda available:", torch.cuda.is_available()) print("Cuda version:", torch.version.cuda) print("CuDNN version:", torch.backends.cudnn.version()) print("Number of GPUs available:", torch.cuda.device_count()) print("NCCL version:", torch.cuda.nccl.version()) except ImportError: print("Torch version:", None) try: import deepspeed print("DeepSpeed version:", deepspeed.__version__) except ImportError: print("DeepSpeed version:", None) try: import tensorflow as tf print("TensorFlow version:", tf.__version__) print("TF GPUs available:", bool(tf.config.list_physical_devices("GPU"))) print("Number of TF GPUs available:", len(tf.config.list_physical_devices("GPU"))) except ImportError: print("TensorFlow version:", None)
23
0
import copy from dataclasses import dataclass from pathlib import Path from typing import Dict, Optional, Union @dataclass class _lowerCamelCase: lowercase_ : Optional[Union[str, Path]] = None lowercase_ : bool = False lowercase_ : bool = False lowercase_ : bool = False lowercase_ : Optional[Dict] = None lowercase_ : Optional[str] = None lowercase_ : bool = False lowercase_ : bool = False lowercase_ : bool = False lowercase_ : bool = True lowercase_ : Optional[int] = None lowercase_ : int = 1 lowercase_ : Optional[Union[str, bool]] = None lowercase_ : bool = False lowercase_ : Optional[Dict] = None lowercase_ : Optional[str] = None def UpperCamelCase ( self) -> "DownloadConfig": """simple docstring""" return self.__class__(**{k: copy.deepcopy(lowerCamelCase) for k, v in self.__dict__.items()})
21
'''simple docstring''' from __future__ import annotations def snake_case_ ( _lowerCAmelCase : list[int | float] , _lowerCAmelCase : int , _lowerCAmelCase : int ) -> int | float: if len(_lowerCAmelCase ) == 0: raise ValueError('''find_max() arg is an empty sequence''' ) if ( left >= len(_lowerCAmelCase ) or left < -len(_lowerCAmelCase ) or right >= len(_lowerCAmelCase ) or right < -len(_lowerCAmelCase ) ): raise IndexError('''list index out of range''' ) if left == right: return nums[left] UpperCAmelCase : List[Any] = (left + right) >> 1 # the middle UpperCAmelCase : Optional[Any] = find_max(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # find max in range[left, mid] UpperCAmelCase : Dict = find_max(_lowerCAmelCase , mid + 1 , _lowerCAmelCase ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
23
0
'''simple docstring''' __SCREAMING_SNAKE_CASE :int = {0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} __SCREAMING_SNAKE_CASE :Any = {0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def UpperCAmelCase_ ( __lowercase : dict[int, list[int]] , __lowercase : int , __lowercase : list[bool] ) -> list[int]: '''simple docstring''' _UpperCAmelCase = True _UpperCAmelCase = [] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(__lowercase , __lowercase , __lowercase ) order.append(__lowercase ) return order def UpperCAmelCase_ ( __lowercase : dict[int, list[int]] , __lowercase : int , __lowercase : list[bool] ) -> list[int]: '''simple docstring''' _UpperCAmelCase = True _UpperCAmelCase = [vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(__lowercase , __lowercase , __lowercase ) return component def UpperCAmelCase_ ( __lowercase : dict[int, list[int]] ) -> list[list[int]]: '''simple docstring''' _UpperCAmelCase = len(__lowercase ) * [False] _UpperCAmelCase = {vert: [] for vert in range(len(__lowercase ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(__lowercase ) _UpperCAmelCase = [] for i, was_visited in enumerate(__lowercase ): if not was_visited: order += topology_sort(__lowercase , __lowercase , __lowercase ) _UpperCAmelCase = [] _UpperCAmelCase = len(__lowercase ) * [False] for i in range(len(__lowercase ) ): _UpperCAmelCase = order[len(__lowercase ) - i - 1] if not visited[vert]: _UpperCAmelCase = find_components(__lowercase , __lowercase , __lowercase ) components_list.append(__lowercase ) return components_list
22
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = 42 lowerCamelCase__ = 42 def __init__( self : Union[str, Any] , __snake_case : UNetaDModel , __snake_case : ScoreSdeVeScheduler ) -> int: super().__init__() self.register_modules(unet=__snake_case , scheduler=__snake_case ) @torch.no_grad() def __call__( self : Optional[int] , __snake_case : int = 1 , __snake_case : int = 2000 , __snake_case : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __snake_case : Optional[str] = "pil" , __snake_case : bool = True , **__snake_case : Optional[int] , ) -> Union[ImagePipelineOutput, Tuple]: UpperCAmelCase : str = self.unet.config.sample_size UpperCAmelCase : Union[str, Any] = (batch_size, 3, img_size, img_size) UpperCAmelCase : int = self.unet UpperCAmelCase : Any = randn_tensor(__snake_case , generator=__snake_case ) * self.scheduler.init_noise_sigma UpperCAmelCase : List[Any] = sample.to(self.device ) self.scheduler.set_timesteps(__snake_case ) self.scheduler.set_sigmas(__snake_case ) for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): UpperCAmelCase : Any = self.scheduler.sigmas[i] * torch.ones(shape[0] , device=self.device ) # correction step for _ in range(self.scheduler.config.correct_steps ): UpperCAmelCase : Union[str, Any] = self.unet(__snake_case , __snake_case ).sample UpperCAmelCase : Optional[Any] = self.scheduler.step_correct(__snake_case , __snake_case , generator=__snake_case ).prev_sample # prediction step UpperCAmelCase : Optional[Any] = model(__snake_case , __snake_case ).sample UpperCAmelCase : List[str] = self.scheduler.step_pred(__snake_case , __snake_case , __snake_case , generator=__snake_case ) UpperCAmelCase , UpperCAmelCase : Optional[Any] = output.prev_sample, output.prev_sample_mean UpperCAmelCase : int = sample_mean.clamp(0 , 1 ) UpperCAmelCase : Union[str, Any] = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCAmelCase : Optional[Any] = self.numpy_to_pil(__snake_case ) if not return_dict: return (sample,) return ImagePipelineOutput(images=__snake_case )
23
0
import unittest import numpy as np from transformers.file_utils import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def __init__(self : Tuple , a__ : int , a__ : Any=7 , a__ : Any=3 , a__ : Union[str, Any]=18 , a__ : str=30 , a__ : Optional[int]=400 , a__ : Any=True , a__ : Tuple=None , a__ : str=True , a__ : str=[0.5, 0.5, 0.5] , a__ : Any=[0.5, 0.5, 0.5] , ): """simple docstring""" __snake_case = size if size is not None else {'''height''': 18, '''width''': 18} __snake_case = parent __snake_case = batch_size __snake_case = num_channels __snake_case = image_size __snake_case = min_resolution __snake_case = max_resolution __snake_case = do_resize __snake_case = size __snake_case = do_normalize __snake_case = image_mean __snake_case = image_std def a (self : str ): """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase , unittest.TestCase ): A_ : List[str] = DPTImageProcessor if is_vision_available() else None def a (self : Optional[int] ): """simple docstring""" __snake_case = DPTImageProcessingTester(self ) @property def a (self : int ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def a (self : Optional[Any] ): """simple docstring""" __snake_case = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(a__ , '''image_mean''' ) ) self.assertTrue(hasattr(a__ , '''image_std''' ) ) self.assertTrue(hasattr(a__ , '''do_normalize''' ) ) self.assertTrue(hasattr(a__ , '''do_resize''' ) ) self.assertTrue(hasattr(a__ , '''size''' ) ) def a (self : Optional[int] ): """simple docstring""" __snake_case = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 18, '''width''': 18} ) __snake_case = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) def a (self : Tuple ): """simple docstring""" __snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=a__ ) for image in image_inputs: self.assertIsInstance(a__ , Image.Image ) # Test not batched input __snake_case = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched __snake_case = image_processing(a__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def a (self : Any ): """simple docstring""" __snake_case = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=a__ , numpify=a__ ) for image in image_inputs: self.assertIsInstance(a__ , np.ndarray ) # Test not batched input __snake_case = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched __snake_case = image_processing(a__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) def a (self : str ): """simple docstring""" __snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=a__ , torchify=a__ ) for image in image_inputs: self.assertIsInstance(a__ , torch.Tensor ) # Test not batched input __snake_case = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , ) # Test batched __snake_case = image_processing(a__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) , )
24
'''simple docstring''' import warnings from contextlib import contextmanager from ....processing_utils import ProcessorMixin class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = """MCTCTFeatureExtractor""" lowerCamelCase__ = """AutoTokenizer""" def __init__( self : Dict , __snake_case : Optional[int] , __snake_case : List[str] ) -> str: super().__init__(__snake_case , __snake_case ) UpperCAmelCase : List[Any] = self.feature_extractor UpperCAmelCase : Union[str, Any] = False def __call__( self : Any , *__snake_case : List[str] , **__snake_case : Any ) -> List[Any]: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*__snake_case , **__snake_case ) if "raw_speech" in kwargs: warnings.warn('''Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.''' ) UpperCAmelCase : int = kwargs.pop('''raw_speech''' ) else: UpperCAmelCase : Union[str, Any] = kwargs.pop('''audio''' , __snake_case ) UpperCAmelCase : Optional[Any] = kwargs.pop('''sampling_rate''' , __snake_case ) UpperCAmelCase : Dict = kwargs.pop('''text''' , __snake_case ) if len(__snake_case ) > 0: UpperCAmelCase : Any = args[0] UpperCAmelCase : Optional[int] = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if audio is not None: UpperCAmelCase : List[str] = self.feature_extractor(__snake_case , *__snake_case , sampling_rate=__snake_case , **__snake_case ) if text is not None: UpperCAmelCase : int = self.tokenizer(__snake_case , **__snake_case ) if text is None: return inputs elif audio is None: return encodings else: UpperCAmelCase : str = encodings['''input_ids'''] return inputs def A ( self : List[Any] , *__snake_case : List[Any] , **__snake_case : List[Any] ) -> str: return self.tokenizer.batch_decode(*__snake_case , **__snake_case ) def A ( self : List[Any] , *__snake_case : int , **__snake_case : Optional[int] ) -> Any: # For backward compatibility if self._in_target_context_manager: return self.current_processor.pad(*__snake_case , **__snake_case ) UpperCAmelCase : List[Any] = kwargs.pop('''input_features''' , __snake_case ) UpperCAmelCase : Optional[Any] = kwargs.pop('''labels''' , __snake_case ) if len(__snake_case ) > 0: UpperCAmelCase : List[str] = args[0] UpperCAmelCase : List[Any] = args[1:] if input_features is not None: UpperCAmelCase : Tuple = self.feature_extractor.pad(__snake_case , *__snake_case , **__snake_case ) if labels is not None: UpperCAmelCase : Optional[int] = self.tokenizer.pad(__snake_case , **__snake_case ) if labels is None: return input_features elif input_features is None: return labels else: UpperCAmelCase : List[str] = labels['''input_ids'''] return input_features def A ( self : Union[str, Any] , *__snake_case : Optional[Any] , **__snake_case : Optional[int] ) -> Optional[Any]: return self.tokenizer.decode(*__snake_case , **__snake_case ) @contextmanager def A ( self : Any ) -> Optional[int]: warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your audio inputs, or in a separate call.''' ) UpperCAmelCase : Dict = True UpperCAmelCase : List[Any] = self.tokenizer yield UpperCAmelCase : Tuple = self.feature_extractor UpperCAmelCase : List[Any] = False
23
0
"""simple docstring""" # Author: OMKAR PATHAK, Nwachukwu Chidiebere # Use a Python dictionary to construct the graph. from __future__ import annotations from pprint import pformat from typing import Generic, TypeVar UpperCAmelCase__ : str = TypeVar('T') class lowerCAmelCase_ (Generic[T] ): """simple docstring""" def __init__(self , SCREAMING_SNAKE_CASE__ = True ) -> None: """simple docstring""" SCREAMING_SNAKE_CASE__ : dict[T, list[T]] = {} # dictionary of lists SCREAMING_SNAKE_CASE__ : Tuple = directed def __magic_name__ (self , SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) -> GraphAdjacencyList[T]: """simple docstring""" if not self.directed: # For undirected graphs # if both source vertex and destination vertex are both present in the # adjacency list, add destination vertex to source vertex list of adjacent # vertices and add source vertex to destination vertex list of adjacent # vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(SCREAMING_SNAKE_CASE__ ) self.adj_list[destination_vertex].append(SCREAMING_SNAKE_CASE__ ) # if only source vertex is present in adjacency list, add destination vertex # to source vertex list of adjacent vertices, then create a new vertex with # destination vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : Any = [source_vertex] # if only destination vertex is present in adjacency list, add source vertex # to destination vertex list of adjacent vertices, then create a new vertex # with source vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif destination_vertex in self.adj_list: self.adj_list[destination_vertex].append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : int = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and assign a list # containing the destination vertex as it's first adjacent vertex also # create a new vertex with destination vertex as key and assign a list # containing the source vertex as it's first adjacent vertex. else: SCREAMING_SNAKE_CASE__ : Optional[Any] = [destination_vertex] SCREAMING_SNAKE_CASE__ : Tuple = [source_vertex] else: # For directed graphs # if both source vertex and destination vertex are present in adjacency # list, add destination vertex to source vertex list of adjacent vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(SCREAMING_SNAKE_CASE__ ) # if only source vertex is present in adjacency list, add destination # vertex to source vertex list of adjacent vertices and create a new vertex # with destination vertex as key, which has no adjacent vertex elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(SCREAMING_SNAKE_CASE__ ) SCREAMING_SNAKE_CASE__ : int = [] # if only destination vertex is present in adjacency list, create a new # vertex with source vertex as key and assign a list containing destination # vertex as first adjacent vertex elif destination_vertex in self.adj_list: SCREAMING_SNAKE_CASE__ : Any = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and a list containing # destination vertex as it's first adjacent vertex. Then create a new vertex # with destination vertex as key, which has no adjacent vertex else: SCREAMING_SNAKE_CASE__ : Any = [destination_vertex] SCREAMING_SNAKE_CASE__ : Optional[int] = [] return self def __repr__(self ) -> str: """simple docstring""" return pformat(self.adj_list )
25
'''simple docstring''' from math import isclose, sqrt def snake_case_ ( _lowerCAmelCase : float , _lowerCAmelCase : float , _lowerCAmelCase : float ) -> tuple[float, float, float]: UpperCAmelCase : Optional[int] = point_y / 4 / point_x UpperCAmelCase : str = 2 * normal_gradient / (1 + normal_gradient * normal_gradient) UpperCAmelCase : Any = (1 - normal_gradient * normal_gradient) / ( 1 + normal_gradient * normal_gradient ) UpperCAmelCase : Union[str, Any] = (sa - ca * incoming_gradient) / (ca + sa * incoming_gradient) # to find the next point, solve the simultaeneous equations: # y^2 + 4x^2 = 100 # y - b = m * (x - a) # ==> A x^2 + B x + C = 0 UpperCAmelCase : Union[str, Any] = outgoing_gradient**2 + 4 UpperCAmelCase : Dict = 2 * outgoing_gradient * (point_y - outgoing_gradient * point_x) UpperCAmelCase : List[Any] = (point_y - outgoing_gradient * point_x) ** 2 - 100 UpperCAmelCase : List[str] = ( -linear_term - sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) UpperCAmelCase : Optional[int] = ( -linear_term + sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) # two solutions, one of which is our input point UpperCAmelCase : Optional[Any] = x_minus if isclose(_lowerCAmelCase , _lowerCAmelCase ) else x_plus UpperCAmelCase : Union[str, Any] = point_y + outgoing_gradient * (next_x - point_x) return next_x, next_y, outgoing_gradient def snake_case_ ( _lowerCAmelCase : float = 1.4 , _lowerCAmelCase : float = -9.6 ) -> int: UpperCAmelCase : int = 0 UpperCAmelCase : float = first_x_coord UpperCAmelCase : float = first_y_coord UpperCAmelCase : float = (1_0.1 - point_y) / (0.0 - point_x) while not (-0.0_1 <= point_x <= 0.0_1 and point_y > 0): UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : str = next_point(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) num_reflections += 1 return num_reflections if __name__ == "__main__": print(F"{solution() = }")
23
0
from math import ceil, sqrt def lowerCAmelCase_ ( snake_case_ = 1000000 ): _A : int = 0 for outer_width in range(3,(limit // 4) + 2 ): if outer_width**2 > limit: _A : List[str] = max(ceil(sqrt(outer_width**2 - limit ) ),1 ) else: _A : Optional[int] = 1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(f"""{solution() = }""")
26
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase__: str = { "configuration_lxmert": ["LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "LxmertConfig"], "tokenization_lxmert": ["LxmertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: int = ["LxmertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: Union[str, Any] = [ "LxmertEncoder", "LxmertForPreTraining", "LxmertForQuestionAnswering", "LxmertModel", "LxmertPreTrainedModel", "LxmertVisualFeatureEncoder", "LxmertXLayer", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: int = [ "TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFLxmertForPreTraining", "TFLxmertMainLayer", "TFLxmertModel", "TFLxmertPreTrainedModel", "TFLxmertVisualFeatureEncoder", ] if TYPE_CHECKING: from .configuration_lxmert import LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, LxmertConfig from .tokenization_lxmert import LxmertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_lxmert_fast import LxmertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lxmert import ( LxmertEncoder, LxmertForPreTraining, LxmertForQuestionAnswering, LxmertModel, LxmertPreTrainedModel, LxmertVisualFeatureEncoder, LxmertXLayer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_lxmert import ( TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFLxmertForPreTraining, TFLxmertMainLayer, TFLxmertModel, TFLxmertPreTrainedModel, TFLxmertVisualFeatureEncoder, ) else: import sys UpperCamelCase__: Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
0
'''simple docstring''' import math import os import sys def lowerCamelCase (_SCREAMING_SNAKE_CASE : str ): __a : str = '' try: with open(_SCREAMING_SNAKE_CASE , 'rb' ) as binary_file: __a : str = binary_file.read() for dat in data: __a : str = F"""{dat:08b}""" result += curr_byte return result except OSError: print('File not accessible' ) sys.exit() def lowerCamelCase (_SCREAMING_SNAKE_CASE : dict[str, str] , _SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : str ): lexicon.pop(_SCREAMING_SNAKE_CASE ) __a : Tuple = last_match_id if math.loga(_SCREAMING_SNAKE_CASE ).is_integer(): for curr_key in lexicon: __a : List[Any] = '0' + lexicon[curr_key] __a : Optional[Any] = bin(_SCREAMING_SNAKE_CASE )[2:] def lowerCamelCase (_SCREAMING_SNAKE_CASE : str ): __a : List[Any] = {'0': '0', '1': '1'} __a , __a : Union[str, Any] = '', '' __a : Dict = len(_SCREAMING_SNAKE_CASE ) for i in range(len(_SCREAMING_SNAKE_CASE ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue __a : Optional[int] = lexicon[curr_string] result += last_match_id add_key_to_lexicon(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) index += 1 __a : List[Any] = '' while curr_string != "" and curr_string not in lexicon: curr_string += "0" if curr_string != "": __a : List[Any] = lexicon[curr_string] result += last_match_id return result def lowerCamelCase (_SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : str ): __a : Any = os.path.getsize(_SCREAMING_SNAKE_CASE ) __a : int = bin(_SCREAMING_SNAKE_CASE )[2:] __a : Dict = len(_SCREAMING_SNAKE_CASE ) return "0" * (length_length - 1) + file_length_binary + compressed def lowerCamelCase (_SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : str ): __a : Any = 8 try: with open(_SCREAMING_SNAKE_CASE , 'wb' ) as opened_file: __a : Optional[int] = [ to_write[i : i + byte_length] for i in range(0 , len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append('10000000' ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array: opened_file.write(int(_SCREAMING_SNAKE_CASE , 2 ).to_bytes(1 , byteorder='big' ) ) except OSError: print('File not accessible' ) sys.exit() def lowerCamelCase (_SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : str ): __a : List[str] = read_file_binary(_SCREAMING_SNAKE_CASE ) __a : Tuple = compress_data(_SCREAMING_SNAKE_CASE ) __a : Union[str, Any] = add_file_length(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) write_file_binary(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
27
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE( A__ , A__ , A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = AltDiffusionPipeline lowerCamelCase__ = TEXT_TO_IMAGE_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS def A ( self : Dict ) -> int: torch.manual_seed(0 ) UpperCAmelCase : str = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) UpperCAmelCase : Dict = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) UpperCAmelCase : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) UpperCAmelCase : Dict = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5002 , ) UpperCAmelCase : List[Any] = CLIPTextModel(__snake_case ) UpperCAmelCase : Optional[Any] = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) UpperCAmelCase : Optional[int] = 77 UpperCAmelCase : Optional[int] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def A ( self : Optional[Any] , __snake_case : Dict , __snake_case : List[str]=0 ) -> Union[str, Any]: if str(__snake_case ).startswith('''mps''' ): UpperCAmelCase : str = torch.manual_seed(__snake_case ) else: UpperCAmelCase : Tuple = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) UpperCAmelCase : Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def A ( self : Union[str, Any] ) -> List[str]: super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def A ( self : Tuple ) -> List[str]: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def A ( self : Tuple ) -> Optional[int]: UpperCAmelCase : List[str] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase : Any = self.get_dummy_components() torch.manual_seed(0 ) UpperCAmelCase : Optional[int] = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase : List[str] = RobertaSeriesModelWithTransformation(__snake_case ) UpperCAmelCase : str = text_encoder UpperCAmelCase : Optional[int] = AltDiffusionPipeline(**__snake_case ) UpperCAmelCase : str = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : Optional[int] = self.get_dummy_inputs(__snake_case ) UpperCAmelCase : Optional[int] = '''A photo of an astronaut''' UpperCAmelCase : List[Any] = alt_pipe(**__snake_case ) UpperCAmelCase : Optional[Any] = output.images UpperCAmelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : List[str] = np.array( [0.5_74_81_62, 0.60_44_71_45, 0.48_82_12_17, 0.50_10_06_36, 0.5_43_11_85, 0.45_76_36_83, 0.49_65_76_96, 0.48_13_27_33, 0.47_57_30_93] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A ( self : int ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase : int = self.get_dummy_components() UpperCAmelCase : int = PNDMScheduler(skip_prk_steps=__snake_case ) torch.manual_seed(0 ) UpperCAmelCase : int = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase : Union[str, Any] = RobertaSeriesModelWithTransformation(__snake_case ) UpperCAmelCase : Union[str, Any] = text_encoder UpperCAmelCase : Optional[int] = AltDiffusionPipeline(**__snake_case ) UpperCAmelCase : Dict = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : int = self.get_dummy_inputs(__snake_case ) UpperCAmelCase : Optional[int] = alt_pipe(**__snake_case ) UpperCAmelCase : Optional[int] = output.images UpperCAmelCase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : Optional[int] = np.array( [0.51_60_50_93, 0.5_70_72_41, 0.47_36_55_07, 0.50_57_88_86, 0.5_63_38_77, 0.4_64_25_03, 0.5_18_20_81, 0.48_76_34_84, 0.49_08_42_37] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : str ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : List[Any] ) -> Any: # make sure here that pndm scheduler skips prk UpperCAmelCase : List[Any] = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''' , safety_checker=__snake_case ) UpperCAmelCase : Tuple = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : List[Any] = '''A painting of a squirrel eating a burger''' UpperCAmelCase : Any = torch.manual_seed(0 ) UpperCAmelCase : Optional[int] = alt_pipe([prompt] , generator=__snake_case , guidance_scale=6.0 , num_inference_steps=20 , output_type='''np''' ) UpperCAmelCase : Dict = output.images UpperCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase : List[str] = np.array([0.10_10, 0.08_00, 0.07_94, 0.08_85, 0.08_43, 0.07_62, 0.07_69, 0.07_29, 0.05_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A ( self : Tuple ) -> int: UpperCAmelCase : int = DDIMScheduler.from_pretrained('''BAAI/AltDiffusion''' , subfolder='''scheduler''' ) UpperCAmelCase : Tuple = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''' , scheduler=__snake_case , safety_checker=__snake_case ) UpperCAmelCase : Dict = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : Tuple = '''A painting of a squirrel eating a burger''' UpperCAmelCase : Optional[int] = torch.manual_seed(0 ) UpperCAmelCase : List[Any] = alt_pipe([prompt] , generator=__snake_case , num_inference_steps=2 , output_type='''numpy''' ) UpperCAmelCase : Dict = output.images UpperCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase : Union[str, Any] = np.array([0.40_19, 0.40_52, 0.38_10, 0.41_19, 0.39_16, 0.39_82, 0.46_51, 0.41_95, 0.53_23] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
23
0
'''simple docstring''' import unittest from transformers import load_tool from transformers.utils import is_torch_available if is_torch_available(): import torch from transformers.testing_utils import require_torch from .test_tools_common import ToolTesterMixin @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase , _a ): """simple docstring""" def A ( self : Any ): """simple docstring""" UpperCamelCase = load_tool('text-to-speech' ) self.tool.setup() def A ( self : Dict ): """simple docstring""" torch.manual_seed(0 ) UpperCamelCase = self.tool('hey' ) UpperCamelCase = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_0_0_5_9_6_6_6_6_8_8_3_2_1_1_5_8_2_9, -0.0_0_0_3_6_5_7_6_4_0_1_9_0_7_9_5_0_6_4, -0.0_0_0_1_3_4_3_9_5_0_2_7_9_9_8_8_3_4_8_5] ) , ) ) def A ( self : Optional[int] ): """simple docstring""" torch.manual_seed(0 ) UpperCamelCase = self.tool('hey' ) UpperCamelCase = result.to_raw() self.assertTrue( torch.allclose( resulting_tensor[:3] , torch.tensor([-0.0_0_0_5_9_6_6_6_6_8_8_3_2_1_1_5_8_2_9, -0.0_0_0_3_6_5_7_6_4_0_1_9_0_7_9_5_0_6_4, -0.0_0_0_1_3_4_3_9_5_0_2_7_9_9_8_8_3_4_8_5] ) , ) )
28
'''simple docstring''' import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] ) -> Optional[int]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[int] ) -> Dict: UpperCAmelCase : Dict = tmp_path / '''cache''' UpperCAmelCase : List[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase : str = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict , _lowerCAmelCase : int ) -> Optional[int]: UpperCAmelCase : Any = tmp_path / '''cache''' UpperCAmelCase : Tuple = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Any = features.copy() if features else default_expected_features UpperCAmelCase : List[Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Dict = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Tuple ) -> Tuple: UpperCAmelCase : Optional[Any] = tmp_path / '''cache''' UpperCAmelCase : Optional[int] = {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''} UpperCAmelCase : int = features.copy() if features else default_expected_features UpperCAmelCase : Any = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict ) -> Union[str, Any]: # jsonl_312_path features are {"col_3": "float64", "col_1": "string", "col_2": "int64"} UpperCAmelCase : Tuple = {'''col_2''': '''int64''', '''col_3''': '''float64''', '''col_1''': '''string'''} UpperCAmelCase : List[str] = features.copy() UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = tmp_path / '''cache''' UpperCAmelCase : List[str] = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ) -> Optional[Any]: UpperCAmelCase : Any = tmp_path / '''cache''' UpperCAmelCase : List[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : List[Any] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , split=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('''path_type''' , [str, list] ) def snake_case_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> Dict: if issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase : str = jsonl_path elif issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase : Dict = [jsonl_path] UpperCAmelCase : int = tmp_path / '''cache''' UpperCAmelCase : Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[int] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : Dict=("train",) ) -> Union[str, Any]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) for split in splits: UpperCAmelCase : List[str] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ) -> Any: UpperCAmelCase : Optional[Any] = tmp_path / '''cache''' UpperCAmelCase : List[str] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase : Optional[int] = JsonDatasetReader({'''train''': jsonl_path} , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] ) -> int: UpperCAmelCase : Dict = tmp_path / '''cache''' UpperCAmelCase : Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[int] = features.copy() if features else default_expected_features UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = JsonDatasetReader({'''train''': jsonl_path} , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict ) -> Union[str, Any]: if split: UpperCAmelCase : Optional[int] = {split: jsonl_path} else: UpperCAmelCase : Any = '''train''' UpperCAmelCase : Any = {'''train''': jsonl_path, '''test''': jsonl_path} UpperCAmelCase : Tuple = tmp_path / '''cache''' UpperCAmelCase : int = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[Any] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def snake_case_ ( _lowerCAmelCase : List[str] ) -> str: return json.load(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Dict ) -> str: return [json.loads(_lowerCAmelCase ) for line in buffer] class SCREAMING_SNAKE_CASE: """simple docstring""" @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def A ( self : Union[str, Any] , __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : Optional[int] ) -> Dict: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case ).write() buffer.seek(0 ) UpperCAmelCase : Union[str, Any] = load_json_function(__snake_case ) assert isinstance(__snake_case , __snake_case ) assert isinstance(exported_content[0] , __snake_case ) assert len(__snake_case ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def A ( self : Optional[int] , __snake_case : Optional[Any] , __snake_case : str , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : Optional[Any] ) -> List[Any]: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , orient=__snake_case ).write() buffer.seek(0 ) UpperCAmelCase : Union[str, Any] = load_json(__snake_case ) assert isinstance(__snake_case , __snake_case ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(__snake_case , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(__snake_case ) == 10 @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def A ( self : str , __snake_case : str , __snake_case : str , __snake_case : int ) -> Any: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase : Any = load_json_function(__snake_case ) assert isinstance(__snake_case , __snake_case ) assert isinstance(exported_content[0] , __snake_case ) assert len(__snake_case ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def A ( self : Any , __snake_case : int , __snake_case : Union[str, Any] , __snake_case : Optional[int] , __snake_case : Optional[Any] , __snake_case : List[str] ) -> Any: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , orient=__snake_case , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase : List[str] = load_json(__snake_case ) assert isinstance(__snake_case , __snake_case ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(__snake_case , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(__snake_case ) == 10 def A ( self : List[Any] , __snake_case : str ) -> Dict: with pytest.raises(__snake_case ): with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , num_proc=0 ) @pytest.mark.parametrize('''compression, extension''' , [('''gzip''', '''gz'''), ('''bz2''', '''bz2'''), ('''xz''', '''xz''')] ) def A ( self : Optional[int] , __snake_case : Any , __snake_case : str , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Dict ) -> Union[str, Any]: UpperCAmelCase : List[str] = tmp_path_factory.mktemp('''data''' ) / F"""test.json.{extension}""" UpperCAmelCase : List[Any] = str(shared_datadir / F"""test_file.json.{extension}""" ) JsonDatasetWriter(__snake_case , __snake_case , compression=__snake_case ).write() with fsspec.open(__snake_case , '''rb''' , compression='''infer''' ) as f: UpperCAmelCase : str = f.read() with fsspec.open(__snake_case , '''rb''' , compression='''infer''' ) as f: UpperCAmelCase : Optional[int] = f.read() assert exported_content == original_content
23
0
import datasets from .nmt_bleu import compute_bleu # From: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py __UpperCAmelCase = '\\n@INPROCEEDINGS{Papineni02bleu:a,\n author = {Kishore Papineni and Salim Roukos and Todd Ward and Wei-jing Zhu},\n title = {BLEU: a Method for Automatic Evaluation of Machine Translation},\n booktitle = {},\n year = {2002},\n pages = {311--318}\n}\n@inproceedings{lin-och-2004-orange,\n title = "{ORANGE}: a Method for Evaluating Automatic Evaluation Metrics for Machine Translation",\n author = "Lin, Chin-Yew and\n Och, Franz Josef",\n booktitle = "{COLING} 2004: Proceedings of the 20th International Conference on Computational Linguistics",\n month = "aug 23{--}aug 27",\n year = "2004",\n address = "Geneva, Switzerland",\n publisher = "COLING",\n url = "https://www.aclweb.org/anthology/C04-1072",\n pages = "501--507",\n}\n' __UpperCAmelCase = '\\nBLEU (bilingual evaluation understudy) is an algorithm for evaluating the quality of text which has been machine-translated from one natural language to another.\nQuality is considered to be the correspondence between a machine\'s output and that of a human: "the closer a machine translation is to a professional human translation,\nthe better it is" – this is the central idea behind BLEU. BLEU was one of the first metrics to claim a high correlation with human judgements of quality, and\nremains one of the most popular automated and inexpensive metrics.\n\nScores are calculated for individual translated segments—generally sentences—by comparing them with a set of good quality reference translations.\nThose scores are then averaged over the whole corpus to reach an estimate of the translation\'s overall quality. Intelligibility or grammatical correctness\nare not taken into account[citation needed].\n\nBLEU\'s output is always a number between 0 and 1. This value indicates how similar the candidate text is to the reference texts, with values closer to 1\nrepresenting more similar texts. Few human translations will attain a score of 1, since this would indicate that the candidate is identical to one of the\nreference translations. For this reason, it is not necessary to attain a score of 1. Because there are more opportunities to match, adding additional\nreference translations will increase the BLEU score.\n' __UpperCAmelCase = '\nComputes BLEU score of translated segments against one or more references.\nArgs:\n predictions: list of translations to score.\n Each translation should be tokenized into a list of tokens.\n references: list of lists of references for each translation.\n Each reference should be tokenized into a list of tokens.\n max_order: Maximum n-gram order to use when computing BLEU score.\n smooth: Whether or not to apply Lin et al. 2004 smoothing.\nReturns:\n \'bleu\': bleu score,\n \'precisions\': geometric mean of n-gram precisions,\n \'brevity_penalty\': brevity penalty,\n \'length_ratio\': ratio of lengths,\n \'translation_length\': translation_length,\n \'reference_length\': reference_length\nExamples:\n\n >>> predictions = [\n ... ["hello", "there", "general", "kenobi"], # tokenized prediction of the first sample\n ... ["foo", "bar", "foobar"] # tokenized prediction of the second sample\n ... ]\n >>> references = [\n ... [["hello", "there", "general", "kenobi"], ["hello", "there", "!"]], # tokenized references for the first sample (2 references)\n ... [["foo", "bar", "foobar"]] # tokenized references for the second sample (1 reference)\n ... ]\n >>> bleu = datasets.load_metric("bleu")\n >>> results = bleu.compute(predictions=predictions, references=references)\n >>> print(results["bleu"])\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCamelCase (datasets.Metric ): '''simple docstring''' def __UpperCAmelCase ( self ) -> Tuple: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('string' , id='token' ) , id='sequence' ), 'references': datasets.Sequence( datasets.Sequence(datasets.Value('string' , id='token' ) , id='sequence' ) , id='references' ), } ) , codebase_urls=['https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py'] , reference_urls=[ 'https://en.wikipedia.org/wiki/BLEU', 'https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213', ] , ) def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase=4 , _UpperCamelCase=False ) -> Union[str, Any]: UpperCAmelCase_ : List[str] = compute_bleu( reference_corpus=_UpperCamelCase , translation_corpus=_UpperCamelCase , max_order=_UpperCamelCase , smooth=_UpperCamelCase ) ((UpperCAmelCase_) , (UpperCAmelCase_) , (UpperCAmelCase_) , (UpperCAmelCase_) , (UpperCAmelCase_) , (UpperCAmelCase_)) : Tuple = score return { "bleu": bleu, "precisions": precisions, "brevity_penalty": bp, "length_ratio": ratio, "translation_length": translation_length, "reference_length": reference_length, }
29
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCamelCase__: Tuple = logging.get_logger(__name__) UpperCamelCase__: Optional[int] = {"vocab_file": "sentencepiece.bpe.model"} UpperCamelCase__: Optional[int] = { "vocab_file": { "moussaKam/mbarthez": "https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez": "https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez-orangesum-title": ( "https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model" ), }, } UpperCamelCase__: Dict = { "moussaKam/mbarthez": 1024, "moussaKam/barthez": 1024, "moussaKam/barthez-orangesum-title": 1024, } UpperCamelCase__: Tuple = "▁" class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = ["""input_ids""", """attention_mask"""] def __init__( self : List[Any] , __snake_case : List[Any] , __snake_case : Tuple="<s>" , __snake_case : List[Any]="</s>" , __snake_case : int="</s>" , __snake_case : Any="<s>" , __snake_case : Optional[int]="<unk>" , __snake_case : Union[str, Any]="<pad>" , __snake_case : Union[str, Any]="<mask>" , __snake_case : Optional[Dict[str, Any]] = None , **__snake_case : Dict , ) -> None: # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase : int = AddedToken(__snake_case , lstrip=__snake_case , rstrip=__snake_case ) if isinstance(__snake_case , __snake_case ) else mask_token UpperCAmelCase : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__snake_case , eos_token=__snake_case , unk_token=__snake_case , sep_token=__snake_case , cls_token=__snake_case , pad_token=__snake_case , mask_token=__snake_case , sp_model_kwargs=self.sp_model_kwargs , **__snake_case , ) UpperCAmelCase : Optional[int] = vocab_file UpperCAmelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__snake_case ) ) UpperCAmelCase : int = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} UpperCAmelCase : Optional[Any] = len(self.sp_model ) - 1 UpperCAmelCase : Optional[int] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def A ( self : Tuple , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase : Union[str, Any] = [self.cls_token_id] UpperCAmelCase : Optional[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A ( self : int , __snake_case : List[int] , __snake_case : Optional[List[int]] = None , __snake_case : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__snake_case , token_ids_a=__snake_case , already_has_special_tokens=__snake_case ) if token_ids_a is None: return [1] + ([0] * len(__snake_case )) + [1] return [1] + ([0] * len(__snake_case )) + [1, 1] + ([0] * len(__snake_case )) + [1] def A ( self : Optional[int] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: UpperCAmelCase : Tuple = [self.sep_token_id] UpperCAmelCase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def A ( self : Dict ) -> Optional[int]: return len(self.sp_model ) def A ( self : List[str] ) -> Dict: UpperCAmelCase : Optional[Any] = {self.convert_ids_to_tokens(__snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def A ( self : Optional[Any] , __snake_case : str ) -> List[str]: return self.sp_model.encode(__snake_case , out_type=__snake_case ) def A ( self : int , __snake_case : int ) -> int: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCAmelCase : Optional[Any] = self.sp_model.PieceToId(__snake_case ) return spm_id if spm_id else self.unk_token_id def A ( self : int , __snake_case : Any ) -> List[Any]: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(__snake_case ) def A ( self : List[Any] , __snake_case : Union[str, Any] ) -> List[str]: UpperCAmelCase : Optional[Any] = [] UpperCAmelCase : int = '''''' UpperCAmelCase : Union[str, Any] = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__snake_case ) + token UpperCAmelCase : str = True UpperCAmelCase : List[str] = [] else: current_sub_tokens.append(__snake_case ) UpperCAmelCase : Optional[int] = False out_string += self.sp_model.decode(__snake_case ) return out_string.strip() def __getstate__( self : Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase : Optional[Any] = self.__dict__.copy() UpperCAmelCase : Any = None return state def __setstate__( self : Optional[int] , __snake_case : Union[str, Any] ) -> List[Any]: UpperCAmelCase : Dict = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): UpperCAmelCase : Optional[Any] = {} UpperCAmelCase : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def A ( self : Optional[int] , __snake_case : str , __snake_case : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__snake_case ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase : Union[str, Any] = os.path.join( __snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __snake_case ) elif not os.path.isfile(self.vocab_file ): with open(__snake_case , '''wb''' ) as fi: UpperCAmelCase : Any = self.sp_model.serialized_model_proto() fi.write(__snake_case ) return (out_vocab_file,)
23
0
def a ( snake_case__: int ): '''simple docstring''' lowercase_ = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(2_7)) print(perfect_cube(4))
30
'''simple docstring''' import tempfile import unittest import numpy as np from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import BertConfig, is_flax_available from transformers.testing_utils import TOKEN, USER, is_staging_test, require_flax if is_flax_available(): import os from flax.core.frozen_dict import unfreeze from flax.traverse_util import flatten_dict from transformers import FlaxBertModel UpperCamelCase__: Tuple = "0.12" # assumed parallelism: 8 @require_flax @is_staging_test class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" @classmethod def A ( cls : Union[str, Any] ) -> int: UpperCAmelCase : Optional[Any] = TOKEN HfFolder.save_token(__snake_case ) @classmethod def A ( cls : List[str] ) -> Tuple: try: delete_repo(token=cls._token , repo_id='''test-model-flax''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-model-flax-org''' ) except HTTPError: pass def A ( self : int ) -> Tuple: UpperCAmelCase : List[Any] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) model.push_to_hub('''test-model-flax''' , use_auth_token=self._token ) UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) UpperCAmelCase : List[Any] = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Tuple = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Union[str, Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''test-model-flax''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(__snake_case , repo_id='''test-model-flax''' , push_to_hub=__snake_case , use_auth_token=self._token ) UpperCAmelCase : str = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) UpperCAmelCase : Any = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : str = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Optional[Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) def A ( self : Optional[Any] ) -> Union[str, Any]: UpperCAmelCase : Dict = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Optional[Any] = FlaxBertModel(__snake_case ) model.push_to_hub('''valid_org/test-model-flax-org''' , use_auth_token=self._token ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) UpperCAmelCase : List[Any] = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : int = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Any = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-model-flax-org''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained( __snake_case , repo_id='''valid_org/test-model-flax-org''' , push_to_hub=__snake_case , use_auth_token=self._token ) UpperCAmelCase : str = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) UpperCAmelCase : Any = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Optional[Any] = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : int = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Union[str, Any]: UpperCAmelCase : str = True UpperCAmelCase : int = flatten_dict(modela.params ) UpperCAmelCase : Dict = flatten_dict(modela.params ) for key in flat_params_a.keys(): if np.sum(np.abs(flat_params_a[key] - flat_params_a[key] ) ) > 1e-4: UpperCAmelCase : Dict = False return models_are_equal @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : Tuple ) -> Union[str, Any]: UpperCAmelCase : List[Any] = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) UpperCAmelCase : int = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(__snake_case , __snake_case ) ) with self.assertRaises(__snake_case ): UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : str = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertTrue(check_models_equal(__snake_case , __snake_case ) ) def A ( self : List[str] ) -> Dict: UpperCAmelCase : Dict = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) UpperCAmelCase : Optional[int] = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(__snake_case , __snake_case ) , max_shard_size='''10KB''' ) with self.assertRaises(__snake_case ): UpperCAmelCase : Any = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertTrue(check_models_equal(__snake_case , __snake_case ) ) def A ( self : Optional[int] ) -> str: UpperCAmelCase : Dict = '''bert''' UpperCAmelCase : int = '''hf-internal-testing/tiny-random-bert-subfolder''' with self.assertRaises(__snake_case ): UpperCAmelCase : Optional[Any] = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertIsNotNone(__snake_case ) def A ( self : Dict ) -> List[Any]: UpperCAmelCase : Optional[int] = '''bert''' UpperCAmelCase : int = '''hf-internal-testing/tiny-random-bert-sharded-subfolder''' with self.assertRaises(__snake_case ): UpperCAmelCase : Dict = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertIsNotNone(__snake_case )
23
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) class lowerCamelCase_ (snake_case__ ): '''simple docstring''' __UpperCamelCase: Optional[Any] = ["pixel_values"] def __init__( self : Tuple , A : bool = True , A : Dict[str, int] = None , A : int = 0.9 , A : PILImageResampling = PILImageResampling.BICUBIC , A : bool = True , A : Dict[str, int] = None , A : Union[int, float] = 1 / 255 , A : bool = True , A : bool = True , A : Optional[Union[float, List[float]]] = None , A : Optional[Union[float, List[float]]] = None , **A : Dict , ): super().__init__(**A ) _UpperCAmelCase : Tuple = size if size is not None else {"shortest_edge": 224} _UpperCAmelCase : List[Any] = get_size_dict(A , default_to_square=A ) _UpperCAmelCase : Any = crop_size if crop_size is not None else {"height": 224, "width": 224} _UpperCAmelCase : Tuple = get_size_dict(A , param_name="crop_size" ) _UpperCAmelCase : Dict = do_resize _UpperCAmelCase : List[str] = size _UpperCAmelCase : Optional[Any] = crop_pct _UpperCAmelCase : Union[str, Any] = resample _UpperCAmelCase : List[str] = do_center_crop _UpperCAmelCase : Any = crop_size _UpperCAmelCase : Tuple = do_rescale _UpperCAmelCase : Union[str, Any] = rescale_factor _UpperCAmelCase : Union[str, Any] = do_normalize _UpperCAmelCase : Optional[Any] = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN _UpperCAmelCase : Optional[int] = image_std if image_std is not None else IMAGENET_DEFAULT_STD def _A ( self : Optional[Any] , A : np.ndarray , A : Dict[str, int] , A : Optional[float] = None , A : PILImageResampling = PILImageResampling.BICUBIC , A : Optional[Union[str, ChannelDimension]] = None , **A : Any , ): _UpperCAmelCase : List[str] = get_size_dict(A , default_to_square=A ) if "shortest_edge" not in size and ("height" not in size or "width" not in size): raise ValueError(F"""size must contain 'height' and 'width' or 'shortest_edge' as keys. Got {size.keys()}""" ) if crop_pct is not None: if "shortest_edge" in size: _UpperCAmelCase : Optional[int] = int(size["shortest_edge"] / crop_pct ) elif "height" in size and "width" in size: if size["height"] == size["width"]: _UpperCAmelCase : Any = int(size["height"] / crop_pct ) else: _UpperCAmelCase : List[Any] = (int(size["height"] / crop_pct ), int(size["width"] / crop_pct )) else: raise ValueError("Invalid size for resize: {}".format(A ) ) _UpperCAmelCase : Optional[int] = get_resize_output_image_size(A , size=A , default_to_square=A ) else: if "shortest_edge" in size: _UpperCAmelCase : Dict = get_resize_output_image_size(A , size=size["shortest_edge"] , default_to_square=A ) elif "height" in size and "width" in size: _UpperCAmelCase : List[str] = (size["height"], size["width"]) else: raise ValueError("Invalid size for resize: {}".format(A ) ) return resize(A , size=A , resample=A , data_format=A , **A ) def _A ( self : Dict , A : np.ndarray , A : Dict[str, int] , A : Optional[Union[str, ChannelDimension]] = None , **A : List[Any] , ): _UpperCAmelCase : Tuple = get_size_dict(A ) if "height" not in size or "width" not in size: raise ValueError(F"""size must contain 'height' and 'width' as keys. Got {size.keys()}""" ) return center_crop(A , size=(size["height"], size["width"]) , data_format=A , **A ) def _A ( self : List[str] , A : np.ndarray , A : Union[int, float] , A : Optional[Union[str, ChannelDimension]] = None , **A : Any , ): return rescale(A , scale=A , data_format=A , **A ) def _A ( self : Optional[int] , A : np.ndarray , A : Union[float, List[float]] , A : Union[float, List[float]] , A : Optional[Union[str, ChannelDimension]] = None , **A : Optional[int] , ): return normalize(A , mean=A , std=A , data_format=A , **A ) def _A ( self : int , A : ImageInput , A : bool = None , A : Dict[str, int] = None , A : int = None , A : PILImageResampling = None , A : bool = None , A : Dict[str, int] = None , A : bool = None , A : float = None , A : bool = None , A : Optional[Union[float, List[float]]] = None , A : Optional[Union[float, List[float]]] = None , A : Optional[Union[str, TensorType]] = None , A : ChannelDimension = ChannelDimension.FIRST , **A : Optional[Any] , ): _UpperCAmelCase : Dict = do_resize if do_resize is not None else self.do_resize _UpperCAmelCase : List[Any] = crop_pct if crop_pct is not None else self.crop_pct _UpperCAmelCase : Dict = resample if resample is not None else self.resample _UpperCAmelCase : Dict = do_center_crop if do_center_crop is not None else self.do_center_crop _UpperCAmelCase : Tuple = do_rescale if do_rescale is not None else self.do_rescale _UpperCAmelCase : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor _UpperCAmelCase : List[Any] = do_normalize if do_normalize is not None else self.do_normalize _UpperCAmelCase : Union[str, Any] = image_mean if image_mean is not None else self.image_mean _UpperCAmelCase : Tuple = image_std if image_std is not None else self.image_std _UpperCAmelCase : List[str] = size if size is not None else self.size _UpperCAmelCase : str = get_size_dict(A , default_to_square=A ) _UpperCAmelCase : Dict = crop_size if crop_size is not None else self.crop_size _UpperCAmelCase : List[Any] = get_size_dict(A , param_name="crop_size" ) _UpperCAmelCase : List[str] = make_list_of_images(A ) if not valid_images(A ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_pct is None: raise ValueError("Crop_pct must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. _UpperCAmelCase : Optional[int] = [to_numpy_array(A ) for image in images] if do_resize: _UpperCAmelCase : Optional[int] = [self.resize(image=A , size=A , crop_pct=A , resample=A ) for image in images] if do_center_crop: _UpperCAmelCase : str = [self.center_crop(image=A , size=A ) for image in images] if do_rescale: _UpperCAmelCase : List[str] = [self.rescale(image=A , scale=A ) for image in images] if do_normalize: _UpperCAmelCase : Union[str, Any] = [self.normalize(image=A , mean=A , std=A ) for image in images] _UpperCAmelCase : Union[str, Any] = [to_channel_dimension_format(A , A ) for image in images] _UpperCAmelCase : str = {"pixel_values": images} return BatchFeature(data=A , tensor_type=A )
31
'''simple docstring''' import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE: """simple docstring""" def __init__( self : Optional[int] , __snake_case : str , __snake_case : Union[str, Any]=2 , __snake_case : Optional[int]=8 , __snake_case : Any=True , __snake_case : Union[str, Any]=True , __snake_case : Dict=True , __snake_case : int=True , __snake_case : List[Any]=99 , __snake_case : str=16 , __snake_case : Tuple=5 , __snake_case : Tuple=2 , __snake_case : str=36 , __snake_case : Dict="gelu" , __snake_case : str=0.0 , __snake_case : Optional[int]=0.0 , __snake_case : Optional[int]=512 , __snake_case : Optional[Any]=16 , __snake_case : int=2 , __snake_case : int=0.02 , __snake_case : str=3 , __snake_case : Dict=4 , __snake_case : str=None , ) -> Optional[int]: UpperCAmelCase : Optional[int] = parent UpperCAmelCase : Tuple = batch_size UpperCAmelCase : List[str] = seq_length UpperCAmelCase : List[Any] = is_training UpperCAmelCase : int = use_input_mask UpperCAmelCase : Any = use_token_type_ids UpperCAmelCase : str = use_labels UpperCAmelCase : Union[str, Any] = vocab_size UpperCAmelCase : List[str] = hidden_size UpperCAmelCase : Optional[Any] = num_hidden_layers UpperCAmelCase : Union[str, Any] = num_attention_heads UpperCAmelCase : Optional[Any] = intermediate_size UpperCAmelCase : Union[str, Any] = hidden_act UpperCAmelCase : int = hidden_dropout_prob UpperCAmelCase : Optional[int] = attention_probs_dropout_prob UpperCAmelCase : Union[str, Any] = max_position_embeddings UpperCAmelCase : str = type_vocab_size UpperCAmelCase : List[str] = type_sequence_label_size UpperCAmelCase : Tuple = initializer_range UpperCAmelCase : Optional[Any] = num_labels UpperCAmelCase : Optional[int] = num_choices UpperCAmelCase : Any = scope def A ( self : Tuple ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase : Optional[int] = None if self.use_input_mask: UpperCAmelCase : Dict = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase : Dict = None if self.use_token_type_ids: UpperCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase : str = None UpperCAmelCase : Tuple = None UpperCAmelCase : int = None if self.use_labels: UpperCAmelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase : str = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase : List[str] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : int ) -> Tuple: return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__snake_case , initializer_range=self.initializer_range , ) def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Optional[Any] = self.get_config() UpperCAmelCase : int = 300 return config def A ( self : Optional[Any] ) -> Any: ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) : Tuple = self.prepare_config_and_inputs() UpperCAmelCase : Dict = True UpperCAmelCase : Tuple = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def A ( self : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : int , __snake_case : Optional[int] , __snake_case : int , __snake_case : Dict , __snake_case : Tuple , __snake_case : Optional[Any] ) -> List[str]: UpperCAmelCase : int = MraModel(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Tuple = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case ) UpperCAmelCase : Optional[int] = model(__snake_case , token_type_ids=__snake_case ) UpperCAmelCase : Dict = model(__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Tuple , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : List[str] , __snake_case : int , __snake_case : Union[str, Any] , __snake_case : Optional[Any] , __snake_case : Any , __snake_case : List[Any] , __snake_case : Optional[Any] , ) -> Tuple: UpperCAmelCase : str = True UpperCAmelCase : Tuple = MraModel(__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Optional[int] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , encoder_hidden_states=__snake_case , encoder_attention_mask=__snake_case , ) UpperCAmelCase : Optional[Any] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , encoder_hidden_states=__snake_case , ) UpperCAmelCase : str = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Tuple , __snake_case : str , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : int ) -> Any: UpperCAmelCase : Dict = MraForMaskedLM(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Optional[int] = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : Tuple , __snake_case : Tuple , __snake_case : Dict , __snake_case : Dict , __snake_case : Any , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Tuple ) -> Optional[int]: UpperCAmelCase : List[str] = MraForQuestionAnswering(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : List[Any] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , start_positions=__snake_case , end_positions=__snake_case , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : str , __snake_case : Optional[int] , __snake_case : List[str] , __snake_case : str , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Union[str, Any] , __snake_case : List[Any] ) -> int: UpperCAmelCase : int = self.num_labels UpperCAmelCase : Union[str, Any] = MraForSequenceClassification(__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : List[str] = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : str , __snake_case : Dict , __snake_case : Optional[Any] , __snake_case : Dict , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Dict ) -> int: UpperCAmelCase : Tuple = self.num_labels UpperCAmelCase : List[str] = MraForTokenClassification(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : str = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : str , __snake_case : int , __snake_case : Any , __snake_case : Tuple , __snake_case : Optional[int] , __snake_case : List[str] , __snake_case : str , __snake_case : Union[str, Any] ) -> Optional[Any]: UpperCAmelCase : Tuple = self.num_choices UpperCAmelCase : int = MraForMultipleChoice(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : str = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[str] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[str] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : str ) -> Dict: UpperCAmelCase : Any = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) : List[str] = config_and_inputs UpperCAmelCase : Any = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = () def A ( self : int ) -> Union[str, Any]: UpperCAmelCase : List[str] = MraModelTester(self ) UpperCAmelCase : Optional[int] = ConfigTester(self , config_class=__snake_case , hidden_size=37 ) def A ( self : Optional[Any] ) -> str: self.config_tester.run_common_tests() def A ( self : Tuple ) -> Optional[Any]: UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__snake_case ) def A ( self : List[Any] ) -> Optional[Any]: UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase : List[Any] = type self.model_tester.create_and_check_model(*__snake_case ) def A ( self : Tuple ) -> Dict: UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__snake_case ) def A ( self : Tuple ) -> List[str]: UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__snake_case ) def A ( self : int ) -> Dict: UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__snake_case ) def A ( self : Dict ) -> Optional[int]: UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__snake_case ) def A ( self : Any ) -> Optional[int]: UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__snake_case ) @slow def A ( self : Dict ) -> Any: for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase : str = MraModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) @unittest.skip(reason='''MRA does not output attentions''' ) def A ( self : str ) -> Optional[Any]: return @require_torch class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" @slow def A ( self : Tuple ) -> List[Any]: UpperCAmelCase : int = MraModel.from_pretrained('''uw-madison/mra-base-512-4''' ) UpperCAmelCase : Optional[Any] = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : List[Any] = model(__snake_case )[0] UpperCAmelCase : Optional[Any] = torch.Size((1, 256, 768) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Any = torch.tensor( [[[-0.01_40, 0.08_30, -0.03_81], [0.15_46, 0.14_02, 0.02_20], [0.11_62, 0.08_51, 0.01_65]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) ) @slow def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Optional[int] = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-512-4''' ) UpperCAmelCase : Dict = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : List[Any] = model(__snake_case )[0] UpperCAmelCase : int = 50265 UpperCAmelCase : int = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Union[str, Any] = torch.tensor( [[[9.25_95, -3.60_38, 11.88_19], [9.38_69, -3.26_93, 11.09_56], [11.85_24, -3.49_38, 13.12_10]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) ) @slow def A ( self : str ) -> List[Any]: UpperCAmelCase : List[Any] = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-4096-8-d3''' ) UpperCAmelCase : List[Any] = torch.arange(4096 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : Tuple = model(__snake_case )[0] UpperCAmelCase : Optional[int] = 50265 UpperCAmelCase : Tuple = torch.Size((1, 4096, vocab_size) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Optional[int] = torch.tensor( [[[5.47_89, -2.35_64, 7.50_64], [7.90_67, -1.33_69, 9.96_68], [9.07_12, -1.81_06, 7.03_80]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) )
23
0
from unittest.mock import Mock, patch from file_transfer.send_file import send_file @patch('socket.socket' ) @patch('builtins.open' ) def SCREAMING_SNAKE_CASE_ ( __A : int , __A : Any ) -> Optional[int]: """simple docstring""" a_ : Any = Mock() a_ : Dict = conn, Mock() a_ : Optional[int] = iter([1, None] ) a_ : List[str] = lambda __A : next(__A ) # ===== invoke ===== send_file(filename='mytext.txt' , testing=__A ) # ===== ensurance ===== sock.assert_called_once() sock.return_value.bind.assert_called_once() sock.return_value.listen.assert_called_once() sock.return_value.accept.assert_called_once() conn.recv.assert_called_once() file.return_value.__enter__.assert_called_once() file.return_value.__enter__.return_value.read.assert_called() conn.send.assert_called_once() conn.close.assert_called_once() sock.return_value.shutdown.assert_called_once() sock.return_value.close.assert_called_once()
32
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : Any ) -> str: UpperCAmelCase : Any = { '''task_specific_params''': { '''summarization''': {'''length_penalty''': 1.0, '''max_length''': 128, '''min_length''': 12, '''num_beams''': 4}, '''summarization_cnn''': {'''length_penalty''': 2.0, '''max_length''': 142, '''min_length''': 56, '''num_beams''': 4}, '''summarization_xsum''': {'''length_penalty''': 1.0, '''max_length''': 62, '''min_length''': 11, '''num_beams''': 6}, } } UpperCAmelCase : int = { '''task_specific_params.summarization.length_penalty''': 1.0, '''task_specific_params.summarization.max_length''': 128, '''task_specific_params.summarization.min_length''': 12, '''task_specific_params.summarization.num_beams''': 4, '''task_specific_params.summarization_cnn.length_penalty''': 2.0, '''task_specific_params.summarization_cnn.max_length''': 142, '''task_specific_params.summarization_cnn.min_length''': 56, '''task_specific_params.summarization_cnn.num_beams''': 4, '''task_specific_params.summarization_xsum.length_penalty''': 1.0, '''task_specific_params.summarization_xsum.max_length''': 62, '''task_specific_params.summarization_xsum.min_length''': 11, '''task_specific_params.summarization_xsum.num_beams''': 6, } self.assertEqual(flatten_dict(__snake_case ) , __snake_case ) def A ( self : int ) -> str: UpperCAmelCase : Any = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(__snake_case ) , x.transpose() ) ) UpperCAmelCase : str = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def A ( self : str ) -> Union[str, Any]: UpperCAmelCase : Any = np.random.randn(3 , 4 ) UpperCAmelCase : List[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , transpose(__snake_case ).numpy() ) ) UpperCAmelCase : Tuple = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Any = torch.tensor(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , transpose(__snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_tf def A ( self : List[str] ) -> Optional[Any]: UpperCAmelCase : int = np.random.randn(3 , 4 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , transpose(__snake_case ).numpy() ) ) UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : str = tf.constant(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , transpose(__snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_flax def A ( self : Tuple ) -> Any: UpperCAmelCase : List[Any] = np.random.randn(3 , 4 ) UpperCAmelCase : List[str] = jnp.array(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , np.asarray(transpose(__snake_case ) ) ) ) UpperCAmelCase : Dict = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , np.asarray(transpose(__snake_case , axes=(1, 2, 0) ) ) ) ) def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , np.reshape(__snake_case , (4, 3) ) ) ) UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , np.reshape(__snake_case , (12, 5) ) ) ) @require_torch def A ( self : Union[str, Any] ) -> int: UpperCAmelCase : Dict = np.random.randn(3 , 4 ) UpperCAmelCase : Optional[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , reshape(__snake_case , (4, 3) ).numpy() ) ) UpperCAmelCase : List[str] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : List[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , reshape(__snake_case , (12, 5) ).numpy() ) ) @require_tf def A ( self : int ) -> List[str]: UpperCAmelCase : List[Any] = np.random.randn(3 , 4 ) UpperCAmelCase : List[str] = tf.constant(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , reshape(__snake_case , (4, 3) ).numpy() ) ) UpperCAmelCase : List[Any] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Optional[Any] = tf.constant(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , reshape(__snake_case , (12, 5) ).numpy() ) ) @require_flax def A ( self : Any ) -> Dict: UpperCAmelCase : Tuple = np.random.randn(3 , 4 ) UpperCAmelCase : Union[str, Any] = jnp.array(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , np.asarray(reshape(__snake_case , (4, 3) ) ) ) ) UpperCAmelCase : Any = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Optional[Any] = jnp.array(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , np.asarray(reshape(__snake_case , (12, 5) ) ) ) ) def A ( self : List[Any] ) -> List[Any]: UpperCAmelCase : Union[str, Any] = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(__snake_case ) , np.squeeze(__snake_case ) ) ) UpperCAmelCase : str = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , np.squeeze(__snake_case , axis=2 ) ) ) @require_torch def A ( self : Optional[int] ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : List[str] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , squeeze(__snake_case ).numpy() ) ) UpperCAmelCase : Any = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : str = torch.tensor(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , squeeze(__snake_case , axis=2 ).numpy() ) ) @require_tf def A ( self : Optional[Any] ) -> Dict: UpperCAmelCase : int = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , squeeze(__snake_case ).numpy() ) ) UpperCAmelCase : List[str] = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , squeeze(__snake_case , axis=2 ).numpy() ) ) @require_flax def A ( self : List[Any] ) -> Dict: UpperCAmelCase : Optional[int] = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , np.asarray(squeeze(__snake_case ) ) ) ) UpperCAmelCase : str = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , np.asarray(squeeze(__snake_case , axis=2 ) ) ) ) def A ( self : Optional[Any] ) -> int: UpperCAmelCase : Optional[Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , np.expand_dims(__snake_case , axis=1 ) ) ) @require_torch def A ( self : List[str] ) -> Tuple: UpperCAmelCase : Tuple = np.random.randn(3 , 4 ) UpperCAmelCase : Tuple = torch.tensor(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , expand_dims(__snake_case , axis=1 ).numpy() ) ) @require_tf def A ( self : List[str] ) -> Union[str, Any]: UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 ) UpperCAmelCase : Any = tf.constant(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , expand_dims(__snake_case , axis=1 ).numpy() ) ) @require_flax def A ( self : Any ) -> List[Any]: UpperCAmelCase : List[str] = np.random.randn(3 , 4 ) UpperCAmelCase : str = jnp.array(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , np.asarray(expand_dims(__snake_case , axis=1 ) ) ) )
23
0
"""simple docstring""" import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor __A : Dict = logging.get_logger(__name__) class _UpperCAmelCase ( _A ): def __init__( self : Optional[int] , *A : Any , **A : Any ) -> None: warnings.warn( '''The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use PerceiverImageProcessor instead.''' , A , ) super().__init__(*A , **A )
33
'''simple docstring''' import argparse import os import re import packaging.version UpperCamelCase__: Union[str, Any] = "examples/" UpperCamelCase__: Optional[Any] = { "examples": (re.compile(r"^check_min_version\(\"[^\"]+\"\)\s*$", re.MULTILINE), "check_min_version(\"VERSION\")\n"), "init": (re.compile(r"^__version__\s+=\s+\"([^\"]+)\"\s*$", re.MULTILINE), "__version__ = \"VERSION\"\n"), "setup": (re.compile(r"^(\s*)version\s*=\s*\"[^\"]+\",", re.MULTILINE), r"\1version=\"VERSION\","), "doc": (re.compile(r"^(\s*)release\s*=\s*\"[^\"]+\"$", re.MULTILINE), "release = \"VERSION\"\n"), } UpperCamelCase__: Optional[int] = { "init": "src/diffusers/__init__.py", "setup": "setup.py", } UpperCamelCase__: List[Any] = "README.md" def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int ) -> Optional[int]: with open(_lowerCAmelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: UpperCAmelCase : Optional[int] = f.read() UpperCAmelCase , UpperCAmelCase : List[Any] = REPLACE_PATTERNS[pattern] UpperCAmelCase : List[Any] = replace.replace('''VERSION''' , _lowerCAmelCase ) UpperCAmelCase : Optional[Any] = re_pattern.sub(_lowerCAmelCase , _lowerCAmelCase ) with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Any ) -> Optional[int]: for folder, directories, fnames in os.walk(_lowerCAmelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase , pattern='''examples''' ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : str=False ) -> List[str]: for pattern, fname in REPLACE_FILES.items(): update_version_in_file(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not patch: update_version_in_examples(_lowerCAmelCase ) def snake_case_ ( ) -> Optional[Any]: UpperCAmelCase : Optional[int] = '''🤗 Transformers currently provides the following architectures''' UpperCAmelCase : Optional[int] = '''1. Want to contribute a new model?''' with open(_lowerCAmelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: UpperCAmelCase : Optional[Any] = f.readlines() # Find the start of the list. UpperCAmelCase : List[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 UpperCAmelCase : Optional[Any] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): UpperCAmelCase : Optional[int] = lines[index].replace( '''https://huggingface.co/docs/diffusers/main/model_doc''' , '''https://huggingface.co/docs/diffusers/model_doc''' , ) index += 1 with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(_lowerCAmelCase ) def snake_case_ ( ) -> Optional[Any]: with open(REPLACE_FILES['''init'''] , '''r''' ) as f: UpperCAmelCase : Union[str, Any] = f.read() UpperCAmelCase : int = REPLACE_PATTERNS['''init'''][0].search(_lowerCAmelCase ).groups()[0] return packaging.version.parse(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : List[str]=False ) -> Any: UpperCAmelCase : Optional[Any] = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: UpperCAmelCase : Optional[int] = default_version.base_version elif patch: UpperCAmelCase : Union[str, Any] = f"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: UpperCAmelCase : Union[str, Any] = f"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. UpperCAmelCase : Dict = input(f"""Which version are you releasing? [{default_version}]""" ) if len(_lowerCAmelCase ) == 0: UpperCAmelCase : Tuple = default_version print(f"""Updating version to {version}.""" ) global_version_update(_lowerCAmelCase , patch=_lowerCAmelCase ) def snake_case_ ( ) -> Any: UpperCAmelCase : List[Any] = get_version() UpperCAmelCase : List[str] = f"""{current_version.major}.{current_version.minor + 1}.0.dev0""" UpperCAmelCase : List[Any] = current_version.base_version # Check with the user we got that right. UpperCAmelCase : Optional[int] = input(f"""Which version are we developing now? [{dev_version}]""" ) if len(_lowerCAmelCase ) == 0: UpperCAmelCase : Dict = dev_version print(f"""Updating version to {version}.""" ) global_version_update(_lowerCAmelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": UpperCamelCase__: Union[str, Any] = argparse.ArgumentParser() parser.add_argument("--post_release", action="store_true", help="Whether this is pre or post release.") parser.add_argument("--patch", action="store_true", help="Whether or not this is a patch release.") UpperCamelCase__: Optional[Any] = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("Nothing to do after a patch :-)") else: post_release_work()
23
0
'''simple docstring''' import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XGLMTokenizer, XGLMTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin A =get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class _a ( __a , unittest.TestCase ): __a : int = XGLMTokenizer __a : Any = XGLMTokenizerFast __a : Any = True __a : Tuple = True def A ( self : Optional[int] ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase = XGLMTokenizer(lowercase , keep_accents=lowercase ) tokenizer.save_pretrained(self.tmpdirname ) def A ( self : Any ): '''simple docstring''' UpperCAmelCase = '''<pad>''' UpperCAmelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowercase ) , lowercase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowercase ) , lowercase ) def A ( self : str ): '''simple docstring''' UpperCAmelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(len(lowercase ) , 1_008 ) def A ( self : str ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size , 1_008 ) def A ( self : List[str] ): '''simple docstring''' UpperCAmelCase = XGLMTokenizer(lowercase , keep_accents=lowercase ) UpperCAmelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(lowercase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowercase ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) UpperCAmelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( lowercase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) UpperCAmelCase = tokenizer.convert_tokens_to_ids(lowercase ) self.assertListEqual( lowercase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) UpperCAmelCase = tokenizer.convert_ids_to_tokens(lowercase ) self.assertListEqual( lowercase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) @cached_property def A ( self : Any ): '''simple docstring''' return XGLMTokenizer.from_pretrained('''facebook/xglm-564M''' ) def A ( self : str ): '''simple docstring''' with tempfile.NamedTemporaryFile() as f: shutil.copyfile(lowercase , f.name ) UpperCAmelCase = XGLMTokenizer(f.name , keep_accents=lowercase ) UpperCAmelCase = pickle.dumps(lowercase ) pickle.loads(lowercase ) def A ( self : List[str] ): '''simple docstring''' if not self.test_rust_tokenizer: return UpperCAmelCase = self.get_tokenizer() UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = '''I was born in 92000, and this is falsé.''' UpperCAmelCase = tokenizer.tokenize(lowercase ) UpperCAmelCase = rust_tokenizer.tokenize(lowercase ) self.assertListEqual(lowercase , lowercase ) UpperCAmelCase = tokenizer.encode(lowercase , add_special_tokens=lowercase ) UpperCAmelCase = rust_tokenizer.encode(lowercase , add_special_tokens=lowercase ) self.assertListEqual(lowercase , lowercase ) UpperCAmelCase = self.get_rust_tokenizer() UpperCAmelCase = tokenizer.encode(lowercase ) UpperCAmelCase = rust_tokenizer.encode(lowercase ) self.assertListEqual(lowercase , lowercase ) @slow def A ( self : List[Any] ): '''simple docstring''' UpperCAmelCase = '''Hello World!''' UpperCAmelCase = [2, 31_227, 4_447, 35] self.assertListEqual(lowercase , self.big_tokenizer.encode(lowercase ) ) @slow def A ( self : int ): '''simple docstring''' UpperCAmelCase = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to unk, such as saoneuhaoesuth''' ) # fmt: off UpperCAmelCase = [2, 1_018, 67, 11, 1_988, 2_617, 5_631, 278, 11, 3_407, 48, 71_630, 28_085, 4, 3_234, 157, 13, 6, 5, 6, 4, 3_526, 768, 15, 659, 57, 298, 3_983, 864, 129, 21, 6, 5, 13_675, 377, 652, 7_580, 10_341, 155, 2_817, 422, 1_666, 7, 1_674, 53, 113, 202_277, 17_892, 33, 60, 87, 4, 3_234, 157, 61, 2_667, 52_376, 19, 88, 23, 735] # fmt: on self.assertListEqual(lowercase , self.big_tokenizer.encode(lowercase ) ) @slow def A ( self : List[str] ): '''simple docstring''' UpperCAmelCase = { '''input_ids''': [[2, 108_825, 1_163, 15, 88_010, 473, 15_898, 157, 13_672, 1_857, 312, 8, 238_021, 1_163, 53, 13_672, 1_857, 312, 8, 53_283, 182_396, 8, 18_566, 16, 36_733, 4_101, 8, 230, 244_017, 122_553, 7, 15, 132_597, 4, 293, 12_511, 7_610, 4, 3_414, 132_597, 9, 4, 32_361, 362, 4, 734, 28_512, 32_569, 18, 4, 32_361, 26_096, 14_982, 73, 18_715, 21_433, 235_261, 15, 492, 12_427, 16, 53, 18_715, 21_433, 65_454, 15, 23_659, 563, 16, 278, 597, 2_843, 595, 7_931, 182_396, 64_186, 22, 886, 595, 132_981, 53, 25_540, 3_449, 43_982, 39_901, 5_951, 878, 330, 4, 27_694, 80_269, 312, 53, 6_517, 11_780, 611, 20_408, 5], [2, 6, 132_597, 67, 42_897, 33, 592, 8, 163_729, 25_540, 361, 136_997, 109_514, 173_230, 7, 501, 60, 102_913, 196, 5_631, 235, 63_243, 473, 6, 231_757, 74, 5_277, 7_905, 53, 3_095, 37_317, 22, 454, 183_874, 5], [2, 268, 31_298, 46_530, 6, 132_935, 43_831, 7, 597, 32, 24, 3_688, 9_865, 5]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] } # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowercase , model_name='''facebook/xglm-564M''' , padding=lowercase , )
34
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake UpperCamelCase__: Tuple = numpy.array([0, 0]) UpperCamelCase__: Union[str, Any] = numpy.array([0.5, 0.8660254]) UpperCamelCase__: Dict = numpy.array([1, 0]) UpperCamelCase__: int = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] , _lowerCAmelCase : int ) -> list[numpy.ndarray]: UpperCAmelCase : Union[str, Any] = initial_vectors for _ in range(_lowerCAmelCase ): UpperCAmelCase : Union[str, Any] = iteration_step(_lowerCAmelCase ) return vectors def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] ) -> list[numpy.ndarray]: UpperCAmelCase : Tuple = [] for i, start_vector in enumerate(vectors[:-1] ): UpperCAmelCase : List[str] = vectors[i + 1] new_vectors.append(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def snake_case_ ( _lowerCAmelCase : numpy.ndarray , _lowerCAmelCase : float ) -> numpy.ndarray: UpperCAmelCase : List[str] = numpy.radians(_lowerCAmelCase ) UpperCAmelCase , UpperCAmelCase : Tuple = numpy.cos(_lowerCAmelCase ), numpy.sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = numpy.array(((c, -s), (s, c)) ) return numpy.dot(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] ) -> None: UpperCAmelCase : List[Any] = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() UpperCAmelCase , UpperCAmelCase : str = zip(*_lowerCAmelCase ) plt.plot(_lowerCAmelCase , _lowerCAmelCase ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__: List[Any] = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
23
0
'''simple docstring''' import argparse import re from pathlib import Path import requests import torch from PIL import Image from torchvision.transforms import CenterCrop, Compose, Normalize, Resize, ToTensor from transformers import ( EfficientFormerConfig, EfficientFormerForImageClassificationWithTeacher, EfficientFormerImageProcessor, ) from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def __snake_case( _lowerCAmelCase , _lowerCAmelCase ) -> str: snake_case__ : Optional[int] = old_name if "patch_embed" in old_name: snake_case__ , snake_case__ , snake_case__ : Optional[int] = old_name.split(""".""" ) if layer == "0": snake_case__ : List[Any] = old_name.replace("""0""" , """convolution1""" ) elif layer == "1": snake_case__ : Tuple = old_name.replace("""1""" , """batchnorm_before""" ) elif layer == "3": snake_case__ : Dict = old_name.replace("""3""" , """convolution2""" ) else: snake_case__ : Optional[int] = old_name.replace("""4""" , """batchnorm_after""" ) if "network" in old_name and re.search(r"""\d\.\d""" , _lowerCAmelCase ): snake_case__ : Tuple = r"""\b\d{2}\b""" if bool(re.search(_lowerCAmelCase , _lowerCAmelCase ) ): snake_case__ : Optional[Any] = re.search(r"""\d\.\d\d.""" , _lowerCAmelCase ).group() else: snake_case__ : List[Any] = re.search(r"""\d\.\d.""" , _lowerCAmelCase ).group() if int(match[0] ) < 6: snake_case__ : Optional[Any] = old_name.replace(_lowerCAmelCase , """""" ) snake_case__ : Any = trimmed_name.replace("""network""" , match[0] + """.meta4D_layers.blocks.""" + match[2:-1] ) snake_case__ : Tuple = """intermediate_stages.""" + trimmed_name else: snake_case__ : Optional[int] = old_name.replace(_lowerCAmelCase , """""" ) if int(match[2] ) < num_meta4D_last_stage: snake_case__ : Dict = trimmed_name.replace("""network""" , """meta4D_layers.blocks.""" + match[2] ) else: snake_case__ : Tuple = str(int(match[2] ) - num_meta4D_last_stage ) snake_case__ : List[str] = trimmed_name.replace("""network""" , """meta3D_layers.blocks.""" + layer_index ) if "norm1" in old_name: snake_case__ : Tuple = trimmed_name.replace("""norm1""" , """layernorm1""" ) elif "norm2" in old_name: snake_case__ : Union[str, Any] = trimmed_name.replace("""norm2""" , """layernorm2""" ) elif "fc1" in old_name: snake_case__ : Optional[int] = trimmed_name.replace("""fc1""" , """linear_in""" ) elif "fc2" in old_name: snake_case__ : Dict = trimmed_name.replace("""fc2""" , """linear_out""" ) snake_case__ : Any = """last_stage.""" + trimmed_name elif "network" in old_name and re.search(r""".\d.""" , _lowerCAmelCase ): snake_case__ : Dict = old_name.replace("""network""" , """intermediate_stages""" ) if "fc" in new_name: snake_case__ : Optional[Any] = new_name.replace("""fc""" , """convolution""" ) elif ("norm1" in new_name) and ("layernorm1" not in new_name): snake_case__ : Optional[Any] = new_name.replace("""norm1""" , """batchnorm_before""" ) elif ("norm2" in new_name) and ("layernorm2" not in new_name): snake_case__ : Optional[int] = new_name.replace("""norm2""" , """batchnorm_after""" ) if "proj" in new_name: snake_case__ : str = new_name.replace("""proj""" , """projection""" ) if "dist_head" in new_name: snake_case__ : Any = new_name.replace("""dist_head""" , """distillation_classifier""" ) elif "head" in new_name: snake_case__ : Any = new_name.replace("""head""" , """classifier""" ) elif "patch_embed" in new_name: snake_case__ : List[Any] = """efficientformer.""" + new_name elif new_name == "norm.weight" or new_name == "norm.bias": snake_case__ : Union[str, Any] = new_name.replace("""norm""" , """layernorm""" ) snake_case__ : int = """efficientformer.""" + new_name else: snake_case__ : Optional[int] = """efficientformer.encoder.""" + new_name return new_name def __snake_case( _lowerCAmelCase , _lowerCAmelCase ) -> Optional[Any]: for key in checkpoint.copy().keys(): snake_case__ : Dict = checkpoint.pop(_lowerCAmelCase ) snake_case__ : int = val return checkpoint def __snake_case( ) -> Any: snake_case__ : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" snake_case__ : List[str] = Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ) return image def __snake_case( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> str: snake_case__ : Dict = torch.load(_lowerCAmelCase , map_location="""cpu""" )["""model"""] snake_case__ : Any = EfficientFormerConfig.from_json_file(_lowerCAmelCase ) snake_case__ : Any = EfficientFormerForImageClassificationWithTeacher(_lowerCAmelCase ) snake_case__ : List[Any] = """_""".join(checkpoint_path.split("""/""" )[-1].split(""".""" )[0].split("""_""" )[:-1] ) snake_case__ : Optional[Any] = config.depths[-1] - config.num_metaad_blocks + 1 snake_case__ : Optional[int] = convert_torch_checkpoint(_lowerCAmelCase , _lowerCAmelCase ) model.load_state_dict(_lowerCAmelCase ) model.eval() snake_case__ : Any = { """bilinear""": PILImageResampling.BILINEAR, """bicubic""": PILImageResampling.BICUBIC, """nearest""": PILImageResampling.NEAREST, } # prepare image snake_case__ : List[Any] = prepare_img() snake_case__ : Dict = 256 snake_case__ : Any = 224 snake_case__ : Any = EfficientFormerImageProcessor( size={"""shortest_edge""": image_size} , crop_size={"""height""": crop_size, """width""": crop_size} , resample=pillow_resamplings["""bicubic"""] , ) snake_case__ : Tuple = processor(images=_lowerCAmelCase , return_tensors="""pt""" ).pixel_values # original processing pipeline snake_case__ : Union[str, Any] = Compose( [ Resize(_lowerCAmelCase , interpolation=pillow_resamplings["""bicubic"""] ), CenterCrop(_lowerCAmelCase ), ToTensor(), Normalize(_lowerCAmelCase , _lowerCAmelCase ), ] ) snake_case__ : Optional[int] = image_transforms(_lowerCAmelCase ).unsqueeze(0 ) assert torch.allclose(_lowerCAmelCase , _lowerCAmelCase ) snake_case__ : Any = model(_lowerCAmelCase ) snake_case__ : Union[str, Any] = outputs.logits snake_case__ : int = (1, 1_000) if "l1" in model_name: snake_case__ : List[Any] = torch.Tensor( [-0.1312, 0.4353, -1.0499, -0.5124, 0.4183, -0.6793, -1.3777, -0.0893, -0.7358, -2.4328] ) assert torch.allclose(logits[0, :10] , _lowerCAmelCase , atol=1e-3 ) assert logits.shape == expected_shape elif "l3" in model_name: snake_case__ : str = torch.Tensor( [-1.3150, -1.5456, -1.2556, -0.8496, -0.7127, -0.7897, -0.9728, -0.3052, 0.3751, -0.3127] ) assert torch.allclose(logits[0, :10] , _lowerCAmelCase , atol=1e-3 ) assert logits.shape == expected_shape elif "l7" in model_name: snake_case__ : List[str] = torch.Tensor( [-1.0283, -1.4131, -0.5644, -1.3115, -0.5785, -1.2049, -0.7528, 0.1992, -0.3822, -0.0878] ) assert logits.shape == expected_shape else: raise ValueError( f"Unknown model checkpoint: {checkpoint_path}. Supported version of efficientformer are l1, l3 and l7" ) # Save Checkpoints Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) model.save_pretrained(_lowerCAmelCase ) print(f"Checkpoint successfuly converted. Model saved at {pytorch_dump_path}" ) processor.save_pretrained(_lowerCAmelCase ) print(f"Processor successfuly saved at {pytorch_dump_path}" ) if push_to_hub: print("""Pushing model to the hub...""" ) model.push_to_hub( repo_id=f"Bearnardd/{pytorch_dump_path}" , commit_message="""Add model""" , use_temp_dir=_lowerCAmelCase , ) processor.push_to_hub( repo_id=f"Bearnardd/{pytorch_dump_path}" , commit_message="""Add image processor""" , use_temp_dir=_lowerCAmelCase , ) if __name__ == "__main__": __a = argparse.ArgumentParser() # Required parameters parser.add_argument( "--pytorch_model_path", default=None, type=str, required=True, help="Path to EfficientFormer pytorch checkpoint.", ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The json file for EfficientFormer model config.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument("--push_to_hub", action="store_true", help="Push model and image processor to the hub") parser.add_argument( "--no-push_to_hub", dest="push_to_hub", action="store_false", help="Do not push model and image processor to the hub", ) parser.set_defaults(push_to_hub=True) __a = parser.parse_args() convert_efficientformer_checkpoint( checkpoint_path=args.pytorch_model_path, efficientformer_config_file=args.config_file, pytorch_dump_path=args.pytorch_dump_path, push_to_hub=args.push_to_hub, )
35
'''simple docstring''' from manim import * class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" def A ( self : Union[str, Any] ) -> List[str]: UpperCAmelCase : Optional[Any] = Rectangle(height=0.5 , width=0.5 ) UpperCAmelCase : str = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) UpperCAmelCase : Tuple = [mem.copy() for i in range(6 )] UpperCAmelCase : Optional[int] = [mem.copy() for i in range(6 )] UpperCAmelCase : Dict = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Any = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Union[str, Any] = VGroup(__snake_case , __snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Optional[Any] = Text('''CPU''' , font_size=24 ) UpperCAmelCase : Union[str, Any] = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__snake_case ) UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(4 )] UpperCAmelCase : Union[str, Any] = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : List[str] = Text('''GPU''' , font_size=24 ) UpperCAmelCase : Dict = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) gpu.move_to([-1, -1, 0] ) self.add(__snake_case ) UpperCAmelCase : int = [mem.copy() for i in range(6 )] UpperCAmelCase : Union[str, Any] = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : List[str] = Text('''Model''' , font_size=24 ) UpperCAmelCase : Tuple = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) model.move_to([3, -1.0, 0] ) self.add(__snake_case ) UpperCAmelCase : Any = [] for i, rect in enumerate(__snake_case ): rect.set_stroke(__snake_case ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) UpperCAmelCase : Dict = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(__snake_case , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=__snake_case ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=__snake_case , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=__snake_case , buff=0.0 ) self.add(__snake_case ) cpu_targs.append(__snake_case ) UpperCAmelCase : int = [mem.copy() for i in range(6 )] UpperCAmelCase : int = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Any = Text('''Loaded Checkpoint''' , font_size=24 ) UpperCAmelCase : Union[str, Any] = Group(__snake_case , __snake_case ).arrange(__snake_case , aligned_edge=__snake_case , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) UpperCAmelCase : Optional[int] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCAmelCase : str = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(__snake_case , __snake_case ) UpperCAmelCase : Tuple = MarkupText( F"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(__snake_case , DOWN * 2.4 , aligned_edge=key_text.get_left() ) UpperCAmelCase : List[Any] = MarkupText( F"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__snake_case ) , Write(__snake_case ) ) self.play(Write(__snake_case , run_time=1 ) , Create(__snake_case , run_time=1 ) ) UpperCAmelCase : Tuple = [] UpperCAmelCase : int = [] for i, rect in enumerate(__snake_case ): UpperCAmelCase : Any = fill.copy().set_fill(__snake_case , opacity=0.7 ) target.move_to(__snake_case ) first_animations.append(GrowFromCenter(__snake_case , run_time=1 ) ) UpperCAmelCase : List[str] = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(__snake_case , run_time=1.5 ) ) self.play(*__snake_case ) self.play(*__snake_case ) self.wait()
23
0
from pathlib import Path from typing import List from transformers import is_torch_available, is_vision_available from transformers.testing_utils import get_tests_dir, is_tool_test from transformers.tools.agent_types import AGENT_TYPE_MAPPING, AgentAudio, AgentImage, AgentText if is_torch_available(): import torch if is_vision_available(): from PIL import Image _snake_case = ["text", "image", "audio"] def A ( _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : str = [] for input_type in input_types: if input_type == "text": inputs.append("Text input" ) elif input_type == "image": inputs.append( Image.open(Path(get_tests_dir("fixtures/tests_samples/COCO" ) ) / "000000039769.png" ).resize((512, 512) ) ) elif input_type == "audio": inputs.append(torch.ones(3_000 ) ) elif isinstance(_lowerCamelCase , _lowerCamelCase ): inputs.append(create_inputs(_lowerCamelCase ) ) else: raise ValueError(F"Invalid type requested: {input_type}" ) return inputs def A ( _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Optional[int] = [] for output in outputs: if isinstance(_lowerCamelCase , (str, AgentText) ): output_types.append("text" ) elif isinstance(_lowerCamelCase , (Image.Image, AgentImage) ): output_types.append("image" ) elif isinstance(_lowerCamelCase , (torch.Tensor, AgentAudio) ): output_types.append("audio" ) else: raise ValueError(F"Invalid output: {output}" ) return output_types @is_tool_test class UpperCAmelCase_ : def snake_case__ ( self): '''simple docstring''' self.assertTrue(hasattr(self.tool, "inputs")) self.assertTrue(hasattr(self.tool, "outputs")) _lowerCAmelCase : List[Any] = self.tool.inputs for _input in inputs: if isinstance(_input, __a): for __input in _input: self.assertTrue(__input in authorized_types) else: self.assertTrue(_input in authorized_types) _lowerCAmelCase : str = self.tool.outputs for _output in outputs: self.assertTrue(_output in authorized_types) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : List[Any] = create_inputs(self.tool.inputs) _lowerCAmelCase : Dict = self.tool(*__a) # There is a single output if len(self.tool.outputs) == 1: _lowerCAmelCase : Dict = [outputs] self.assertListEqual(output_types(__a), self.tool.outputs) def snake_case__ ( self): '''simple docstring''' self.assertTrue(hasattr(self.tool, "description")) self.assertTrue(hasattr(self.tool, "default_checkpoint")) self.assertTrue(self.tool.description.startswith("This is a tool that")) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Dict = create_inputs(self.tool.inputs) _lowerCAmelCase : Any = self.tool(*__a) if not isinstance(__a, __a): _lowerCAmelCase : str = [outputs] self.assertEqual(len(__a), len(self.tool.outputs)) for output, output_type in zip(__a, self.tool.outputs): _lowerCAmelCase : Any = AGENT_TYPE_MAPPING[output_type] self.assertTrue(isinstance(__a, __a)) def snake_case__ ( self): '''simple docstring''' _lowerCAmelCase : Tuple = create_inputs(self.tool.inputs) _lowerCAmelCase : Tuple = [] for _input, input_type in zip(__a, self.tool.inputs): if isinstance(__a, __a): _inputs.append([AGENT_TYPE_MAPPING[_input_type](_input) for _input_type in input_type]) else: _inputs.append(AGENT_TYPE_MAPPING[input_type](_input)) # Should not raise an error _lowerCAmelCase : Dict = self.tool(*__a) if not isinstance(__a, __a): _lowerCAmelCase : Any = [outputs] self.assertEqual(len(__a), len(self.tool.outputs))
36
'''simple docstring''' import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( "The converted tokenizer will be the `slow` tokenizer. To use the fast, update your `tokenizers` library and re-run the tokenizer conversion" ) UpperCamelCase__: str = None UpperCamelCase__: int = { "7B": 11008, "13B": 13824, "30B": 17920, "65B": 22016, "70B": 28672, } UpperCamelCase__: List[Any] = { "7B": 1, "7Bf": 1, "13B": 2, "13Bf": 2, "30B": 4, "65B": 8, "70B": 8, "70Bf": 8, } def snake_case_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple=1 , _lowerCAmelCase : List[Any]=256 ) -> Optional[Any]: return multiple_of * ((int(ffn_dim_multiplier * int(8 * n / 3 ) ) + multiple_of - 1) // multiple_of) def snake_case_ ( _lowerCAmelCase : List[str] ) -> str: with open(_lowerCAmelCase , '''r''' ) as f: return json.load(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any ) -> Optional[Any]: with open(_lowerCAmelCase , '''w''' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any=True ) -> List[Any]: os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : int = os.path.join(_lowerCAmelCase , '''tmp''' ) os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : List[str] = read_json(os.path.join(_lowerCAmelCase , '''params.json''' ) ) UpperCAmelCase : str = NUM_SHARDS[model_size] UpperCAmelCase : Any = params['''n_layers'''] UpperCAmelCase : str = params['''n_heads'''] UpperCAmelCase : Any = n_heads // num_shards UpperCAmelCase : List[str] = params['''dim'''] UpperCAmelCase : Optional[Any] = dim // n_heads UpperCAmelCase : str = 1_0_0_0_0.0 UpperCAmelCase : Optional[int] = 1.0 / (base ** (torch.arange(0 , _lowerCAmelCase , 2 ).float() / dims_per_head)) if "n_kv_heads" in params: UpperCAmelCase : Tuple = params['''n_kv_heads'''] # for GQA / MQA UpperCAmelCase : Optional[int] = n_heads_per_shard // num_key_value_heads UpperCAmelCase : Optional[Any] = dim // num_key_value_heads else: # compatibility with other checkpoints UpperCAmelCase : List[str] = n_heads UpperCAmelCase : Optional[int] = n_heads_per_shard UpperCAmelCase : List[str] = dim # permute for sliced rotary def permute(_lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]=n_heads , _lowerCAmelCase : int=dim , _lowerCAmelCase : Dict=dim ): return w.view(_lowerCAmelCase , dima // n_heads // 2 , 2 , _lowerCAmelCase ).transpose(1 , 2 ).reshape(_lowerCAmelCase , _lowerCAmelCase ) print(f"""Fetching all parameters from the checkpoint at {input_base_path}.""" ) # Load weights if model_size == "7B": # Not sharded # (The sharded implementation would also work, but this is simpler.) UpperCAmelCase : int = torch.load(os.path.join(_lowerCAmelCase , '''consolidated.00.pth''' ) , map_location='''cpu''' ) else: # Sharded UpperCAmelCase : Optional[Any] = [ torch.load(os.path.join(_lowerCAmelCase , f"""consolidated.{i:02d}.pth""" ) , map_location='''cpu''' ) for i in range(_lowerCAmelCase ) ] UpperCAmelCase : Any = 0 UpperCAmelCase : str = {'''weight_map''': {}} for layer_i in range(_lowerCAmelCase ): UpperCAmelCase : Optional[Any] = f"""pytorch_model-{layer_i + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase : Optional[int] = { f"""model.layers.{layer_i}.self_attn.q_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wq.weight"""] ), f"""model.layers.{layer_i}.self_attn.k_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wk.weight"""] ), f"""model.layers.{layer_i}.self_attn.v_proj.weight""": loaded[f"""layers.{layer_i}.attention.wv.weight"""], f"""model.layers.{layer_i}.self_attn.o_proj.weight""": loaded[f"""layers.{layer_i}.attention.wo.weight"""], f"""model.layers.{layer_i}.mlp.gate_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w1.weight"""], f"""model.layers.{layer_i}.mlp.down_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w2.weight"""], f"""model.layers.{layer_i}.mlp.up_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w3.weight"""], f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[f"""layers.{layer_i}.attention_norm.weight"""], f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[f"""layers.{layer_i}.ffn_norm.weight"""], } else: # Sharded # Note that attention.w{q,k,v,o}, feed_fordward.w[1,2,3], attention_norm.weight and ffn_norm.weight share # the same storage object, saving attention_norm and ffn_norm will save other weights too, which is # redundant as other weights will be stitched from multiple shards. To avoid that, they are cloned. UpperCAmelCase : List[str] = { f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.attention_norm.weight""" ].clone(), f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.ffn_norm.weight""" ].clone(), } UpperCAmelCase : Union[str, Any] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wq.weight"""].view(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase : Optional[Any] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wk.weight"""].view( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) UpperCAmelCase : str = torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wv.weight"""].view( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = torch.cat( [loaded[i][f"""layers.{layer_i}.attention.wo.weight"""] for i in range(_lowerCAmelCase )] , dim=1 ) UpperCAmelCase : Any = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w1.weight"""] for i in range(_lowerCAmelCase )] , dim=0 ) UpperCAmelCase : str = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w2.weight"""] for i in range(_lowerCAmelCase )] , dim=1 ) UpperCAmelCase : Tuple = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w3.weight"""] for i in range(_lowerCAmelCase )] , dim=0 ) UpperCAmelCase : Any = inv_freq for k, v in state_dict.items(): UpperCAmelCase : List[Any] = filename param_count += v.numel() torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase : Optional[int] = f"""pytorch_model-{n_layers + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase : str = { '''model.embed_tokens.weight''': loaded['''tok_embeddings.weight'''], '''model.norm.weight''': loaded['''norm.weight'''], '''lm_head.weight''': loaded['''output.weight'''], } else: UpperCAmelCase : Any = { '''model.norm.weight''': loaded[0]['''norm.weight'''], '''model.embed_tokens.weight''': torch.cat( [loaded[i]['''tok_embeddings.weight'''] for i in range(_lowerCAmelCase )] , dim=1 ), '''lm_head.weight''': torch.cat([loaded[i]['''output.weight'''] for i in range(_lowerCAmelCase )] , dim=0 ), } for k, v in state_dict.items(): UpperCAmelCase : Optional[int] = filename param_count += v.numel() torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , _lowerCAmelCase ) ) # Write configs UpperCAmelCase : Union[str, Any] = {'''total_size''': param_count * 2} write_json(_lowerCAmelCase , os.path.join(_lowerCAmelCase , '''pytorch_model.bin.index.json''' ) ) UpperCAmelCase : int = params['''ffn_dim_multiplier'''] if '''ffn_dim_multiplier''' in params else 1 UpperCAmelCase : Tuple = params['''multiple_of'''] if '''multiple_of''' in params else 256 UpperCAmelCase : Any = LlamaConfig( hidden_size=_lowerCAmelCase , intermediate_size=compute_intermediate_size(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) , num_attention_heads=params['''n_heads'''] , num_hidden_layers=params['''n_layers'''] , rms_norm_eps=params['''norm_eps'''] , num_key_value_heads=_lowerCAmelCase , ) config.save_pretrained(_lowerCAmelCase ) # Make space so we can load the model properly now. del state_dict del loaded gc.collect() print('''Loading the checkpoint in a Llama model.''' ) UpperCAmelCase : Optional[int] = LlamaForCausalLM.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa , low_cpu_mem_usage=_lowerCAmelCase ) # Avoid saving this as part of the config. del model.config._name_or_path print('''Saving in the Transformers format.''' ) model.save_pretrained(_lowerCAmelCase , safe_serialization=_lowerCAmelCase ) shutil.rmtree(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any] ) -> List[str]: # Initialize the tokenizer based on the `spm` model UpperCAmelCase : Dict = LlamaTokenizer if LlamaTokenizerFast is None else LlamaTokenizerFast print(f"""Saving a {tokenizer_class.__name__} to {tokenizer_path}.""" ) UpperCAmelCase : List[Any] = tokenizer_class(_lowerCAmelCase ) tokenizer.save_pretrained(_lowerCAmelCase ) def snake_case_ ( ) -> List[Any]: UpperCAmelCase : int = argparse.ArgumentParser() parser.add_argument( '''--input_dir''' , help='''Location of LLaMA weights, which contains tokenizer.model and model folders''' , ) parser.add_argument( '''--model_size''' , choices=['''7B''', '''7Bf''', '''13B''', '''13Bf''', '''30B''', '''65B''', '''70B''', '''70Bf''', '''tokenizer_only'''] , ) parser.add_argument( '''--output_dir''' , help='''Location to write HF model and tokenizer''' , ) parser.add_argument('''--safe_serialization''' , type=_lowerCAmelCase , help='''Whether or not to save using `safetensors`.''' ) UpperCAmelCase : List[Any] = parser.parse_args() if args.model_size != "tokenizer_only": write_model( model_path=args.output_dir , input_base_path=os.path.join(args.input_dir , args.model_size ) , model_size=args.model_size , safe_serialization=args.safe_serialization , ) UpperCAmelCase : Optional[int] = os.path.join(args.input_dir , '''tokenizer.model''' ) write_tokenizer(args.output_dir , _lowerCAmelCase ) if __name__ == "__main__": main()
23
0
'''simple docstring''' import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class lowerCAmelCase_( SCREAMING_SNAKE_CASE_ ): '''simple docstring''' def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> Any: with open(__UpperCAmelCase ,encoding="""utf-8""" ) as input_file: lowerCAmelCase__ : List[str] = re.compile(R"""(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)""" ) lowerCAmelCase__ : Any = input_file.read() lowerCAmelCase__ : Any = regexp.search(__UpperCAmelCase ) return match def UpperCAmelCase_ ( self ,__UpperCAmelCase ) -> Optional[int]: with open(__UpperCAmelCase ,encoding="""utf-8""" ) as input_file: lowerCAmelCase__ : str = re.compile(R"""#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()""" ,re.DOTALL ) lowerCAmelCase__ : Optional[int] = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` lowerCAmelCase__ : Any = regexp.finditer(__UpperCAmelCase ) lowerCAmelCase__ : Union[str, Any] = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def UpperCAmelCase_ ( self ) -> List[Any]: lowerCAmelCase__ : str = Path("""./datasets""" ) lowerCAmelCase__ : Dict = list(dataset_paths.absolute().glob("""**/*.py""" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(__UpperCAmelCase ) ): raise AssertionError(F"""open(...) must use utf-8 encoding in {dataset}""" ) def UpperCAmelCase_ ( self ) -> Tuple: lowerCAmelCase__ : Union[str, Any] = Path("""./datasets""" ) lowerCAmelCase__ : Optional[int] = list(dataset_paths.absolute().glob("""**/*.py""" ) ) for dataset in dataset_files: if self._no_print_statements(str(__UpperCAmelCase ) ): raise AssertionError(F"""print statement found in {dataset}. Use datasets.logger/logging instead.""" )
37
'''simple docstring''' from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Optional[int] = tau * frequency / samplerate UpperCAmelCase : List[Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = cos(_lowerCAmelCase ) UpperCAmelCase : int = _sin / (2 * q_factor) UpperCAmelCase : Any = (1 - _cos) / 2 UpperCAmelCase : List[Any] = 1 - _cos UpperCAmelCase : Union[str, Any] = 1 + alpha UpperCAmelCase : Any = -2 * _cos UpperCAmelCase : Dict = 1 - alpha UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Any = tau * frequency / samplerate UpperCAmelCase : Tuple = sin(_lowerCAmelCase ) UpperCAmelCase : Tuple = cos(_lowerCAmelCase ) UpperCAmelCase : Dict = _sin / (2 * q_factor) UpperCAmelCase : int = (1 + _cos) / 2 UpperCAmelCase : List[Any] = -1 - _cos UpperCAmelCase : Tuple = 1 + alpha UpperCAmelCase : List[str] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha UpperCAmelCase : Tuple = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Optional[int] = tau * frequency / samplerate UpperCAmelCase : Optional[int] = sin(_lowerCAmelCase ) UpperCAmelCase : Tuple = cos(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = _sin / (2 * q_factor) UpperCAmelCase : Union[str, Any] = _sin / 2 UpperCAmelCase : Any = 0 UpperCAmelCase : int = -ba UpperCAmelCase : Optional[Any] = 1 + alpha UpperCAmelCase : List[Any] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha UpperCAmelCase : int = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : List[str] = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : str = cos(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = _sin / (2 * q_factor) UpperCAmelCase : List[str] = 1 - alpha UpperCAmelCase : Any = -2 * _cos UpperCAmelCase : Optional[int] = 1 + alpha UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([ba, ba, ba] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : Optional[Any] = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = cos(_lowerCAmelCase ) UpperCAmelCase : Dict = _sin / (2 * q_factor) UpperCAmelCase : str = 10 ** (gain_db / 40) UpperCAmelCase : int = 1 + alpha * big_a UpperCAmelCase : Union[str, Any] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha * big_a UpperCAmelCase : Union[str, Any] = 1 + alpha / big_a UpperCAmelCase : Tuple = -2 * _cos UpperCAmelCase : Any = 1 - alpha / big_a UpperCAmelCase : Optional[Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : Any = tau * frequency / samplerate UpperCAmelCase : Optional[int] = sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = cos(_lowerCAmelCase ) UpperCAmelCase : str = _sin / (2 * q_factor) UpperCAmelCase : List[str] = 10 ** (gain_db / 40) UpperCAmelCase : Optional[int] = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase : int = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase : int = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase : Optional[int] = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase : str = 2 * sqrt(_lowerCAmelCase ) * alpha UpperCAmelCase : Dict = big_a * (pmc + aaa) UpperCAmelCase : Any = 2 * big_a * mpc UpperCAmelCase : Union[str, Any] = big_a * (pmc - aaa) UpperCAmelCase : Optional[int] = ppmc + aaa UpperCAmelCase : Optional[Any] = -2 * pmpc UpperCAmelCase : Optional[Any] = ppmc - aaa UpperCAmelCase : int = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : int = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = cos(_lowerCAmelCase ) UpperCAmelCase : Any = _sin / (2 * q_factor) UpperCAmelCase : int = 10 ** (gain_db / 40) UpperCAmelCase : List[str] = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase : Union[str, Any] = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase : Optional[Any] = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase : Union[str, Any] = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase : List[str] = 2 * sqrt(_lowerCAmelCase ) * alpha UpperCAmelCase : Any = big_a * (ppmc + aaa) UpperCAmelCase : str = -2 * big_a * pmpc UpperCAmelCase : List[Any] = big_a * (ppmc - aaa) UpperCAmelCase : Optional[Any] = pmc + aaa UpperCAmelCase : Any = 2 * mpc UpperCAmelCase : str = pmc - aaa UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt
23
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from pathlib import Path import torch from ...utils import is_npu_available, is_xpu_available from .config_args import ClusterConfig, default_json_config_file from .config_utils import SubcommandHelpFormatter UpperCAmelCase_ : Any = '''Create a default config file for Accelerate with only a few flags set.''' def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Optional[int]="no" , __magic_name__ : str = default_json_config_file , __magic_name__ : bool = False ) -> str: """simple docstring""" UpperCamelCase :Any = Path(__magic_name__ ) path.parent.mkdir(parents=__magic_name__ , exist_ok=__magic_name__ ) if path.exists(): print( f"""Configuration already exists at {save_location}, will not override. Run `accelerate config` manually or pass a different `save_location`.""" ) return False UpperCamelCase :Dict = mixed_precision.lower() if mixed_precision not in ["no", "fp16", "bf16", "fp8"]: raise ValueError( f"""`mixed_precision` should be one of 'no', 'fp16', 'bf16', or 'fp8'. Received {mixed_precision}""" ) UpperCamelCase :Optional[Any] = { """compute_environment""": """LOCAL_MACHINE""", """mixed_precision""": mixed_precision, } if torch.cuda.is_available(): UpperCamelCase :Union[str, Any] = torch.cuda.device_count() UpperCamelCase :List[Any] = num_gpus UpperCamelCase :Dict = False if num_gpus > 1: UpperCamelCase :Any = """MULTI_GPU""" else: UpperCamelCase :Any = """NO""" elif is_xpu_available() and use_xpu: UpperCamelCase :Optional[Any] = torch.xpu.device_count() UpperCamelCase :Optional[int] = num_xpus UpperCamelCase :int = False if num_xpus > 1: UpperCamelCase :Union[str, Any] = """MULTI_XPU""" else: UpperCamelCase :Union[str, Any] = """NO""" elif is_npu_available(): UpperCamelCase :List[Any] = torch.npu.device_count() UpperCamelCase :Optional[Any] = num_npus UpperCamelCase :Tuple = False if num_npus > 1: UpperCamelCase :Optional[Any] = """MULTI_NPU""" else: UpperCamelCase :List[Any] = """NO""" else: UpperCamelCase :Any = 0 UpperCamelCase :Optional[Any] = True UpperCamelCase :Optional[Any] = 1 UpperCamelCase :List[str] = """NO""" UpperCamelCase :int = ClusterConfig(**__magic_name__ ) config.to_json_file(__magic_name__ ) return path def SCREAMING_SNAKE_CASE_ ( __magic_name__ : Dict , __magic_name__ : Tuple ) -> List[str]: """simple docstring""" UpperCamelCase :Dict = parser.add_parser("""default""" , parents=__magic_name__ , help=__magic_name__ , formatter_class=__magic_name__ ) parser.add_argument( """--config_file""" , default=__magic_name__ , help=( """The path to use to store the config file. Will default to a file named default_config.yaml in the cache """ """location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have """ """such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed """ """with 'huggingface'.""" ) , dest="""save_location""" , ) parser.add_argument( """--mixed_precision""" , choices=["""no""", """fp16""", """bf16"""] , type=__magic_name__ , help="""Whether or not to use mixed precision training. """ """Choose between FP16 and BF16 (bfloat16) training. """ """BF16 training is only supported on Nvidia Ampere GPUs and PyTorch 1.10 or later.""" , default="""no""" , ) parser.set_defaults(func=__magic_name__ ) return parser def SCREAMING_SNAKE_CASE_ ( __magic_name__ : List[Any] ) -> List[str]: """simple docstring""" UpperCamelCase :Optional[Any] = write_basic_config(args.mixed_precision , args.save_location ) if config_file: print(f"""accelerate configuration saved at {config_file}""" )
38
'''simple docstring''' from __future__ import annotations def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : str ) -> bool: UpperCAmelCase : str = get_failure_array(_lowerCAmelCase ) # 2) Step through text searching for pattern UpperCAmelCase , UpperCAmelCase : Optional[Any] = 0, 0 # index into text, pattern while i < len(_lowerCAmelCase ): if pattern[j] == text[i]: if j == (len(_lowerCAmelCase ) - 1): return True j += 1 # if this is a prefix in our pattern # just go back far enough to continue elif j > 0: UpperCAmelCase : Optional[Any] = failure[j - 1] continue i += 1 return False def snake_case_ ( _lowerCAmelCase : str ) -> list[int]: UpperCAmelCase : Optional[Any] = [0] UpperCAmelCase : str = 0 UpperCAmelCase : List[str] = 1 while j < len(_lowerCAmelCase ): if pattern[i] == pattern[j]: i += 1 elif i > 0: UpperCAmelCase : Union[str, Any] = failure[i - 1] continue j += 1 failure.append(_lowerCAmelCase ) return failure if __name__ == "__main__": # Test 1) UpperCamelCase__: str = "abc1abc12" UpperCamelCase__: str = "alskfjaldsabc1abc1abc12k23adsfabcabc" UpperCamelCase__: Any = "alskfjaldsk23adsfabcabc" assert kmp(pattern, texta) and not kmp(pattern, texta) # Test 2) UpperCamelCase__: Tuple = "ABABX" UpperCamelCase__: Union[str, Any] = "ABABZABABYABABX" assert kmp(pattern, text) # Test 3) UpperCamelCase__: Any = "AAAB" UpperCamelCase__: str = "ABAAAAAB" assert kmp(pattern, text) # Test 4) UpperCamelCase__: int = "abcdabcy" UpperCamelCase__: Any = "abcxabcdabxabcdabcdabcy" assert kmp(pattern, text) # Test 5) UpperCamelCase__: List[str] = "aabaabaaa" assert get_failure_array(pattern) == [0, 1, 0, 1, 2, 3, 4, 5, 2]
23
0
import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def __A ( __lowerCAmelCase=None , __lowerCAmelCase=None )-> Dict: """simple docstring""" return field(default_factory=lambda: default , metadata=__lowerCAmelCase ) @dataclass class __lowerCamelCase : """simple docstring""" UpperCamelCase__ = field( metadata={"help": "The csv file to plot."} , ) UpperCamelCase__ = field( default=snake_case__ , metadata={"help": "Whether to plot along batch size or sequence length. Defaults to sequence length."} , ) UpperCamelCase__ = field( default=snake_case__ , metadata={"help": "Whether the csv file has time results or memory results. Defaults to memory results."} , ) UpperCamelCase__ = field( default=snake_case__ , metadata={"help": "Disable logarithmic scale when plotting"} , ) UpperCamelCase__ = field( default=snake_case__ , metadata={ "help": "Whether the csv file has training results or inference results. Defaults to inference results." } , ) UpperCamelCase__ = field( default=snake_case__ , metadata={"help": "Filename under which the plot will be saved. If unused no plot is saved."} , ) UpperCamelCase__ = list_field( default=snake_case__ , metadata={"help": "List of model names that are used instead of the ones in the csv file."}) def __A ( __lowerCAmelCase )-> List[str]: """simple docstring""" try: int(__lowerCAmelCase ) return True except ValueError: return False def __A ( __lowerCAmelCase )-> Optional[int]: """simple docstring""" try: float(__lowerCAmelCase ) return True except ValueError: return False class __lowerCamelCase : """simple docstring""" def __init__( self , UpperCAmelCase ): """simple docstring""" _UpperCAmelCase = args _UpperCAmelCase = defaultdict(lambda: {"bsz": [], "seq_len": [], "result": {}} ) with open(self.args.csv_file , newline='' ) as csv_file: _UpperCAmelCase = csv.DictReader(UpperCAmelCase ) for row in reader: _UpperCAmelCase = row['model'] self.result_dict[model_name]["bsz"].append(int(row['batch_size'] ) ) self.result_dict[model_name]["seq_len"].append(int(row['sequence_length'] ) ) if can_convert_to_int(row['result'] ): # value is not None _UpperCAmelCase = int(row['result'] ) elif can_convert_to_float(row['result'] ): # value is not None _UpperCAmelCase = float(row['result'] ) def UpperCamelCase ( self ): """simple docstring""" _UpperCAmelCase , _UpperCAmelCase = plt.subplots() _UpperCAmelCase = 'Time usage' if self.args.is_time else 'Memory usage' _UpperCAmelCase = title_str + ' for training' if self.args.is_train else title_str + ' for inference' if not self.args.no_log_scale: # set logarithm scales ax.set_xscale('log' ) ax.set_yscale('log' ) for axis in [ax.xaxis, ax.yaxis]: axis.set_major_formatter(ScalarFormatter() ) for model_name_idx, model_name in enumerate(self.result_dict.keys() ): _UpperCAmelCase = sorted(set(self.result_dict[model_name]['bsz'] ) ) _UpperCAmelCase = sorted(set(self.result_dict[model_name]['seq_len'] ) ) _UpperCAmelCase = self.result_dict[model_name]['result'] ((_UpperCAmelCase) , (_UpperCAmelCase)) = ( (batch_sizes, sequence_lengths) if self.args.plot_along_batch else (sequence_lengths, batch_sizes) ) _UpperCAmelCase = ( model_name if self.args.short_model_names is None else self.args.short_model_names[model_name_idx] ) for inner_loop_value in inner_loop_array: if self.args.plot_along_batch: _UpperCAmelCase = np.asarray( [results[(x, inner_loop_value)] for x in x_axis_array if (x, inner_loop_value) in results] , dtype=UpperCAmelCase , ) else: _UpperCAmelCase = np.asarray( [results[(inner_loop_value, x)] for x in x_axis_array if (inner_loop_value, x) in results] , dtype=np.floataa , ) ((_UpperCAmelCase) , (_UpperCAmelCase)) = ( ('batch_size', 'len') if self.args.plot_along_batch else ('in #tokens', 'bsz') ) _UpperCAmelCase = np.asarray(UpperCAmelCase , UpperCAmelCase )[: len(UpperCAmelCase )] plt.scatter( UpperCAmelCase , UpperCAmelCase , label=F"""{label_model_name} - {inner_loop_label}: {inner_loop_value}""" ) plt.plot(UpperCAmelCase , UpperCAmelCase , '--' ) title_str += F""" {label_model_name} vs.""" _UpperCAmelCase = title_str[:-4] _UpperCAmelCase = 'Time in s' if self.args.is_time else 'Memory in MB' # plot plt.title(UpperCAmelCase ) plt.xlabel(UpperCAmelCase ) plt.ylabel(UpperCAmelCase ) plt.legend() if self.args.figure_png_file is not None: plt.savefig(self.args.figure_png_file ) else: plt.show() def __A ( )-> List[Any]: """simple docstring""" _UpperCAmelCase = HfArgumentParser(__lowerCAmelCase ) _UpperCAmelCase = parser.parse_args_into_dataclasses()[0] _UpperCAmelCase = Plot(args=__lowerCAmelCase ) plot.plot() if __name__ == "__main__": main()
39
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from transformers.models.wavaveca.modeling_wavaveca import WavaVecaForSequenceClassification logging.set_verbosity_info() UpperCamelCase__: int = logging.get_logger(__name__) UpperCamelCase__: Dict = { "post_extract_proj": "feature_projection.projection", "encoder.pos_conv.0": "encoder.pos_conv_embed.conv", "self_attn.k_proj": "encoder.layers.*.attention.k_proj", "self_attn.v_proj": "encoder.layers.*.attention.v_proj", "self_attn.q_proj": "encoder.layers.*.attention.q_proj", "self_attn.out_proj": "encoder.layers.*.attention.out_proj", "self_attn_layer_norm": "encoder.layers.*.layer_norm", "fc1": "encoder.layers.*.feed_forward.intermediate_dense", "fc2": "encoder.layers.*.feed_forward.output_dense", "final_layer_norm": "encoder.layers.*.final_layer_norm", "encoder.layer_norm": "encoder.layer_norm", "adapter_layer": "encoder.layers.*.adapter_layer", "w2v_model.layer_norm": "feature_projection.layer_norm", "quantizer.weight_proj": "quantizer.weight_proj", "quantizer.vars": "quantizer.codevectors", "project_q": "project_q", "final_proj": "project_hid", "w2v_encoder.proj": "lm_head", "mask_emb": "masked_spec_embed", "pooling_layer.linear": "projector", "pooling_layer.projection": "classifier", } UpperCamelCase__: Optional[Any] = [ "lm_head", "quantizer.weight_proj", "quantizer.codevectors", "project_q", "project_hid", "projector", "classifier", ] def snake_case_ ( _lowerCAmelCase : str ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = {} with open(_lowerCAmelCase , '''r''' ) as file: for line_number, line in enumerate(_lowerCAmelCase ): UpperCAmelCase : List[str] = line.strip() if line: UpperCAmelCase : str = line.split() UpperCAmelCase : Union[str, Any] = line_number UpperCAmelCase : List[Any] = words[0] UpperCAmelCase : Union[str, Any] = value return result def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : str ) -> int: for attribute in key.split('''.''' ): UpperCAmelCase : Dict = getattr(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Dict = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(_lowerCAmelCase ): UpperCAmelCase : Any = PARAM_MAPPING[full_name.split('''.''' )[-1]] UpperCAmelCase : Dict = '''param''' if weight_type is not None and weight_type != "param": UpperCAmelCase : Optional[int] = getattr(_lowerCAmelCase , _lowerCAmelCase ).shape elif weight_type is not None and weight_type == "param": UpperCAmelCase : List[Any] = hf_pointer for attribute in hf_param_name.split('''.''' ): UpperCAmelCase : Optional[Any] = getattr(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : int = shape_pointer.shape # let's reduce dimension UpperCAmelCase : Union[str, Any] = value[0] else: UpperCAmelCase : List[Any] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": UpperCAmelCase : int = value elif weight_type == "weight_g": UpperCAmelCase : str = value elif weight_type == "weight_v": UpperCAmelCase : Dict = value elif weight_type == "bias": UpperCAmelCase : str = value elif weight_type == "param": for attribute in hf_param_name.split('''.''' ): UpperCAmelCase : int = getattr(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = value else: UpperCAmelCase : Tuple = value logger.info(f"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Any , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : List[Any] ) -> List[Any]: UpperCAmelCase : List[str] = None for param_key in PARAM_MAPPING.keys(): if full_name.endswith(_lowerCAmelCase ): UpperCAmelCase : List[str] = PARAM_MAPPING[full_name.split('''.''' )[-1]] UpperCAmelCase : Any = '''param''' if weight_type is not None and weight_type != "param": UpperCAmelCase : Optional[int] = '''.'''.join([key, weight_type] ) elif weight_type is not None and weight_type == "param": UpperCAmelCase : Optional[int] = '''.'''.join([key, hf_param_name] ) else: UpperCAmelCase : List[Any] = key UpperCAmelCase : Tuple = value if '''lm_head''' in full_key else value[0] UpperCamelCase__: Tuple = { "W_a": "linear_1.weight", "W_b": "linear_2.weight", "b_a": "linear_1.bias", "b_b": "linear_2.bias", "ln_W": "norm.weight", "ln_b": "norm.bias", } def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : List[str] , _lowerCAmelCase : Any=None , _lowerCAmelCase : Optional[Any]=None ) -> int: UpperCAmelCase : List[Any] = False for key, mapped_key in MAPPING.items(): UpperCAmelCase : int = '''wav2vec2.''' + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: UpperCAmelCase : Optional[Any] = True if "*" in mapped_key: UpperCAmelCase : Tuple = name.split(_lowerCAmelCase )[0].split('''.''' )[-2] UpperCAmelCase : List[Any] = mapped_key.replace('''*''' , _lowerCAmelCase ) if "weight_g" in name: UpperCAmelCase : str = '''weight_g''' elif "weight_v" in name: UpperCAmelCase : int = '''weight_v''' elif "bias" in name: UpperCAmelCase : int = '''bias''' elif "weight" in name: # TODO: don't match quantizer.weight_proj UpperCAmelCase : List[str] = '''weight''' else: UpperCAmelCase : Dict = None if hf_dict is not None: rename_dict(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) else: set_recursively(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return is_used return is_used def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Any: UpperCAmelCase : Dict = [] UpperCAmelCase : Dict = fairseq_model.state_dict() UpperCAmelCase : Union[str, Any] = hf_model.wavaveca.feature_extractor for name, value in fairseq_dict.items(): UpperCAmelCase : Dict = False if "conv_layers" in name: load_conv_layer( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , hf_model.config.feat_extract_norm == '''group''' , ) UpperCAmelCase : Any = True else: UpperCAmelCase : Optional[Any] = load_wavaveca_layer(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not is_used: unused_weights.append(_lowerCAmelCase ) logger.warning(f"""Unused weights: {unused_weights}""" ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] ) -> Union[str, Any]: UpperCAmelCase : Any = full_name.split('''conv_layers.''' )[-1] UpperCAmelCase : Optional[int] = name.split('''.''' ) UpperCAmelCase : Tuple = int(items[0] ) UpperCAmelCase : Tuple = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) UpperCAmelCase : Tuple = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) UpperCAmelCase : Union[str, Any] = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.""" ) UpperCAmelCase : Union[str, Any] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.""" ) UpperCAmelCase : List[str] = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(_lowerCAmelCase ) @torch.no_grad() def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Dict=None , _lowerCAmelCase : List[Any]=None , _lowerCAmelCase : int=True , _lowerCAmelCase : Optional[int]=False ) -> Dict: if config_path is not None: UpperCAmelCase : List[str] = WavaVecaConfig.from_pretrained(_lowerCAmelCase ) else: UpperCAmelCase : List[Any] = WavaVecaConfig() if is_seq_class: UpperCAmelCase : Optional[Any] = read_txt_into_dict(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = idalabel UpperCAmelCase : Optional[Any] = WavaVecaForSequenceClassification(_lowerCAmelCase ) UpperCAmelCase : Dict = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase , ) feature_extractor.save_pretrained(_lowerCAmelCase ) elif is_finetuned: if dict_path: UpperCAmelCase : Dict = Dictionary.load(_lowerCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq UpperCAmelCase : Any = target_dict.pad_index UpperCAmelCase : Tuple = target_dict.bos_index UpperCAmelCase : Optional[int] = target_dict.eos_index UpperCAmelCase : Union[str, Any] = len(target_dict.symbols ) UpperCAmelCase : Dict = os.path.join(_lowerCAmelCase , '''vocab.json''' ) if not os.path.isdir(_lowerCAmelCase ): logger.error('''--pytorch_dump_folder_path ({}) should be a directory'''.format(_lowerCAmelCase ) ) return os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : List[Any] = target_dict.indices # fairseq has the <pad> and <s> switched UpperCAmelCase : List[str] = 0 UpperCAmelCase : List[str] = 1 with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as vocab_handle: json.dump(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = WavaVecaCTCTokenizer( _lowerCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token='''|''' , do_lower_case=_lowerCAmelCase , ) UpperCAmelCase : int = True if config.feat_extract_norm == '''layer''' else False UpperCAmelCase : int = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=16000 , padding_value=0 , do_normalize=_lowerCAmelCase , return_attention_mask=_lowerCAmelCase , ) UpperCAmelCase : str = WavaVecaProcessor(feature_extractor=_lowerCAmelCase , tokenizer=_lowerCAmelCase ) processor.save_pretrained(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = WavaVecaForCTC(_lowerCAmelCase ) else: UpperCAmelCase : Dict = WavaVecaForPreTraining(_lowerCAmelCase ) if is_finetuned or is_seq_class: UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : List[str] = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) else: UpperCAmelCase : Optional[Any] = argparse.Namespace(task='''audio_pretraining''' ) UpperCAmelCase : List[Any] = fairseq.tasks.setup_task(_lowerCAmelCase ) UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : int = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] , task=_lowerCAmelCase ) UpperCAmelCase : Optional[int] = model[0].eval() recursively_load_weights(_lowerCAmelCase , _lowerCAmelCase , not is_finetuned ) hf_wavavec.save_pretrained(_lowerCAmelCase ) if __name__ == "__main__": UpperCamelCase__: Dict = argparse.ArgumentParser() parser.add_argument("--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model.") parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to fairseq checkpoint") parser.add_argument("--dict_path", default=None, type=str, help="Path to dict of fine-tuned model") parser.add_argument("--config_path", default=None, type=str, help="Path to hf config.json of model to convert") parser.add_argument( "--not_finetuned", action="store_true", help="Whether the model to convert is a fine-tuned model or not" ) parser.add_argument( "--is_seq_class", action="store_true", help="Whether the model to convert is a fine-tuned sequence classification model or not", ) UpperCamelCase__: Any = parser.parse_args() UpperCamelCase__: int = not args.not_finetuned and not args.is_seq_class convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, is_finetuned, args.is_seq_class, )
23
0
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import torch from ..models.speechta import SpeechTaForTextToSpeech, SpeechTaHifiGan, SpeechTaProcessor from ..utils import is_datasets_available from .base import PipelineTool if is_datasets_available(): from datasets import load_dataset class _A ( _a ): """simple docstring""" UpperCAmelCase : Union[str, Any] = """microsoft/speecht5_tts""" UpperCAmelCase : Optional[Any] = ( """This is a tool that reads an English text out loud. It takes an input named `text` which should contain the """ """text to read (in English) and returns a waveform object containing the sound.""" ) UpperCAmelCase : str = """text_reader""" UpperCAmelCase : str = SpeechTaProcessor UpperCAmelCase : Tuple = SpeechTaForTextToSpeech UpperCAmelCase : Tuple = SpeechTaHifiGan UpperCAmelCase : Optional[Any] = ["""text"""] UpperCAmelCase : List[Any] = ["""audio"""] def __snake_case ( self : Tuple): if self.post_processor is None: a : Tuple = "microsoft/speecht5_hifigan" super().setup() def __snake_case ( self : List[Any] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : List[Any]=None): a : Any = self.pre_processor(text=__UpperCAmelCase , return_tensors="pt" , truncation=__UpperCAmelCase) if speaker_embeddings is None: if not is_datasets_available(): raise ImportError("Datasets needs to be installed if not passing speaker embeddings.") a : List[str] = load_dataset("Matthijs/cmu-arctic-xvectors" , split="validation") a : Optional[int] = torch.tensor(embeddings_dataset[7305]["xvector"]).unsqueeze(0) return {"input_ids": inputs["input_ids"], "speaker_embeddings": speaker_embeddings} def __snake_case ( self : Union[str, Any] , __UpperCAmelCase : Optional[int]): with torch.no_grad(): return self.model.generate_speech(**__UpperCAmelCase) def __snake_case ( self : Tuple , __UpperCAmelCase : str): with torch.no_grad(): return self.post_processor(__UpperCAmelCase).cpu().detach()
40
'''simple docstring''' import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : str ) -> int: with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights UpperCAmelCase : Optional[Any] = FlaxDiffusionPipeline.from_pretrained( '''hf-internal-testing/tiny-stable-diffusion-pipe''' , safety_checker=__snake_case , cache_dir=__snake_case ) UpperCAmelCase : str = [t[-1] for t in os.walk(os.path.join(__snake_case , os.listdir(__snake_case )[0] , '''snapshots''' ) )] UpperCAmelCase : str = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith('''.bin''' ) for f in files ) @slow @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : List[str] ) -> Dict: UpperCAmelCase , UpperCAmelCase : str = FlaxStableDiffusionPipeline.from_pretrained( '''hf-internal-testing/tiny-stable-diffusion-pipe''' , safety_checker=__snake_case ) UpperCAmelCase : List[Any] = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : List[str] = jax.random.PRNGKey(0 ) UpperCAmelCase : Optional[Any] = 4 UpperCAmelCase : Optional[Any] = jax.device_count() UpperCAmelCase : Tuple = num_samples * [prompt] UpperCAmelCase : int = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Tuple = replicate(__snake_case ) UpperCAmelCase : Any = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Optional[Any] = shard(__snake_case ) UpperCAmelCase : Optional[int] = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1E-3 assert np.abs(np.abs(__snake_case , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5E-1 UpperCAmelCase : Union[str, Any] = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(__snake_case ) == num_samples def A ( self : List[Any] ) -> List[str]: UpperCAmelCase , UpperCAmelCase : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''flax''' , safety_checker=__snake_case ) UpperCAmelCase : Dict = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : Optional[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase : Any = 50 UpperCAmelCase : Union[str, Any] = jax.device_count() UpperCAmelCase : int = num_samples * [prompt] UpperCAmelCase : Union[str, Any] = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Dict = replicate(__snake_case ) UpperCAmelCase : int = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Tuple = shard(__snake_case ) UpperCAmelCase : Tuple = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5E-1 def A ( self : int ) -> Dict: UpperCAmelCase , UpperCAmelCase : List[str] = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__snake_case ) UpperCAmelCase : Dict = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : Union[str, Any] = jax.random.PRNGKey(0 ) UpperCAmelCase : List[str] = 50 UpperCAmelCase : Union[str, Any] = jax.device_count() UpperCAmelCase : List[Any] = num_samples * [prompt] UpperCAmelCase : int = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Tuple = replicate(__snake_case ) UpperCAmelCase : List[Any] = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Optional[int] = shard(__snake_case ) UpperCAmelCase : Optional[Any] = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def A ( self : int ) -> Any: UpperCAmelCase , UpperCAmelCase : Dict = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa ) UpperCAmelCase : List[str] = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : List[str] = jax.random.PRNGKey(0 ) UpperCAmelCase : Union[str, Any] = 50 UpperCAmelCase : Optional[int] = jax.device_count() UpperCAmelCase : List[str] = num_samples * [prompt] UpperCAmelCase : Dict = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : Tuple = replicate(__snake_case ) UpperCAmelCase : Any = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : str = shard(__snake_case ) UpperCAmelCase : Optional[int] = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def A ( self : Tuple ) -> Optional[Any]: UpperCAmelCase : int = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , set_alpha_to_one=__snake_case , steps_offset=1 , ) UpperCAmelCase , UpperCAmelCase : Tuple = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , scheduler=__snake_case , safety_checker=__snake_case , ) UpperCAmelCase : Tuple = scheduler.create_state() UpperCAmelCase : Dict = scheduler_state UpperCAmelCase : str = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : int = jax.random.PRNGKey(0 ) UpperCAmelCase : Union[str, Any] = 50 UpperCAmelCase : Optional[Any] = jax.device_count() UpperCAmelCase : Any = num_samples * [prompt] UpperCAmelCase : Dict = pipeline.prepare_inputs(__snake_case ) # shard inputs and rng UpperCAmelCase : str = replicate(__snake_case ) UpperCAmelCase : List[str] = jax.random.split(__snake_case , __snake_case ) UpperCAmelCase : Optional[int] = shard(__snake_case ) UpperCAmelCase : Dict = pipeline(__snake_case , __snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1E-3 assert np.abs((np.abs(__snake_case , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5E-1 def A ( self : Any ) -> Tuple: UpperCAmelCase : List[Any] = ( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) UpperCAmelCase : Union[str, Any] = jax.device_count() UpperCAmelCase : List[Any] = num_samples * [prompt] UpperCAmelCase : str = jax.random.split(jax.random.PRNGKey(0 ) , __snake_case ) UpperCAmelCase , UpperCAmelCase : Tuple = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__snake_case , ) UpperCAmelCase : Dict = replicate(__snake_case ) UpperCAmelCase : Optional[Any] = pipeline.prepare_inputs(__snake_case ) UpperCAmelCase : List[str] = shard(__snake_case ) UpperCAmelCase : Any = pipeline(__snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase : Optional[int] = images[2, 0, 256, 10:17, 1] # With memory efficient attention UpperCAmelCase , UpperCAmelCase : Any = FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__snake_case , use_memory_efficient_attention=__snake_case , ) UpperCAmelCase : int = replicate(__snake_case ) UpperCAmelCase : int = pipeline.prepare_inputs(__snake_case ) UpperCAmelCase : List[Any] = shard(__snake_case ) UpperCAmelCase : Optional[Any] = pipeline(__snake_case , __snake_case , __snake_case , jit=__snake_case ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase : int = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
23
0
'''simple docstring''' import math import sys def SCREAMING_SNAKE_CASE_ (UpperCamelCase ) -> int: if number != int(UpperCamelCase ): raise ValueError("""the value of input must be a natural number""" ) if number < 0: raise ValueError("""the value of input must not be a negative number""" ) if number == 0: return 1 lowerCamelCase__ : Tuple = [-1] * (number + 1) lowerCamelCase__ : Optional[Any] = 0 for i in range(1 , number + 1 ): lowerCamelCase__ : Optional[Any] = sys.maxsize lowerCamelCase__ : Optional[Any] = int(math.sqrt(UpperCamelCase ) ) for j in range(1 , root + 1 ): lowerCamelCase__ : Optional[Any] = 1 + answers[i - (j**2)] lowerCamelCase__ : Optional[Any] = min(UpperCamelCase , UpperCamelCase ) lowerCamelCase__ : str = answer return answers[number] if __name__ == "__main__": import doctest doctest.testmod()
41
'''simple docstring''' import random from .binary_exp_mod import bin_exp_mod def snake_case_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]=1000 ) -> int: if n < 2: return False if n % 2 == 0: return n == 2 # this means n is odd UpperCAmelCase : str = n - 1 UpperCAmelCase : List[Any] = 0 while d % 2 == 0: d /= 2 exp += 1 # n - 1=d*(2**exp) UpperCAmelCase : List[str] = 0 while count < prec: UpperCAmelCase : int = random.randint(2 , n - 1 ) UpperCAmelCase : List[str] = bin_exp_mod(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if b != 1: UpperCAmelCase : int = True for _ in range(_lowerCAmelCase ): if b == n - 1: UpperCAmelCase : Dict = False break UpperCAmelCase : str = b * b b %= n if flag: return False count += 1 return True if __name__ == "__main__": UpperCamelCase__: Optional[int] = abs(int(input("Enter bound : ").strip())) print("Here's the list of primes:") print(", ".join(str(i) for i in range(n + 1) if is_prime_big(i)))
23
0
'''simple docstring''' import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class __UpperCAmelCase ( ctypes.Structure ): # _fields is a specific attr expected by ctypes __lowercase = [("""size""", ctypes.c_int), ("""visible""", ctypes.c_byte)] def SCREAMING_SNAKE_CASE__ ( ) -> List[Any]: if os.name == "nt": _snake_case = CursorInfo() _snake_case = ctypes.windll.kernelaa.GetStdHandle(-11 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(__A , ctypes.byref(__A ) ) _snake_case = False ctypes.windll.kernelaa.SetConsoleCursorInfo(__A , ctypes.byref(__A ) ) elif os.name == "posix": sys.stdout.write('\033[?25l' ) sys.stdout.flush() def SCREAMING_SNAKE_CASE__ ( ) -> List[str]: if os.name == "nt": _snake_case = CursorInfo() _snake_case = ctypes.windll.kernelaa.GetStdHandle(-11 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(__A , ctypes.byref(__A ) ) _snake_case = True ctypes.windll.kernelaa.SetConsoleCursorInfo(__A , ctypes.byref(__A ) ) elif os.name == "posix": sys.stdout.write('\033[?25h' ) sys.stdout.flush() @contextmanager def SCREAMING_SNAKE_CASE__ ( ) -> Union[str, Any]: try: hide_cursor() yield finally: show_cursor()
42
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__: Optional[int] = logging.get_logger(__name__) def snake_case_ ( _lowerCAmelCase : Optional[int] ) -> Optional[int]: UpperCAmelCase : Tuple = DPTConfig(embedding_type='''hybrid''' ) if "large" in checkpoint_url: UpperCAmelCase : Tuple = 1024 UpperCAmelCase : List[Any] = 4096 UpperCAmelCase : str = 24 UpperCAmelCase : List[Any] = 16 UpperCAmelCase : str = [5, 11, 17, 23] UpperCAmelCase : List[Any] = [256, 512, 1024, 1024] UpperCAmelCase : Tuple = (1, 384, 384) if "nyu" or "midas" in checkpoint_url: UpperCAmelCase : Optional[Any] = 768 UpperCAmelCase : Tuple = [1, 1, 1, 0.5] UpperCAmelCase : int = [256, 512, 768, 768] UpperCAmelCase : Any = 150 UpperCAmelCase : Tuple = 16 UpperCAmelCase : Any = (1, 384, 384) UpperCAmelCase : Optional[Any] = False UpperCAmelCase : Tuple = '''project''' if "ade" in checkpoint_url: UpperCAmelCase : Any = True UpperCAmelCase : str = 768 UpperCAmelCase : Optional[int] = [1, 1, 1, 0.5] UpperCAmelCase : List[Any] = 150 UpperCAmelCase : List[Any] = 16 UpperCAmelCase : str = '''huggingface/label-files''' UpperCAmelCase : Tuple = '''ade20k-id2label.json''' UpperCAmelCase : Any = json.load(open(cached_download(hf_hub_url(_lowerCAmelCase , _lowerCAmelCase , repo_type='''dataset''' ) ) , '''r''' ) ) UpperCAmelCase : Optional[Any] = {int(_lowerCAmelCase ): v for k, v in idalabel.items()} UpperCAmelCase : List[Any] = idalabel UpperCAmelCase : Optional[int] = {v: k for k, v in idalabel.items()} UpperCAmelCase : Union[str, Any] = [1, 150, 480, 480] return config, expected_shape def snake_case_ ( _lowerCAmelCase : Union[str, Any] ) -> int: UpperCAmelCase : List[str] = ['''pretrained.model.head.weight''', '''pretrained.model.head.bias'''] for k in ignore_keys: state_dict.pop(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Tuple ) -> Any: if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): UpperCAmelCase : Tuple = name.replace('''pretrained.model''' , '''dpt.encoder''' ) if "pretrained.model" in name: UpperCAmelCase : Union[str, Any] = name.replace('''pretrained.model''' , '''dpt.embeddings''' ) if "patch_embed" in name: UpperCAmelCase : int = name.replace('''patch_embed''' , '''''' ) if "pos_embed" in name: UpperCAmelCase : Tuple = name.replace('''pos_embed''' , '''position_embeddings''' ) if "attn.proj" in name: UpperCAmelCase : Any = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "proj" in name and "project" not in name: UpperCAmelCase : str = name.replace('''proj''' , '''projection''' ) if "blocks" in name: UpperCAmelCase : Any = name.replace('''blocks''' , '''layer''' ) if "mlp.fc1" in name: UpperCAmelCase : Optional[int] = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: UpperCAmelCase : Optional[Any] = name.replace('''mlp.fc2''' , '''output.dense''' ) if "norm1" in name and "backbone" not in name: UpperCAmelCase : Dict = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name and "backbone" not in name: UpperCAmelCase : Tuple = name.replace('''norm2''' , '''layernorm_after''' ) if "scratch.output_conv" in name: UpperCAmelCase : Tuple = name.replace('''scratch.output_conv''' , '''head''' ) if "scratch" in name: UpperCAmelCase : str = name.replace('''scratch''' , '''neck''' ) if "layer1_rn" in name: UpperCAmelCase : Dict = name.replace('''layer1_rn''' , '''convs.0''' ) if "layer2_rn" in name: UpperCAmelCase : int = name.replace('''layer2_rn''' , '''convs.1''' ) if "layer3_rn" in name: UpperCAmelCase : Tuple = name.replace('''layer3_rn''' , '''convs.2''' ) if "layer4_rn" in name: UpperCAmelCase : int = name.replace('''layer4_rn''' , '''convs.3''' ) if "refinenet" in name: UpperCAmelCase : List[str] = int(name[len('''neck.refinenet''' ) : len('''neck.refinenet''' ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 UpperCAmelCase : str = name.replace(f"""refinenet{layer_idx}""" , f"""fusion_stage.layers.{abs(layer_idx-4 )}""" ) if "out_conv" in name: UpperCAmelCase : List[str] = name.replace('''out_conv''' , '''projection''' ) if "resConfUnit1" in name: UpperCAmelCase : Union[str, Any] = name.replace('''resConfUnit1''' , '''residual_layer1''' ) if "resConfUnit2" in name: UpperCAmelCase : Any = name.replace('''resConfUnit2''' , '''residual_layer2''' ) if "conv1" in name: UpperCAmelCase : Optional[int] = name.replace('''conv1''' , '''convolution1''' ) if "conv2" in name: UpperCAmelCase : Tuple = name.replace('''conv2''' , '''convolution2''' ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: UpperCAmelCase : Dict = name.replace('''pretrained.act_postprocess1.0.project.0''' , '''neck.reassemble_stage.readout_projects.0.0''' ) if "pretrained.act_postprocess2.0.project.0" in name: UpperCAmelCase : int = name.replace('''pretrained.act_postprocess2.0.project.0''' , '''neck.reassemble_stage.readout_projects.1.0''' ) if "pretrained.act_postprocess3.0.project.0" in name: UpperCAmelCase : Any = name.replace('''pretrained.act_postprocess3.0.project.0''' , '''neck.reassemble_stage.readout_projects.2.0''' ) if "pretrained.act_postprocess4.0.project.0" in name: UpperCAmelCase : Optional[Any] = name.replace('''pretrained.act_postprocess4.0.project.0''' , '''neck.reassemble_stage.readout_projects.3.0''' ) # resize blocks if "pretrained.act_postprocess1.3" in name: UpperCAmelCase : List[Any] = name.replace('''pretrained.act_postprocess1.3''' , '''neck.reassemble_stage.layers.0.projection''' ) if "pretrained.act_postprocess1.4" in name: UpperCAmelCase : Any = name.replace('''pretrained.act_postprocess1.4''' , '''neck.reassemble_stage.layers.0.resize''' ) if "pretrained.act_postprocess2.3" in name: UpperCAmelCase : Optional[int] = name.replace('''pretrained.act_postprocess2.3''' , '''neck.reassemble_stage.layers.1.projection''' ) if "pretrained.act_postprocess2.4" in name: UpperCAmelCase : str = name.replace('''pretrained.act_postprocess2.4''' , '''neck.reassemble_stage.layers.1.resize''' ) if "pretrained.act_postprocess3.3" in name: UpperCAmelCase : List[str] = name.replace('''pretrained.act_postprocess3.3''' , '''neck.reassemble_stage.layers.2.projection''' ) if "pretrained.act_postprocess4.3" in name: UpperCAmelCase : Tuple = name.replace('''pretrained.act_postprocess4.3''' , '''neck.reassemble_stage.layers.3.projection''' ) if "pretrained.act_postprocess4.4" in name: UpperCAmelCase : int = name.replace('''pretrained.act_postprocess4.4''' , '''neck.reassemble_stage.layers.3.resize''' ) if "pretrained" in name: UpperCAmelCase : Optional[int] = name.replace('''pretrained''' , '''dpt''' ) if "bn" in name: UpperCAmelCase : Dict = name.replace('''bn''' , '''batch_norm''' ) if "head" in name: UpperCAmelCase : Any = name.replace('''head''' , '''head.head''' ) if "encoder.norm" in name: UpperCAmelCase : Optional[int] = name.replace('''encoder.norm''' , '''layernorm''' ) if "auxlayer" in name: UpperCAmelCase : Union[str, Any] = name.replace('''auxlayer''' , '''auxiliary_head.head''' ) if "backbone" in name: UpperCAmelCase : List[Any] = name.replace('''backbone''' , '''backbone.bit.encoder''' ) if ".." in name: UpperCAmelCase : Optional[int] = name.replace('''..''' , '''.''' ) if "stem.conv" in name: UpperCAmelCase : Optional[Any] = name.replace('''stem.conv''' , '''bit.embedder.convolution''' ) if "blocks" in name: UpperCAmelCase : Optional[int] = name.replace('''blocks''' , '''layers''' ) if "convolution" in name and "backbone" in name: UpperCAmelCase : List[Any] = name.replace('''convolution''' , '''conv''' ) if "layer" in name and "backbone" in name: UpperCAmelCase : List[str] = name.replace('''layer''' , '''layers''' ) if "backbone.bit.encoder.bit" in name: UpperCAmelCase : List[Any] = name.replace('''backbone.bit.encoder.bit''' , '''backbone.bit''' ) if "embedder.conv" in name: UpperCAmelCase : List[Any] = name.replace('''embedder.conv''' , '''embedder.convolution''' ) if "backbone.bit.encoder.stem.norm" in name: UpperCAmelCase : Tuple = name.replace('''backbone.bit.encoder.stem.norm''' , '''backbone.bit.embedder.norm''' ) return name def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : List[Any] ) -> Optional[Any]: for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase : Optional[int] = state_dict.pop(f"""dpt.encoder.layer.{i}.attn.qkv.weight""" ) UpperCAmelCase : Tuple = state_dict.pop(f"""dpt.encoder.layer.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase : Tuple = in_proj_weight[: config.hidden_size, :] UpperCAmelCase : int = in_proj_bias[: config.hidden_size] UpperCAmelCase : List[str] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase : List[str] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase : str = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase : Union[str, Any] = in_proj_bias[-config.hidden_size :] def snake_case_ ( ) -> List[str]: UpperCAmelCase : Optional[int] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' UpperCAmelCase : Optional[int] = Image.open(requests.get(_lowerCAmelCase , stream=_lowerCAmelCase ).raw ) return im @torch.no_grad() def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[str] ) -> Any: UpperCAmelCase , UpperCAmelCase : int = get_dpt_config(_lowerCAmelCase ) # load original state_dict from URL # state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu") UpperCAmelCase : List[Any] = torch.load(_lowerCAmelCase , map_location='''cpu''' ) # remove certain keys remove_ignore_keys_(_lowerCAmelCase ) # rename keys for key in state_dict.copy().keys(): UpperCAmelCase : Any = state_dict.pop(_lowerCAmelCase ) UpperCAmelCase : List[Any] = val # read in qkv matrices read_in_q_k_v(_lowerCAmelCase , _lowerCAmelCase ) # load HuggingFace model UpperCAmelCase : Optional[Any] = DPTForSemanticSegmentation(_lowerCAmelCase ) if '''ade''' in checkpoint_url else DPTForDepthEstimation(_lowerCAmelCase ) model.load_state_dict(_lowerCAmelCase ) model.eval() # Check outputs on an image UpperCAmelCase : int = 480 if '''ade''' in checkpoint_url else 384 UpperCAmelCase : List[Any] = DPTImageProcessor(size=_lowerCAmelCase ) UpperCAmelCase : Dict = prepare_img() UpperCAmelCase : Optional[int] = image_processor(_lowerCAmelCase , return_tensors='''pt''' ) # forward pass UpperCAmelCase : Any = model(**_lowerCAmelCase ).logits if '''ade''' in checkpoint_url else model(**_lowerCAmelCase ).predicted_depth if show_prediction: UpperCAmelCase : Dict = ( torch.nn.functional.interpolate( outputs.unsqueeze(1 ) , size=(image.size[1], image.size[0]) , mode='''bicubic''' , align_corners=_lowerCAmelCase , ) .squeeze() .cpu() .numpy() ) Image.fromarray((prediction / prediction.max()) * 255 ).show() if pytorch_dump_folder_path is not None: Path(_lowerCAmelCase ).mkdir(exist_ok=_lowerCAmelCase ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(_lowerCAmelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_lowerCAmelCase ) if push_to_hub: model.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) image_processor.push_to_hub('''ybelkada/dpt-hybrid-midas''' ) if __name__ == "__main__": UpperCamelCase__: Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( "--checkpoint_url", default="https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt", type=str, help="URL of the original DPT checkpoint you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, required=False, help="Path to the output PyTorch model directory.", ) parser.add_argument( "--push_to_hub", action="store_true", ) parser.add_argument( "--model_name", default="dpt-large", type=str, help="Name of the model, in case you're pushing to the hub.", ) parser.add_argument( "--show_prediction", action="store_true", ) UpperCamelCase__: Tuple = parser.parse_args() convert_dpt_checkpoint( args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name, args.show_prediction )
23
0
import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() __lowercase = logging.get_logger('''transformers.models.encodec''') __lowercase = { '''quantizer.vq.layers.*._codebook.inited''': '''quantizer.layers.*.codebook.inited''', '''quantizer.vq.layers.*._codebook.cluster_size''': '''quantizer.layers.*.codebook.cluster_size''', '''quantizer.vq.layers.*._codebook.embed''': '''quantizer.layers.*.codebook.embed''', '''quantizer.vq.layers.*._codebook.embed_avg''': '''quantizer.layers.*.codebook.embed_avg''', } __lowercase = { '''encoder.model.0.conv.conv''': '''encoder.layers.0.conv''', '''encoder.model.1.block.1.conv.conv''': '''encoder.layers.1.block.1.conv''', '''encoder.model.1.block.3.conv.conv''': '''encoder.layers.1.block.3.conv''', '''encoder.model.1.shortcut.conv.conv''': '''encoder.layers.1.shortcut.conv''', '''encoder.model.3.conv.conv''': '''encoder.layers.3.conv''', '''encoder.model.4.block.1.conv.conv''': '''encoder.layers.4.block.1.conv''', '''encoder.model.4.block.3.conv.conv''': '''encoder.layers.4.block.3.conv''', '''encoder.model.4.shortcut.conv.conv''': '''encoder.layers.4.shortcut.conv''', '''encoder.model.6.conv.conv''': '''encoder.layers.6.conv''', '''encoder.model.7.block.1.conv.conv''': '''encoder.layers.7.block.1.conv''', '''encoder.model.7.block.3.conv.conv''': '''encoder.layers.7.block.3.conv''', '''encoder.model.7.shortcut.conv.conv''': '''encoder.layers.7.shortcut.conv''', '''encoder.model.9.conv.conv''': '''encoder.layers.9.conv''', '''encoder.model.10.block.1.conv.conv''': '''encoder.layers.10.block.1.conv''', '''encoder.model.10.block.3.conv.conv''': '''encoder.layers.10.block.3.conv''', '''encoder.model.10.shortcut.conv.conv''': '''encoder.layers.10.shortcut.conv''', '''encoder.model.12.conv.conv''': '''encoder.layers.12.conv''', '''encoder.model.13.lstm''': '''encoder.layers.13.lstm''', '''encoder.model.15.conv.conv''': '''encoder.layers.15.conv''', } __lowercase = { '''encoder.model.0.conv.norm''': '''encoder.layers.0.norm''', '''encoder.model.1.block.1.conv.norm''': '''encoder.layers.1.block.1.norm''', '''encoder.model.1.block.3.conv.norm''': '''encoder.layers.1.block.3.norm''', '''encoder.model.1.shortcut.conv.norm''': '''encoder.layers.1.shortcut.norm''', '''encoder.model.3.conv.norm''': '''encoder.layers.3.norm''', '''encoder.model.4.block.1.conv.norm''': '''encoder.layers.4.block.1.norm''', '''encoder.model.4.block.3.conv.norm''': '''encoder.layers.4.block.3.norm''', '''encoder.model.4.shortcut.conv.norm''': '''encoder.layers.4.shortcut.norm''', '''encoder.model.6.conv.norm''': '''encoder.layers.6.norm''', '''encoder.model.7.block.1.conv.norm''': '''encoder.layers.7.block.1.norm''', '''encoder.model.7.block.3.conv.norm''': '''encoder.layers.7.block.3.norm''', '''encoder.model.7.shortcut.conv.norm''': '''encoder.layers.7.shortcut.norm''', '''encoder.model.9.conv.norm''': '''encoder.layers.9.norm''', '''encoder.model.10.block.1.conv.norm''': '''encoder.layers.10.block.1.norm''', '''encoder.model.10.block.3.conv.norm''': '''encoder.layers.10.block.3.norm''', '''encoder.model.10.shortcut.conv.norm''': '''encoder.layers.10.shortcut.norm''', '''encoder.model.12.conv.norm''': '''encoder.layers.12.norm''', '''encoder.model.15.conv.norm''': '''encoder.layers.15.norm''', } __lowercase = { '''decoder.model.0.conv.conv''': '''decoder.layers.0.conv''', '''decoder.model.1.lstm''': '''decoder.layers.1.lstm''', '''decoder.model.3.convtr.convtr''': '''decoder.layers.3.conv''', '''decoder.model.4.block.1.conv.conv''': '''decoder.layers.4.block.1.conv''', '''decoder.model.4.block.3.conv.conv''': '''decoder.layers.4.block.3.conv''', '''decoder.model.4.shortcut.conv.conv''': '''decoder.layers.4.shortcut.conv''', '''decoder.model.6.convtr.convtr''': '''decoder.layers.6.conv''', '''decoder.model.7.block.1.conv.conv''': '''decoder.layers.7.block.1.conv''', '''decoder.model.7.block.3.conv.conv''': '''decoder.layers.7.block.3.conv''', '''decoder.model.7.shortcut.conv.conv''': '''decoder.layers.7.shortcut.conv''', '''decoder.model.9.convtr.convtr''': '''decoder.layers.9.conv''', '''decoder.model.10.block.1.conv.conv''': '''decoder.layers.10.block.1.conv''', '''decoder.model.10.block.3.conv.conv''': '''decoder.layers.10.block.3.conv''', '''decoder.model.10.shortcut.conv.conv''': '''decoder.layers.10.shortcut.conv''', '''decoder.model.12.convtr.convtr''': '''decoder.layers.12.conv''', '''decoder.model.13.block.1.conv.conv''': '''decoder.layers.13.block.1.conv''', '''decoder.model.13.block.3.conv.conv''': '''decoder.layers.13.block.3.conv''', '''decoder.model.13.shortcut.conv.conv''': '''decoder.layers.13.shortcut.conv''', '''decoder.model.15.conv.conv''': '''decoder.layers.15.conv''', } __lowercase = { '''decoder.model.0.conv.norm''': '''decoder.layers.0.norm''', '''decoder.model.3.convtr.norm''': '''decoder.layers.3.norm''', '''decoder.model.4.block.1.conv.norm''': '''decoder.layers.4.block.1.norm''', '''decoder.model.4.block.3.conv.norm''': '''decoder.layers.4.block.3.norm''', '''decoder.model.4.shortcut.conv.norm''': '''decoder.layers.4.shortcut.norm''', '''decoder.model.6.convtr.norm''': '''decoder.layers.6.norm''', '''decoder.model.7.block.1.conv.norm''': '''decoder.layers.7.block.1.norm''', '''decoder.model.7.block.3.conv.norm''': '''decoder.layers.7.block.3.norm''', '''decoder.model.7.shortcut.conv.norm''': '''decoder.layers.7.shortcut.norm''', '''decoder.model.9.convtr.norm''': '''decoder.layers.9.norm''', '''decoder.model.10.block.1.conv.norm''': '''decoder.layers.10.block.1.norm''', '''decoder.model.10.block.3.conv.norm''': '''decoder.layers.10.block.3.norm''', '''decoder.model.10.shortcut.conv.norm''': '''decoder.layers.10.shortcut.norm''', '''decoder.model.12.convtr.norm''': '''decoder.layers.12.norm''', '''decoder.model.13.block.1.conv.norm''': '''decoder.layers.13.block.1.norm''', '''decoder.model.13.block.3.conv.norm''': '''decoder.layers.13.block.3.norm''', '''decoder.model.13.shortcut.conv.norm''': '''decoder.layers.13.shortcut.norm''', '''decoder.model.15.conv.norm''': '''decoder.layers.15.norm''', } __lowercase = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } __lowercase = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } __lowercase = [] __lowercase = [] def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' for attribute in key.split('''.''' ): __UpperCamelCase :Optional[int] = getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) if weight_type is not None: __UpperCamelCase :List[Any] = getattr(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ).shape else: __UpperCamelCase :Optional[Any] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": __UpperCamelCase :List[str] = value elif weight_type == "weight_g": __UpperCamelCase :str = value elif weight_type == "weight_v": __UpperCamelCase :Dict = value elif weight_type == "bias": __UpperCamelCase :Dict = value elif weight_type == "running_mean": __UpperCamelCase :Dict = value elif weight_type == "running_var": __UpperCamelCase :Union[str, Any] = value elif weight_type == "num_batches_tracked": __UpperCamelCase :Any = value elif weight_type == "weight_ih_l0": __UpperCamelCase :str = value elif weight_type == "weight_hh_l0": __UpperCamelCase :int = value elif weight_type == "bias_ih_l0": __UpperCamelCase :List[str] = value elif weight_type == "bias_hh_l0": __UpperCamelCase :Tuple = value elif weight_type == "weight_ih_l1": __UpperCamelCase :List[Any] = value elif weight_type == "weight_hh_l1": __UpperCamelCase :List[str] = value elif weight_type == "bias_ih_l1": __UpperCamelCase :Dict = value elif weight_type == "bias_hh_l1": __UpperCamelCase :Tuple = value else: __UpperCamelCase :int = value logger.info(f"""{key + ('.' + weight_type if weight_type is not None else '')} was initialized from {full_name}.""" ) def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' for key in ignore_keys: if key.endswith('''.*''' ): if name.startswith(key[:-1] ): return True elif ".*." in key: __UpperCamelCase , __UpperCamelCase :Dict = key.split('''.*.''' ) if prefix in name and suffix in name: return True elif key in name: return True return False def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCamelCase :Dict = [] if model_name == "encodec_24khz" or "encodec_32khz": __UpperCamelCase :Any = MAPPING_24K elif model_name == "encodec_48khz": __UpperCamelCase :Optional[int] = MAPPING_48K else: raise ValueError(f"""Unsupported model: {model_name}""" ) for name, value in orig_dict.items(): if should_ignore(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): logger.info(f"""{name} was ignored""" ) continue __UpperCamelCase :Optional[Any] = False for key, mapped_key in MAPPING.items(): if "*" in key: __UpperCamelCase , __UpperCamelCase :Tuple = key.split('''.*.''' ) if prefix in name and suffix in name: __UpperCamelCase :Union[str, Any] = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith('''embed''' ) and name.endswith('''embed_avg''' ): continue __UpperCamelCase :Optional[Any] = True if "*" in mapped_key: __UpperCamelCase :Tuple = name.split(SCREAMING_SNAKE_CASE )[0].split('''.''' )[-2] __UpperCamelCase :Optional[int] = mapped_key.replace('''*''' , SCREAMING_SNAKE_CASE ) if "weight_g" in name: __UpperCamelCase :Dict = '''weight_g''' elif "weight_v" in name: __UpperCamelCase :List[Any] = '''weight_v''' elif "weight_ih_l0" in name: __UpperCamelCase :Optional[int] = '''weight_ih_l0''' elif "weight_hh_l0" in name: __UpperCamelCase :Optional[Any] = '''weight_hh_l0''' elif "bias_ih_l0" in name: __UpperCamelCase :List[Any] = '''bias_ih_l0''' elif "bias_hh_l0" in name: __UpperCamelCase :Optional[int] = '''bias_hh_l0''' elif "weight_ih_l1" in name: __UpperCamelCase :List[str] = '''weight_ih_l1''' elif "weight_hh_l1" in name: __UpperCamelCase :Any = '''weight_hh_l1''' elif "bias_ih_l1" in name: __UpperCamelCase :List[Any] = '''bias_ih_l1''' elif "bias_hh_l1" in name: __UpperCamelCase :Optional[int] = '''bias_hh_l1''' elif "bias" in name: __UpperCamelCase :int = '''bias''' elif "weight" in name: __UpperCamelCase :int = '''weight''' elif "running_mean" in name: __UpperCamelCase :List[str] = '''running_mean''' elif "running_var" in name: __UpperCamelCase :str = '''running_var''' elif "num_batches_tracked" in name: __UpperCamelCase :List[str] = '''num_batches_tracked''' else: __UpperCamelCase :int = None set_recursively(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) continue if not is_used: unused_weights.append(SCREAMING_SNAKE_CASE ) logger.warning(f"""Unused weights: {unused_weights}""" ) @torch.no_grad() def lowerCamelCase ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE=None , SCREAMING_SNAKE_CASE=None , ): '''simple docstring''' if config_path is not None: __UpperCamelCase :Any = EncodecConfig.from_pretrained(SCREAMING_SNAKE_CASE ) else: __UpperCamelCase :Optional[int] = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": __UpperCamelCase :List[str] = [8, 5, 4, 4] __UpperCamelCase :Dict = [2.2] __UpperCamelCase :int = 64 __UpperCamelCase :Any = 32_000 __UpperCamelCase :List[str] = 2_048 __UpperCamelCase :str = False __UpperCamelCase :Any = False __UpperCamelCase :Tuple = False elif model_name == "encodec_48khz": __UpperCamelCase :List[str] = [8, 5, 4, 2] __UpperCamelCase :Union[str, Any] = [3.0, 6.0, 12.0, 24.0] __UpperCamelCase :Tuple = 48_000 __UpperCamelCase :Dict = 2 __UpperCamelCase :Optional[Any] = False __UpperCamelCase :List[Any] = '''time_group_norm''' __UpperCamelCase :int = True __UpperCamelCase :int = 1.0 __UpperCamelCase :str = 0.01 else: raise ValueError(f"""Unknown model name: {model_name}""" ) __UpperCamelCase :str = EncodecModel(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Union[str, Any] = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(SCREAMING_SNAKE_CASE ) __UpperCamelCase :Union[str, Any] = torch.load(SCREAMING_SNAKE_CASE ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights __UpperCamelCase :List[str] = original_checkpoint['''best_state'''] recursively_load_weights(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) model.save_pretrained(SCREAMING_SNAKE_CASE ) if repo_id: print('''Pushing to the hub...''' ) feature_extractor.push_to_hub(SCREAMING_SNAKE_CASE ) model.push_to_hub(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase = argparse.ArgumentParser() parser.add_argument( '''--model''', default='''encodec_24khz''', type=str, help='''The model to convert. Should be one of \'encodec_24khz\', \'encodec_32khz\', \'encodec_48khz\'.''', ) parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to original checkpoint''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) __lowercase = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
43
'''simple docstring''' # coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import sys import transformers UpperCamelCase__: Optional[int] = "3" print("Python version:", sys.version) print("transformers version:", transformers.__version__) try: import torch print("Torch version:", torch.__version__) print("Cuda available:", torch.cuda.is_available()) print("Cuda version:", torch.version.cuda) print("CuDNN version:", torch.backends.cudnn.version()) print("Number of GPUs available:", torch.cuda.device_count()) print("NCCL version:", torch.cuda.nccl.version()) except ImportError: print("Torch version:", None) try: import deepspeed print("DeepSpeed version:", deepspeed.__version__) except ImportError: print("DeepSpeed version:", None) try: import tensorflow as tf print("TensorFlow version:", tf.__version__) print("TF GPUs available:", bool(tf.config.list_physical_devices("GPU"))) print("Number of TF GPUs available:", len(tf.config.list_physical_devices("GPU"))) except ImportError: print("TensorFlow version:", None)
23
0
"""simple docstring""" import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def SCREAMING_SNAKE_CASE ( _lowerCamelCase : Tuple ) -> Dict: _lowerCAmelCase : str = [ """encoder.version""", """decoder.version""", """model.encoder.version""", """model.decoder.version""", """_float_tensor""", """decoder.output_projection.weight""", ] for k in ignore_keys: state_dict.pop(_lowerCamelCase ,_lowerCamelCase ) def SCREAMING_SNAKE_CASE ( _lowerCamelCase : Tuple ) -> Union[str, Any]: _lowerCAmelCase , _lowerCAmelCase : List[Any] = emb.weight.shape _lowerCAmelCase : Any = nn.Linear(_lowerCamelCase ,_lowerCamelCase ,bias=_lowerCamelCase ) _lowerCAmelCase : Optional[Any] = emb.weight.data return lin_layer def SCREAMING_SNAKE_CASE ( _lowerCamelCase : Dict ,_lowerCamelCase : Any="facebook/mbart-large-en-ro" ,_lowerCamelCase : List[Any]=False ,_lowerCamelCase : Any=False ) -> int: _lowerCAmelCase : Dict = torch.load(_lowerCamelCase ,map_location="""cpu""" )["""model"""] remove_ignore_keys_(_lowerCamelCase ) _lowerCAmelCase : List[str] = state_dict["""encoder.embed_tokens.weight"""].shape[0] _lowerCAmelCase : Any = MBartConfig.from_pretrained(_lowerCamelCase ,vocab_size=_lowerCamelCase ) if mbart_aa and finetuned: _lowerCAmelCase : Any = """relu""" _lowerCAmelCase : Tuple = state_dict["""decoder.embed_tokens.weight"""] _lowerCAmelCase : Optional[int] = MBartForConditionalGeneration(_lowerCamelCase ) model.model.load_state_dict(_lowerCamelCase ) if finetuned: _lowerCAmelCase : str = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": _a : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( 'fairseq_path', type=str, help='bart.large, bart.large.cnn or a path to a model.pt on local filesystem.' ) parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--hf_config', default='facebook/mbart-large-cc25', type=str, help='Which huggingface architecture to use: mbart-large', ) parser.add_argument('--mbart_50', action='store_true', help='whether the model is mMART-50 checkpoint') parser.add_argument('--finetuned', action='store_true', help='whether the model is a fine-tuned checkpoint') _a : Optional[Any] = parser.parse_args() _a : List[str] = convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
44
'''simple docstring''' from __future__ import annotations def snake_case_ ( _lowerCAmelCase : list[int | float] , _lowerCAmelCase : int , _lowerCAmelCase : int ) -> int | float: if len(_lowerCAmelCase ) == 0: raise ValueError('''find_max() arg is an empty sequence''' ) if ( left >= len(_lowerCAmelCase ) or left < -len(_lowerCAmelCase ) or right >= len(_lowerCAmelCase ) or right < -len(_lowerCAmelCase ) ): raise IndexError('''list index out of range''' ) if left == right: return nums[left] UpperCAmelCase : List[Any] = (left + right) >> 1 # the middle UpperCAmelCase : Optional[Any] = find_max(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # find max in range[left, mid] UpperCAmelCase : Dict = find_max(_lowerCAmelCase , mid + 1 , _lowerCAmelCase ) # find max in range[mid + 1, right] return left_max if left_max >= right_max else right_max if __name__ == "__main__": import doctest doctest.testmod(verbose=True)
23
0
"""simple docstring""" import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, get_image_size, is_torch_available, is_torch_tensor, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_torch_available(): import torch if is_vision_available(): import PIL lowercase_ = logging.get_logger(__name__) def lowercase ( lowerCAmelCase__ : np.ndarray , lowerCAmelCase__ : Union[int, Iterable[int]] , lowerCAmelCase__ : bool , lowerCAmelCase__ : int ) -> Tuple[int, int]: def constraint_to_multiple_of(lowerCAmelCase__ : Dict , lowerCAmelCase__ : Dict , lowerCAmelCase__ : List[str]=0 , lowerCAmelCase__ : List[str]=None ): __a = round(val / multiple ) * multiple if max_val is not None and x > max_val: __a = math.floor(val / multiple ) * multiple if x < min_val: __a = math.ceil(val / multiple ) * multiple return x __a = (output_size, output_size) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else output_size __a , __a = get_image_size(lowerCAmelCase__ ) __a , __a = output_size # determine new height and width __a = output_height / input_height __a = output_width / input_width if keep_aspect_ratio: # scale as little as possible if abs(1 - scale_width ) < abs(1 - scale_height ): # fit width __a = scale_width else: # fit height __a = scale_height __a = constraint_to_multiple_of(scale_height * input_height , multiple=lowerCAmelCase__ ) __a = constraint_to_multiple_of(scale_width * input_width , multiple=lowerCAmelCase__ ) return (new_height, new_width) class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): '''simple docstring''' __UpperCAmelCase : Dict = ['pixel_values'] def __init__( self , _a = True , _a = None , _a = PILImageResampling.BILINEAR , _a = False , _a = 1 , _a = True , _a = 1 / 255 , _a = True , _a = None , _a = None , **_a , ): super().__init__(**_a ) __a = size if size is not None else {'''height''': 384, '''width''': 384} __a = get_size_dict(_a ) __a = do_resize __a = size __a = keep_aspect_ratio __a = ensure_multiple_of __a = resample __a = do_rescale __a = rescale_factor __a = do_normalize __a = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __a = image_std if image_std is not None else IMAGENET_STANDARD_STD def __UpperCAmelCase ( self , _a , _a , _a = False , _a = 1 , _a = PILImageResampling.BICUBIC , _a = None , **_a , ): __a = get_size_dict(_a ) if "height" not in size or "width" not in size: raise ValueError(f'''The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}''' ) __a = get_resize_output_image_size( _a , output_size=(size['''height'''], size['''width''']) , keep_aspect_ratio=_a , multiple=_a , ) return resize(_a , size=_a , resample=_a , data_format=_a , **_a ) def __UpperCAmelCase ( self , _a , _a , _a = None , **_a , ): return rescale(_a , scale=_a , data_format=_a , **_a ) def __UpperCAmelCase ( self , _a , _a , _a , _a = None , **_a , ): return normalize(_a , mean=_a , std=_a , data_format=_a , **_a ) def __UpperCAmelCase ( self , _a , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = None , _a = ChannelDimension.FIRST , **_a , ): __a = do_resize if do_resize is not None else self.do_resize __a = size if size is not None else self.size __a = get_size_dict(_a ) __a = keep_aspect_ratio if keep_aspect_ratio is not None else self.keep_aspect_ratio __a = ensure_multiple_of if ensure_multiple_of is not None else self.ensure_multiple_of __a = resample if resample is not None else self.resample __a = do_rescale if do_rescale is not None else self.do_rescale __a = rescale_factor if rescale_factor is not None else self.rescale_factor __a = do_normalize if do_normalize is not None else self.do_normalize __a = image_mean if image_mean is not None else self.image_mean __a = image_std if image_std is not None else self.image_std __a = make_list_of_images(_a ) if not valid_images(_a ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. __a = [to_numpy_array(_a ) for image in images] if do_resize: __a = [self.resize(image=_a , size=_a , resample=_a ) for image in images] if do_rescale: __a = [self.rescale(image=_a , scale=_a ) for image in images] if do_normalize: __a = [self.normalize(image=_a , mean=_a , std=_a ) for image in images] __a = [to_channel_dimension_format(_a , _a ) for image in images] __a = {'''pixel_values''': images} return BatchFeature(data=_a , tensor_type=_a ) def __UpperCAmelCase ( self , _a , _a = None ): __a = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(_a ) != len(_a ): raise ValueError( '''Make sure that you pass in as many target sizes as the batch dimension of the logits''' ) if is_torch_tensor(_a ): __a = target_sizes.numpy() __a = [] for idx in range(len(_a ) ): __a = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='''bilinear''' , align_corners=_a ) __a = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(_a ) else: __a = logits.argmax(dim=1 ) __a = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
45
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = 42 lowerCamelCase__ = 42 def __init__( self : Union[str, Any] , __snake_case : UNetaDModel , __snake_case : ScoreSdeVeScheduler ) -> int: super().__init__() self.register_modules(unet=__snake_case , scheduler=__snake_case ) @torch.no_grad() def __call__( self : Optional[int] , __snake_case : int = 1 , __snake_case : int = 2000 , __snake_case : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __snake_case : Optional[str] = "pil" , __snake_case : bool = True , **__snake_case : Optional[int] , ) -> Union[ImagePipelineOutput, Tuple]: UpperCAmelCase : str = self.unet.config.sample_size UpperCAmelCase : Union[str, Any] = (batch_size, 3, img_size, img_size) UpperCAmelCase : int = self.unet UpperCAmelCase : Any = randn_tensor(__snake_case , generator=__snake_case ) * self.scheduler.init_noise_sigma UpperCAmelCase : List[Any] = sample.to(self.device ) self.scheduler.set_timesteps(__snake_case ) self.scheduler.set_sigmas(__snake_case ) for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): UpperCAmelCase : Any = self.scheduler.sigmas[i] * torch.ones(shape[0] , device=self.device ) # correction step for _ in range(self.scheduler.config.correct_steps ): UpperCAmelCase : Union[str, Any] = self.unet(__snake_case , __snake_case ).sample UpperCAmelCase : Optional[Any] = self.scheduler.step_correct(__snake_case , __snake_case , generator=__snake_case ).prev_sample # prediction step UpperCAmelCase : Optional[Any] = model(__snake_case , __snake_case ).sample UpperCAmelCase : List[str] = self.scheduler.step_pred(__snake_case , __snake_case , __snake_case , generator=__snake_case ) UpperCAmelCase , UpperCAmelCase : Optional[Any] = output.prev_sample, output.prev_sample_mean UpperCAmelCase : int = sample_mean.clamp(0 , 1 ) UpperCAmelCase : Union[str, Any] = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCAmelCase : Optional[Any] = self.numpy_to_pil(__snake_case ) if not return_dict: return (sample,) return ImagePipelineOutput(images=__snake_case )
23
0
"""simple docstring""" import math import os import unittest from transformers import MegatronBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, ) class lowercase : def __init__( self , lowercase , lowercase=13 , lowercase=7 , lowercase=True , lowercase=True , lowercase=True , lowercase=True , lowercase=99 , lowercase=64 , lowercase=32 , lowercase=5 , lowercase=4 , lowercase=37 , lowercase="gelu" , lowercase=0.1 , lowercase=0.1 , lowercase=512 , lowercase=16 , lowercase=2 , lowercase=0.02 , lowercase=3 , lowercase=4 , lowercase=None , ) -> List[Any]: lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = seq_length lowerCAmelCase = is_training lowerCAmelCase = use_input_mask lowerCAmelCase = use_token_type_ids lowerCAmelCase = use_labels lowerCAmelCase = vocab_size lowerCAmelCase = hidden_size lowerCAmelCase = embedding_size lowerCAmelCase = num_hidden_layers lowerCAmelCase = num_attention_heads lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_act lowerCAmelCase = hidden_dropout_prob lowerCAmelCase = attention_probs_dropout_prob lowerCAmelCase = max_position_embeddings lowerCAmelCase = type_vocab_size lowerCAmelCase = type_sequence_label_size lowerCAmelCase = initializer_range lowerCAmelCase = num_labels lowerCAmelCase = num_choices lowerCAmelCase = scope def _snake_case ( self ) -> int: lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowerCAmelCase = None if self.use_input_mask: lowerCAmelCase = random_attention_mask([self.batch_size, self.seq_length] ) lowerCAmelCase = None if self.use_token_type_ids: lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowerCAmelCase = None lowerCAmelCase = None lowerCAmelCase = None if self.use_labels: lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowerCAmelCase = ids_tensor([self.batch_size] , self.num_choices ) lowerCAmelCase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _snake_case ( self ) -> List[str]: return MegatronBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , embedding_size=self.embedding_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=lowercase , initializer_range=self.initializer_range , ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) -> Optional[int]: lowerCAmelCase = MegatronBertModel(config=lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase ) lowerCAmelCase = model(lowercase , token_type_ids=lowercase ) lowerCAmelCase = model(lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) -> Dict: lowerCAmelCase = MegatronBertForMaskedLM(config=lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) -> Any: lowerCAmelCase = MegatronBertForCausalLM(config=lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) -> List[Any]: lowerCAmelCase = MegatronBertForNextSentencePrediction(config=lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model( lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) -> Tuple: lowerCAmelCase = MegatronBertForPreTraining(config=lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model( lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase , next_sentence_label=lowercase , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) -> List[Any]: lowerCAmelCase = MegatronBertForQuestionAnswering(config=lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model( lowercase , attention_mask=lowercase , token_type_ids=lowercase , start_positions=lowercase , end_positions=lowercase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) -> List[Any]: lowerCAmelCase = self.num_labels lowerCAmelCase = MegatronBertForSequenceClassification(lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) -> Union[str, Any]: lowerCAmelCase = self.num_labels lowerCAmelCase = MegatronBertForTokenClassification(config=lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = model(lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _snake_case ( self , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase , lowercase ) -> Optional[Any]: lowerCAmelCase = self.num_choices lowerCAmelCase = MegatronBertForMultipleChoice(config=lowercase ) model.to(lowercase ) model.eval() lowerCAmelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() lowerCAmelCase = model( lowercase , attention_mask=lowercase , token_type_ids=lowercase , labels=lowercase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _snake_case ( self ) -> int: lowerCAmelCase = self.prepare_config_and_inputs() ( ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ( lowerCAmelCase ) , ) = config_and_inputs lowerCAmelCase = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class lowercase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): _SCREAMING_SNAKE_CASE = ( ( MegatronBertModel, MegatronBertForMaskedLM, MegatronBertForCausalLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, ) if is_torch_available() else () ) _SCREAMING_SNAKE_CASE = ( { 'feature-extraction': MegatronBertModel, 'fill-mask': MegatronBertForMaskedLM, 'question-answering': MegatronBertForQuestionAnswering, 'text-classification': MegatronBertForSequenceClassification, 'text-generation': MegatronBertForCausalLM, 'token-classification': MegatronBertForTokenClassification, 'zero-shot': MegatronBertForSequenceClassification, } if is_torch_available() else {} ) _SCREAMING_SNAKE_CASE = True # test_resize_embeddings = False _SCREAMING_SNAKE_CASE = False def _snake_case ( self , lowercase , lowercase , lowercase=False ) -> int: lowerCAmelCase = super()._prepare_for_class(lowercase , lowercase , return_labels=lowercase ) if return_labels: if model_class in get_values(lowercase ): lowerCAmelCase = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=lowercase ) lowerCAmelCase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowercase ) return inputs_dict def _snake_case ( self ) -> List[Any]: lowerCAmelCase = MegatronBertModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=lowercase , hidden_size=37 ) def _snake_case ( self ) -> Tuple: self.config_tester.run_common_tests() def _snake_case ( self ) -> str: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_model(*lowercase ) def _snake_case ( self ) -> List[str]: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_masked_lm(*lowercase ) def _snake_case ( self ) -> Union[str, Any]: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_multiple_choice(*lowercase ) def _snake_case ( self ) -> int: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_next_sequence_prediction(*lowercase ) def _snake_case ( self ) -> int: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_pretraining(*lowercase ) def _snake_case ( self ) -> Dict: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_question_answering(*lowercase ) def _snake_case ( self ) -> Optional[int]: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_sequence_classification(*lowercase ) def _snake_case ( self ) -> int: lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_token_classification(*lowercase ) def UpperCAmelCase__ ( SCREAMING_SNAKE_CASE : List[Any] ): '''simple docstring''' return torch.tensor( SCREAMING_SNAKE_CASE , dtype=torch.long , device=SCREAMING_SNAKE_CASE , ) SCREAMING_SNAKE_CASE__ = 1e-4 @require_torch @require_sentencepiece @require_tokenizers class lowercase ( unittest.TestCase ): @slow @unittest.skip("""Model is not available.""" ) def _snake_case ( self ) -> Any: lowerCAmelCase = """nvidia/megatron-bert-uncased-345m""" if "MYDIR" in os.environ: lowerCAmelCase = os.path.join(os.environ["""MYDIR"""] , lowercase ) lowerCAmelCase = MegatronBertModel.from_pretrained(lowercase ) model.to(lowercase ) model.half() lowerCAmelCase = _long_tensor([[101, 7_110, 1_005, 1_056, 2_023, 11_333, 17_413, 1_029, 102]] ) with torch.no_grad(): lowerCAmelCase = model(lowercase )[0] lowerCAmelCase = torch.Size((1, 9, 1_024) ) self.assertEqual(output.shape , lowercase ) lowerCAmelCase = [-0.6_040, -0.2_517, -0.1_025, 0.3_420, -0.6_758, -0.0_017, -0.1_089, -0.1_990, 0.5_728] for ii in range(3 ): for jj in range(3 ): lowerCAmelCase = output[0, ii, jj] lowerCAmelCase = expected[3 * ii + jj] lowerCAmelCase = """ii={} jj={} a={} b={}""".format(lowercase , lowercase , lowercase , lowercase ) self.assertTrue(math.isclose(lowercase , lowercase , rel_tol=lowercase , abs_tol=lowercase ) , msg=lowercase )
46
'''simple docstring''' import warnings from contextlib import contextmanager from ....processing_utils import ProcessorMixin class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = """MCTCTFeatureExtractor""" lowerCamelCase__ = """AutoTokenizer""" def __init__( self : Dict , __snake_case : Optional[int] , __snake_case : List[str] ) -> str: super().__init__(__snake_case , __snake_case ) UpperCAmelCase : List[Any] = self.feature_extractor UpperCAmelCase : Union[str, Any] = False def __call__( self : Any , *__snake_case : List[str] , **__snake_case : Any ) -> List[Any]: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*__snake_case , **__snake_case ) if "raw_speech" in kwargs: warnings.warn('''Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.''' ) UpperCAmelCase : int = kwargs.pop('''raw_speech''' ) else: UpperCAmelCase : Union[str, Any] = kwargs.pop('''audio''' , __snake_case ) UpperCAmelCase : Optional[Any] = kwargs.pop('''sampling_rate''' , __snake_case ) UpperCAmelCase : Dict = kwargs.pop('''text''' , __snake_case ) if len(__snake_case ) > 0: UpperCAmelCase : Any = args[0] UpperCAmelCase : Optional[int] = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if audio is not None: UpperCAmelCase : List[str] = self.feature_extractor(__snake_case , *__snake_case , sampling_rate=__snake_case , **__snake_case ) if text is not None: UpperCAmelCase : int = self.tokenizer(__snake_case , **__snake_case ) if text is None: return inputs elif audio is None: return encodings else: UpperCAmelCase : str = encodings['''input_ids'''] return inputs def A ( self : List[Any] , *__snake_case : List[Any] , **__snake_case : List[Any] ) -> str: return self.tokenizer.batch_decode(*__snake_case , **__snake_case ) def A ( self : List[Any] , *__snake_case : int , **__snake_case : Optional[int] ) -> Any: # For backward compatibility if self._in_target_context_manager: return self.current_processor.pad(*__snake_case , **__snake_case ) UpperCAmelCase : List[Any] = kwargs.pop('''input_features''' , __snake_case ) UpperCAmelCase : Optional[Any] = kwargs.pop('''labels''' , __snake_case ) if len(__snake_case ) > 0: UpperCAmelCase : List[str] = args[0] UpperCAmelCase : List[Any] = args[1:] if input_features is not None: UpperCAmelCase : Tuple = self.feature_extractor.pad(__snake_case , *__snake_case , **__snake_case ) if labels is not None: UpperCAmelCase : Optional[int] = self.tokenizer.pad(__snake_case , **__snake_case ) if labels is None: return input_features elif input_features is None: return labels else: UpperCAmelCase : List[str] = labels['''input_ids'''] return input_features def A ( self : Union[str, Any] , *__snake_case : Optional[Any] , **__snake_case : Optional[int] ) -> Optional[Any]: return self.tokenizer.decode(*__snake_case , **__snake_case ) @contextmanager def A ( self : Any ) -> Optional[int]: warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your audio inputs, or in a separate call.''' ) UpperCAmelCase : Dict = True UpperCAmelCase : List[Any] = self.tokenizer yield UpperCAmelCase : Tuple = self.feature_extractor UpperCAmelCase : List[Any] = False
23
0
'''simple docstring''' lowerCamelCase : Dict = "Alexander Joslin" import operator as op from .stack import Stack def _lowerCAmelCase ( _UpperCamelCase : str ) -> int: """simple docstring""" _SCREAMING_SNAKE_CASE ={'*': op.mul, '/': op.truediv, '+': op.add, '-': op.sub} _SCREAMING_SNAKE_CASE =Stack() _SCREAMING_SNAKE_CASE =Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(_UpperCamelCase ) ) elif i in operators: # RULE 2 operator_stack.push(_UpperCamelCase ) elif i == ")": # RULE 4 _SCREAMING_SNAKE_CASE =operator_stack.peek() operator_stack.pop() _SCREAMING_SNAKE_CASE =operand_stack.peek() operand_stack.pop() _SCREAMING_SNAKE_CASE =operand_stack.peek() operand_stack.pop() _SCREAMING_SNAKE_CASE =operators[opr](_UpperCamelCase , _UpperCamelCase ) operand_stack.push(_UpperCamelCase ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase : Optional[Any] = "(5 + ((4 * 2) * (2 + 3)))" # answer = 45 print(f'''{equation} = {dijkstras_two_stack_algorithm(equation)}''')
47
'''simple docstring''' from math import isclose, sqrt def snake_case_ ( _lowerCAmelCase : float , _lowerCAmelCase : float , _lowerCAmelCase : float ) -> tuple[float, float, float]: UpperCAmelCase : Optional[int] = point_y / 4 / point_x UpperCAmelCase : str = 2 * normal_gradient / (1 + normal_gradient * normal_gradient) UpperCAmelCase : Any = (1 - normal_gradient * normal_gradient) / ( 1 + normal_gradient * normal_gradient ) UpperCAmelCase : Union[str, Any] = (sa - ca * incoming_gradient) / (ca + sa * incoming_gradient) # to find the next point, solve the simultaeneous equations: # y^2 + 4x^2 = 100 # y - b = m * (x - a) # ==> A x^2 + B x + C = 0 UpperCAmelCase : Union[str, Any] = outgoing_gradient**2 + 4 UpperCAmelCase : Dict = 2 * outgoing_gradient * (point_y - outgoing_gradient * point_x) UpperCAmelCase : List[Any] = (point_y - outgoing_gradient * point_x) ** 2 - 100 UpperCAmelCase : List[str] = ( -linear_term - sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) UpperCAmelCase : Optional[int] = ( -linear_term + sqrt(linear_term**2 - 4 * quadratic_term * constant_term ) ) / (2 * quadratic_term) # two solutions, one of which is our input point UpperCAmelCase : Optional[Any] = x_minus if isclose(_lowerCAmelCase , _lowerCAmelCase ) else x_plus UpperCAmelCase : Union[str, Any] = point_y + outgoing_gradient * (next_x - point_x) return next_x, next_y, outgoing_gradient def snake_case_ ( _lowerCAmelCase : float = 1.4 , _lowerCAmelCase : float = -9.6 ) -> int: UpperCAmelCase : int = 0 UpperCAmelCase : float = first_x_coord UpperCAmelCase : float = first_y_coord UpperCAmelCase : float = (1_0.1 - point_y) / (0.0 - point_x) while not (-0.0_1 <= point_x <= 0.0_1 and point_y > 0): UpperCAmelCase , UpperCAmelCase , UpperCAmelCase : str = next_point(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) num_reflections += 1 return num_reflections if __name__ == "__main__": print(F"{solution() = }")
23
0
import unittest from transformers import is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class UpperCamelCase__ : '''simple docstring''' @staticmethod def _lowercase ( *UpperCamelCase__ , **UpperCamelCase__ ) -> Union[str, Any]: pass @is_pipeline_test @require_vision class UpperCamelCase__ (unittest.TestCase ): '''simple docstring''' @require_torch def _lowercase ( self ) -> Union[str, Any]: lowerCamelCase : Tuple = pipeline( model="hf-internal-testing/tiny-random-clip-zero-shot-image-classification" , ) lowerCamelCase : Tuple = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowerCamelCase : List[Any] = image_classifier(UpperCamelCase__ , candidate_labels=["a", "b", "c"] ) # The floating scores are so close, we enter floating error approximation and the order is not guaranteed across # python and torch versions. self.assertIn( nested_simplify(UpperCamelCase__ ) , [ [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "b"}, {"score": 0.333, "label": "c"}], [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "c"}, {"score": 0.333, "label": "b"}], ] , ) lowerCamelCase : List[Any] = image_classifier([image] * 5 , candidate_labels=["A", "B", "C"] , batch_size=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ) , [ [ {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, ], [ {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, ], [ {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, ], [ {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, ], [ {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, ], ] , ) @require_tf def _lowercase ( self ) -> int: lowerCamelCase : str = pipeline( model="hf-internal-testing/tiny-random-clip-zero-shot-image-classification" , framework="tf" ) lowerCamelCase : int = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowerCamelCase : Dict = image_classifier(UpperCamelCase__ , candidate_labels=["a", "b", "c"] ) self.assertEqual( nested_simplify(UpperCamelCase__ ) , [{"score": 0.333, "label": "a"}, {"score": 0.333, "label": "b"}, {"score": 0.333, "label": "c"}] , ) lowerCamelCase : str = image_classifier([image] * 5 , candidate_labels=["A", "B", "C"] , batch_size=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ) , [ [ {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, ], [ {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, ], [ {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, ], [ {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, ], [ {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, {"score": 0.333, "label": ANY(UpperCamelCase__ )}, ], ] , ) @slow @require_torch def _lowercase ( self ) -> Optional[Any]: lowerCamelCase : int = pipeline( task="zero-shot-image-classification" , model="openai/clip-vit-base-patch32" , ) # This is an image of 2 cats with remotes and no planes lowerCamelCase : Union[str, Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowerCamelCase : Optional[int] = image_classifier(UpperCamelCase__ , candidate_labels=["cat", "plane", "remote"] ) self.assertEqual( nested_simplify(UpperCamelCase__ ) , [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ] , ) lowerCamelCase : Tuple = image_classifier([image] * 5 , candidate_labels=["cat", "plane", "remote"] , batch_size=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ) , [ [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ] * 5 , ) @slow @require_tf def _lowercase ( self ) -> Tuple: lowerCamelCase : str = pipeline( task="zero-shot-image-classification" , model="openai/clip-vit-base-patch32" , framework="tf" ) # This is an image of 2 cats with remotes and no planes lowerCamelCase : Any = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowerCamelCase : Tuple = image_classifier(UpperCamelCase__ , candidate_labels=["cat", "plane", "remote"] ) self.assertEqual( nested_simplify(UpperCamelCase__ ) , [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ] , ) lowerCamelCase : int = image_classifier([image] * 5 , candidate_labels=["cat", "plane", "remote"] , batch_size=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ) , [ [ {"score": 0.511, "label": "remote"}, {"score": 0.485, "label": "cat"}, {"score": 0.004, "label": "plane"}, ], ] * 5 , )
48
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase__: str = { "configuration_lxmert": ["LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "LxmertConfig"], "tokenization_lxmert": ["LxmertTokenizer"], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: int = ["LxmertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: Union[str, Any] = [ "LxmertEncoder", "LxmertForPreTraining", "LxmertForQuestionAnswering", "LxmertModel", "LxmertPreTrainedModel", "LxmertVisualFeatureEncoder", "LxmertXLayer", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__: int = [ "TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFLxmertForPreTraining", "TFLxmertMainLayer", "TFLxmertModel", "TFLxmertPreTrainedModel", "TFLxmertVisualFeatureEncoder", ] if TYPE_CHECKING: from .configuration_lxmert import LXMERT_PRETRAINED_CONFIG_ARCHIVE_MAP, LxmertConfig from .tokenization_lxmert import LxmertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_lxmert_fast import LxmertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_lxmert import ( LxmertEncoder, LxmertForPreTraining, LxmertForQuestionAnswering, LxmertModel, LxmertPreTrainedModel, LxmertVisualFeatureEncoder, LxmertXLayer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_lxmert import ( TF_LXMERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFLxmertForPreTraining, TFLxmertMainLayer, TFLxmertModel, TFLxmertPreTrainedModel, TFLxmertVisualFeatureEncoder, ) else: import sys UpperCamelCase__: Tuple = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
23
0
import numpy as np import torch from torch.utils.data import Dataset from utils import logger class _A ( __UpperCAmelCase ): def __init__( self : Dict , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : List[str]): '''simple docstring''' __a = params __a = np.array(__SCREAMING_SNAKE_CASE) __a = np.array([len(__SCREAMING_SNAKE_CASE) for t in data]) self.check() self.remove_long_sequences() self.remove_empty_sequences() self.remove_unknown_sequences() self.check() self.print_statistics() def __getitem__( self : List[Any] , __SCREAMING_SNAKE_CASE : Optional[int]): '''simple docstring''' return (self.token_ids[index], self.lengths[index]) def __len__( self : Union[str, Any]): '''simple docstring''' return len(self.lengths) def _lowerCamelCase ( self : str): '''simple docstring''' assert len(self.token_ids) == len(self.lengths) assert all(self.lengths[i] == len(self.token_ids[i]) for i in range(len(self.lengths))) def _lowerCamelCase ( self : Tuple): '''simple docstring''' __a = self.params.max_model_input_size __a = self.lengths > max_len logger.info(F'Splitting {sum(__SCREAMING_SNAKE_CASE)} too long sequences.') def divide_chunks(__SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : Any): return [l[i : i + n] for i in range(0 , len(__SCREAMING_SNAKE_CASE) , __SCREAMING_SNAKE_CASE)] __a = [] __a = [] if self.params.mlm: __a , __a = self.params.special_tok_ids['''cls_token'''], self.params.special_tok_ids['''sep_token'''] else: __a , __a = self.params.special_tok_ids['''bos_token'''], self.params.special_tok_ids['''eos_token'''] for seq_, len_ in zip(self.token_ids , self.lengths): assert (seq_[0] == cls_id) and (seq_[-1] == sep_id), seq_ if len_ <= max_len: new_tok_ids.append(seq_) new_lengths.append(len_) else: __a = [] for sub_s in divide_chunks(seq_ , max_len - 2): if sub_s[0] != cls_id: __a = np.insert(__SCREAMING_SNAKE_CASE , 0 , __SCREAMING_SNAKE_CASE) if sub_s[-1] != sep_id: __a = np.insert(__SCREAMING_SNAKE_CASE , len(__SCREAMING_SNAKE_CASE) , __SCREAMING_SNAKE_CASE) assert len(__SCREAMING_SNAKE_CASE) <= max_len assert (sub_s[0] == cls_id) and (sub_s[-1] == sep_id), sub_s sub_seqs.append(__SCREAMING_SNAKE_CASE) new_tok_ids.extend(__SCREAMING_SNAKE_CASE) new_lengths.extend([len(__SCREAMING_SNAKE_CASE) for l in sub_seqs]) __a = np.array(__SCREAMING_SNAKE_CASE) __a = np.array(__SCREAMING_SNAKE_CASE) def _lowerCamelCase ( self : Dict): '''simple docstring''' __a = len(self) __a = self.lengths > 11 __a = self.token_ids[indices] __a = self.lengths[indices] __a = len(self) logger.info(F'Remove {init_size - new_size} too short (<=11 tokens) sequences.') def _lowerCamelCase ( self : int): '''simple docstring''' if "unk_token" not in self.params.special_tok_ids: return else: __a = self.params.special_tok_ids['''unk_token'''] __a = len(self) __a = np.array([np.count_nonzero(a == unk_token_id) for a in self.token_ids]) __a = (unk_occs / self.lengths) < 0.5 __a = self.token_ids[indices] __a = self.lengths[indices] __a = len(self) logger.info(F'Remove {init_size - new_size} sequences with a high level of unknown tokens (50%).') def _lowerCamelCase ( self : Any): '''simple docstring''' if not self.params.is_master: return logger.info(F'{len(self)} sequences') # data_len = sum(self.lengths) # nb_unique_tokens = len(Counter(list(chain(*self.token_ids)))) # logger.info(f'{data_len} tokens ({nb_unique_tokens} unique)') # unk_idx = self.params.special_tok_ids['unk_token'] # nb_unknown = sum([(t==unk_idx).sum() for t in self.token_ids]) # logger.info(f'{nb_unknown} unknown tokens (covering {100*nb_unknown/data_len:.2f}% of the data)') def _lowerCamelCase ( self : List[Any] , __SCREAMING_SNAKE_CASE : List[Any]): '''simple docstring''' __a = [t[0] for t in batch] __a = [t[1] for t in batch] assert len(__SCREAMING_SNAKE_CASE) == len(__SCREAMING_SNAKE_CASE) # Max for paddings __a = max(__SCREAMING_SNAKE_CASE) # Pad token ids if self.params.mlm: __a = self.params.special_tok_ids['''pad_token'''] else: __a = self.params.special_tok_ids['''unk_token'''] __a = [list(t.astype(__SCREAMING_SNAKE_CASE)) + [pad_idx] * (max_seq_len_ - len(__SCREAMING_SNAKE_CASE)) for t in token_ids] assert len(tk_) == len(__SCREAMING_SNAKE_CASE) assert all(len(__SCREAMING_SNAKE_CASE) == max_seq_len_ for t in tk_) __a = torch.tensor(tk_) # (bs, max_seq_len_) __a = torch.tensor(__SCREAMING_SNAKE_CASE) # (bs) return tk_t, lg_t
49
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class SCREAMING_SNAKE_CASE( A__ , A__ , A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = AltDiffusionPipeline lowerCamelCase__ = TEXT_TO_IMAGE_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_BATCH_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS lowerCamelCase__ = TEXT_TO_IMAGE_IMAGE_PARAMS def A ( self : Dict ) -> int: torch.manual_seed(0 ) UpperCAmelCase : str = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) UpperCAmelCase : Dict = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule='''scaled_linear''' , clip_sample=__snake_case , set_alpha_to_one=__snake_case , ) torch.manual_seed(0 ) UpperCAmelCase : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) UpperCAmelCase : Dict = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5002 , ) UpperCAmelCase : List[Any] = CLIPTextModel(__snake_case ) UpperCAmelCase : Optional[Any] = XLMRobertaTokenizer.from_pretrained('''hf-internal-testing/tiny-xlm-roberta''' ) UpperCAmelCase : Optional[int] = 77 UpperCAmelCase : Optional[int] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def A ( self : Optional[Any] , __snake_case : Dict , __snake_case : List[str]=0 ) -> Union[str, Any]: if str(__snake_case ).startswith('''mps''' ): UpperCAmelCase : str = torch.manual_seed(__snake_case ) else: UpperCAmelCase : Tuple = torch.Generator(device=__snake_case ).manual_seed(__snake_case ) UpperCAmelCase : Dict = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def A ( self : Union[str, Any] ) -> List[str]: super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def A ( self : Tuple ) -> List[str]: super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def A ( self : Tuple ) -> Optional[int]: UpperCAmelCase : List[str] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase : Any = self.get_dummy_components() torch.manual_seed(0 ) UpperCAmelCase : Optional[int] = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase : List[str] = RobertaSeriesModelWithTransformation(__snake_case ) UpperCAmelCase : str = text_encoder UpperCAmelCase : Optional[int] = AltDiffusionPipeline(**__snake_case ) UpperCAmelCase : str = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : Optional[int] = self.get_dummy_inputs(__snake_case ) UpperCAmelCase : Optional[int] = '''A photo of an astronaut''' UpperCAmelCase : List[Any] = alt_pipe(**__snake_case ) UpperCAmelCase : Optional[Any] = output.images UpperCAmelCase : str = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : List[str] = np.array( [0.5_74_81_62, 0.60_44_71_45, 0.48_82_12_17, 0.50_10_06_36, 0.5_43_11_85, 0.45_76_36_83, 0.49_65_76_96, 0.48_13_27_33, 0.47_57_30_93] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A ( self : int ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase : int = self.get_dummy_components() UpperCAmelCase : int = PNDMScheduler(skip_prk_steps=__snake_case ) torch.manual_seed(0 ) UpperCAmelCase : int = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase : Union[str, Any] = RobertaSeriesModelWithTransformation(__snake_case ) UpperCAmelCase : Union[str, Any] = text_encoder UpperCAmelCase : Optional[int] = AltDiffusionPipeline(**__snake_case ) UpperCAmelCase : Dict = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : int = self.get_dummy_inputs(__snake_case ) UpperCAmelCase : Optional[int] = alt_pipe(**__snake_case ) UpperCAmelCase : Optional[int] = output.images UpperCAmelCase : int = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase : Optional[int] = np.array( [0.51_60_50_93, 0.5_70_72_41, 0.47_36_55_07, 0.50_57_88_86, 0.5_63_38_77, 0.4_64_25_03, 0.5_18_20_81, 0.48_76_34_84, 0.49_08_42_37] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : str ) -> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def A ( self : List[Any] ) -> Any: # make sure here that pndm scheduler skips prk UpperCAmelCase : List[Any] = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''' , safety_checker=__snake_case ) UpperCAmelCase : Tuple = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : List[Any] = '''A painting of a squirrel eating a burger''' UpperCAmelCase : Any = torch.manual_seed(0 ) UpperCAmelCase : Optional[int] = alt_pipe([prompt] , generator=__snake_case , guidance_scale=6.0 , num_inference_steps=20 , output_type='''np''' ) UpperCAmelCase : Dict = output.images UpperCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase : List[str] = np.array([0.10_10, 0.08_00, 0.07_94, 0.08_85, 0.08_43, 0.07_62, 0.07_69, 0.07_29, 0.05_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def A ( self : Tuple ) -> int: UpperCAmelCase : int = DDIMScheduler.from_pretrained('''BAAI/AltDiffusion''' , subfolder='''scheduler''' ) UpperCAmelCase : Tuple = AltDiffusionPipeline.from_pretrained('''BAAI/AltDiffusion''' , scheduler=__snake_case , safety_checker=__snake_case ) UpperCAmelCase : Dict = alt_pipe.to(__snake_case ) alt_pipe.set_progress_bar_config(disable=__snake_case ) UpperCAmelCase : Tuple = '''A painting of a squirrel eating a burger''' UpperCAmelCase : Optional[int] = torch.manual_seed(0 ) UpperCAmelCase : List[Any] = alt_pipe([prompt] , generator=__snake_case , num_inference_steps=2 , output_type='''numpy''' ) UpperCAmelCase : Dict = output.images UpperCAmelCase : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase : Union[str, Any] = np.array([0.40_19, 0.40_52, 0.38_10, 0.41_19, 0.39_16, 0.39_82, 0.46_51, 0.41_95, 0.53_23] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
23
0
from __future__ import annotations def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> tuple[int, int]: if b == 0: return (1, 0) ((lowerCamelCase__) , (lowerCamelCase__)) : Tuple = extended_euclid(_UpperCAmelCase , a % b ) lowerCamelCase__ : str = a // b return (y, x - k * y) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> int: ((lowerCamelCase__) , (lowerCamelCase__)) : List[Any] = extended_euclid(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : Union[str, Any] = na * na lowerCamelCase__ : Optional[Any] = ra * x * na + ra * y * na return (n % m + m) % m def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> int: ((lowerCamelCase__) , (lowerCamelCase__)) : List[Any] = extended_euclid(_UpperCAmelCase , _UpperCAmelCase ) if b < 0: lowerCamelCase__ : List[str] = (b % n + n) % n return b def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> int: lowerCamelCase__ , lowerCamelCase__ : Tuple = invert_modulo(_UpperCAmelCase , _UpperCAmelCase ), invert_modulo(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : str = na * na lowerCamelCase__ : str = ra * x * na + ra * y * na return (n % m + m) % m if __name__ == "__main__": from doctest import testmod testmod(name="""chinese_remainder_theorem""", verbose=True) testmod(name="""chinese_remainder_theorem2""", verbose=True) testmod(name="""invert_modulo""", verbose=True) testmod(name="""extended_euclid""", verbose=True)
50
'''simple docstring''' import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] ) -> Optional[int]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[int] ) -> Dict: UpperCAmelCase : Dict = tmp_path / '''cache''' UpperCAmelCase : List[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase : str = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict , _lowerCAmelCase : int ) -> Optional[int]: UpperCAmelCase : Any = tmp_path / '''cache''' UpperCAmelCase : Tuple = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Any = features.copy() if features else default_expected_features UpperCAmelCase : List[Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Dict = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Tuple ) -> Tuple: UpperCAmelCase : Optional[Any] = tmp_path / '''cache''' UpperCAmelCase : Optional[int] = {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''} UpperCAmelCase : int = features.copy() if features else default_expected_features UpperCAmelCase : Any = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict ) -> Union[str, Any]: # jsonl_312_path features are {"col_3": "float64", "col_1": "string", "col_2": "int64"} UpperCAmelCase : Tuple = {'''col_2''': '''int64''', '''col_3''': '''float64''', '''col_1''': '''string'''} UpperCAmelCase : List[str] = features.copy() UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = tmp_path / '''cache''' UpperCAmelCase : List[str] = JsonDatasetReader(_lowerCAmelCase , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ) -> Optional[Any]: UpperCAmelCase : Any = tmp_path / '''cache''' UpperCAmelCase : List[Any] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : List[Any] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase , split=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('''path_type''' , [str, list] ) def snake_case_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : str , _lowerCAmelCase : Any ) -> Dict: if issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase : str = jsonl_path elif issubclass(_lowerCAmelCase , _lowerCAmelCase ): UpperCAmelCase : Dict = [jsonl_path] UpperCAmelCase : int = tmp_path / '''cache''' UpperCAmelCase : Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[int] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_dataset(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict , _lowerCAmelCase : Dict=("train",) ) -> Union[str, Any]: assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) for split in splits: UpperCAmelCase : List[str] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ) -> Any: UpperCAmelCase : Optional[Any] = tmp_path / '''cache''' UpperCAmelCase : List[str] = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase : Optional[int] = JsonDatasetReader({'''train''': jsonl_path} , cache_dir=_lowerCAmelCase , keep_in_memory=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Union[str, Any] ) -> int: UpperCAmelCase : Dict = tmp_path / '''cache''' UpperCAmelCase : Dict = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[int] = features.copy() if features else default_expected_features UpperCAmelCase : Union[str, Any] = ( Features({feature: Value(_lowerCAmelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase : Tuple = JsonDatasetReader({'''train''': jsonl_path} , features=_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase ) @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Dict ) -> Union[str, Any]: if split: UpperCAmelCase : Optional[int] = {split: jsonl_path} else: UpperCAmelCase : Any = '''train''' UpperCAmelCase : Any = {'''train''': jsonl_path, '''test''': jsonl_path} UpperCAmelCase : Tuple = tmp_path / '''cache''' UpperCAmelCase : int = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} UpperCAmelCase : Optional[Any] = JsonDatasetReader(_lowerCAmelCase , cache_dir=_lowerCAmelCase ).read() _check_json_datasetdict(_lowerCAmelCase , _lowerCAmelCase , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def snake_case_ ( _lowerCAmelCase : List[str] ) -> str: return json.load(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Dict ) -> str: return [json.loads(_lowerCAmelCase ) for line in buffer] class SCREAMING_SNAKE_CASE: """simple docstring""" @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def A ( self : Union[str, Any] , __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : Optional[int] ) -> Dict: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case ).write() buffer.seek(0 ) UpperCAmelCase : Union[str, Any] = load_json_function(__snake_case ) assert isinstance(__snake_case , __snake_case ) assert isinstance(exported_content[0] , __snake_case ) assert len(__snake_case ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def A ( self : Optional[int] , __snake_case : Optional[Any] , __snake_case : str , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : Optional[Any] ) -> List[Any]: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , orient=__snake_case ).write() buffer.seek(0 ) UpperCAmelCase : Union[str, Any] = load_json(__snake_case ) assert isinstance(__snake_case , __snake_case ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(__snake_case , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(__snake_case ) == 10 @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def A ( self : str , __snake_case : str , __snake_case : str , __snake_case : int ) -> Any: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase : Any = load_json_function(__snake_case ) assert isinstance(__snake_case , __snake_case ) assert isinstance(exported_content[0] , __snake_case ) assert len(__snake_case ) == 10 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def A ( self : Any , __snake_case : int , __snake_case : Union[str, Any] , __snake_case : Optional[int] , __snake_case : Optional[Any] , __snake_case : List[str] ) -> Any: with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , lines=__snake_case , orient=__snake_case , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase : List[str] = load_json(__snake_case ) assert isinstance(__snake_case , __snake_case ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(__snake_case , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(__snake_case ) == 10 def A ( self : List[Any] , __snake_case : str ) -> Dict: with pytest.raises(__snake_case ): with io.BytesIO() as buffer: JsonDatasetWriter(__snake_case , __snake_case , num_proc=0 ) @pytest.mark.parametrize('''compression, extension''' , [('''gzip''', '''gz'''), ('''bz2''', '''bz2'''), ('''xz''', '''xz''')] ) def A ( self : Optional[int] , __snake_case : Any , __snake_case : str , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Dict ) -> Union[str, Any]: UpperCAmelCase : List[str] = tmp_path_factory.mktemp('''data''' ) / F"""test.json.{extension}""" UpperCAmelCase : List[Any] = str(shared_datadir / F"""test_file.json.{extension}""" ) JsonDatasetWriter(__snake_case , __snake_case , compression=__snake_case ).write() with fsspec.open(__snake_case , '''rb''' , compression='''infer''' ) as f: UpperCAmelCase : str = f.read() with fsspec.open(__snake_case , '''rb''' , compression='''infer''' ) as f: UpperCAmelCase : Optional[int] = f.read() assert exported_content == original_content
23
0
def A (__A : bytes ) -> str: """simple docstring""" return "".join([hex(__A )[2:].zfill(2 ).upper() for byte in list(__A )] ) def A (__A : str ) -> bytes: """simple docstring""" if (len(__A ) % 2) != 0: raise ValueError( '''Base16 encoded data is invalid: Data does not have an even number of hex digits.''' ) # Check the character set - the standard base16 alphabet # is uppercase according to RFC3548 section 6 if not set(__A ) <= set('''0123456789ABCDEF''' ): raise ValueError( '''Base16 encoded data is invalid: Data is not uppercase hex or it contains invalid characters.''' ) # For every two hexadecimal digits (= a byte), turn it into an integer. # Then, string the result together into bytes, and return it. return bytes(int(data[i] + data[i + 1] , 16 ) for i in range(0 , len(__A ) , 2 ) ) if __name__ == "__main__": import doctest doctest.testmod()
51
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCamelCase__: Tuple = logging.get_logger(__name__) UpperCamelCase__: Optional[int] = {"vocab_file": "sentencepiece.bpe.model"} UpperCamelCase__: Optional[int] = { "vocab_file": { "moussaKam/mbarthez": "https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez": "https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model", "moussaKam/barthez-orangesum-title": ( "https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model" ), }, } UpperCamelCase__: Dict = { "moussaKam/mbarthez": 1024, "moussaKam/barthez": 1024, "moussaKam/barthez-orangesum-title": 1024, } UpperCamelCase__: Tuple = "▁" class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" lowerCamelCase__ = VOCAB_FILES_NAMES lowerCamelCase__ = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ = ["""input_ids""", """attention_mask"""] def __init__( self : List[Any] , __snake_case : List[Any] , __snake_case : Tuple="<s>" , __snake_case : List[Any]="</s>" , __snake_case : int="</s>" , __snake_case : Any="<s>" , __snake_case : Optional[int]="<unk>" , __snake_case : Union[str, Any]="<pad>" , __snake_case : Union[str, Any]="<mask>" , __snake_case : Optional[Dict[str, Any]] = None , **__snake_case : Dict , ) -> None: # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase : int = AddedToken(__snake_case , lstrip=__snake_case , rstrip=__snake_case ) if isinstance(__snake_case , __snake_case ) else mask_token UpperCAmelCase : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__snake_case , eos_token=__snake_case , unk_token=__snake_case , sep_token=__snake_case , cls_token=__snake_case , pad_token=__snake_case , mask_token=__snake_case , sp_model_kwargs=self.sp_model_kwargs , **__snake_case , ) UpperCAmelCase : Optional[int] = vocab_file UpperCAmelCase : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__snake_case ) ) UpperCAmelCase : int = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} UpperCAmelCase : Optional[Any] = len(self.sp_model ) - 1 UpperCAmelCase : Optional[int] = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def A ( self : Tuple , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase : Union[str, Any] = [self.cls_token_id] UpperCAmelCase : Optional[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def A ( self : int , __snake_case : List[int] , __snake_case : Optional[List[int]] = None , __snake_case : bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__snake_case , token_ids_a=__snake_case , already_has_special_tokens=__snake_case ) if token_ids_a is None: return [1] + ([0] * len(__snake_case )) + [1] return [1] + ([0] * len(__snake_case )) + [1, 1] + ([0] * len(__snake_case )) + [1] def A ( self : Optional[int] , __snake_case : List[int] , __snake_case : Optional[List[int]] = None ) -> List[int]: UpperCAmelCase : Tuple = [self.sep_token_id] UpperCAmelCase : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def A ( self : Dict ) -> Optional[int]: return len(self.sp_model ) def A ( self : List[str] ) -> Dict: UpperCAmelCase : Optional[Any] = {self.convert_ids_to_tokens(__snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def A ( self : Optional[Any] , __snake_case : str ) -> List[str]: return self.sp_model.encode(__snake_case , out_type=__snake_case ) def A ( self : int , __snake_case : int ) -> int: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCAmelCase : Optional[Any] = self.sp_model.PieceToId(__snake_case ) return spm_id if spm_id else self.unk_token_id def A ( self : int , __snake_case : Any ) -> List[Any]: if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(__snake_case ) def A ( self : List[Any] , __snake_case : Union[str, Any] ) -> List[str]: UpperCAmelCase : Optional[Any] = [] UpperCAmelCase : int = '''''' UpperCAmelCase : Union[str, Any] = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__snake_case ) + token UpperCAmelCase : str = True UpperCAmelCase : List[str] = [] else: current_sub_tokens.append(__snake_case ) UpperCAmelCase : Optional[int] = False out_string += self.sp_model.decode(__snake_case ) return out_string.strip() def __getstate__( self : Union[str, Any] ) -> Union[str, Any]: UpperCAmelCase : Optional[Any] = self.__dict__.copy() UpperCAmelCase : Any = None return state def __setstate__( self : Optional[int] , __snake_case : Union[str, Any] ) -> List[Any]: UpperCAmelCase : Dict = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): UpperCAmelCase : Optional[Any] = {} UpperCAmelCase : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def A ( self : Optional[int] , __snake_case : str , __snake_case : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__snake_case ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase : Union[str, Any] = os.path.join( __snake_case , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __snake_case ) elif not os.path.isfile(self.vocab_file ): with open(__snake_case , '''wb''' ) as fi: UpperCAmelCase : Any = self.sp_model.serialized_model_proto() fi.write(__snake_case ) return (out_vocab_file,)
23
0
import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def A_ ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> List[Any]: UpperCamelCase : Dict = ("dense.weight", "attention.self.query", "attention.self.key", "attention.self.value") UpperCamelCase : Any = ( ("layer.", "layer_"), ("word_embeddings.weight", "word_embeddings"), ("position_embeddings.weight", "position_embeddings"), ("token_type_embeddings.weight", "token_type_embeddings"), (".", "/"), ("LayerNorm/weight", "LayerNorm/gamma"), ("LayerNorm/bias", "LayerNorm/beta"), ("weight", "kernel"), ) if not os.path.isdir(_lowerCAmelCase ): os.makedirs(_lowerCAmelCase ) UpperCamelCase : str = model.state_dict() def to_tf_var_name(_lowerCAmelCase ): for patt, repl in iter(_lowerCAmelCase ): UpperCamelCase : str = name.replace(_lowerCAmelCase , _lowerCAmelCase ) return F"""bert/{name}""" def create_tf_var(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): UpperCamelCase : str = tf.dtypes.as_dtype(tensor.dtype ) UpperCamelCase : str = tf.get_variable(dtype=_lowerCAmelCase , shape=tensor.shape , name=_lowerCAmelCase , initializer=tf.zeros_initializer() ) session.run(tf.variables_initializer([tf_var] ) ) session.run(_lowerCAmelCase ) return tf_var tf.reset_default_graph() with tf.Session() as session: for var_name in state_dict: UpperCamelCase : str = to_tf_var_name(_lowerCAmelCase ) UpperCamelCase : Union[str, Any] = state_dict[var_name].numpy() if any(x in var_name for x in tensors_to_transpose ): UpperCamelCase : Any = torch_tensor.T UpperCamelCase : Any = create_tf_var(tensor=_lowerCAmelCase , name=_lowerCAmelCase , session=_lowerCAmelCase ) tf.keras.backend.set_value(_lowerCAmelCase , _lowerCAmelCase ) UpperCamelCase : List[str] = session.run(_lowerCAmelCase ) print(F"""Successfully created {tf_name}: {np.allclose(_lowerCAmelCase , _lowerCAmelCase )}""" ) UpperCamelCase : Optional[Any] = tf.train.Saver(tf.trainable_variables() ) saver.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , model_name.replace("-" , "_" ) + ".ckpt" ) ) def A_ ( _lowerCAmelCase=None ) -> Dict: UpperCamelCase : str = argparse.ArgumentParser() parser.add_argument("--model_name" , type=_lowerCAmelCase , required=_lowerCAmelCase , help="model name e.g. bert-base-uncased" ) parser.add_argument( "--cache_dir" , type=_lowerCAmelCase , default=_lowerCAmelCase , required=_lowerCAmelCase , help="Directory containing pytorch model" ) parser.add_argument("--pytorch_model_path" , type=_lowerCAmelCase , required=_lowerCAmelCase , help="/path/to/<pytorch-model-name>.bin" ) parser.add_argument("--tf_cache_dir" , type=_lowerCAmelCase , required=_lowerCAmelCase , help="Directory in which to save tensorflow model" ) UpperCamelCase : List[Any] = parser.parse_args(_lowerCAmelCase ) UpperCamelCase : Dict = BertModel.from_pretrained( pretrained_model_name_or_path=args.model_name , state_dict=torch.load(args.pytorch_model_path ) , cache_dir=args.cache_dir , ) convert_pytorch_checkpoint_to_tf(model=_lowerCAmelCase , ckpt_dir=args.tf_cache_dir , model_name=args.model_name ) if __name__ == "__main__": main()
52
'''simple docstring''' import tempfile import unittest import numpy as np from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import BertConfig, is_flax_available from transformers.testing_utils import TOKEN, USER, is_staging_test, require_flax if is_flax_available(): import os from flax.core.frozen_dict import unfreeze from flax.traverse_util import flatten_dict from transformers import FlaxBertModel UpperCamelCase__: Tuple = "0.12" # assumed parallelism: 8 @require_flax @is_staging_test class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" @classmethod def A ( cls : Union[str, Any] ) -> int: UpperCAmelCase : Optional[Any] = TOKEN HfFolder.save_token(__snake_case ) @classmethod def A ( cls : List[str] ) -> Tuple: try: delete_repo(token=cls._token , repo_id='''test-model-flax''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-model-flax-org''' ) except HTTPError: pass def A ( self : int ) -> Tuple: UpperCAmelCase : List[Any] = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) model.push_to_hub('''test-model-flax''' , use_auth_token=self._token ) UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) UpperCAmelCase : List[Any] = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Tuple = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Union[str, Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''test-model-flax''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(__snake_case , repo_id='''test-model-flax''' , push_to_hub=__snake_case , use_auth_token=self._token ) UpperCAmelCase : str = FlaxBertModel.from_pretrained(F"""{USER}/test-model-flax""" ) UpperCAmelCase : Any = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : str = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Optional[Any] = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) def A ( self : Optional[Any] ) -> Union[str, Any]: UpperCAmelCase : Dict = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) UpperCAmelCase : Optional[Any] = FlaxBertModel(__snake_case ) model.push_to_hub('''valid_org/test-model-flax-org''' , use_auth_token=self._token ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) UpperCAmelCase : List[Any] = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : int = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : Any = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-model-flax-org''' ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained( __snake_case , repo_id='''valid_org/test-model-flax-org''' , push_to_hub=__snake_case , use_auth_token=self._token ) UpperCAmelCase : str = FlaxBertModel.from_pretrained('''valid_org/test-model-flax-org''' ) UpperCAmelCase : Any = flatten_dict(unfreeze(model.params ) ) UpperCAmelCase : Optional[Any] = flatten_dict(unfreeze(new_model.params ) ) for key in base_params.keys(): UpperCAmelCase : int = (base_params[key] - new_params[key]).sum().item() self.assertLessEqual(__snake_case , 1E-3 , msg=F"""{key} not identical""" ) def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Any ) -> Union[str, Any]: UpperCAmelCase : str = True UpperCAmelCase : int = flatten_dict(modela.params ) UpperCAmelCase : Dict = flatten_dict(modela.params ) for key in flat_params_a.keys(): if np.sum(np.abs(flat_params_a[key] - flat_params_a[key] ) ) > 1e-4: UpperCAmelCase : Dict = False return models_are_equal @require_flax class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : Tuple ) -> Union[str, Any]: UpperCAmelCase : List[Any] = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) UpperCAmelCase : int = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(__snake_case , __snake_case ) ) with self.assertRaises(__snake_case ): UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : str = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertTrue(check_models_equal(__snake_case , __snake_case ) ) def A ( self : List[str] ) -> Dict: UpperCAmelCase : Dict = BertConfig.from_pretrained('''hf-internal-testing/tiny-bert-flax-only''' ) UpperCAmelCase : Dict = FlaxBertModel(__snake_case ) UpperCAmelCase : Optional[int] = '''bert''' with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(os.path.join(__snake_case , __snake_case ) , max_shard_size='''10KB''' ) with self.assertRaises(__snake_case ): UpperCAmelCase : Any = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertTrue(check_models_equal(__snake_case , __snake_case ) ) def A ( self : Optional[int] ) -> str: UpperCAmelCase : Dict = '''bert''' UpperCAmelCase : int = '''hf-internal-testing/tiny-random-bert-subfolder''' with self.assertRaises(__snake_case ): UpperCAmelCase : Optional[Any] = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Tuple = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertIsNotNone(__snake_case ) def A ( self : Dict ) -> List[Any]: UpperCAmelCase : Optional[int] = '''bert''' UpperCAmelCase : int = '''hf-internal-testing/tiny-random-bert-sharded-subfolder''' with self.assertRaises(__snake_case ): UpperCAmelCase : Dict = FlaxBertModel.from_pretrained(__snake_case ) UpperCAmelCase : Union[str, Any] = FlaxBertModel.from_pretrained(__snake_case , subfolder=__snake_case ) self.assertIsNotNone(__snake_case )
23
0
'''simple docstring''' import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def lowercase__ ( __lowercase : str ) -> int: """simple docstring""" __UpperCamelCase , __UpperCamelCase = image.size __UpperCamelCase , __UpperCamelCase = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 __UpperCamelCase = image.resize((w, h) , resample=PIL_INTERPOLATION['lanczos'] ) __UpperCamelCase = np.array(__lowercase ).astype(np.floataa ) / 2_5_5.0 __UpperCamelCase = image[None].transpose(0 , 3 , 1 , 2 ) __UpperCamelCase = torch.from_numpy(__lowercase ) return 2.0 * image - 1.0 class snake_case ( __lowerCamelCase ): """simple docstring""" def __init__( self : str , __A : VQModel , __A : UNetaDModel , __A : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): super().__init__() self.register_modules(vqvae=__A , unet=__A , scheduler=__A ) @torch.no_grad() def __call__( self : Any , __A : Union[torch.Tensor, PIL.Image.Image] = None , __A : Optional[int] = 1 , __A : Optional[int] = 1_0_0 , __A : Optional[float] = 0.0 , __A : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __A : Optional[str] = "pil" , __A : bool = True , ): if isinstance(__A , PIL.Image.Image ): __UpperCamelCase = 1 elif isinstance(__A , torch.Tensor ): __UpperCamelCase = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(__A )}''' ) if isinstance(__A , PIL.Image.Image ): __UpperCamelCase = preprocess(__A ) __UpperCamelCase , __UpperCamelCase = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __UpperCamelCase = (batch_size, self.unet.config.in_channels // 2, height, width) __UpperCamelCase = next(self.unet.parameters() ).dtype __UpperCamelCase = randn_tensor(__A , generator=__A , device=self.device , dtype=__A ) __UpperCamelCase = image.to(device=self.device , dtype=__A ) # set timesteps and move to the correct device self.scheduler.set_timesteps(__A , device=self.device ) __UpperCamelCase = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __UpperCamelCase = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __UpperCamelCase = 'eta' in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __UpperCamelCase = {} if accepts_eta: __UpperCamelCase = eta for t in self.progress_bar(__A ): # concat latents and low resolution image in the channel dimension. __UpperCamelCase = torch.cat([latents, image] , dim=1 ) __UpperCamelCase = self.scheduler.scale_model_input(__A , __A ) # predict the noise residual __UpperCamelCase = self.unet(__A , __A ).sample # compute the previous noisy sample x_t -> x_t-1 __UpperCamelCase = self.scheduler.step(__A , __A , __A , **__A ).prev_sample # decode the image latents with the VQVAE __UpperCamelCase = self.vqvae.decode(__A ).sample __UpperCamelCase = torch.clamp(__A , -1.0 , 1.0 ) __UpperCamelCase = image / 2 + 0.5 __UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __UpperCamelCase = self.numpy_to_pil(__A ) if not return_dict: return (image,) return ImagePipelineOutput(images=__A )
53
'''simple docstring''' import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE: """simple docstring""" def __init__( self : Optional[int] , __snake_case : str , __snake_case : Union[str, Any]=2 , __snake_case : Optional[int]=8 , __snake_case : Any=True , __snake_case : Union[str, Any]=True , __snake_case : Dict=True , __snake_case : int=True , __snake_case : List[Any]=99 , __snake_case : str=16 , __snake_case : Tuple=5 , __snake_case : Tuple=2 , __snake_case : str=36 , __snake_case : Dict="gelu" , __snake_case : str=0.0 , __snake_case : Optional[int]=0.0 , __snake_case : Optional[int]=512 , __snake_case : Optional[Any]=16 , __snake_case : int=2 , __snake_case : int=0.02 , __snake_case : str=3 , __snake_case : Dict=4 , __snake_case : str=None , ) -> Optional[int]: UpperCAmelCase : Optional[int] = parent UpperCAmelCase : Tuple = batch_size UpperCAmelCase : List[str] = seq_length UpperCAmelCase : List[Any] = is_training UpperCAmelCase : int = use_input_mask UpperCAmelCase : Any = use_token_type_ids UpperCAmelCase : str = use_labels UpperCAmelCase : Union[str, Any] = vocab_size UpperCAmelCase : List[str] = hidden_size UpperCAmelCase : Optional[Any] = num_hidden_layers UpperCAmelCase : Union[str, Any] = num_attention_heads UpperCAmelCase : Optional[Any] = intermediate_size UpperCAmelCase : Union[str, Any] = hidden_act UpperCAmelCase : int = hidden_dropout_prob UpperCAmelCase : Optional[int] = attention_probs_dropout_prob UpperCAmelCase : Union[str, Any] = max_position_embeddings UpperCAmelCase : str = type_vocab_size UpperCAmelCase : List[str] = type_sequence_label_size UpperCAmelCase : Tuple = initializer_range UpperCAmelCase : Optional[Any] = num_labels UpperCAmelCase : Optional[int] = num_choices UpperCAmelCase : Any = scope def A ( self : Tuple ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase : Optional[int] = None if self.use_input_mask: UpperCAmelCase : Dict = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase : Dict = None if self.use_token_type_ids: UpperCAmelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase : str = None UpperCAmelCase : Tuple = None UpperCAmelCase : int = None if self.use_labels: UpperCAmelCase : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase : str = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase : List[str] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : int ) -> Tuple: return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__snake_case , initializer_range=self.initializer_range , ) def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Optional[Any] = self.get_config() UpperCAmelCase : int = 300 return config def A ( self : Optional[Any] ) -> Any: ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) : Tuple = self.prepare_config_and_inputs() UpperCAmelCase : Dict = True UpperCAmelCase : Tuple = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) UpperCAmelCase : str = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def A ( self : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : int , __snake_case : Optional[int] , __snake_case : int , __snake_case : Dict , __snake_case : Tuple , __snake_case : Optional[Any] ) -> List[str]: UpperCAmelCase : int = MraModel(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Tuple = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case ) UpperCAmelCase : Optional[int] = model(__snake_case , token_type_ids=__snake_case ) UpperCAmelCase : Dict = model(__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Tuple , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : List[str] , __snake_case : int , __snake_case : Union[str, Any] , __snake_case : Optional[Any] , __snake_case : Any , __snake_case : List[Any] , __snake_case : Optional[Any] , ) -> Tuple: UpperCAmelCase : str = True UpperCAmelCase : Tuple = MraModel(__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Optional[int] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , encoder_hidden_states=__snake_case , encoder_attention_mask=__snake_case , ) UpperCAmelCase : Optional[Any] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , encoder_hidden_states=__snake_case , ) UpperCAmelCase : str = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Tuple , __snake_case : str , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : Tuple , __snake_case : Optional[Any] , __snake_case : List[str] , __snake_case : int ) -> Any: UpperCAmelCase : Dict = MraForMaskedLM(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : Optional[int] = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : Tuple , __snake_case : Tuple , __snake_case : Dict , __snake_case : Dict , __snake_case : Any , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Tuple ) -> Optional[int]: UpperCAmelCase : List[str] = MraForQuestionAnswering(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : List[Any] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , start_positions=__snake_case , end_positions=__snake_case , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : str , __snake_case : Optional[int] , __snake_case : List[str] , __snake_case : str , __snake_case : int , __snake_case : Optional[Any] , __snake_case : Union[str, Any] , __snake_case : List[Any] ) -> int: UpperCAmelCase : int = self.num_labels UpperCAmelCase : Union[str, Any] = MraForSequenceClassification(__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : List[str] = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : str , __snake_case : Dict , __snake_case : Optional[Any] , __snake_case : Dict , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Union[str, Any] , __snake_case : Dict ) -> int: UpperCAmelCase : Tuple = self.num_labels UpperCAmelCase : List[str] = MraForTokenClassification(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : str = model(__snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : str , __snake_case : int , __snake_case : Any , __snake_case : Tuple , __snake_case : Optional[int] , __snake_case : List[str] , __snake_case : str , __snake_case : Union[str, Any] ) -> Optional[Any]: UpperCAmelCase : Tuple = self.num_choices UpperCAmelCase : int = MraForMultipleChoice(config=__snake_case ) model.to(__snake_case ) model.eval() UpperCAmelCase : str = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[str] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() UpperCAmelCase : List[str] = model( __snake_case , attention_mask=__snake_case , token_type_ids=__snake_case , labels=__snake_case , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : str ) -> Dict: UpperCAmelCase : Any = self.prepare_config_and_inputs() ( ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ( UpperCAmelCase ) , ) : List[str] = config_and_inputs UpperCAmelCase : Any = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE( A__ , unittest.TestCase ): """simple docstring""" lowerCamelCase__ = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = False lowerCamelCase__ = () def A ( self : int ) -> Union[str, Any]: UpperCAmelCase : List[str] = MraModelTester(self ) UpperCAmelCase : Optional[int] = ConfigTester(self , config_class=__snake_case , hidden_size=37 ) def A ( self : Optional[Any] ) -> str: self.config_tester.run_common_tests() def A ( self : Tuple ) -> Optional[Any]: UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__snake_case ) def A ( self : List[Any] ) -> Optional[Any]: UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase : List[Any] = type self.model_tester.create_and_check_model(*__snake_case ) def A ( self : Tuple ) -> Dict: UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__snake_case ) def A ( self : Tuple ) -> List[str]: UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__snake_case ) def A ( self : int ) -> Dict: UpperCAmelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__snake_case ) def A ( self : Dict ) -> Optional[int]: UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__snake_case ) def A ( self : Any ) -> Optional[int]: UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__snake_case ) @slow def A ( self : Dict ) -> Any: for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase : str = MraModel.from_pretrained(__snake_case ) self.assertIsNotNone(__snake_case ) @unittest.skip(reason='''MRA does not output attentions''' ) def A ( self : str ) -> Optional[Any]: return @require_torch class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" @slow def A ( self : Tuple ) -> List[Any]: UpperCAmelCase : int = MraModel.from_pretrained('''uw-madison/mra-base-512-4''' ) UpperCAmelCase : Optional[Any] = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : List[Any] = model(__snake_case )[0] UpperCAmelCase : Optional[Any] = torch.Size((1, 256, 768) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Any = torch.tensor( [[[-0.01_40, 0.08_30, -0.03_81], [0.15_46, 0.14_02, 0.02_20], [0.11_62, 0.08_51, 0.01_65]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) ) @slow def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Optional[int] = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-512-4''' ) UpperCAmelCase : Dict = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : List[Any] = model(__snake_case )[0] UpperCAmelCase : int = 50265 UpperCAmelCase : int = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Union[str, Any] = torch.tensor( [[[9.25_95, -3.60_38, 11.88_19], [9.38_69, -3.26_93, 11.09_56], [11.85_24, -3.49_38, 13.12_10]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) ) @slow def A ( self : str ) -> List[Any]: UpperCAmelCase : List[Any] = MraForMaskedLM.from_pretrained('''uw-madison/mra-base-4096-8-d3''' ) UpperCAmelCase : List[Any] = torch.arange(4096 ).unsqueeze(0 ) with torch.no_grad(): UpperCAmelCase : Tuple = model(__snake_case )[0] UpperCAmelCase : Optional[int] = 50265 UpperCAmelCase : Tuple = torch.Size((1, 4096, vocab_size) ) self.assertEqual(output.shape , __snake_case ) UpperCAmelCase : Optional[int] = torch.tensor( [[[5.47_89, -2.35_64, 7.50_64], [7.90_67, -1.33_69, 9.96_68], [9.07_12, -1.81_06, 7.03_80]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __snake_case , atol=1E-4 ) )
23
0
"""simple docstring""" from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available a__ : str = {'''configuration_van''': ['''VAN_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''VanConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Tuple = [ '''VAN_PRETRAINED_MODEL_ARCHIVE_LIST''', '''VanForImageClassification''', '''VanModel''', '''VanPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_van import VAN_PRETRAINED_CONFIG_ARCHIVE_MAP, VanConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_van import ( VAN_PRETRAINED_MODEL_ARCHIVE_LIST, VanForImageClassification, VanModel, VanPreTrainedModel, ) else: import sys a__ : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
54
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class SCREAMING_SNAKE_CASE( unittest.TestCase ): """simple docstring""" def A ( self : Any ) -> str: UpperCAmelCase : Any = { '''task_specific_params''': { '''summarization''': {'''length_penalty''': 1.0, '''max_length''': 128, '''min_length''': 12, '''num_beams''': 4}, '''summarization_cnn''': {'''length_penalty''': 2.0, '''max_length''': 142, '''min_length''': 56, '''num_beams''': 4}, '''summarization_xsum''': {'''length_penalty''': 1.0, '''max_length''': 62, '''min_length''': 11, '''num_beams''': 6}, } } UpperCAmelCase : int = { '''task_specific_params.summarization.length_penalty''': 1.0, '''task_specific_params.summarization.max_length''': 128, '''task_specific_params.summarization.min_length''': 12, '''task_specific_params.summarization.num_beams''': 4, '''task_specific_params.summarization_cnn.length_penalty''': 2.0, '''task_specific_params.summarization_cnn.max_length''': 142, '''task_specific_params.summarization_cnn.min_length''': 56, '''task_specific_params.summarization_cnn.num_beams''': 4, '''task_specific_params.summarization_xsum.length_penalty''': 1.0, '''task_specific_params.summarization_xsum.max_length''': 62, '''task_specific_params.summarization_xsum.min_length''': 11, '''task_specific_params.summarization_xsum.num_beams''': 6, } self.assertEqual(flatten_dict(__snake_case ) , __snake_case ) def A ( self : int ) -> str: UpperCAmelCase : Any = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(__snake_case ) , x.transpose() ) ) UpperCAmelCase : str = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def A ( self : str ) -> Union[str, Any]: UpperCAmelCase : Any = np.random.randn(3 , 4 ) UpperCAmelCase : List[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , transpose(__snake_case ).numpy() ) ) UpperCAmelCase : Tuple = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Any = torch.tensor(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , transpose(__snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_tf def A ( self : List[str] ) -> Optional[Any]: UpperCAmelCase : int = np.random.randn(3 , 4 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , transpose(__snake_case ).numpy() ) ) UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : str = tf.constant(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , transpose(__snake_case , axes=(1, 2, 0) ).numpy() ) ) @require_flax def A ( self : Tuple ) -> Any: UpperCAmelCase : List[Any] = np.random.randn(3 , 4 ) UpperCAmelCase : List[str] = jnp.array(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case ) , np.asarray(transpose(__snake_case ) ) ) ) UpperCAmelCase : Dict = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(transpose(__snake_case , axes=(1, 2, 0) ) , np.asarray(transpose(__snake_case , axes=(1, 2, 0) ) ) ) ) def A ( self : Optional[Any] ) -> Any: UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , np.reshape(__snake_case , (4, 3) ) ) ) UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , np.reshape(__snake_case , (12, 5) ) ) ) @require_torch def A ( self : Union[str, Any] ) -> int: UpperCAmelCase : Dict = np.random.randn(3 , 4 ) UpperCAmelCase : Optional[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , reshape(__snake_case , (4, 3) ).numpy() ) ) UpperCAmelCase : List[str] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : List[Any] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , reshape(__snake_case , (12, 5) ).numpy() ) ) @require_tf def A ( self : int ) -> List[str]: UpperCAmelCase : List[Any] = np.random.randn(3 , 4 ) UpperCAmelCase : List[str] = tf.constant(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , reshape(__snake_case , (4, 3) ).numpy() ) ) UpperCAmelCase : List[Any] = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Optional[Any] = tf.constant(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , reshape(__snake_case , (12, 5) ).numpy() ) ) @require_flax def A ( self : Any ) -> Dict: UpperCAmelCase : Tuple = np.random.randn(3 , 4 ) UpperCAmelCase : Union[str, Any] = jnp.array(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (4, 3) ) , np.asarray(reshape(__snake_case , (4, 3) ) ) ) ) UpperCAmelCase : Any = np.random.randn(3 , 4 , 5 ) UpperCAmelCase : Optional[Any] = jnp.array(__snake_case ) self.assertTrue(np.allclose(reshape(__snake_case , (12, 5) ) , np.asarray(reshape(__snake_case , (12, 5) ) ) ) ) def A ( self : List[Any] ) -> List[Any]: UpperCAmelCase : Union[str, Any] = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(__snake_case ) , np.squeeze(__snake_case ) ) ) UpperCAmelCase : str = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , np.squeeze(__snake_case , axis=2 ) ) ) @require_torch def A ( self : Optional[int] ) -> Union[str, Any]: UpperCAmelCase : Optional[int] = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : List[str] = torch.tensor(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , squeeze(__snake_case ).numpy() ) ) UpperCAmelCase : Any = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : str = torch.tensor(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , squeeze(__snake_case , axis=2 ).numpy() ) ) @require_tf def A ( self : Optional[Any] ) -> Dict: UpperCAmelCase : int = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , squeeze(__snake_case ).numpy() ) ) UpperCAmelCase : List[str] = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : Optional[int] = tf.constant(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , squeeze(__snake_case , axis=2 ).numpy() ) ) @require_flax def A ( self : List[Any] ) -> Dict: UpperCAmelCase : Optional[int] = np.random.randn(1 , 3 , 4 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case ) , np.asarray(squeeze(__snake_case ) ) ) ) UpperCAmelCase : str = np.random.randn(1 , 4 , 1 , 5 ) UpperCAmelCase : int = jnp.array(__snake_case ) self.assertTrue(np.allclose(squeeze(__snake_case , axis=2 ) , np.asarray(squeeze(__snake_case , axis=2 ) ) ) ) def A ( self : Optional[Any] ) -> int: UpperCAmelCase : Optional[Any] = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , np.expand_dims(__snake_case , axis=1 ) ) ) @require_torch def A ( self : List[str] ) -> Tuple: UpperCAmelCase : Tuple = np.random.randn(3 , 4 ) UpperCAmelCase : Tuple = torch.tensor(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , expand_dims(__snake_case , axis=1 ).numpy() ) ) @require_tf def A ( self : List[str] ) -> Union[str, Any]: UpperCAmelCase : Union[str, Any] = np.random.randn(3 , 4 ) UpperCAmelCase : Any = tf.constant(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , expand_dims(__snake_case , axis=1 ).numpy() ) ) @require_flax def A ( self : Any ) -> List[Any]: UpperCAmelCase : List[str] = np.random.randn(3 , 4 ) UpperCAmelCase : str = jnp.array(__snake_case ) self.assertTrue(np.allclose(expand_dims(__snake_case , axis=1 ) , np.asarray(expand_dims(__snake_case , axis=1 ) ) ) )
23
0
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging a_ : Union[str, Any] = logging.get_logger(__name__) a_ : Optional[Any] = { """microsoft/git-base""": """https://huggingface.co/microsoft/git-base/resolve/main/config.json""", } class snake_case ( lowercase ): """simple docstring""" _lowerCamelCase = "git_vision_model" def __init__( self , UpperCamelCase=768 , UpperCamelCase=3072 , UpperCamelCase=12 , UpperCamelCase=12 , UpperCamelCase=3 , UpperCamelCase=224 , UpperCamelCase=16 , UpperCamelCase="quick_gelu" , UpperCamelCase=1e-5 , UpperCamelCase=0.0 , UpperCamelCase=0.02 , **UpperCamelCase , ): """simple docstring""" super().__init__(**UpperCamelCase ) lowerCamelCase_ = hidden_size lowerCamelCase_ = intermediate_size lowerCamelCase_ = num_hidden_layers lowerCamelCase_ = num_attention_heads lowerCamelCase_ = num_channels lowerCamelCase_ = patch_size lowerCamelCase_ = image_size lowerCamelCase_ = initializer_range lowerCamelCase_ = attention_dropout lowerCamelCase_ = layer_norm_eps lowerCamelCase_ = hidden_act @classmethod def snake_case ( cls , UpperCamelCase , **UpperCamelCase ): """simple docstring""" cls._set_token_in_kwargs(UpperCamelCase ) lowerCamelCase_ ,lowerCamelCase_ = cls.get_config_dict(UpperCamelCase , **UpperCamelCase ) # get the vision config dict if we are loading from GITConfig if config_dict.get("model_type" ) == "git": lowerCamelCase_ = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' f'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''' ) return cls.from_dict(UpperCamelCase , **UpperCamelCase ) class snake_case ( lowercase ): """simple docstring""" _lowerCamelCase = "git" def __init__( self , UpperCamelCase=None , UpperCamelCase=3_0522 , UpperCamelCase=768 , UpperCamelCase=6 , UpperCamelCase=12 , UpperCamelCase=3072 , UpperCamelCase="gelu" , UpperCamelCase=0.1 , UpperCamelCase=0.1 , UpperCamelCase=1024 , UpperCamelCase=0.02 , UpperCamelCase=1e-12 , UpperCamelCase=0 , UpperCamelCase="absolute" , UpperCamelCase=True , UpperCamelCase=False , UpperCamelCase=101 , UpperCamelCase=102 , UpperCamelCase=None , **UpperCamelCase , ): """simple docstring""" super().__init__(bos_token_id=UpperCamelCase , eos_token_id=UpperCamelCase , pad_token_id=UpperCamelCase , **UpperCamelCase ) if vision_config is None: lowerCamelCase_ = {} logger.info("vision_config is None. initializing the GitVisionConfig with default values." ) lowerCamelCase_ = GitVisionConfig(**UpperCamelCase ) lowerCamelCase_ = vocab_size lowerCamelCase_ = hidden_size lowerCamelCase_ = num_hidden_layers lowerCamelCase_ = num_attention_heads lowerCamelCase_ = hidden_act lowerCamelCase_ = intermediate_size lowerCamelCase_ = hidden_dropout_prob lowerCamelCase_ = attention_probs_dropout_prob lowerCamelCase_ = max_position_embeddings lowerCamelCase_ = initializer_range lowerCamelCase_ = layer_norm_eps lowerCamelCase_ = position_embedding_type lowerCamelCase_ = use_cache lowerCamelCase_ = tie_word_embeddings lowerCamelCase_ = num_image_with_embedding lowerCamelCase_ = bos_token_id lowerCamelCase_ = eos_token_id def snake_case ( self ): """simple docstring""" lowerCamelCase_ = copy.deepcopy(self.__dict__ ) lowerCamelCase_ = self.vision_config.to_dict() lowerCamelCase_ = self.__class__.model_type return output
55
'''simple docstring''' import argparse import os import re import packaging.version UpperCamelCase__: Union[str, Any] = "examples/" UpperCamelCase__: Optional[Any] = { "examples": (re.compile(r"^check_min_version\(\"[^\"]+\"\)\s*$", re.MULTILINE), "check_min_version(\"VERSION\")\n"), "init": (re.compile(r"^__version__\s+=\s+\"([^\"]+)\"\s*$", re.MULTILINE), "__version__ = \"VERSION\"\n"), "setup": (re.compile(r"^(\s*)version\s*=\s*\"[^\"]+\",", re.MULTILINE), r"\1version=\"VERSION\","), "doc": (re.compile(r"^(\s*)release\s*=\s*\"[^\"]+\"$", re.MULTILINE), "release = \"VERSION\"\n"), } UpperCamelCase__: Optional[int] = { "init": "src/diffusers/__init__.py", "setup": "setup.py", } UpperCamelCase__: List[Any] = "README.md" def snake_case_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : List[Any] , _lowerCAmelCase : int ) -> Optional[int]: with open(_lowerCAmelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: UpperCAmelCase : Optional[int] = f.read() UpperCAmelCase , UpperCAmelCase : List[Any] = REPLACE_PATTERNS[pattern] UpperCAmelCase : List[Any] = replace.replace('''VERSION''' , _lowerCAmelCase ) UpperCAmelCase : Optional[Any] = re_pattern.sub(_lowerCAmelCase , _lowerCAmelCase ) with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Any ) -> Optional[int]: for folder, directories, fnames in os.walk(_lowerCAmelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase , pattern='''examples''' ) def snake_case_ ( _lowerCAmelCase : Any , _lowerCAmelCase : str=False ) -> List[str]: for pattern, fname in REPLACE_FILES.items(): update_version_in_file(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if not patch: update_version_in_examples(_lowerCAmelCase ) def snake_case_ ( ) -> Optional[Any]: UpperCAmelCase : Optional[int] = '''🤗 Transformers currently provides the following architectures''' UpperCAmelCase : Optional[int] = '''1. Want to contribute a new model?''' with open(_lowerCAmelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: UpperCAmelCase : Optional[Any] = f.readlines() # Find the start of the list. UpperCAmelCase : List[Any] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 UpperCAmelCase : Optional[Any] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): UpperCAmelCase : Optional[int] = lines[index].replace( '''https://huggingface.co/docs/diffusers/main/model_doc''' , '''https://huggingface.co/docs/diffusers/model_doc''' , ) index += 1 with open(_lowerCAmelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(_lowerCAmelCase ) def snake_case_ ( ) -> Optional[Any]: with open(REPLACE_FILES['''init'''] , '''r''' ) as f: UpperCAmelCase : Union[str, Any] = f.read() UpperCAmelCase : int = REPLACE_PATTERNS['''init'''][0].search(_lowerCAmelCase ).groups()[0] return packaging.version.parse(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : List[str]=False ) -> Any: UpperCAmelCase : Optional[Any] = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: UpperCAmelCase : Optional[int] = default_version.base_version elif patch: UpperCAmelCase : Union[str, Any] = f"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: UpperCAmelCase : Union[str, Any] = f"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. UpperCAmelCase : Dict = input(f"""Which version are you releasing? [{default_version}]""" ) if len(_lowerCAmelCase ) == 0: UpperCAmelCase : Tuple = default_version print(f"""Updating version to {version}.""" ) global_version_update(_lowerCAmelCase , patch=_lowerCAmelCase ) def snake_case_ ( ) -> Any: UpperCAmelCase : List[Any] = get_version() UpperCAmelCase : List[str] = f"""{current_version.major}.{current_version.minor + 1}.0.dev0""" UpperCAmelCase : List[Any] = current_version.base_version # Check with the user we got that right. UpperCAmelCase : Optional[int] = input(f"""Which version are we developing now? [{dev_version}]""" ) if len(_lowerCAmelCase ) == 0: UpperCAmelCase : Dict = dev_version print(f"""Updating version to {version}.""" ) global_version_update(_lowerCAmelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": UpperCamelCase__: Union[str, Any] = argparse.ArgumentParser() parser.add_argument("--post_release", action="store_true", help="Whether this is pre or post release.") parser.add_argument("--patch", action="store_true", help="Whether or not this is a patch release.") UpperCamelCase__: Optional[Any] = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print("Nothing to do after a patch :-)") else: post_release_work()
23
0
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from timm import create_model from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import BitConfig, BitForImageClassification, BitImageProcessor from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() a : List[Any] = logging.get_logger(__name__) def __magic_name__ ( __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' snake_case_ = '''huggingface/label-files''' snake_case_ = '''imagenet-1k-id2label.json''' snake_case_ = json.load(open(hf_hub_download(__UpperCAmelCase, __UpperCAmelCase, repo_type='''dataset''' ), '''r''' ) ) snake_case_ = {int(__UpperCAmelCase ): v for k, v in idalabel.items()} snake_case_ = {v: k for k, v in idalabel.items()} snake_case_ = '''std_conv''' if '''bit''' in model_name else False # note that when using BiT as backbone for ViT-hybrid checkpoints, # one needs to additionally set config.layer_type = "bottleneck", config.stem_type = "same", # config.conv_layer = "std_conv_same" snake_case_ = BitConfig( conv_layer=__UpperCAmelCase, num_labels=1000, idalabel=__UpperCAmelCase, labelaid=__UpperCAmelCase, ) return config def __magic_name__ ( __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' if "stem.conv" in name: snake_case_ = name.replace('''stem.conv''', '''bit.embedder.convolution''' ) if "blocks" in name: snake_case_ = name.replace('''blocks''', '''layers''' ) if "head.fc" in name: snake_case_ = name.replace('''head.fc''', '''classifier.1''' ) if name.startswith('''norm''' ): snake_case_ = '''bit.''' + name if "bit" not in name and "classifier" not in name: snake_case_ = '''bit.encoder.''' + name return name def __magic_name__ ( ) -> Tuple: '''simple docstring''' snake_case_ = '''http://images.cocodataset.org/val2017/000000039769.jpg''' snake_case_ = Image.open(requests.get(__UpperCAmelCase, stream=__UpperCAmelCase ).raw ) return im @torch.no_grad() def __magic_name__ ( __UpperCAmelCase, __UpperCAmelCase, __UpperCAmelCase=False ) -> str: '''simple docstring''' snake_case_ = get_config(__UpperCAmelCase ) # load original model from timm snake_case_ = create_model(__UpperCAmelCase, pretrained=__UpperCAmelCase ) timm_model.eval() # load state_dict of original model snake_case_ = timm_model.state_dict() for key in state_dict.copy().keys(): snake_case_ = state_dict.pop(__UpperCAmelCase ) snake_case_ = val.squeeze() if '''head''' in key else val # load HuggingFace model snake_case_ = BitForImageClassification(__UpperCAmelCase ) model.eval() model.load_state_dict(__UpperCAmelCase ) # create image processor snake_case_ = create_transform(**resolve_data_config({}, model=__UpperCAmelCase ) ) snake_case_ = transform.transforms snake_case_ = { '''bilinear''': PILImageResampling.BILINEAR, '''bicubic''': PILImageResampling.BICUBIC, '''nearest''': PILImageResampling.NEAREST, } snake_case_ = BitImageProcessor( do_resize=__UpperCAmelCase, size={'''shortest_edge''': timm_transforms[0].size}, resample=pillow_resamplings[timm_transforms[0].interpolation.value], do_center_crop=__UpperCAmelCase, crop_size={'''height''': timm_transforms[1].size[0], '''width''': timm_transforms[1].size[1]}, do_normalize=__UpperCAmelCase, image_mean=timm_transforms[-1].mean.tolist(), image_std=timm_transforms[-1].std.tolist(), ) snake_case_ = prepare_img() snake_case_ = transform(__UpperCAmelCase ).unsqueeze(0 ) snake_case_ = processor(__UpperCAmelCase, return_tensors='''pt''' ).pixel_values # verify pixel values assert torch.allclose(__UpperCAmelCase, __UpperCAmelCase ) # verify logits with torch.no_grad(): snake_case_ = model(__UpperCAmelCase ) snake_case_ = outputs.logits print('''Logits:''', logits[0, :3] ) print('''Predicted class:''', model.config.idalabel[logits.argmax(-1 ).item()] ) snake_case_ = timm_model(__UpperCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__UpperCAmelCase, outputs.logits, atol=1e-3 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: Path(__UpperCAmelCase ).mkdir(exist_ok=__UpperCAmelCase ) print(F"Saving model {model_name} and processor to {pytorch_dump_folder_path}" ) model.save_pretrained(__UpperCAmelCase ) processor.save_pretrained(__UpperCAmelCase ) if push_to_hub: print(F"Pushing model {model_name} and processor to the hub" ) model.push_to_hub(F"ybelkada/{model_name}" ) processor.push_to_hub(F"ybelkada/{model_name}" ) if __name__ == "__main__": a : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='resnetv2_50x1_bitm', type=str, help='Name of the BiT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model to the hub.', ) a : Any = parser.parse_args() convert_bit_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
56
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake UpperCamelCase__: Tuple = numpy.array([0, 0]) UpperCamelCase__: Union[str, Any] = numpy.array([0.5, 0.8660254]) UpperCamelCase__: Dict = numpy.array([1, 0]) UpperCamelCase__: int = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] , _lowerCAmelCase : int ) -> list[numpy.ndarray]: UpperCAmelCase : Union[str, Any] = initial_vectors for _ in range(_lowerCAmelCase ): UpperCAmelCase : Union[str, Any] = iteration_step(_lowerCAmelCase ) return vectors def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] ) -> list[numpy.ndarray]: UpperCAmelCase : Tuple = [] for i, start_vector in enumerate(vectors[:-1] ): UpperCAmelCase : List[str] = vectors[i + 1] new_vectors.append(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def snake_case_ ( _lowerCAmelCase : numpy.ndarray , _lowerCAmelCase : float ) -> numpy.ndarray: UpperCAmelCase : List[str] = numpy.radians(_lowerCAmelCase ) UpperCAmelCase , UpperCAmelCase : Tuple = numpy.cos(_lowerCAmelCase ), numpy.sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = numpy.array(((c, -s), (s, c)) ) return numpy.dot(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : list[numpy.ndarray] ) -> None: UpperCAmelCase : List[Any] = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() UpperCAmelCase , UpperCAmelCase : str = zip(*_lowerCAmelCase ) plt.plot(_lowerCAmelCase , _lowerCAmelCase ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__: List[Any] = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
23
0
"""simple docstring""" import datasets from .evaluate import evaluate A : Optional[Any] = "\\n@article{hendrycks2021cuad,\n title={CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review},\n author={Dan Hendrycks and Collin Burns and Anya Chen and Spencer Ball},\n journal={arXiv preprint arXiv:2103.06268},\n year={2021}\n}\n" A : Optional[int] = "\nThis metric wrap the official scoring script for version 1 of the Contract\nUnderstanding Atticus Dataset (CUAD).\nContract Understanding Atticus Dataset (CUAD) v1 is a corpus of more than 13,000 labels in 510\ncommercial legal contracts that have been manually labeled to identify 41 categories of important\nclauses that lawyers look for when reviewing contracts in connection with corporate transactions.\n" A : Optional[int] = "\nComputes CUAD scores (EM, F1, AUPR, Precision@80%Recall, and Precision@90%Recall).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair as given in the references (see below)\n - 'prediction_text': list of possible texts for the answer, as a list of strings\n depending on a threshold on the confidence probability of each prediction.\n references: List of question-answers dictionaries with the following key-values:\n - 'id': id of the question-answer pair (see above),\n - 'answers': a Dict in the CUAD dataset format\n {\n 'text': list of possible texts for the answer, as a list of strings\n 'answer_start': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n 'exact_match': Exact match (the normalized answer exactly match the gold answer)\n 'f1': The F-score of predicted tokens versus the gold answer\n 'aupr': Area Under the Precision-Recall curve\n 'prec_at_80_recall': Precision at 80% recall\n 'prec_at_90_recall': Precision at 90% recall\nExamples:\n >>> predictions = [{'prediction_text': ['The seller:', 'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.'], 'id': 'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties'}]\n >>> references = [{'answers': {'answer_start': [143, 49], 'text': ['The seller:', 'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.']}, 'id': 'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties'}]\n >>> cuad_metric = datasets.load_metric(\"cuad\")\n >>> results = cuad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {'exact_match': 100.0, 'f1': 100.0, 'aupr': 0.0, 'prec_at_80_recall': 1.0, 'prec_at_90_recall': 1.0}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class _UpperCamelCase ( datasets.Metric ): '''simple docstring''' def snake_case ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": { "id": datasets.Value("string" ), "prediction_text": datasets.features.Sequence(datasets.Value("string" ) ), }, "references": { "id": datasets.Value("string" ), "answers": datasets.features.Sequence( { "text": datasets.Value("string" ), "answer_start": datasets.Value("int32" ), } ), }, } ) , codebase_urls=["https://www.atticusprojectai.org/cuad"] , reference_urls=["https://www.atticusprojectai.org/cuad"] , ) def snake_case ( self , __a , __a ): __lowerCAmelCase = {prediction["id"]: prediction["prediction_text"] for prediction in predictions} __lowerCAmelCase = [ { "paragraphs": [ { "qas": [ { "answers": [{"text": answer_text} for answer_text in ref["answers"]["text"]], "id": ref["id"], } for ref in references ] } ] } ] __lowerCAmelCase = evaluate(dataset=__a , predictions=__a ) return score
57
'''simple docstring''' from manim import * class SCREAMING_SNAKE_CASE( A__ ): """simple docstring""" def A ( self : Union[str, Any] ) -> List[str]: UpperCAmelCase : Optional[Any] = Rectangle(height=0.5 , width=0.5 ) UpperCAmelCase : str = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) UpperCAmelCase : Tuple = [mem.copy() for i in range(6 )] UpperCAmelCase : Optional[int] = [mem.copy() for i in range(6 )] UpperCAmelCase : Dict = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Any = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Union[str, Any] = VGroup(__snake_case , __snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Optional[Any] = Text('''CPU''' , font_size=24 ) UpperCAmelCase : Union[str, Any] = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__snake_case ) UpperCAmelCase : Optional[Any] = [mem.copy() for i in range(4 )] UpperCAmelCase : Union[str, Any] = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : List[str] = Text('''GPU''' , font_size=24 ) UpperCAmelCase : Dict = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) gpu.move_to([-1, -1, 0] ) self.add(__snake_case ) UpperCAmelCase : int = [mem.copy() for i in range(6 )] UpperCAmelCase : Union[str, Any] = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : List[str] = Text('''Model''' , font_size=24 ) UpperCAmelCase : Tuple = Group(__snake_case , __snake_case ).arrange(__snake_case , buff=0.5 , aligned_edge=__snake_case ) model.move_to([3, -1.0, 0] ) self.add(__snake_case ) UpperCAmelCase : Any = [] for i, rect in enumerate(__snake_case ): rect.set_stroke(__snake_case ) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) UpperCAmelCase : Dict = Rectangle(height=0.46 / 4 , width=0.46 / 3 ).set_stroke(width=0.0 ).set_fill(__snake_case , opacity=0.7 ) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT ) , buff=0.02 , direction=__snake_case ) cpu_target.set_x(cpu_target.get_x() + 0.1 ) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=__snake_case , buff=0.0 ) else: cpu_target.next_to(cpu_targs[i - 1] , direction=__snake_case , buff=0.0 ) self.add(__snake_case ) cpu_targs.append(__snake_case ) UpperCAmelCase : int = [mem.copy() for i in range(6 )] UpperCAmelCase : int = VGroup(*__snake_case ).arrange(__snake_case , buff=0 ) UpperCAmelCase : Any = Text('''Loaded Checkpoint''' , font_size=24 ) UpperCAmelCase : Union[str, Any] = Group(__snake_case , __snake_case ).arrange(__snake_case , aligned_edge=__snake_case , buff=0.4 ) checkpoint.move_to([3, 0.5, 0] ) UpperCAmelCase : Optional[int] = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCAmelCase : str = MarkupText( F"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=18 , ) key_text.move_to([-5, 2.4, 0] ) self.add(__snake_case , __snake_case ) UpperCAmelCase : Tuple = MarkupText( F"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=18 , ) blue_text.next_to(__snake_case , DOWN * 2.4 , aligned_edge=key_text.get_left() ) UpperCAmelCase : List[Any] = MarkupText( F"""Next, a <i><span fgcolor=\"{BLUE}\">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor=\"{BLUE}\">single shard</span>.""" , font_size=24 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__snake_case ) , Write(__snake_case ) ) self.play(Write(__snake_case , run_time=1 ) , Create(__snake_case , run_time=1 ) ) UpperCAmelCase : Tuple = [] UpperCAmelCase : int = [] for i, rect in enumerate(__snake_case ): UpperCAmelCase : Any = fill.copy().set_fill(__snake_case , opacity=0.7 ) target.move_to(__snake_case ) first_animations.append(GrowFromCenter(__snake_case , run_time=1 ) ) UpperCAmelCase : List[str] = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1] ) else: cpu_target.target.move_to(cpu_right_col_base[i - 5] ) second_animations.append(MoveToTarget(__snake_case , run_time=1.5 ) ) self.play(*__snake_case ) self.play(*__snake_case ) self.wait()
23
0
'''simple docstring''' import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # and perform gradient accumulation # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## lowercase_ = 16 lowercase_ = 32 def lowerCamelCase ( __lowerCamelCase : Accelerator , __lowerCamelCase : int = 16 ) ->Optional[int]: _SCREAMING_SNAKE_CASE = AutoTokenizer.from_pretrained("""bert-base-cased""" ) _SCREAMING_SNAKE_CASE = load_dataset("""glue""" , """mrpc""" ) def tokenize_function(__lowerCamelCase : int ): # max_length=None => use the model max length (it's actually the default) _SCREAMING_SNAKE_CASE = tokenizer(examples["""sentence1"""] , examples["""sentence2"""] , truncation=__lowerCamelCase , max_length=__lowerCamelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): _SCREAMING_SNAKE_CASE = datasets.map( __lowerCamelCase , batched=__lowerCamelCase , remove_columns=["""idx""", """sentence1""", """sentence2"""] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _SCREAMING_SNAKE_CASE = tokenized_datasets.rename_column("""label""" , """labels""" ) def collate_fn(__lowerCamelCase : List[str] ): # On TPU it's best to pad everything to the same length or training will be very slow. _SCREAMING_SNAKE_CASE = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": _SCREAMING_SNAKE_CASE = 16 elif accelerator.mixed_precision != "no": _SCREAMING_SNAKE_CASE = 8 else: _SCREAMING_SNAKE_CASE = None return tokenizer.pad( __lowerCamelCase , padding="""longest""" , max_length=__lowerCamelCase , pad_to_multiple_of=__lowerCamelCase , return_tensors="""pt""" , ) # Instantiate dataloaders. _SCREAMING_SNAKE_CASE = DataLoader( tokenized_datasets["""train"""] , shuffle=__lowerCamelCase , collate_fn=__lowerCamelCase , batch_size=__lowerCamelCase ) _SCREAMING_SNAKE_CASE = DataLoader( tokenized_datasets["""validation"""] , shuffle=__lowerCamelCase , collate_fn=__lowerCamelCase , batch_size=__lowerCamelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", None) == "1": from accelerate.test_utils.training import mocked_dataloaders lowercase_ = mocked_dataloaders # noqa: F811 def lowerCamelCase ( __lowerCamelCase : List[str] , __lowerCamelCase : Optional[int] ) ->str: # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""" , __lowerCamelCase ) == "1": _SCREAMING_SNAKE_CASE = 2 # New Code # _SCREAMING_SNAKE_CASE = int(args.gradient_accumulation_steps ) # Initialize accelerator _SCREAMING_SNAKE_CASE = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , gradient_accumulation_steps=__lowerCamelCase ) if accelerator.distributed_type == DistributedType.TPU and gradient_accumulation_steps > 1: raise NotImplementedError( """Gradient accumulation on TPUs is currently not supported. Pass `gradient_accumulation_steps=1`""" ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _SCREAMING_SNAKE_CASE = config["""lr"""] _SCREAMING_SNAKE_CASE = int(config["""num_epochs"""] ) _SCREAMING_SNAKE_CASE = int(config["""seed"""] ) _SCREAMING_SNAKE_CASE = int(config["""batch_size"""] ) _SCREAMING_SNAKE_CASE = evaluate.load("""glue""" , """mrpc""" ) set_seed(__lowerCamelCase ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = get_dataloaders(__lowerCamelCase , __lowerCamelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _SCREAMING_SNAKE_CASE = AutoModelForSequenceClassification.from_pretrained("""bert-base-cased""" , return_dict=__lowerCamelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). _SCREAMING_SNAKE_CASE = model.to(accelerator.device ) # Instantiate optimizer _SCREAMING_SNAKE_CASE = AdamW(params=model.parameters() , lr=__lowerCamelCase ) # Instantiate scheduler _SCREAMING_SNAKE_CASE = get_linear_schedule_with_warmup( optimizer=__lowerCamelCase , num_warmup_steps=100 , num_training_steps=(len(__lowerCamelCase ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = accelerator.prepare( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # Now we train the model for epoch in range(__lowerCamelCase ): model.train() for step, batch in enumerate(__lowerCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(__lowerCamelCase ): _SCREAMING_SNAKE_CASE = model(**__lowerCamelCase ) _SCREAMING_SNAKE_CASE = output.loss accelerator.backward(__lowerCamelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__lowerCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _SCREAMING_SNAKE_CASE = model(**__lowerCamelCase ) _SCREAMING_SNAKE_CASE = outputs.logits.argmax(dim=-1 ) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = accelerator.gather_for_metrics((predictions, batch["""labels"""]) ) metric.add_batch( predictions=__lowerCamelCase , references=__lowerCamelCase , ) _SCREAMING_SNAKE_CASE = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'epoch {epoch}:' , __lowerCamelCase ) def lowerCamelCase ( ) ->Any: _SCREAMING_SNAKE_CASE = argparse.ArgumentParser(description="""Simple example of training script.""" ) parser.add_argument( """--mixed_precision""" , type=__lowerCamelCase , default=__lowerCamelCase , choices=["""no""", """fp16""", """bf16""", """fp8"""] , help="""Whether to use mixed precision. Choose""" """between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.""" """and an Nvidia Ampere GPU.""" , ) # New Code # parser.add_argument( """--gradient_accumulation_steps""" , type=__lowerCamelCase , default=1 , help="""The number of minibatches to be ran before gradients are accumulated.""" , ) parser.add_argument("""--cpu""" , action="""store_true""" , help="""If passed, will train on the CPU.""" ) _SCREAMING_SNAKE_CASE = parser.parse_args() _SCREAMING_SNAKE_CASE = {"""lr""": 2e-5, """num_epochs""": 3, """seed""": 42, """batch_size""": 16} training_function(__lowerCamelCase , __lowerCamelCase ) if __name__ == "__main__": main()
58
'''simple docstring''' import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( "The converted tokenizer will be the `slow` tokenizer. To use the fast, update your `tokenizers` library and re-run the tokenizer conversion" ) UpperCamelCase__: str = None UpperCamelCase__: int = { "7B": 11008, "13B": 13824, "30B": 17920, "65B": 22016, "70B": 28672, } UpperCamelCase__: List[Any] = { "7B": 1, "7Bf": 1, "13B": 2, "13Bf": 2, "30B": 4, "65B": 8, "70B": 8, "70Bf": 8, } def snake_case_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple=1 , _lowerCAmelCase : List[Any]=256 ) -> Optional[Any]: return multiple_of * ((int(ffn_dim_multiplier * int(8 * n / 3 ) ) + multiple_of - 1) // multiple_of) def snake_case_ ( _lowerCAmelCase : List[str] ) -> str: with open(_lowerCAmelCase , '''r''' ) as f: return json.load(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any ) -> Optional[Any]: with open(_lowerCAmelCase , '''w''' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Any , _lowerCAmelCase : Any=True ) -> List[Any]: os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : int = os.path.join(_lowerCAmelCase , '''tmp''' ) os.makedirs(_lowerCAmelCase , exist_ok=_lowerCAmelCase ) UpperCAmelCase : List[str] = read_json(os.path.join(_lowerCAmelCase , '''params.json''' ) ) UpperCAmelCase : str = NUM_SHARDS[model_size] UpperCAmelCase : Any = params['''n_layers'''] UpperCAmelCase : str = params['''n_heads'''] UpperCAmelCase : Any = n_heads // num_shards UpperCAmelCase : List[str] = params['''dim'''] UpperCAmelCase : Optional[Any] = dim // n_heads UpperCAmelCase : str = 1_0_0_0_0.0 UpperCAmelCase : Optional[int] = 1.0 / (base ** (torch.arange(0 , _lowerCAmelCase , 2 ).float() / dims_per_head)) if "n_kv_heads" in params: UpperCAmelCase : Tuple = params['''n_kv_heads'''] # for GQA / MQA UpperCAmelCase : Optional[int] = n_heads_per_shard // num_key_value_heads UpperCAmelCase : Optional[Any] = dim // num_key_value_heads else: # compatibility with other checkpoints UpperCAmelCase : List[str] = n_heads UpperCAmelCase : Optional[int] = n_heads_per_shard UpperCAmelCase : List[str] = dim # permute for sliced rotary def permute(_lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any]=n_heads , _lowerCAmelCase : int=dim , _lowerCAmelCase : Dict=dim ): return w.view(_lowerCAmelCase , dima // n_heads // 2 , 2 , _lowerCAmelCase ).transpose(1 , 2 ).reshape(_lowerCAmelCase , _lowerCAmelCase ) print(f"""Fetching all parameters from the checkpoint at {input_base_path}.""" ) # Load weights if model_size == "7B": # Not sharded # (The sharded implementation would also work, but this is simpler.) UpperCAmelCase : int = torch.load(os.path.join(_lowerCAmelCase , '''consolidated.00.pth''' ) , map_location='''cpu''' ) else: # Sharded UpperCAmelCase : Optional[Any] = [ torch.load(os.path.join(_lowerCAmelCase , f"""consolidated.{i:02d}.pth""" ) , map_location='''cpu''' ) for i in range(_lowerCAmelCase ) ] UpperCAmelCase : Any = 0 UpperCAmelCase : str = {'''weight_map''': {}} for layer_i in range(_lowerCAmelCase ): UpperCAmelCase : Optional[Any] = f"""pytorch_model-{layer_i + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase : Optional[int] = { f"""model.layers.{layer_i}.self_attn.q_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wq.weight"""] ), f"""model.layers.{layer_i}.self_attn.k_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wk.weight"""] ), f"""model.layers.{layer_i}.self_attn.v_proj.weight""": loaded[f"""layers.{layer_i}.attention.wv.weight"""], f"""model.layers.{layer_i}.self_attn.o_proj.weight""": loaded[f"""layers.{layer_i}.attention.wo.weight"""], f"""model.layers.{layer_i}.mlp.gate_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w1.weight"""], f"""model.layers.{layer_i}.mlp.down_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w2.weight"""], f"""model.layers.{layer_i}.mlp.up_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w3.weight"""], f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[f"""layers.{layer_i}.attention_norm.weight"""], f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[f"""layers.{layer_i}.ffn_norm.weight"""], } else: # Sharded # Note that attention.w{q,k,v,o}, feed_fordward.w[1,2,3], attention_norm.weight and ffn_norm.weight share # the same storage object, saving attention_norm and ffn_norm will save other weights too, which is # redundant as other weights will be stitched from multiple shards. To avoid that, they are cloned. UpperCAmelCase : List[str] = { f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.attention_norm.weight""" ].clone(), f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.ffn_norm.weight""" ].clone(), } UpperCAmelCase : Union[str, Any] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wq.weight"""].view(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase : Optional[Any] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wk.weight"""].view( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , ) UpperCAmelCase : str = torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wv.weight"""].view( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for i in range(_lowerCAmelCase ) ] , dim=0 , ).reshape(_lowerCAmelCase , _lowerCAmelCase ) UpperCAmelCase : Optional[int] = torch.cat( [loaded[i][f"""layers.{layer_i}.attention.wo.weight"""] for i in range(_lowerCAmelCase )] , dim=1 ) UpperCAmelCase : Any = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w1.weight"""] for i in range(_lowerCAmelCase )] , dim=0 ) UpperCAmelCase : str = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w2.weight"""] for i in range(_lowerCAmelCase )] , dim=1 ) UpperCAmelCase : Tuple = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w3.weight"""] for i in range(_lowerCAmelCase )] , dim=0 ) UpperCAmelCase : Any = inv_freq for k, v in state_dict.items(): UpperCAmelCase : List[Any] = filename param_count += v.numel() torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , _lowerCAmelCase ) ) UpperCAmelCase : Optional[int] = f"""pytorch_model-{n_layers + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase : str = { '''model.embed_tokens.weight''': loaded['''tok_embeddings.weight'''], '''model.norm.weight''': loaded['''norm.weight'''], '''lm_head.weight''': loaded['''output.weight'''], } else: UpperCAmelCase : Any = { '''model.norm.weight''': loaded[0]['''norm.weight'''], '''model.embed_tokens.weight''': torch.cat( [loaded[i]['''tok_embeddings.weight'''] for i in range(_lowerCAmelCase )] , dim=1 ), '''lm_head.weight''': torch.cat([loaded[i]['''output.weight'''] for i in range(_lowerCAmelCase )] , dim=0 ), } for k, v in state_dict.items(): UpperCAmelCase : Optional[int] = filename param_count += v.numel() torch.save(_lowerCAmelCase , os.path.join(_lowerCAmelCase , _lowerCAmelCase ) ) # Write configs UpperCAmelCase : Union[str, Any] = {'''total_size''': param_count * 2} write_json(_lowerCAmelCase , os.path.join(_lowerCAmelCase , '''pytorch_model.bin.index.json''' ) ) UpperCAmelCase : int = params['''ffn_dim_multiplier'''] if '''ffn_dim_multiplier''' in params else 1 UpperCAmelCase : Tuple = params['''multiple_of'''] if '''multiple_of''' in params else 256 UpperCAmelCase : Any = LlamaConfig( hidden_size=_lowerCAmelCase , intermediate_size=compute_intermediate_size(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) , num_attention_heads=params['''n_heads'''] , num_hidden_layers=params['''n_layers'''] , rms_norm_eps=params['''norm_eps'''] , num_key_value_heads=_lowerCAmelCase , ) config.save_pretrained(_lowerCAmelCase ) # Make space so we can load the model properly now. del state_dict del loaded gc.collect() print('''Loading the checkpoint in a Llama model.''' ) UpperCAmelCase : Optional[int] = LlamaForCausalLM.from_pretrained(_lowerCAmelCase , torch_dtype=torch.floataa , low_cpu_mem_usage=_lowerCAmelCase ) # Avoid saving this as part of the config. del model.config._name_or_path print('''Saving in the Transformers format.''' ) model.save_pretrained(_lowerCAmelCase , safe_serialization=_lowerCAmelCase ) shutil.rmtree(_lowerCAmelCase ) def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any] ) -> List[str]: # Initialize the tokenizer based on the `spm` model UpperCAmelCase : Dict = LlamaTokenizer if LlamaTokenizerFast is None else LlamaTokenizerFast print(f"""Saving a {tokenizer_class.__name__} to {tokenizer_path}.""" ) UpperCAmelCase : List[Any] = tokenizer_class(_lowerCAmelCase ) tokenizer.save_pretrained(_lowerCAmelCase ) def snake_case_ ( ) -> List[Any]: UpperCAmelCase : int = argparse.ArgumentParser() parser.add_argument( '''--input_dir''' , help='''Location of LLaMA weights, which contains tokenizer.model and model folders''' , ) parser.add_argument( '''--model_size''' , choices=['''7B''', '''7Bf''', '''13B''', '''13Bf''', '''30B''', '''65B''', '''70B''', '''70Bf''', '''tokenizer_only'''] , ) parser.add_argument( '''--output_dir''' , help='''Location to write HF model and tokenizer''' , ) parser.add_argument('''--safe_serialization''' , type=_lowerCAmelCase , help='''Whether or not to save using `safetensors`.''' ) UpperCAmelCase : List[Any] = parser.parse_args() if args.model_size != "tokenizer_only": write_model( model_path=args.output_dir , input_base_path=os.path.join(args.input_dir , args.model_size ) , model_size=args.model_size , safe_serialization=args.safe_serialization , ) UpperCAmelCase : Optional[int] = os.path.join(args.input_dir , '''tokenizer.model''' ) write_tokenizer(args.output_dir , _lowerCAmelCase ) if __name__ == "__main__": main()
23
0
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
59
'''simple docstring''' from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Optional[int] = tau * frequency / samplerate UpperCAmelCase : List[Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = cos(_lowerCAmelCase ) UpperCAmelCase : int = _sin / (2 * q_factor) UpperCAmelCase : Any = (1 - _cos) / 2 UpperCAmelCase : List[Any] = 1 - _cos UpperCAmelCase : Union[str, Any] = 1 + alpha UpperCAmelCase : Any = -2 * _cos UpperCAmelCase : Dict = 1 - alpha UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Any = tau * frequency / samplerate UpperCAmelCase : Tuple = sin(_lowerCAmelCase ) UpperCAmelCase : Tuple = cos(_lowerCAmelCase ) UpperCAmelCase : Dict = _sin / (2 * q_factor) UpperCAmelCase : int = (1 + _cos) / 2 UpperCAmelCase : List[Any] = -1 - _cos UpperCAmelCase : Tuple = 1 + alpha UpperCAmelCase : List[str] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha UpperCAmelCase : Tuple = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : Optional[int] = tau * frequency / samplerate UpperCAmelCase : Optional[int] = sin(_lowerCAmelCase ) UpperCAmelCase : Tuple = cos(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = _sin / (2 * q_factor) UpperCAmelCase : Union[str, Any] = _sin / 2 UpperCAmelCase : Any = 0 UpperCAmelCase : int = -ba UpperCAmelCase : Optional[Any] = 1 + alpha UpperCAmelCase : List[Any] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha UpperCAmelCase : int = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 / sqrt(2 ) ) -> IIRFilter: UpperCAmelCase : List[str] = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : str = cos(_lowerCAmelCase ) UpperCAmelCase : Optional[Any] = _sin / (2 * q_factor) UpperCAmelCase : List[str] = 1 - alpha UpperCAmelCase : Any = -2 * _cos UpperCAmelCase : Optional[int] = 1 + alpha UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([ba, ba, ba] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : Optional[Any] = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Optional[int] = cos(_lowerCAmelCase ) UpperCAmelCase : Dict = _sin / (2 * q_factor) UpperCAmelCase : str = 10 ** (gain_db / 40) UpperCAmelCase : int = 1 + alpha * big_a UpperCAmelCase : Union[str, Any] = -2 * _cos UpperCAmelCase : Optional[Any] = 1 - alpha * big_a UpperCAmelCase : Union[str, Any] = 1 + alpha / big_a UpperCAmelCase : Tuple = -2 * _cos UpperCAmelCase : Any = 1 - alpha / big_a UpperCAmelCase : Optional[Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : Any = tau * frequency / samplerate UpperCAmelCase : Optional[int] = sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = cos(_lowerCAmelCase ) UpperCAmelCase : str = _sin / (2 * q_factor) UpperCAmelCase : List[str] = 10 ** (gain_db / 40) UpperCAmelCase : Optional[int] = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase : int = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase : int = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase : Optional[int] = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase : str = 2 * sqrt(_lowerCAmelCase ) * alpha UpperCAmelCase : Dict = big_a * (pmc + aaa) UpperCAmelCase : Any = 2 * big_a * mpc UpperCAmelCase : Union[str, Any] = big_a * (pmc - aaa) UpperCAmelCase : Optional[int] = ppmc + aaa UpperCAmelCase : Optional[Any] = -2 * pmpc UpperCAmelCase : Optional[Any] = ppmc - aaa UpperCAmelCase : int = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt def snake_case_ ( _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : float , _lowerCAmelCase : float = 1 / sqrt(2 ) , ) -> IIRFilter: UpperCAmelCase : int = tau * frequency / samplerate UpperCAmelCase : Union[str, Any] = sin(_lowerCAmelCase ) UpperCAmelCase : Union[str, Any] = cos(_lowerCAmelCase ) UpperCAmelCase : Any = _sin / (2 * q_factor) UpperCAmelCase : int = 10 ** (gain_db / 40) UpperCAmelCase : List[str] = (big_a + 1) - (big_a - 1) * _cos UpperCAmelCase : Union[str, Any] = (big_a + 1) + (big_a - 1) * _cos UpperCAmelCase : Optional[Any] = (big_a - 1) - (big_a + 1) * _cos UpperCAmelCase : Union[str, Any] = (big_a - 1) + (big_a + 1) * _cos UpperCAmelCase : List[str] = 2 * sqrt(_lowerCAmelCase ) * alpha UpperCAmelCase : Any = big_a * (ppmc + aaa) UpperCAmelCase : str = -2 * big_a * pmpc UpperCAmelCase : List[Any] = big_a * (ppmc - aaa) UpperCAmelCase : Optional[Any] = pmc + aaa UpperCAmelCase : Any = 2 * mpc UpperCAmelCase : str = pmc - aaa UpperCAmelCase : Union[str, Any] = IIRFilter(2 ) filt.set_coefficients([aa, aa, aa] , [ba, ba, ba] ) return filt
23
0
"""simple docstring""" def _snake_case ( ): lowerCAmelCase : str = [] lowerCAmelCase : List[Any] = 1 while len(_snake_case ) < 1E6: constant.append(str(_snake_case ) ) i += 1 lowerCAmelCase : int = ''''''.join(_snake_case ) return ( int(constant[0] ) * int(constant[9] ) * int(constant[99] ) * int(constant[999] ) * int(constant[9999] ) * int(constant[99999] ) * int(constant[999999] ) ) if __name__ == "__main__": print(solution())
60
'''simple docstring''' from __future__ import annotations def snake_case_ ( _lowerCAmelCase : str , _lowerCAmelCase : str ) -> bool: UpperCAmelCase : str = get_failure_array(_lowerCAmelCase ) # 2) Step through text searching for pattern UpperCAmelCase , UpperCAmelCase : Optional[Any] = 0, 0 # index into text, pattern while i < len(_lowerCAmelCase ): if pattern[j] == text[i]: if j == (len(_lowerCAmelCase ) - 1): return True j += 1 # if this is a prefix in our pattern # just go back far enough to continue elif j > 0: UpperCAmelCase : Optional[Any] = failure[j - 1] continue i += 1 return False def snake_case_ ( _lowerCAmelCase : str ) -> list[int]: UpperCAmelCase : Optional[Any] = [0] UpperCAmelCase : str = 0 UpperCAmelCase : List[str] = 1 while j < len(_lowerCAmelCase ): if pattern[i] == pattern[j]: i += 1 elif i > 0: UpperCAmelCase : Union[str, Any] = failure[i - 1] continue j += 1 failure.append(_lowerCAmelCase ) return failure if __name__ == "__main__": # Test 1) UpperCamelCase__: str = "abc1abc12" UpperCamelCase__: str = "alskfjaldsabc1abc1abc12k23adsfabcabc" UpperCamelCase__: Any = "alskfjaldsk23adsfabcabc" assert kmp(pattern, texta) and not kmp(pattern, texta) # Test 2) UpperCamelCase__: Tuple = "ABABX" UpperCamelCase__: Union[str, Any] = "ABABZABABYABABX" assert kmp(pattern, text) # Test 3) UpperCamelCase__: Any = "AAAB" UpperCamelCase__: str = "ABAAAAAB" assert kmp(pattern, text) # Test 4) UpperCamelCase__: int = "abcdabcy" UpperCamelCase__: Any = "abcxabcdabxabcdabcdabcy" assert kmp(pattern, text) # Test 5) UpperCamelCase__: List[str] = "aabaabaaa" assert get_failure_array(pattern) == [0, 1, 0, 1, 2, 3, 4, 5, 2]
23
0