code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
from __future__ import annotations import requests def a(lowercase__ ): '''simple docstring''' snake_case_ = f"""https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty""" return requests.get(lowercase__ ).json() def a(lowercase__ = 10 ): '''simple docstring''' snake_case_ = 'https://hacker-news.firebaseio.com/v0/topstories.json?print=pretty' snake_case_ = requests.get(lowercase__ ).json()[:max_stories] return [get_hackernews_story(lowercase__ ) for story_id in story_ids] def a(lowercase__ = 10 ): '''simple docstring''' snake_case_ = hackernews_top_stories(lowercase__ ) return "\n".join('* [{title}]({url})'.format(**lowercase__ ) for story in stories ) if __name__ == "__main__": print(hackernews_top_stories_as_markdown())
46
from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput A = logging.get_logger(__name__) # pylint: disable=invalid-name class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ): """simple docstring""" @register_to_config def __init__( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None ): """simple docstring""" super().__init__() snake_case_ = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" snake_case_ = torch.zeros(__UpperCamelCase , __UpperCamelCase ) else: snake_case_ = None snake_case_ = torch.nn.Parameter(__UpperCamelCase ) class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" super().__init__() self.register_modules( vqvae=__UpperCamelCase , transformer=__UpperCamelCase , text_encoder=__UpperCamelCase , tokenizer=__UpperCamelCase , scheduler=__UpperCamelCase , learned_classifier_free_sampling_embeddings=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = len(__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else 1 # get prompt text embeddings snake_case_ = self.tokenizer( __UpperCamelCase , padding='max_length' , max_length=self.tokenizer.model_max_length , return_tensors='pt' , ) snake_case_ = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: snake_case_ = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( 'The following part of your input was truncated because CLIP can only handle sequences up to' f""" {self.tokenizer.model_max_length} tokens: {removed_text}""" ) snake_case_ = text_input_ids[:, : self.tokenizer.model_max_length] snake_case_ = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 snake_case_ = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate text embeddings for each generation per prompt snake_case_ = prompt_embeds.repeat_interleave(__UpperCamelCase , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: snake_case_ = self.learned_classifier_free_sampling_embeddings.embeddings snake_case_ = negative_prompt_embeds.unsqueeze(0 ).repeat(__UpperCamelCase , 1 , 1 ) else: snake_case_ = [''] * batch_size snake_case_ = text_input_ids.shape[-1] snake_case_ = self.tokenizer( __UpperCamelCase , padding='max_length' , max_length=__UpperCamelCase , truncation=__UpperCamelCase , return_tensors='pt' , ) snake_case_ = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings snake_case_ = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method snake_case_ = negative_prompt_embeds.shape[1] snake_case_ = negative_prompt_embeds.repeat(1 , __UpperCamelCase , 1 ) snake_case_ = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __UpperCamelCase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes snake_case_ = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self , __UpperCamelCase , __UpperCamelCase = 1_00 , __UpperCamelCase = 5.0 , __UpperCamelCase = 1.0 , __UpperCamelCase = 1 , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = "pil" , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = 1 , ): """simple docstring""" if isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = 1 elif isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = len(__UpperCamelCase ) else: raise ValueError(f"""`prompt` has to be of type `str` or `list` but is {type(__UpperCamelCase )}""" ) snake_case_ = batch_size * num_images_per_prompt snake_case_ = guidance_scale > 1.0 snake_case_ = self._encode_prompt(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__UpperCamelCase , __UpperCamelCase ) or callback_steps <= 0) ): raise ValueError( f"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" f""" {type(__UpperCamelCase )}.""" ) # get the initial completely masked latents unless the user supplied it snake_case_ = (batch_size, self.transformer.num_latent_pixels) if latents is None: snake_case_ = self.transformer.num_vector_embeds - 1 snake_case_ = torch.full(__UpperCamelCase , __UpperCamelCase ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {latents_shape}""" ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( 'Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0,' f""" {self.transformer.num_vector_embeds - 1} (inclusive).""" ) snake_case_ = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(__UpperCamelCase , device=self.device ) snake_case_ = self.scheduler.timesteps.to(self.device ) snake_case_ = latents for i, t in enumerate(self.progress_bar(__UpperCamelCase ) ): # expand the sample if we are doing classifier free guidance snake_case_ = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` snake_case_ = self.transformer(__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , timestep=__UpperCamelCase ).sample if do_classifier_free_guidance: snake_case_ , snake_case_ = model_output.chunk(2 ) snake_case_ = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(__UpperCamelCase , dim=1 , keepdim=__UpperCamelCase ) snake_case_ = self.truncate(__UpperCamelCase , __UpperCamelCase ) # remove `log(0)`'s (`-inf`s) snake_case_ = model_output.clamp(-70 ) # compute the previous noisy sample x_t -> x_t-1 snake_case_ = self.scheduler.step(__UpperCamelCase , timestep=__UpperCamelCase , sample=__UpperCamelCase , generator=__UpperCamelCase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = self.vqvae.config.vq_embed_dim snake_case_ = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) snake_case_ = self.vqvae.quantize.get_codebook_entry(__UpperCamelCase , shape=__UpperCamelCase ) snake_case_ = self.vqvae.decode(__UpperCamelCase , force_not_quantize=__UpperCamelCase ).sample snake_case_ = (image / 2 + 0.5).clamp(0 , 1 ) snake_case_ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case_ = self.numpy_to_pil(__UpperCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ , snake_case_ = torch.sort(__UpperCamelCase , 1 , descending=__UpperCamelCase ) snake_case_ = torch.exp(__UpperCamelCase ) snake_case_ = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out snake_case_ = torch.full_like(keep_mask[:, 0:1, :] , __UpperCamelCase ) snake_case_ = torch.cat((all_true, keep_mask) , dim=1 ) snake_case_ = keep_mask[:, :-1, :] snake_case_ = keep_mask.gather(1 , indices.argsort(1 ) ) snake_case_ = log_p_x_0.clone() snake_case_ = -torch.inf # -inf = log(0) return rv
46
1
import copy import os import cva import numpy as np from matplotlib import pyplot as plt class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self ): """simple docstring""" snake_case_ = '' snake_case_ = '' snake_case_ = [] snake_case_ = 0 snake_case_ = 2_56 snake_case_ = 0 snake_case_ = 0 snake_case_ = 0 snake_case_ = 0 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = cva.imread(__UpperCamelCase , 0 ) snake_case_ = copy.deepcopy(self.img ) snake_case_ , snake_case_ , snake_case_ = plt.hist(self.img.ravel() , 2_56 , [0, 2_56] , label='x' ) snake_case_ = np.sum(__UpperCamelCase ) for i in range(len(__UpperCamelCase ) ): snake_case_ = x[i] / self.k self.sk += prk snake_case_ = (self.L - 1) * self.sk if self.rem != 0: snake_case_ = int(last % last ) snake_case_ = int(last + 1 if self.rem >= 0.5 else last ) self.last_list.append(__UpperCamelCase ) snake_case_ = int(np.ma.count(self.img ) / self.img[1].size ) snake_case_ = self.img[1].size for i in range(self.number_of_cols ): for j in range(self.number_of_rows ): snake_case_ = self.img[j][i] if num != self.last_list[num]: snake_case_ = self.last_list[num] cva.imwrite('output_data/output.jpg' , self.img ) def __lowerCAmelCase ( self ): """simple docstring""" plt.hist(self.img.ravel() , 2_56 , [0, 2_56] ) def __lowerCAmelCase ( self ): """simple docstring""" cva.imshow('Output-Image' , self.img ) cva.imshow('Input-Image' , self.original_image ) cva.waitKey(50_00 ) cva.destroyAllWindows() if __name__ == "__main__": A = os.path.join(os.path.basename(__file__), 'image_data/input.jpg') A = ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
46
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'hidden_sizes' ) ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'neck_hidden_sizes' ) ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'num_attention_heads' ) ) class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=3 , __UpperCamelCase=6_40 , __UpperCamelCase=4 , __UpperCamelCase="silu" , __UpperCamelCase=3 , __UpperCamelCase=32 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.02 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=10 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = last_hidden_size snake_case_ = num_attention_heads snake_case_ = hidden_act snake_case_ = conv_kernel_size snake_case_ = output_stride snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = classifier_dropout_prob snake_case_ = use_labels snake_case_ = is_training snake_case_ = num_labels snake_case_ = initializer_range snake_case_ = scope def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) snake_case_ = self.get_config() return config, pixel_values, labels, pixel_labels def __lowerCAmelCase ( self ): """simple docstring""" return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MobileViTModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MobileViTForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MobileViTForSemanticSegmentation(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __A = ( { """feature-extraction""": MobileViTModel, """image-classification""": MobileViTForImageClassification, """image-segmentation""": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTModelTester(self ) snake_case_ = MobileViTConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='MobileViT does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MobileViT does not support input and output embeddings' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MobileViT does not output attentions' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" def check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.hidden_states snake_case_ = 5 self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. snake_case_ = 2 for i in range(len(__UpperCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = MobileViTModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def a(): '''simple docstring''' snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return MobileViTImageProcessor.from_pretrained('apple/mobilevit-xx-small' ) if is_vision_available() else None @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForImageClassification.from_pretrained('apple/mobilevit-xx-small' ).to(__UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([-1.9364, -1.2327, -0.4653] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = model.to(__UpperCamelCase ) snake_case_ = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) snake_case_ = outputs.logits # verify the logits snake_case_ = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [ [[6.9713, 6.9786, 7.2422], [7.2893, 7.2825, 7.4446], [7.6580, 7.8797, 7.9420]], [[-10.6869, -10.3250, -10.3471], [-10.4228, -9.9868, -9.7132], [-11.0405, -11.0221, -10.7318]], [[-3.3089, -2.8539, -2.6740], [-3.2706, -2.5621, -2.5108], [-3.2534, -2.6615, -2.6651]], ] , device=__UpperCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = model.to(__UpperCamelCase ) snake_case_ = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) snake_case_ = outputs.logits.detach().cpu() snake_case_ = image_processor.post_process_semantic_segmentation(outputs=__UpperCamelCase , target_sizes=[(50, 60)] ) snake_case_ = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , __UpperCamelCase ) snake_case_ = image_processor.post_process_semantic_segmentation(outputs=__UpperCamelCase ) snake_case_ = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , __UpperCamelCase )
46
1
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
46
from ..utils import DummyObject, requires_backends class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] )
46
1
from typing import Any def a(lowercase__ ): '''simple docstring''' if not input_list: return [] snake_case_ = [input_list.count(lowercase__ ) for value in input_list] snake_case_ = max(lowercase__ ) # Gets the maximum count in the input list. # Gets values of modes return sorted({input_list[i] for i, value in enumerate(lowercase__ ) if value == y} ) if __name__ == "__main__": import doctest doctest.testmod()
46
from __future__ import annotations import os import tempfile import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import is_tensorflow_text_available, is_tf_available from transformers.testing_utils import require_tensorflow_text, require_tf, slow from ..test_modeling_tf_common import floats_tensor from .test_framework_agnostic import GenerationIntegrationTestsMixin if is_tf_available(): import tensorflow as tf from transformers import ( AutoTokenizer, TFAutoModelForCausalLM, TFAutoModelForSeqaSeqLM, TFAutoModelForSpeechSeqaSeq, TFAutoModelForVisionaSeq, TFBartForConditionalGeneration, TFLogitsProcessorList, TFMinLengthLogitsProcessor, tf_top_k_top_p_filtering, ) if is_tensorflow_text_available(): import tensorflow_text as text @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = tf.convert_to_tensor( [ [ 8.222_0991, # 3rd highest value; idx. 0 -0.562_0044, 5.2322_9752, 4.038_6393, -6.879_8378, -0.5478_5802, -3.201_2153, 2.9277_7176, 1.8817_1953, 7.3534_1276, # 5th highest value; idx. 9 8.4320_7833, # 2nd highest value; idx. 10 -9.8571_1836, -5.9620_9236, -1.1303_9161, -7.111_5294, -0.836_9633, -5.318_6408, 7.0642_7407, 0.8136_9344, -0.8202_3817, -5.917_9796, 0.5881_3443, -6.9977_8438, 4.7155_1189, -0.1877_1637, 7.4402_0759, # 4th highest value; idx. 25 9.3845_0987, # 1st highest value; idx. 26 2.1266_2941, -9.3256_2038, 2.3565_2522, ], # cummulative prob of 5 highest values <= 0.6 [ 0.5842_5518, 4.5313_9238, -5.5751_0464, -6.2803_0699, -7.1952_9503, -4.0212_2551, 1.3933_7037, -6.0670_7057, 1.5948_0517, -9.64_3119, 0.0390_7799, 0.6723_1762, -8.8820_6726, 6.2711_5922, # 4th highest value; idx. 13 2.2852_0723, 4.8276_7506, 4.3042_1368, 8.827_5313, # 2nd highest value; idx. 17 5.4402_9958, # 5th highest value; idx. 18 -4.473_5794, 7.3857_9536, # 3rd highest value; idx. 20 -2.9105_1663, 2.6194_6077, -2.567_4762, -9.4895_9302, -4.0292_2645, -1.3541_6918, 9.6770_2323, # 1st highest value; idx. 27 -5.8947_8553, 1.8537_0467, ], # cummulative prob of 5 highest values <= 0.6 ] , dtype=tf.floataa , ) snake_case_ = tf.convert_to_tensor( [[0, 0], [0, 9], [0, 10], [0, 25], [0, 26], [1, 13], [1, 17], [1, 18], [1, 20], [1, 27]] , dtype=tf.intaa , ) # expected non filtered idx as noted above snake_case_ = tf.convert_to_tensor( [8.22_2099, 7.353_4126, 8.43_2078, 7.440_2075, 9.3_8451, 6.27_1159, 8.82_7531, 5.440_2995, 7.385_7956, 9.67_7023] , dtype=tf.floataa , ) # expected non filtered values as noted above snake_case_ = tf_top_k_top_p_filtering(__UpperCamelCase , top_k=10 , top_p=0.6 , min_tokens_to_keep=4 ) snake_case_ = output[output != -float('inf' )] snake_case_ = tf.cast( tf.where(tf.not_equal(__UpperCamelCase , tf.constant(-float('inf' ) , dtype=tf.floataa ) ) ) , dtype=tf.intaa , ) tf.debugging.assert_near(__UpperCamelCase , __UpperCamelCase , rtol=1E-12 ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase , __snake_case ): """simple docstring""" if is_tf_available(): __A = { """AutoModelForCausalLM""": TFAutoModelForCausalLM, """AutoModelForSpeechSeq2Seq""": TFAutoModelForSpeechSeqaSeq, """AutoModelForSeq2SeqLM""": TFAutoModelForSeqaSeqLM, """AutoModelForVision2Seq""": TFAutoModelForVisionaSeq, """LogitsProcessorList""": TFLogitsProcessorList, """MinLengthLogitsProcessor""": TFMinLengthLogitsProcessor, """create_tensor_fn""": tf.convert_to_tensor, """floats_tensor""": floats_tensor, """return_tensors""": """tf""", } @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 2 snake_case_ = 2 class SCREAMING_SNAKE_CASE ( tf.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super(__UpperCamelCase , self ).__init__() snake_case_ = model @tf.function( input_signature=( tf.TensorSpec((None, input_length) , tf.intaa , name='input_ids' ), tf.TensorSpec((None, input_length) , tf.intaa , name='attention_mask' ), ) , jit_compile=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.model.generate( input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase , max_new_tokens=__UpperCamelCase , return_dict_in_generate=__UpperCamelCase , ) return {"sequences": outputs["sequences"]} snake_case_ = [[2, 0], [1_02, 1_03]] snake_case_ = [[1, 0], [1, 1]] snake_case_ = DummyModel(model=__UpperCamelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__UpperCamelCase , __UpperCamelCase , signatures={'serving_default': dummy_model.serving} ) snake_case_ = tf.saved_model.load(__UpperCamelCase ).signatures['serving_default'] for batch_size in range(1 , len(__UpperCamelCase ) + 1 ): snake_case_ = { 'input_ids': tf.constant(dummy_input_ids[:batch_size] ), 'attention_mask': tf.constant(dummy_attention_masks[:batch_size] ), } snake_case_ = serving_func(**__UpperCamelCase )['sequences'] snake_case_ = test_model.generate(**__UpperCamelCase , max_new_tokens=__UpperCamelCase ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 1 snake_case_ = 2 class SCREAMING_SNAKE_CASE ( tf.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super(__UpperCamelCase , self ).__init__() snake_case_ = model @tf.function( input_signature=( tf.TensorSpec((batch_size, None) , tf.intaa , name='input_ids' ), tf.TensorSpec((batch_size, None) , tf.intaa , name='attention_mask' ), ) , jit_compile=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.model.generate( input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase , max_new_tokens=__UpperCamelCase , return_dict_in_generate=__UpperCamelCase , ) return {"sequences": outputs["sequences"]} snake_case_ = [[2], [1_02, 1_03]] snake_case_ = [[1], [1, 1]] snake_case_ = DummyModel(model=__UpperCamelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__UpperCamelCase , __UpperCamelCase , signatures={'serving_default': dummy_model.serving} ) snake_case_ = tf.saved_model.load(__UpperCamelCase ).signatures['serving_default'] for input_row in range(len(__UpperCamelCase ) ): snake_case_ = { 'input_ids': tf.constant([dummy_input_ids[input_row]] ), 'attention_mask': tf.constant([dummy_attention_masks[input_row]] ), } snake_case_ = serving_func(**__UpperCamelCase )['sequences'] snake_case_ = test_model.generate(**__UpperCamelCase , max_new_tokens=__UpperCamelCase ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @slow @require_tensorflow_text def __lowerCAmelCase ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: # file needed to load the TF tokenizer hf_hub_download(repo_id='google/flan-t5-small' , filename='spiece.model' , local_dir=__UpperCamelCase ) class SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__() snake_case_ = text.SentencepieceTokenizer( model=tf.io.gfile.GFile(os.path.join(__UpperCamelCase , 'spiece.model' ) , 'rb' ).read() ) snake_case_ = TFAutoModelForSeqaSeqLM.from_pretrained('hf-internal-testing/tiny-random-t5' ) def __lowerCAmelCase ( self , __UpperCamelCase , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" snake_case_ = self.tokenizer.tokenize(__UpperCamelCase ) snake_case_ , snake_case_ = text.pad_model_inputs( __UpperCamelCase , max_seq_length=64 , pad_value=self.model.config.pad_token_id ) snake_case_ = self.model.generate(input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase ) return self.tokenizer.detokenize(__UpperCamelCase ) snake_case_ = CompleteSentenceTransformer() snake_case_ = tf.keras.layers.Input(shape=(1,) , dtype=tf.string , name='inputs' ) snake_case_ = complete_model(__UpperCamelCase ) snake_case_ = tf.keras.Model(__UpperCamelCase , __UpperCamelCase ) keras_model.save(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = { 'do_sample': True, 'num_beams': 1, 'top_p': 0.7, 'top_k': 10, 'temperature': 0.7, } snake_case_ = 14 snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 'Hello, my dog is cute and' snake_case_ = tokenizer(__UpperCamelCase , return_tensors='tf' ) snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 6_38 # forces the generation to happen on CPU, to avoid GPU-related quirks with tf.device(':/CPU:0' ): tf.random.set_seed(0 ) snake_case_ = model.generate(**__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) snake_case_ = [6_38, 1_98] with tf.device(':/CPU:0' ): tf.random.set_seed(0 ) snake_case_ = model.generate(**__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = 'Hugging Face is a technology company based in New York and Paris.' snake_case_ = bart_tokenizer(__UpperCamelCase , return_tensors='tf' ).input_ids snake_case_ = TFBartForConditionalGeneration.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = bart_model.generate(__UpperCamelCase ).numpy() class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase=None , **__UpperCamelCase ): """simple docstring""" return super().call(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = FakeBart.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = bart_model.generate(__UpperCamelCase , foo='bar' ).numpy() self.assertTrue(np.array_equal(__UpperCamelCase , __UpperCamelCase ) ) class SCREAMING_SNAKE_CASE ( bart_model.model.encoder.__class__ ): """simple docstring""" def __lowerCAmelCase ( self , __UpperCamelCase , **__UpperCamelCase ): """simple docstring""" return super().call(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = FakeEncoder(bart_model.config , bart_model.model.shared ) snake_case_ = fake_encoder # Normal generation still works (the output will be different because the encoder weights are different) snake_case_ = bart_model.generate(__UpperCamelCase ).numpy() with self.assertRaises(__UpperCamelCase ): # FakeEncoder.call() accepts **kwargs -> no filtering -> value error due to unexpected input "foo" bart_model.generate(__UpperCamelCase , foo='bar' )
46
1
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm A = logging.get_logger(__name__) @dataclass class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = [ """no_inference""", """no_cuda""", """no_tpu""", """no_speed""", """no_memory""", """no_env_print""", """no_multi_process""", ] def __init__( self , **__UpperCamelCase ): """simple docstring""" for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: snake_case_ = deprecated_arg[3:] setattr(self , __UpperCamelCase , not kwargs.pop(__UpperCamelCase ) ) logger.warning( f"""{deprecated_arg} is depreciated. Please use --no_{positive_arg} or""" f""" {positive_arg}={kwargs[positive_arg]}""" ) snake_case_ = kwargs.pop('torchscript' , self.torchscript ) snake_case_ = kwargs.pop('torch_xla_tpu_print_metrics' , self.torch_xla_tpu_print_metrics ) snake_case_ = kwargs.pop('fp16_opt_level' , self.fpaa_opt_level ) super().__init__(**__UpperCamelCase ) __A = field(default=__snake_case , metadata={"""help""": """Trace the models using torchscript"""} ) __A = field(default=__snake_case , metadata={"""help""": """Print Xla/PyTorch tpu metrics"""} ) __A = field( default="""O1""" , metadata={ """help""": ( """For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3']. """ """See details at https://nvidia.github.io/apex/amp.html""" ) } , ) @cached_property def __lowerCAmelCase ( self ): """simple docstring""" requires_backends(self , ['torch'] ) logger.info('PyTorch: setting up devices' ) if not self.cuda: snake_case_ = torch.device('cpu' ) snake_case_ = 0 elif is_torch_tpu_available(): snake_case_ = xm.xla_device() snake_case_ = 0 else: snake_case_ = torch.device('cuda' if torch.cuda.is_available() else 'cpu' ) snake_case_ = torch.cuda.device_count() return device, n_gpu @property def __lowerCAmelCase ( self ): """simple docstring""" return is_torch_tpu_available() and self.tpu @property def __lowerCAmelCase ( self ): """simple docstring""" requires_backends(self , ['torch'] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def __lowerCAmelCase ( self ): """simple docstring""" requires_backends(self , ['torch'] ) return self._setup_devices[0] @property def __lowerCAmelCase ( self ): """simple docstring""" requires_backends(self , ['torch'] ) return self._setup_devices[1] @property def __lowerCAmelCase ( self ): """simple docstring""" return self.n_gpu > 0
46
import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=2 , __UpperCamelCase=8 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=99 , __UpperCamelCase=16 , __UpperCamelCase=5 , __UpperCamelCase=2 , __UpperCamelCase=36 , __UpperCamelCase="gelu" , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=5_12 , __UpperCamelCase=16 , __UpperCamelCase=2 , __UpperCamelCase=0.02 , __UpperCamelCase=3 , __UpperCamelCase=4 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_input_mask snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_labels snake_case_ = num_choices snake_case_ = scope def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None if self.use_token_type_ids: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowerCAmelCase ( self ): """simple docstring""" return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.get_config() snake_case_ = 3_00 return config def __lowerCAmelCase ( self ): """simple docstring""" ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = self.prepare_config_and_inputs() snake_case_ = True snake_case_ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase ) snake_case_ = model(__UpperCamelCase , token_type_ids=__UpperCamelCase ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" snake_case_ = True snake_case_ = MraModel(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , encoder_attention_mask=__UpperCamelCase , ) snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , ) snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraForMaskedLM(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraForQuestionAnswering(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , start_positions=__UpperCamelCase , end_positions=__UpperCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MraForSequenceClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MraForTokenClassification(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_choices snake_case_ = MraForMultipleChoice(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = config_and_inputs snake_case_ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , unittest.TestCase ): """simple docstring""" __A = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) __A = False __A = False __A = False __A = False __A = () def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , hidden_size=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: snake_case_ = type self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = MraModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) @unittest.skip(reason='MRA does not output attentions' ) def __lowerCAmelCase ( self ): """simple docstring""" return @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraModel.from_pretrained('uw-madison/mra-base-512-4' ) snake_case_ = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = torch.Size((1, 2_56, 7_68) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[-0.0140, 0.0830, -0.0381], [0.1546, 0.1402, 0.0220], [0.1162, 0.0851, 0.0165]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraForMaskedLM.from_pretrained('uw-madison/mra-base-512-4' ) snake_case_ = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = 5_02_65 snake_case_ = torch.Size((1, 2_56, vocab_size) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[9.2595, -3.6038, 11.8819], [9.3869, -3.2693, 11.0956], [11.8524, -3.4938, 13.1210]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraForMaskedLM.from_pretrained('uw-madison/mra-base-4096-8-d3' ) snake_case_ = torch.arange(40_96 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = 5_02_65 snake_case_ = torch.Size((1, 40_96, vocab_size) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[5.4789, -2.3564, 7.5064], [7.9067, -1.3369, 9.9668], [9.0712, -1.8106, 7.0380]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) )
46
1
from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def a(lowercase__ ): '''simple docstring''' # A local function to see if a dot lands in the circle. def is_in_circle(lowercase__ , lowercase__ ) -> bool: snake_case_ = sqrt((x**2) + (y**2) ) # Our circle has a radius of 1, so a distance # greater than 1 would land outside the circle. return distance_from_centre <= 1 # The proportion of guesses that landed in the circle snake_case_ = mean( int(is_in_circle(uniform(-1.0 , 1.0 ) , uniform(-1.0 , 1.0 ) ) ) for _ in range(lowercase__ ) ) # The ratio of the area for circle to square is pi/4. snake_case_ = proportion * 4 print(f"""The estimated value of pi is {pi_estimate}""" ) print(f"""The numpy value of pi is {pi}""" ) print(f"""The total error is {abs(pi - pi_estimate )}""" ) def a(lowercase__ , lowercase__ , lowercase__ = 0.0 , lowercase__ = 1.0 , ): '''simple docstring''' return mean( function_to_integrate(uniform(lowercase__ , lowercase__ ) ) for _ in range(lowercase__ ) ) * (max_value - min_value) def a(lowercase__ , lowercase__ = 0.0 , lowercase__ = 1.0 ): '''simple docstring''' def identity_function(lowercase__ ) -> float: return x snake_case_ = area_under_curve_estimator( lowercase__ , lowercase__ , lowercase__ , lowercase__ ) snake_case_ = (max_value * max_value - min_value * min_value) / 2 print('******************' ) print(f"""Estimating area under y=x where x varies from {min_value} to {max_value}""" ) print(f"""Estimated value is {estimated_value}""" ) print(f"""Expected value is {expected_value}""" ) print(f"""Total error is {abs(estimated_value - expected_value )}""" ) print('******************' ) def a(lowercase__ ): '''simple docstring''' def function_to_integrate(lowercase__ ) -> float: return sqrt(4.0 - x * x ) snake_case_ = area_under_curve_estimator( lowercase__ , lowercase__ , 0.0 , 2.0 ) print('******************' ) print('Estimating pi using area_under_curve_estimator' ) print(f"""Estimated value is {estimated_value}""" ) print(f"""Expected value is {pi}""" ) print(f"""Total error is {abs(estimated_value - pi )}""" ) print('******************' ) if __name__ == "__main__": import doctest doctest.testmod()
46
import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def a(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file snake_case_ = TapasConfig.from_json_file(lowercase__ ) # set absolute/relative position embeddings parameter snake_case_ = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "WTQ": # run_task_main.py hparams snake_case_ = 4 snake_case_ = True # hparam_utils.py hparams snake_case_ = 0.66_4694 snake_case_ = 0.20_7951 snake_case_ = 0.12_1194 snake_case_ = True snake_case_ = True snake_case_ = False snake_case_ = 0.035_2513 snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams snake_case_ = 4 snake_case_ = False # hparam_utils.py hparams snake_case_ = 36.4519 snake_case_ = 0.90_3421 snake_case_ = 222.088 snake_case_ = True snake_case_ = True snake_case_ = True snake_case_ = 0.76_3141 snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "TABFACT": snake_case_ = TapasForSequenceClassification(config=lowercase__ ) elif task == "MLM": snake_case_ = TapasForMaskedLM(config=lowercase__ ) elif task == "INTERMEDIATE_PRETRAINING": snake_case_ = TapasModel(config=lowercase__ ) else: raise ValueError(f"""Task {task} not supported.""" ) print(f"""Building PyTorch model from configuration: {config}""" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(lowercase__ , lowercase__ , lowercase__ ) # Save pytorch-model (weights and configuration) print(f"""Save PyTorch model to {pytorch_dump_path}""" ) model.save_pretrained(lowercase__ ) # Save tokenizer files print(f"""Save tokenizer files to {pytorch_dump_path}""" ) snake_case_ = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + 'vocab.txt' , model_max_length=512 ) tokenizer.save_pretrained(lowercase__ ) print('Used relative position embeddings:' , model.config.reset_position_index_per_cell ) if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( '--task', default='SQA', type=str, help='Model task for which to convert a checkpoint. Defaults to SQA.' ) parser.add_argument( '--reset_position_index_per_cell', default=False, action='store_true', help='Whether to use relative position embeddings or not. Defaults to True.', ) parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--tapas_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained TAPAS model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) A = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
46
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A = { 'configuration_roberta': ['ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RobertaConfig', 'RobertaOnnxConfig'], 'tokenization_roberta': ['RobertaTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['RobertaTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'RobertaForCausalLM', 'RobertaForMaskedLM', 'RobertaForMultipleChoice', 'RobertaForQuestionAnswering', 'RobertaForSequenceClassification', 'RobertaForTokenClassification', 'RobertaModel', 'RobertaPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFRobertaForCausalLM', 'TFRobertaForMaskedLM', 'TFRobertaForMultipleChoice', 'TFRobertaForQuestionAnswering', 'TFRobertaForSequenceClassification', 'TFRobertaForTokenClassification', 'TFRobertaMainLayer', 'TFRobertaModel', 'TFRobertaPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'FlaxRobertaForCausalLM', 'FlaxRobertaForMaskedLM', 'FlaxRobertaForMultipleChoice', 'FlaxRobertaForQuestionAnswering', 'FlaxRobertaForSequenceClassification', 'FlaxRobertaForTokenClassification', 'FlaxRobertaModel', 'FlaxRobertaPreTrainedModel', ] if TYPE_CHECKING: from .configuration_roberta import ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaConfig, RobertaOnnxConfig from .tokenization_roberta import RobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roberta_fast import RobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta import ( ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaForCausalLM, RobertaForMaskedLM, RobertaForMultipleChoice, RobertaForQuestionAnswering, RobertaForSequenceClassification, RobertaForTokenClassification, RobertaModel, RobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta import ( TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForMultipleChoice, TFRobertaForQuestionAnswering, TFRobertaForSequenceClassification, TFRobertaForTokenClassification, TFRobertaMainLayer, TFRobertaModel, TFRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, FlaxRobertaPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=3 , __UpperCamelCase=16 , __UpperCamelCase=[1, 2, 1] , __UpperCamelCase=[2, 2, 4] , __UpperCamelCase=2 , __UpperCamelCase=2.0 , __UpperCamelCase=True , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.1 , __UpperCamelCase="gelu" , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=0.02 , __UpperCamelCase=1E-5 , __UpperCamelCase=True , __UpperCamelCase=None , __UpperCamelCase=True , __UpperCamelCase=10 , __UpperCamelCase=8 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = embed_dim snake_case_ = depths snake_case_ = num_heads snake_case_ = window_size snake_case_ = mlp_ratio snake_case_ = qkv_bias snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = drop_path_rate snake_case_ = hidden_act snake_case_ = use_absolute_embeddings snake_case_ = patch_norm snake_case_ = layer_norm_eps snake_case_ = initializer_range snake_case_ = is_training snake_case_ = scope snake_case_ = use_labels snake_case_ = type_sequence_label_size snake_case_ = encoder_stride def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = self.get_config() return config, pixel_values, labels def __lowerCAmelCase ( self ): """simple docstring""" return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = SwinvaModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) snake_case_ = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) snake_case_ = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = SwinvaForMaskedImageModeling(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images snake_case_ = 1 snake_case_ = SwinvaForMaskedImageModeling(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.type_sequence_label_size snake_case_ = SwinvaForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) __A = ( {"""feature-extraction""": SwinvaModel, """image-classification""": SwinvaForImageClassification} if is_torch_available() else {} ) __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = SwinvaModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , embed_dim=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase , nn.Linear ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = True for model_class in self.all_model_classes: snake_case_ = True snake_case_ = False snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.attentions snake_case_ = len(self.model_tester.depths ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # check that output_attentions also work using config del inputs_dict["output_attentions"] snake_case_ = True snake_case_ = config.window_size**2 snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) snake_case_ = len(__UpperCamelCase ) # Check attention is always last and order is fine snake_case_ = True snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): snake_case_ = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states snake_case_ = 2 self.assertEqual(out_len + added_hidden_states , len(__UpperCamelCase ) ) snake_case_ = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.hidden_states snake_case_ = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # Swinv2 has a different seq_length snake_case_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) snake_case_ = outputs.reshaped_hidden_states self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) snake_case_ , snake_case_ , snake_case_ , snake_case_ = reshaped_hidden_states[0].shape snake_case_ = ( reshaped_hidden_states[0].view(__UpperCamelCase , __UpperCamelCase , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) snake_case_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) snake_case_ = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = SwinvaModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = _config_zero_init(__UpperCamelCase ) for model_class in self.all_model_classes: snake_case_ = model_class(config=__UpperCamelCase ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @require_vision @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( __UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([-0.3947, -0.4306, 0.0026] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) )
46
1
A = '\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' A = [{'type': 'code', 'content': INSTALL_CONTENT}] A = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
46
import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline A = argparse.ArgumentParser('Stable Diffusion script with intel optimization', add_help=False) parser.add_argument('--dpm', action='store_true', help='Enable DPMSolver or not') parser.add_argument('--steps', default=None, type=int, help='Num inference steps') A = parser.parse_args() A = 'cpu' A = 'a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings' A = 'path-to-your-trained-model' A = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: A = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) A = pipe.to(device) # to channels last A = pipe.unet.to(memory_format=torch.channels_last) A = pipe.vae.to(memory_format=torch.channels_last) A = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: A = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex A = torch.randn(2, 4, 64, 64) A = torch.rand(1) * 999 A = torch.randn(2, 77, 768) A = (sample, timestep, encoder_hidden_status) try: A = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: A = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) A = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) A = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: A = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute A = 666 A = torch.Generator(device).manual_seed(seed) A = {'generator': generator} if args.steps is not None: A = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): A = pipe(prompt, **generate_kwargs).images[0] # save image image.save('generated.png')
46
1
from math import sqrt def a(lowercase__ ): '''simple docstring''' snake_case_ = 0 for i in range(1 , int(sqrt(lowercase__ ) + 1 ) ): if n % i == 0 and i != sqrt(lowercase__ ): total += i + n // i elif i == sqrt(lowercase__ ): total += i return total - n def a(lowercase__ = 10000 ): '''simple docstring''' snake_case_ = sum( i for i in range(1 , lowercase__ ) if sum_of_divisors(sum_of_divisors(lowercase__ ) ) == i and sum_of_divisors(lowercase__ ) != i ) return total if __name__ == "__main__": print(solution(int(str(input()).strip())))
46
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { 'microsoft/unispeech-sat-base-100h-libri-ft': ( 'https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json' ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = """unispeech-sat""" def __init__( self , __UpperCamelCase=32 , __UpperCamelCase=7_68 , __UpperCamelCase=12 , __UpperCamelCase=12 , __UpperCamelCase=30_72 , __UpperCamelCase="gelu" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.02 , __UpperCamelCase=1E-5 , __UpperCamelCase="group" , __UpperCamelCase="gelu" , __UpperCamelCase=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , __UpperCamelCase=(5, 2, 2, 2, 2, 2, 2) , __UpperCamelCase=(10, 3, 3, 3, 3, 2, 2) , __UpperCamelCase=False , __UpperCamelCase=1_28 , __UpperCamelCase=16 , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=0.05 , __UpperCamelCase=10 , __UpperCamelCase=2 , __UpperCamelCase=0.0 , __UpperCamelCase=10 , __UpperCamelCase=0 , __UpperCamelCase=3_20 , __UpperCamelCase=2 , __UpperCamelCase=0.1 , __UpperCamelCase=1_00 , __UpperCamelCase=2_56 , __UpperCamelCase=2_56 , __UpperCamelCase=0.1 , __UpperCamelCase="mean" , __UpperCamelCase=False , __UpperCamelCase=False , __UpperCamelCase=2_56 , __UpperCamelCase=(5_12, 5_12, 5_12, 5_12, 15_00) , __UpperCamelCase=(5, 3, 3, 1, 1) , __UpperCamelCase=(1, 2, 3, 1, 1) , __UpperCamelCase=5_12 , __UpperCamelCase=0 , __UpperCamelCase=1 , __UpperCamelCase=2 , __UpperCamelCase=5_04 , **__UpperCamelCase , ): """simple docstring""" super().__init__(**__UpperCamelCase , pad_token_id=__UpperCamelCase , bos_token_id=__UpperCamelCase , eos_token_id=__UpperCamelCase ) snake_case_ = hidden_size snake_case_ = feat_extract_norm snake_case_ = feat_extract_activation snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = conv_bias snake_case_ = num_conv_pos_embeddings snake_case_ = num_conv_pos_embedding_groups snake_case_ = len(self.conv_dim ) snake_case_ = num_hidden_layers snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = num_attention_heads snake_case_ = hidden_dropout snake_case_ = attention_dropout snake_case_ = activation_dropout snake_case_ = feat_proj_dropout snake_case_ = final_dropout snake_case_ = layerdrop snake_case_ = layer_norm_eps snake_case_ = initializer_range snake_case_ = vocab_size snake_case_ = num_clusters snake_case_ = do_stable_layer_norm snake_case_ = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( 'Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==' ' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =' f""" {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,""" f""" `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 snake_case_ = apply_spec_augment snake_case_ = mask_time_prob snake_case_ = mask_time_length snake_case_ = mask_time_min_masks snake_case_ = mask_feature_prob snake_case_ = mask_feature_length snake_case_ = mask_feature_min_masks # parameters for pretraining with codevector quantized representations snake_case_ = num_codevectors_per_group snake_case_ = num_codevector_groups snake_case_ = contrastive_logits_temperature snake_case_ = feat_quantizer_dropout snake_case_ = num_negatives snake_case_ = codevector_dim snake_case_ = proj_codevector_dim snake_case_ = diversity_loss_weight # ctc loss snake_case_ = ctc_loss_reduction snake_case_ = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. snake_case_ = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = xvector_output_dim @property def __lowerCAmelCase ( self ): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
46
1
import importlib import json import os import sys import tempfile import unittest from pathlib import Path import transformers import transformers.models.auto from transformers.models.auto.configuration_auto import CONFIG_MAPPING, AutoConfig from transformers.models.bert.configuration_bert import BertConfig from transformers.models.roberta.configuration_roberta import RobertaConfig from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 A = get_tests_dir('fixtures/dummy-config.json') class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = 0 def __lowerCAmelCase ( self ): """simple docstring""" self.assertIsNotNone(transformers.models.auto.__spec__ ) self.assertIsNotNone(importlib.util.find_spec('transformers.models.auto' ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = AutoConfig.from_pretrained('bert-base-uncased' ) self.assertIsInstance(__UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = AutoConfig.from_pretrained(__UpperCamelCase ) self.assertIsInstance(__UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = AutoConfig.from_pretrained(__UpperCamelCase ) self.assertIsInstance(__UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = AutoConfig.for_model('roberta' ) self.assertIsInstance(__UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: # This model name contains bert and roberta, but roberta ends up being picked. snake_case_ = os.path.join(__UpperCamelCase , 'fake-roberta' ) os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) with open(os.path.join(__UpperCamelCase , 'config.json' ) , 'w' ) as f: f.write(json.dumps({} ) ) snake_case_ = AutoConfig.from_pretrained(__UpperCamelCase ) self.assertEqual(type(__UpperCamelCase ) , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" try: AutoConfig.register('custom' , __UpperCamelCase ) # Wrong model type will raise an error with self.assertRaises(__UpperCamelCase ): AutoConfig.register('model' , __UpperCamelCase ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(__UpperCamelCase ): AutoConfig.register('bert' , __UpperCamelCase ) # Now that the config is registered, it can be used as any other config with the auto-API snake_case_ = CustomConfig() with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__UpperCamelCase ) snake_case_ = AutoConfig.from_pretrained(__UpperCamelCase ) self.assertIsInstance(__UpperCamelCase , __UpperCamelCase ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] def __lowerCAmelCase ( self ): """simple docstring""" with self.assertRaisesRegex( __UpperCamelCase , 'bert-base is not a local folder and is not a valid model identifier' ): snake_case_ = AutoConfig.from_pretrained('bert-base' ) def __lowerCAmelCase ( self ): """simple docstring""" with self.assertRaisesRegex( __UpperCamelCase , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): snake_case_ = AutoConfig.from_pretrained(__UpperCamelCase , revision='aaaaaa' ) def __lowerCAmelCase ( self ): """simple docstring""" with self.assertRaisesRegex( __UpperCamelCase , 'hf-internal-testing/no-config-test-repo does not appear to have a file named config.json.' , ): snake_case_ = AutoConfig.from_pretrained('hf-internal-testing/no-config-test-repo' ) def __lowerCAmelCase ( self ): """simple docstring""" with self.assertRaises(__UpperCamelCase ): snake_case_ = AutoConfig.from_pretrained('hf-internal-testing/test_dynamic_model' ) # If remote code is disabled, we can't load this config. with self.assertRaises(__UpperCamelCase ): snake_case_ = AutoConfig.from_pretrained('hf-internal-testing/test_dynamic_model' , trust_remote_code=__UpperCamelCase ) snake_case_ = AutoConfig.from_pretrained('hf-internal-testing/test_dynamic_model' , trust_remote_code=__UpperCamelCase ) self.assertEqual(config.__class__.__name__ , 'NewModelConfig' ) # Test config can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(__UpperCamelCase ) snake_case_ = AutoConfig.from_pretrained(__UpperCamelCase , trust_remote_code=__UpperCamelCase ) self.assertEqual(reloaded_config.__class__.__name__ , 'NewModelConfig' ) def __lowerCAmelCase ( self ): """simple docstring""" class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = """new-model""" try: AutoConfig.register('new-model' , __UpperCamelCase ) # If remote code is not set, the default is to use local snake_case_ = AutoConfig.from_pretrained('hf-internal-testing/test_dynamic_model' ) self.assertEqual(config.__class__.__name__ , 'NewModelConfigLocal' ) # If remote code is disabled, we load the local one. snake_case_ = AutoConfig.from_pretrained('hf-internal-testing/test_dynamic_model' , trust_remote_code=__UpperCamelCase ) self.assertEqual(config.__class__.__name__ , 'NewModelConfigLocal' ) # If remote is enabled, we load from the Hub snake_case_ = AutoConfig.from_pretrained('hf-internal-testing/test_dynamic_model' , trust_remote_code=__UpperCamelCase ) self.assertEqual(config.__class__.__name__ , 'NewModelConfig' ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"]
46
class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = name snake_case_ = val def __str__( self ): """simple docstring""" return f"""{self.__class__.__name__}({self.name}, {self.val})""" def __lt__( self , __UpperCamelCase ): """simple docstring""" return self.val < other.val class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" snake_case_ = {} snake_case_ = {} snake_case_ = self.build_heap(__UpperCamelCase ) def __getitem__( self , __UpperCamelCase ): """simple docstring""" return self.get_value(__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return (idx - 1) // 2 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return idx * 2 + 1 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return idx * 2 + 2 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return self.heap_dict[key] def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = len(__UpperCamelCase ) - 1 snake_case_ = self.get_parent_idx(__UpperCamelCase ) for idx, i in enumerate(__UpperCamelCase ): snake_case_ = idx snake_case_ = i.val for i in range(__UpperCamelCase , -1 , -1 ): self.sift_down(__UpperCamelCase , __UpperCamelCase ) return array def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" while True: snake_case_ = self.get_left_child_idx(__UpperCamelCase ) # noqa: E741 snake_case_ = self.get_right_child_idx(__UpperCamelCase ) snake_case_ = idx if l < len(__UpperCamelCase ) and array[l] < array[idx]: snake_case_ = l if r < len(__UpperCamelCase ) and array[r] < array[smallest]: snake_case_ = r if smallest != idx: snake_case_ , snake_case_ = array[smallest], array[idx] ( ( snake_case_ ) , ( snake_case_ ) , ) = ( self.idx_of_element[array[smallest]], self.idx_of_element[array[idx]], ) snake_case_ = smallest else: break def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = self.get_parent_idx(__UpperCamelCase ) while p >= 0 and self.heap[p] > self.heap[idx]: snake_case_ , snake_case_ = self.heap[idx], self.heap[p] snake_case_ , snake_case_ = ( self.idx_of_element[self.heap[idx]], self.idx_of_element[self.heap[p]], ) snake_case_ = p snake_case_ = self.get_parent_idx(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" return self.heap[0] def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.heap[-1], self.heap[0] snake_case_ , snake_case_ = ( self.idx_of_element[self.heap[-1]], self.idx_of_element[self.heap[0]], ) snake_case_ = self.heap.pop() del self.idx_of_element[x] self.sift_down(0 , self.heap ) return x def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" self.heap.append(__UpperCamelCase ) snake_case_ = len(self.heap ) - 1 snake_case_ = node.val self.sift_up(len(self.heap ) - 1 ) def __lowerCAmelCase ( self ): """simple docstring""" return len(self.heap ) == 0 def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" assert ( self.heap[self.idx_of_element[node]].val > new_value ), "newValue must be less that current value" snake_case_ = new_value snake_case_ = new_value self.sift_up(self.idx_of_element[node] ) A = Node('R', -1) A = Node('B', 6) A = Node('A', 3) A = Node('X', 1) A = Node('E', 4) # Use one of these two ways to generate Min-Heap # Generating Min-Heap from array A = MinHeap([r, b, a, x, e]) # Generating Min-Heap by Insert method # myMinHeap.insert(a) # myMinHeap.insert(b) # myMinHeap.insert(x) # myMinHeap.insert(r) # myMinHeap.insert(e) # Before print('Min Heap - before decrease key') for i in my_min_heap.heap: print(i) print('Min Heap - After decrease key of node [B -> -17]') my_min_heap.decrease_key(b, -17) # After for i in my_min_heap.heap: print(i) if __name__ == "__main__": import doctest doctest.testmod()
46
1
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=7 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=99 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=4 , __UpperCamelCase=37 , __UpperCamelCase="gelu" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=5_12 , __UpperCamelCase=16 , __UpperCamelCase=2 , __UpperCamelCase=0.02 , __UpperCamelCase=3 , __UpperCamelCase=4 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = 13 snake_case_ = 7 snake_case_ = True snake_case_ = True snake_case_ = True snake_case_ = True snake_case_ = 99 snake_case_ = 3_84 snake_case_ = 2 snake_case_ = 4 snake_case_ = 37 snake_case_ = 'gelu' snake_case_ = 0.1 snake_case_ = 0.1 snake_case_ = 5_12 snake_case_ = 16 snake_case_ = 2 snake_case_ = 0.02 snake_case_ = 3 snake_case_ = 4 snake_case_ = 1_28 snake_case_ = 2 snake_case_ = 9 snake_case_ = 1 snake_case_ = None def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None if self.use_token_type_ids: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=__UpperCamelCase , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = TFConvBertModel(config=__UpperCamelCase ) snake_case_ = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} snake_case_ = [input_ids, input_mask] snake_case_ = model(__UpperCamelCase ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = TFConvBertForMaskedLM(config=__UpperCamelCase ) snake_case_ = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = TFConvBertForSequenceClassification(config=__UpperCamelCase ) snake_case_ = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_choices snake_case_ = TFConvBertForMultipleChoice(config=__UpperCamelCase ) snake_case_ = tf.tile(tf.expand_dims(__UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) snake_case_ = tf.tile(tf.expand_dims(__UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) snake_case_ = tf.tile(tf.expand_dims(__UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) snake_case_ = { 'input_ids': multiple_choice_inputs_ids, 'attention_mask': multiple_choice_input_mask, 'token_type_ids': multiple_choice_token_type_ids, } snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = TFConvBertForTokenClassification(config=__UpperCamelCase ) snake_case_ = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = TFConvBertForQuestionAnswering(config=__UpperCamelCase ) snake_case_ = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = config_and_inputs snake_case_ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) __A = ( { """feature-extraction""": TFConvBertModel, """fill-mask""": TFConvBertForMaskedLM, """question-answering""": TFConvBertForQuestionAnswering, """text-classification""": TFConvBertForSequenceClassification, """token-classification""": TFConvBertForTokenClassification, """zero-shot""": TFConvBertForSequenceClassification, } if is_tf_available() else {} ) __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFConvBertModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , hidden_size=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = True snake_case_ = True if hasattr(__UpperCamelCase , 'use_cache' ): snake_case_ = True snake_case_ = getattr(self.model_tester , 'encoder_seq_length' , self.model_tester.seq_length ) snake_case_ = getattr(self.model_tester , 'key_length' , __UpperCamelCase ) for model_class in self.all_model_classes: snake_case_ = self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) snake_case_ = model_class(__UpperCamelCase ) snake_case_ = len(model(__UpperCamelCase ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__UpperCamelCase , saved_model=__UpperCamelCase ) snake_case_ = os.path.join(__UpperCamelCase , 'saved_model' , '1' ) snake_case_ = tf.keras.models.load_model(__UpperCamelCase ) snake_case_ = model(__UpperCamelCase ) if self.is_encoder_decoder: snake_case_ = outputs['encoder_hidden_states'] snake_case_ = outputs['encoder_attentions'] else: snake_case_ = outputs['hidden_states'] snake_case_ = outputs['attentions'] self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) snake_case_ = getattr( self.model_tester , 'expected_num_hidden_layers' , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFConvBertModel.from_pretrained('YituTech/conv-bert-base' ) self.assertIsNotNone(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = True snake_case_ = getattr(self.model_tester , 'decoder_seq_length' , self.model_tester.seq_length ) snake_case_ = getattr(self.model_tester , 'encoder_seq_length' , self.model_tester.seq_length ) snake_case_ = getattr(self.model_tester , 'key_length' , __UpperCamelCase ) snake_case_ = getattr(self.model_tester , 'key_length' , __UpperCamelCase ) def check_decoder_attentions_output(__UpperCamelCase ): snake_case_ = len(__UpperCamelCase ) self.assertEqual(out_len % 2 , 0 ) snake_case_ = outputs.decoder_attentions self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(__UpperCamelCase ): snake_case_ = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: snake_case_ = True snake_case_ = False snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = len(__UpperCamelCase ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) if self.is_encoder_decoder: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_decoder_attentions_output(__UpperCamelCase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) # Check attention is always last and order is fine snake_case_ = True snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(__UpperCamelCase ) ) self.assertEqual(model.config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFConvBertModel.from_pretrained('YituTech/conv-bert-base' ) snake_case_ = tf.constant([[0, 1, 2, 3, 4, 5]] ) snake_case_ = model(__UpperCamelCase )[0] snake_case_ = [1, 6, 7_68] self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = tf.constant( [ [ [-0.0347_5493, -0.468_6034, -0.3063_8832], [0.2263_7248, -0.2698_8646, -0.742_3424], [0.1032_4868, -0.4501_3508, -0.5828_0784], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 )
46
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) A = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['PerceiverFeatureExtractor'] A = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
1
from PIL import Image def a(lowercase__ ): '''simple docstring''' snake_case_ , snake_case_ = image.size snake_case_ = 0 snake_case_ = image.load() for i in range(lowercase__ ): for j in range(lowercase__ ): snake_case_ = pixels[j, i] mean += pixel mean //= width * height for j in range(lowercase__ ): for i in range(lowercase__ ): snake_case_ = 255 if pixels[i, j] > mean else 0 return image if __name__ == "__main__": A = mean_threshold(Image.open('path_to_image').convert('L')) image.save('output_image_path')
46
def a(lowercase__ , lowercase__ ): '''simple docstring''' if not isinstance(lowercase__ , lowercase__ ): raise ValueError('iterations must be defined as integers' ) if not isinstance(lowercase__ , lowercase__ ) or not number >= 1: raise ValueError( 'starting number must be\n and integer and be more than 0' ) if not iterations >= 1: raise ValueError('Iterations must be done more than 0 times to play FizzBuzz' ) snake_case_ = '' while number <= iterations: if number % 3 == 0: out += "Fizz" if number % 5 == 0: out += "Buzz" if 0 not in (number % 3, number % 5): out += str(lowercase__ ) # print(out) number += 1 out += " " return out if __name__ == "__main__": import doctest doctest.testmod()
46
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A = { 'configuration_electra': ['ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ElectraConfig', 'ElectraOnnxConfig'], 'tokenization_electra': ['ElectraTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['ElectraTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST', 'ElectraForCausalLM', 'ElectraForMaskedLM', 'ElectraForMultipleChoice', 'ElectraForPreTraining', 'ElectraForQuestionAnswering', 'ElectraForSequenceClassification', 'ElectraForTokenClassification', 'ElectraModel', 'ElectraPreTrainedModel', 'load_tf_weights_in_electra', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFElectraForMaskedLM', 'TFElectraForMultipleChoice', 'TFElectraForPreTraining', 'TFElectraForQuestionAnswering', 'TFElectraForSequenceClassification', 'TFElectraForTokenClassification', 'TFElectraModel', 'TFElectraPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'FlaxElectraForCausalLM', 'FlaxElectraForMaskedLM', 'FlaxElectraForMultipleChoice', 'FlaxElectraForPreTraining', 'FlaxElectraForQuestionAnswering', 'FlaxElectraForSequenceClassification', 'FlaxElectraForTokenClassification', 'FlaxElectraModel', 'FlaxElectraPreTrainedModel', ] if TYPE_CHECKING: from .configuration_electra import ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ElectraConfig, ElectraOnnxConfig from .tokenization_electra import ElectraTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_electra_fast import ElectraTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_electra import ( ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, ElectraForCausalLM, ElectraForMaskedLM, ElectraForMultipleChoice, ElectraForPreTraining, ElectraForQuestionAnswering, ElectraForSequenceClassification, ElectraForTokenClassification, ElectraModel, ElectraPreTrainedModel, load_tf_weights_in_electra, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_electra import ( TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, TFElectraForMaskedLM, TFElectraForMultipleChoice, TFElectraForPreTraining, TFElectraForQuestionAnswering, TFElectraForSequenceClassification, TFElectraForTokenClassification, TFElectraModel, TFElectraPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_electra import ( FlaxElectraForCausalLM, FlaxElectraForMaskedLM, FlaxElectraForMultipleChoice, FlaxElectraForPreTraining, FlaxElectraForQuestionAnswering, FlaxElectraForSequenceClassification, FlaxElectraForTokenClassification, FlaxElectraModel, FlaxElectraPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = 1.5 snake_case_ = int(factor * num_class_images ) snake_case_ = ClipClient( url='https://knn.laion.ai/knn-service' , indice_name='laion_400m' , num_images=lowercase__ , aesthetic_weight=0.1 ) os.makedirs(f"""{class_data_dir}/images""" , exist_ok=lowercase__ ) if len(list(Path(f"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: snake_case_ = client.query(text=lowercase__ ) if len(lowercase__ ) >= factor * num_class_images or num_images > 1e4: break else: snake_case_ = int(factor * num_images ) snake_case_ = ClipClient( url='https://knn.laion.ai/knn-service' , indice_name='laion_400m' , num_images=lowercase__ , aesthetic_weight=0.1 , ) snake_case_ = 0 snake_case_ = 0 snake_case_ = tqdm(desc='downloading real regularization images' , total=lowercase__ ) with open(f"""{class_data_dir}/caption.txt""" , 'w' ) as fa, open(f"""{class_data_dir}/urls.txt""" , 'w' ) as fa, open( f"""{class_data_dir}/images.txt""" , 'w' ) as fa: while total < num_class_images: snake_case_ = class_images[count] count += 1 try: snake_case_ = requests.get(images['url'] ) if img.status_code == 200: snake_case_ = Image.open(BytesIO(img.content ) ) with open(f"""{class_data_dir}/images/{total}.jpg""" , 'wb' ) as f: f.write(img.content ) fa.write(images['caption'] + '\n' ) fa.write(images['url'] + '\n' ) fa.write(f"""{class_data_dir}/images/{total}.jpg""" + '\n' ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def a(): '''simple docstring''' snake_case_ = argparse.ArgumentParser('' , add_help=lowercase__ ) parser.add_argument('--class_prompt' , help='text prompt to retrieve images' , required=lowercase__ , type=lowercase__ ) parser.add_argument('--class_data_dir' , help='path to save images' , required=lowercase__ , type=lowercase__ ) parser.add_argument('--num_class_images' , help='number of images to download' , default=200 , type=lowercase__ ) return parser.parse_args() if __name__ == "__main__": A = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
46
1
import inspect import unittest from transformers import ConvNextVaConfig from transformers.models.auto import get_values from transformers.models.auto.modeling_auto import MODEL_FOR_BACKBONE_MAPPING_NAMES, MODEL_MAPPING_NAMES from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextVaBackbone, ConvNextVaForImageClassification, ConvNextVaModel from transformers.models.convnextva.modeling_convnextva import CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=32 , __UpperCamelCase=3 , __UpperCamelCase=4 , __UpperCamelCase=[10, 20, 30, 40] , __UpperCamelCase=[2, 2, 3, 2] , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=37 , __UpperCamelCase="gelu" , __UpperCamelCase=10 , __UpperCamelCase=0.02 , __UpperCamelCase=["stage2", "stage3", "stage4"] , __UpperCamelCase=[2, 3, 4] , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = num_channels snake_case_ = num_stages snake_case_ = hidden_sizes snake_case_ = depths snake_case_ = is_training snake_case_ = use_labels snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = num_labels snake_case_ = initializer_range snake_case_ = out_features snake_case_ = out_indices snake_case_ = scope def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.num_labels ) snake_case_ = self.get_config() return config, pixel_values, labels def __lowerCAmelCase ( self ): """simple docstring""" return ConvNextVaConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = ConvNextVaModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = ConvNextVaForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = ConvNextVaBackbone(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None snake_case_ = None snake_case_ = ConvNextVaBackbone(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values, 'labels': labels} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( ( ConvNextVaModel, ConvNextVaForImageClassification, ConvNextVaBackbone, ) if is_torch_available() else () ) __A = ( {"""feature-extraction""": ConvNextVaModel, """image-classification""": ConvNextVaForImageClassification} if is_torch_available() else {} ) __A = False __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ConvNextVaModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase , hidden_size=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowerCAmelCase ( self ): """simple docstring""" return @unittest.skip(reason='ConvNextV2 does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='ConvNextV2 does not support input and output embeddings' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='ConvNextV2 does not use feedforward chunking' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" if not self.model_tester.is_training: return for model_class in self.all_model_classes: snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_with_labels() snake_case_ = True if model_class.__name__ in [ *get_values(__UpperCamelCase ), *get_values(__UpperCamelCase ), ]: continue snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.train() snake_case_ = self._prepare_for_class(__UpperCamelCase , __UpperCamelCase , return_labels=__UpperCamelCase ) snake_case_ = model(**__UpperCamelCase ).loss loss.backward() def __lowerCAmelCase ( self ): """simple docstring""" if not self.model_tester.is_training: return for model_class in self.all_model_classes: snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_with_labels() snake_case_ = False snake_case_ = True if ( model_class.__name__ in [*get_values(__UpperCamelCase ), *get_values(__UpperCamelCase )] or not model_class.supports_gradient_checkpointing ): continue snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.gradient_checkpointing_enable() model.train() snake_case_ = self._prepare_for_class(__UpperCamelCase , __UpperCamelCase , return_labels=__UpperCamelCase ) snake_case_ = model(**__UpperCamelCase ).loss loss.backward() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" def check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states snake_case_ = self.model_tester.num_stages self.assertEqual(len(__UpperCamelCase ) , expected_num_stages + 1 ) # ConvNextV2's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = ConvNextVaModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def a(): '''simple docstring''' snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return AutoImageProcessor.from_pretrained('facebook/convnextv2-tiny-1k-224' ) if is_vision_available() else None @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ConvNextVaForImageClassification.from_pretrained('facebook/convnextv2-tiny-1k-224' ).to(__UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = prepare_img() snake_case_ = preprocessor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([0.9996, 0.1966, -0.4386] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) )
46
# coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import platform import sys A = '3' print('Python version:', sys.version) print('OS platform:', platform.platform()) print('OS architecture:', platform.machine()) try: import torch print('Torch version:', torch.__version__) print('Cuda available:', torch.cuda.is_available()) print('Cuda version:', torch.version.cuda) print('CuDNN version:', torch.backends.cudnn.version()) print('Number of GPUs available:', torch.cuda.device_count()) except ImportError: print('Torch version:', None) try: import transformers print('transformers version:', transformers.__version__) except ImportError: print('transformers version:', None)
46
1
from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput A = logging.get_logger(__name__) # pylint: disable=invalid-name class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ): """simple docstring""" @register_to_config def __init__( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None ): """simple docstring""" super().__init__() snake_case_ = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" snake_case_ = torch.zeros(__UpperCamelCase , __UpperCamelCase ) else: snake_case_ = None snake_case_ = torch.nn.Parameter(__UpperCamelCase ) class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" super().__init__() self.register_modules( vqvae=__UpperCamelCase , transformer=__UpperCamelCase , text_encoder=__UpperCamelCase , tokenizer=__UpperCamelCase , scheduler=__UpperCamelCase , learned_classifier_free_sampling_embeddings=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = len(__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else 1 # get prompt text embeddings snake_case_ = self.tokenizer( __UpperCamelCase , padding='max_length' , max_length=self.tokenizer.model_max_length , return_tensors='pt' , ) snake_case_ = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: snake_case_ = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( 'The following part of your input was truncated because CLIP can only handle sequences up to' f""" {self.tokenizer.model_max_length} tokens: {removed_text}""" ) snake_case_ = text_input_ids[:, : self.tokenizer.model_max_length] snake_case_ = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 snake_case_ = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate text embeddings for each generation per prompt snake_case_ = prompt_embeds.repeat_interleave(__UpperCamelCase , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: snake_case_ = self.learned_classifier_free_sampling_embeddings.embeddings snake_case_ = negative_prompt_embeds.unsqueeze(0 ).repeat(__UpperCamelCase , 1 , 1 ) else: snake_case_ = [''] * batch_size snake_case_ = text_input_ids.shape[-1] snake_case_ = self.tokenizer( __UpperCamelCase , padding='max_length' , max_length=__UpperCamelCase , truncation=__UpperCamelCase , return_tensors='pt' , ) snake_case_ = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings snake_case_ = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method snake_case_ = negative_prompt_embeds.shape[1] snake_case_ = negative_prompt_embeds.repeat(1 , __UpperCamelCase , 1 ) snake_case_ = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __UpperCamelCase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes snake_case_ = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self , __UpperCamelCase , __UpperCamelCase = 1_00 , __UpperCamelCase = 5.0 , __UpperCamelCase = 1.0 , __UpperCamelCase = 1 , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = "pil" , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = 1 , ): """simple docstring""" if isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = 1 elif isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = len(__UpperCamelCase ) else: raise ValueError(f"""`prompt` has to be of type `str` or `list` but is {type(__UpperCamelCase )}""" ) snake_case_ = batch_size * num_images_per_prompt snake_case_ = guidance_scale > 1.0 snake_case_ = self._encode_prompt(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__UpperCamelCase , __UpperCamelCase ) or callback_steps <= 0) ): raise ValueError( f"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" f""" {type(__UpperCamelCase )}.""" ) # get the initial completely masked latents unless the user supplied it snake_case_ = (batch_size, self.transformer.num_latent_pixels) if latents is None: snake_case_ = self.transformer.num_vector_embeds - 1 snake_case_ = torch.full(__UpperCamelCase , __UpperCamelCase ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {latents_shape}""" ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( 'Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0,' f""" {self.transformer.num_vector_embeds - 1} (inclusive).""" ) snake_case_ = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(__UpperCamelCase , device=self.device ) snake_case_ = self.scheduler.timesteps.to(self.device ) snake_case_ = latents for i, t in enumerate(self.progress_bar(__UpperCamelCase ) ): # expand the sample if we are doing classifier free guidance snake_case_ = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` snake_case_ = self.transformer(__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , timestep=__UpperCamelCase ).sample if do_classifier_free_guidance: snake_case_ , snake_case_ = model_output.chunk(2 ) snake_case_ = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(__UpperCamelCase , dim=1 , keepdim=__UpperCamelCase ) snake_case_ = self.truncate(__UpperCamelCase , __UpperCamelCase ) # remove `log(0)`'s (`-inf`s) snake_case_ = model_output.clamp(-70 ) # compute the previous noisy sample x_t -> x_t-1 snake_case_ = self.scheduler.step(__UpperCamelCase , timestep=__UpperCamelCase , sample=__UpperCamelCase , generator=__UpperCamelCase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = self.vqvae.config.vq_embed_dim snake_case_ = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) snake_case_ = self.vqvae.quantize.get_codebook_entry(__UpperCamelCase , shape=__UpperCamelCase ) snake_case_ = self.vqvae.decode(__UpperCamelCase , force_not_quantize=__UpperCamelCase ).sample snake_case_ = (image / 2 + 0.5).clamp(0 , 1 ) snake_case_ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case_ = self.numpy_to_pil(__UpperCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ , snake_case_ = torch.sort(__UpperCamelCase , 1 , descending=__UpperCamelCase ) snake_case_ = torch.exp(__UpperCamelCase ) snake_case_ = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out snake_case_ = torch.full_like(keep_mask[:, 0:1, :] , __UpperCamelCase ) snake_case_ = torch.cat((all_true, keep_mask) , dim=1 ) snake_case_ = keep_mask[:, :-1, :] snake_case_ = keep_mask.gather(1 , indices.argsort(1 ) ) snake_case_ = log_p_x_0.clone() snake_case_ = -torch.inf # -inf = log(0) return rv
46
import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') A = logging.getLogger(__name__) @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) __A = field( default=__snake_case , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) __A = field( default=__snake_case , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) __A = field( default=__snake_case , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) __A = field( default=__snake_case , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) __A = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) __A = field( default=__snake_case , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = field(default=__snake_case , metadata={"""help""": """The input training data file (a text file)."""} ) __A = field( default=__snake_case , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) __A = field( default=__snake_case , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) __A = field( default=__snake_case , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) __A = field( default=__snake_case , metadata={ """help""": ( """The maximum total input sequence length after tokenization. If passed, sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) __A = field( default=__snake_case , metadata={ """help""": ( """Whether to pad all samples to the maximum sentence length. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch. More """ """efficient on GPU but very bad for TPU.""" ) } , ) __A = field( default=__snake_case , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) __A = field( default=__snake_case , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def __lowerCAmelCase ( self ): """simple docstring""" if self.train_file is not None: snake_case_ = self.train_file.split('.' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: snake_case_ = self.validation_file.split('.' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = 42 __A = True __A = None __A = None def __call__( self , __UpperCamelCase ): """simple docstring""" snake_case_ = 'label' if 'label' in features[0].keys() else 'labels' snake_case_ = [feature.pop(__UpperCamelCase ) for feature in features] snake_case_ = len(__UpperCamelCase ) snake_case_ = len(features[0]['input_ids'] ) snake_case_ = [ [{k: v[i] for k, v in feature.items()} for i in range(__UpperCamelCase )] for feature in features ] snake_case_ = list(chain(*__UpperCamelCase ) ) snake_case_ = self.tokenizer.pad( __UpperCamelCase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='pt' , ) # Un-flatten snake_case_ = {k: v.view(__UpperCamelCase , __UpperCamelCase , -1 ) for k, v in batch.items()} # Add back labels snake_case_ = torch.tensor(__UpperCamelCase , dtype=torch.intaa ) return batch def a(): '''simple docstring''' # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. snake_case_ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. snake_case_ , snake_case_ , snake_case_ = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: snake_case_ , snake_case_ , snake_case_ = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_swag' , lowercase__ , lowercase__ ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() snake_case_ = training_args.get_process_log_level() logger.setLevel(lowercase__ ) datasets.utils.logging.set_verbosity(lowercase__ ) transformers.utils.logging.set_verbosity(lowercase__ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + f"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(f"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. snake_case_ = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: snake_case_ = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. """ 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: snake_case_ = {} if data_args.train_file is not None: snake_case_ = data_args.train_file if data_args.validation_file is not None: snake_case_ = data_args.validation_file snake_case_ = data_args.train_file.split('.' )[-1] snake_case_ = load_dataset( lowercase__ , data_files=lowercase__ , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. snake_case_ = load_dataset( 'swag' , 'regular' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case_ = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) snake_case_ = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) snake_case_ = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=lowercase__ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. snake_case_ = [f"""ending{i}""" for i in range(4 )] snake_case_ = 'sent1' snake_case_ = 'sent2' if data_args.max_seq_length is None: snake_case_ = tokenizer.model_max_length if max_seq_length > 1024: logger.warning( 'The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value' ' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can' ' override this default with `--block_size xxx`.' ) snake_case_ = 1024 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( f"""The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the""" f"""model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.""" ) snake_case_ = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(lowercase__ ): snake_case_ = [[context] * 4 for context in examples[context_name]] snake_case_ = examples[question_header_name] snake_case_ = [ [f"""{header} {examples[end][i]}""" for end in ending_names] for i, header in enumerate(lowercase__ ) ] # Flatten out snake_case_ = list(chain(*lowercase__ ) ) snake_case_ = list(chain(*lowercase__ ) ) # Tokenize snake_case_ = tokenizer( lowercase__ , lowercase__ , truncation=lowercase__ , max_length=lowercase__ , padding='max_length' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(lowercase__ ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('--do_train requires a train dataset' ) snake_case_ = raw_datasets['train'] if data_args.max_train_samples is not None: snake_case_ = min(len(lowercase__ ) , data_args.max_train_samples ) snake_case_ = train_dataset.select(range(lowercase__ ) ) with training_args.main_process_first(desc='train dataset map pre-processing' ): snake_case_ = train_dataset.map( lowercase__ , batched=lowercase__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('--do_eval requires a validation dataset' ) snake_case_ = raw_datasets['validation'] if data_args.max_eval_samples is not None: snake_case_ = min(len(lowercase__ ) , data_args.max_eval_samples ) snake_case_ = eval_dataset.select(range(lowercase__ ) ) with training_args.main_process_first(desc='validation dataset map pre-processing' ): snake_case_ = eval_dataset.map( lowercase__ , batched=lowercase__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator snake_case_ = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=lowercase__ , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(lowercase__ ): snake_case_ , snake_case_ = eval_predictions snake_case_ = np.argmax(lowercase__ , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer snake_case_ = Trainer( model=lowercase__ , args=lowercase__ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=lowercase__ , data_collator=lowercase__ , compute_metrics=lowercase__ , ) # Training if training_args.do_train: snake_case_ = None if training_args.resume_from_checkpoint is not None: snake_case_ = training_args.resume_from_checkpoint elif last_checkpoint is not None: snake_case_ = last_checkpoint snake_case_ = trainer.train(resume_from_checkpoint=lowercase__ ) trainer.save_model() # Saves the tokenizer too for easy upload snake_case_ = train_result.metrics snake_case_ = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowercase__ ) ) snake_case_ = min(lowercase__ , len(lowercase__ ) ) trainer.log_metrics('train' , lowercase__ ) trainer.save_metrics('train' , lowercase__ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) snake_case_ = trainer.evaluate() snake_case_ = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(lowercase__ ) snake_case_ = min(lowercase__ , len(lowercase__ ) ) trainer.log_metrics('eval' , lowercase__ ) trainer.save_metrics('eval' , lowercase__ ) snake_case_ = { 'finetuned_from': model_args.model_name_or_path, 'tasks': 'multiple-choice', 'dataset_tags': 'swag', 'dataset_args': 'regular', 'dataset': 'SWAG', 'language': 'en', } if training_args.push_to_hub: trainer.push_to_hub(**lowercase__ ) else: trainer.create_model_card(**lowercase__ ) def a(lowercase__ ): '''simple docstring''' # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
46
1
import json import os import unittest from transformers import MgpstrTokenizer from transformers.models.mgp_str.tokenization_mgp_str import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class SCREAMING_SNAKE_CASE ( __snake_case , unittest.TestCase ): """simple docstring""" __A = MgpstrTokenizer __A = False __A = {} __A = False def __lowerCAmelCase ( self ): """simple docstring""" super().setUp() # fmt: off snake_case_ = ['[GO]', '[s]', '0', '1', '2', '3', '4', '5', '6', '7', '8', '9', 'a', 'b', 'c', 'd', 'e', 'f', 'g', 'h', 'i', 'j', 'k', 'l', 'm', 'n', 'o', 'p', 'q', 'r', 's', 't', 'u', 'v', 'w', 'x', 'y', 'z'] # fmt: on snake_case_ = dict(zip(__UpperCamelCase , range(len(__UpperCamelCase ) ) ) ) snake_case_ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as fp: fp.write(json.dumps(__UpperCamelCase ) + '\n' ) def __lowerCAmelCase ( self , **__UpperCamelCase ): """simple docstring""" return MgpstrTokenizer.from_pretrained(self.tmpdirname , **__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = 'tester' snake_case_ = 'tester' return input_text, output_text @unittest.skip('MGP-STR always lower cases letters.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.get_tokenizers(do_lower_case=__UpperCamelCase ) for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): snake_case_ = '[SPECIAL_TOKEN]' tokenizer.add_special_tokens({'cls_token': special_token} ) snake_case_ = tokenizer.encode([special_token] , add_special_tokens=__UpperCamelCase ) self.assertEqual(len(__UpperCamelCase ) , 1 ) snake_case_ = tokenizer.decode(__UpperCamelCase , skip_special_tokens=__UpperCamelCase ) self.assertTrue(special_token not in decoded ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): snake_case_ , snake_case_ = self.get_input_output_texts(__UpperCamelCase ) snake_case_ = tokenizer.tokenize(__UpperCamelCase ) snake_case_ = tokenizer.convert_tokens_to_ids(__UpperCamelCase ) snake_case_ = tokenizer.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) snake_case_ = tokenizer.convert_ids_to_tokens(__UpperCamelCase ) self.assertNotEqual(len(__UpperCamelCase ) , 0 ) snake_case_ = tokenizer.decode(__UpperCamelCase ) self.assertIsInstance(__UpperCamelCase , __UpperCamelCase ) self.assertEqual(text_a.replace(' ' , '' ) , __UpperCamelCase ) @unittest.skip('MGP-STR tokenizer only handles one sequence.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip('inputs cannot be pretokenized in MgpstrTokenizer' ) def __lowerCAmelCase ( self ): """simple docstring""" pass
46
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_speech_available, is_torch_available A = { 'configuration_audio_spectrogram_transformer': [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ASTConfig', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'ASTForAudioClassification', 'ASTModel', 'ASTPreTrainedModel', ] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['ASTFeatureExtractor'] if TYPE_CHECKING: from .configuration_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ASTConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ASTForAudioClassification, ASTModel, ASTPreTrainedModel, ) try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_audio_spectrogram_transformer import ASTFeatureExtractor else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
1
from typing import List, Optional, Union import numpy as np import PIL import torch from PIL import Image from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) A = logging.get_logger(__name__) # pylint: disable=invalid-name A = '\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Img2ImgPipeline, KandinskyV22PriorPipeline\n >>> from diffusers.utils import load_image\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16\n ... )\n >>> pipe_prior.to("cuda")\n\n >>> prompt = "A red cartoon frog, 4k"\n >>> image_emb, zero_image_emb = pipe_prior(prompt, return_dict=False)\n\n >>> pipe = KandinskyV22Img2ImgPipeline.from_pretrained(\n ... "kandinsky-community/kandinsky-2-2-decoder", torch_dtype=torch.float16\n ... )\n >>> pipe.to("cuda")\n\n >>> init_image = load_image(\n ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main"\n ... "/kandinsky/frog.png"\n ... )\n\n >>> image = pipe(\n ... image=init_image,\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=100,\n ... strength=0.2,\n ... ).images\n\n >>> image[0].save("red_frog.png")\n ```\n' def a(lowercase__ , lowercase__ , lowercase__=8 ): '''simple docstring''' snake_case_ = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 snake_case_ = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor def a(lowercase__ , lowercase__=512 , lowercase__=512 ): '''simple docstring''' snake_case_ = pil_image.resize((w, h) , resample=Image.BICUBIC , reducing_gap=1 ) snake_case_ = np.array(pil_image.convert('RGB' ) ) snake_case_ = arr.astype(np.floataa ) / 127.5 - 1 snake_case_ = np.transpose(lowercase__ , [2, 0, 1] ) snake_case_ = torch.from_numpy(lowercase__ ).unsqueeze(0 ) return image class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" super().__init__() self.register_modules( unet=__UpperCamelCase , scheduler=__UpperCamelCase , movq=__UpperCamelCase , ) snake_case_ = 2 ** (len(self.movq.config.block_out_channels ) - 1) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = min(int(num_inference_steps * strength ) , __UpperCamelCase ) snake_case_ = max(num_inference_steps - init_timestep , 0 ) snake_case_ = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None ): """simple docstring""" if not isinstance(__UpperCamelCase , (torch.Tensor, PIL.Image.Image, list) ): raise ValueError( f"""`image` has to be of type `torch.Tensor`, `PIL.Image.Image` or list but is {type(__UpperCamelCase )}""" ) snake_case_ = image.to(device=__UpperCamelCase , dtype=__UpperCamelCase ) snake_case_ = batch_size * num_images_per_prompt if image.shape[1] == 4: snake_case_ = image else: if isinstance(__UpperCamelCase , __UpperCamelCase ) and len(__UpperCamelCase ) != batch_size: raise ValueError( f"""You have passed a list of generators of length {len(__UpperCamelCase )}, but requested an effective batch""" f""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) elif isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = [ self.movq.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(__UpperCamelCase ) ] snake_case_ = torch.cat(__UpperCamelCase , dim=0 ) else: snake_case_ = self.movq.encode(__UpperCamelCase ).latent_dist.sample(__UpperCamelCase ) snake_case_ = self.movq.config.scaling_factor * init_latents snake_case_ = torch.cat([init_latents] , dim=0 ) snake_case_ = init_latents.shape snake_case_ = randn_tensor(__UpperCamelCase , generator=__UpperCamelCase , device=__UpperCamelCase , dtype=__UpperCamelCase ) # get latents snake_case_ = self.scheduler.add_noise(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = init_latents return latents def __lowerCAmelCase ( self , __UpperCamelCase=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) snake_case_ = torch.device(f"""cuda:{gpu_id}""" ) snake_case_ = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(__UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase=0 ): """simple docstring""" if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) snake_case_ = torch.device(f"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=__UpperCamelCase ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) snake_case_ = None for cpu_offloaded_model in [self.unet, self.movq]: snake_case_ , snake_case_ = cpu_offload_with_hook(__UpperCamelCase , __UpperCamelCase , prev_module_hook=__UpperCamelCase ) # We'll offload the last model manually. snake_case_ = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def __lowerCAmelCase ( self ): """simple docstring""" if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(__UpperCamelCase , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(__UpperCamelCase ) def __call__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 5_12 , __UpperCamelCase = 5_12 , __UpperCamelCase = 1_00 , __UpperCamelCase = 4.0 , __UpperCamelCase = 0.3 , __UpperCamelCase = 1 , __UpperCamelCase = None , __UpperCamelCase = "pil" , __UpperCamelCase = True , ): """simple docstring""" snake_case_ = self._execution_device snake_case_ = guidance_scale > 1.0 if isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = torch.cat(__UpperCamelCase , dim=0 ) snake_case_ = image_embeds.shape[0] if isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = torch.cat(__UpperCamelCase , dim=0 ) if do_classifier_free_guidance: snake_case_ = image_embeds.repeat_interleave(__UpperCamelCase , dim=0 ) snake_case_ = negative_image_embeds.repeat_interleave(__UpperCamelCase , dim=0 ) snake_case_ = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=__UpperCamelCase ) if not isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = [image] if not all(isinstance(__UpperCamelCase , (PIL.Image.Image, torch.Tensor) ) for i in image ): raise ValueError( f"""Input is in incorrect format: {[type(__UpperCamelCase ) for i in image]}. Currently, we only support PIL image and pytorch tensor""" ) snake_case_ = torch.cat([prepare_image(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) for i in image] , dim=0 ) snake_case_ = image.to(dtype=image_embeds.dtype , device=__UpperCamelCase ) snake_case_ = self.movq.encode(__UpperCamelCase )['latents'] snake_case_ = latents.repeat_interleave(__UpperCamelCase , dim=0 ) self.scheduler.set_timesteps(__UpperCamelCase , device=__UpperCamelCase ) snake_case_ , snake_case_ = self.get_timesteps(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = timesteps[:1].repeat(batch_size * num_images_per_prompt ) snake_case_ , snake_case_ = downscale_height_and_width(__UpperCamelCase , __UpperCamelCase , self.movq_scale_factor ) snake_case_ = self.prepare_latents( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , image_embeds.dtype , __UpperCamelCase , __UpperCamelCase ) for i, t in enumerate(self.progress_bar(__UpperCamelCase ) ): # expand the latents if we are doing classifier free guidance snake_case_ = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents snake_case_ = {'image_embeds': image_embeds} snake_case_ = self.unet( sample=__UpperCamelCase , timestep=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , added_cond_kwargs=__UpperCamelCase , return_dict=__UpperCamelCase , )[0] if do_classifier_free_guidance: snake_case_ , snake_case_ = noise_pred.split(latents.shape[1] , dim=1 ) snake_case_ , snake_case_ = noise_pred.chunk(2 ) snake_case_ , snake_case_ = variance_pred.chunk(2 ) snake_case_ = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) snake_case_ = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): snake_case_ , snake_case_ = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 snake_case_ = self.scheduler.step( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , generator=__UpperCamelCase , )[0] # post-processing snake_case_ = self.movq.decode(__UpperCamelCase , force_not_quantize=__UpperCamelCase )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(f"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: snake_case_ = image * 0.5 + 0.5 snake_case_ = image.clamp(0 , 1 ) snake_case_ = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": snake_case_ = self.numpy_to_pil(__UpperCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__UpperCamelCase )
46
import operator as op def a(lowercase__ ): '''simple docstring''' snake_case_ = [] snake_case_ = lambda lowercase__ , lowercase__ : int(x / y ) # noqa: E731 integer division operation snake_case_ = { '^': op.pow, '*': op.mul, '/': div, '+': op.add, '-': op.sub, } # operators & their respective operation # print table header print('Symbol'.center(8 ) , 'Action'.center(12 ) , 'Stack' , sep=' | ' ) print('-' * (30 + len(lowercase__ )) ) for x in post_fix: if x.isdigit(): # if x in digit stack.append(lowercase__ ) # append x to stack # output in tabular format print(x.rjust(8 ) , ('push(' + x + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) else: snake_case_ = stack.pop() # pop stack # output in tabular format print(''.rjust(8 ) , ('pop(' + b + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) snake_case_ = stack.pop() # pop stack # output in tabular format print(''.rjust(8 ) , ('pop(' + a + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) stack.append( str(opr[x](int(lowercase__ ) , int(lowercase__ ) ) ) ) # evaluate the 2 values popped from stack & push result to stack # output in tabular format print( x.rjust(8 ) , ('push(' + a + x + b + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' , ) return int(stack[0] ) if __name__ == "__main__": A = input('\n\nEnter a Postfix Equation (space separated) = ').split(' ') print('\n\tResult = ', solve(Postfix))
46
1
def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = len(lowercase__ ) snake_case_ = [[0] * n for i in range(lowercase__ )] for i in range(lowercase__ ): snake_case_ = y_points[i] for i in range(2 , lowercase__ ): for j in range(lowercase__ , lowercase__ ): snake_case_ = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
46
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices A = logging.get_logger(__name__) A = { 'google/bit-50': 'https://huggingface.co/google/bit-50/resolve/main/config.json', } class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ): """simple docstring""" __A = """bit""" __A = ["""preactivation""", """bottleneck"""] __A = ["""SAME""", """VALID"""] def __init__( self , __UpperCamelCase=3 , __UpperCamelCase=64 , __UpperCamelCase=[2_56, 5_12, 10_24, 20_48] , __UpperCamelCase=[3, 4, 6, 3] , __UpperCamelCase="preactivation" , __UpperCamelCase="relu" , __UpperCamelCase=None , __UpperCamelCase=32 , __UpperCamelCase=0.0 , __UpperCamelCase=False , __UpperCamelCase=32 , __UpperCamelCase=1 , __UpperCamelCase=None , __UpperCamelCase=None , **__UpperCamelCase , ): """simple docstring""" super().__init__(**__UpperCamelCase ) if layer_type not in self.layer_types: raise ValueError(f"""layer_type={layer_type} is not one of {','.join(self.layer_types )}""" ) if global_padding is not None: if global_padding.upper() in self.supported_padding: snake_case_ = global_padding.upper() else: raise ValueError(f"""Padding strategy {global_padding} not supported""" ) snake_case_ = num_channels snake_case_ = embedding_size snake_case_ = hidden_sizes snake_case_ = depths snake_case_ = layer_type snake_case_ = hidden_act snake_case_ = global_padding snake_case_ = num_groups snake_case_ = drop_path_rate snake_case_ = embedding_dynamic_padding snake_case_ = output_stride snake_case_ = width_factor snake_case_ = ['stem'] + [f"""stage{idx}""" for idx in range(1 , len(__UpperCamelCase ) + 1 )] snake_case_ , snake_case_ = get_aligned_output_features_output_indices( out_features=__UpperCamelCase , out_indices=__UpperCamelCase , stage_names=self.stage_names )
46
1
from __future__ import annotations from typing import Any class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase = 6 ): """simple docstring""" snake_case_ = None snake_case_ = None self.create_linked_list(__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = Node() snake_case_ = current_node snake_case_ = current_node snake_case_ = current_node for _ in range(1 , __UpperCamelCase ): snake_case_ = Node() snake_case_ = current_node snake_case_ = previous_node snake_case_ = current_node snake_case_ = self.front snake_case_ = previous_node def __lowerCAmelCase ( self ): """simple docstring""" return ( self.front == self.rear and self.front is not None and self.front.data is None ) def __lowerCAmelCase ( self ): """simple docstring""" self.check_can_perform_operation() return self.front.data if self.front else None def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" if self.rear is None: return self.check_is_full() if not self.is_empty(): snake_case_ = self.rear.next if self.rear: snake_case_ = data def __lowerCAmelCase ( self ): """simple docstring""" self.check_can_perform_operation() if self.rear is None or self.front is None: return None if self.front == self.rear: snake_case_ = self.front.data snake_case_ = None return data snake_case_ = self.front snake_case_ = old_front.next snake_case_ = old_front.data snake_case_ = None return data def __lowerCAmelCase ( self ): """simple docstring""" if self.is_empty(): raise Exception('Empty Queue' ) def __lowerCAmelCase ( self ): """simple docstring""" if self.rear and self.rear.next == self.front: raise Exception('Full Queue' ) class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self ): """simple docstring""" snake_case_ = None snake_case_ = None snake_case_ = None if __name__ == "__main__": import doctest doctest.testmod()
46
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = VQModel( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=3 , ) return model @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModel(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.dummy_uncond_unet snake_case_ = DDIMScheduler() snake_case_ = self.dummy_vq_model snake_case_ = LDMPipeline(unet=__UpperCamelCase , vqvae=__UpperCamelCase , scheduler=__UpperCamelCase ) ldm.to(__UpperCamelCase ) ldm.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=2 , output_type='numpy' ).images snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=2 , output_type='numpy' , return_dict=__UpperCamelCase )[0] snake_case_ = image[0, -3:, -3:, -1] snake_case_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) snake_case_ = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) snake_case_ = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(__UpperCamelCase ) ldm.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=5 , output_type='numpy' ).images snake_case_ = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) snake_case_ = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) snake_case_ = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
46
1
from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging A = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = ["""pixel_values"""] def __init__( self , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = PILImageResampling.BICUBIC , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = True , __UpperCamelCase = 1 / 2_55 , __UpperCamelCase = True , __UpperCamelCase = IMAGENET_DEFAULT_MEAN , __UpperCamelCase = IMAGENET_DEFAULT_STD , **__UpperCamelCase , ): """simple docstring""" super().__init__(**__UpperCamelCase ) snake_case_ = size if size is not None else {'shortest_edge': 2_24} snake_case_ = get_size_dict(__UpperCamelCase , default_to_square=__UpperCamelCase ) snake_case_ = crop_size if crop_size is not None else {'height': 2_24, 'width': 2_24} snake_case_ = get_size_dict(__UpperCamelCase , param_name='crop_size' ) snake_case_ = do_resize snake_case_ = size snake_case_ = resample snake_case_ = do_center_crop snake_case_ = crop_size snake_case_ = do_rescale snake_case_ = rescale_factor snake_case_ = do_normalize snake_case_ = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN snake_case_ = image_std if image_std is not None else IMAGENET_DEFAULT_STD def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = PILImageResampling.BICUBIC , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" snake_case_ = get_size_dict(__UpperCamelCase , default_to_square=__UpperCamelCase ) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: snake_case_ = int((2_56 / 2_24) * size['shortest_edge'] ) snake_case_ = get_resize_output_image_size(__UpperCamelCase , size=__UpperCamelCase , default_to_square=__UpperCamelCase ) snake_case_ = {'height': output_size[0], 'width': output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( f"""Size dict must have keys 'height' and 'width' or 'shortest_edge'. Got {size_dict.keys()}""" ) return resize( __UpperCamelCase , size=(size_dict['height'], size_dict['width']) , resample=__UpperCamelCase , data_format=__UpperCamelCase , **__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" snake_case_ = get_size_dict(__UpperCamelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""Size dict must have keys 'height' and 'width'. Got {size.keys()}""" ) return center_crop(__UpperCamelCase , size=(size['height'], size['width']) , data_format=__UpperCamelCase , **__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" return rescale(__UpperCamelCase , scale=__UpperCamelCase , data_format=__UpperCamelCase , **__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" return normalize(__UpperCamelCase , mean=__UpperCamelCase , std=__UpperCamelCase , data_format=__UpperCamelCase , **__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = ChannelDimension.FIRST , **__UpperCamelCase , ): """simple docstring""" snake_case_ = do_resize if do_resize is not None else self.do_resize snake_case_ = resample if resample is not None else self.resample snake_case_ = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ = do_rescale if do_rescale is not None else self.do_rescale snake_case_ = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ = do_normalize if do_normalize is not None else self.do_normalize snake_case_ = image_mean if image_mean is not None else self.image_mean snake_case_ = image_std if image_std is not None else self.image_std snake_case_ = size if size is not None else self.size snake_case_ = get_size_dict(__UpperCamelCase , default_to_square=__UpperCamelCase ) snake_case_ = crop_size if crop_size is not None else self.crop_size snake_case_ = get_size_dict(__UpperCamelCase , param_name='crop_size' ) snake_case_ = make_list_of_images(__UpperCamelCase ) if not valid_images(__UpperCamelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. snake_case_ = [to_numpy_array(__UpperCamelCase ) for image in images] if do_resize: snake_case_ = [self.resize(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) for image in images] if do_center_crop: snake_case_ = [self.center_crop(__UpperCamelCase , __UpperCamelCase ) for image in images] if do_rescale: snake_case_ = [self.rescale(__UpperCamelCase , __UpperCamelCase ) for image in images] if do_normalize: snake_case_ = [self.normalize(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) for image in images] snake_case_ = [to_channel_dimension_format(__UpperCamelCase , __UpperCamelCase ) for image in images] snake_case_ = {'pixel_values': images} return BatchFeature(data=__UpperCamelCase , tensor_type=__UpperCamelCase )
46
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class SCREAMING_SNAKE_CASE : """simple docstring""" __A = LEDConfig __A = {} __A = """gelu""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=7 , __UpperCamelCase=True , __UpperCamelCase=False , __UpperCamelCase=99 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=4 , __UpperCamelCase=37 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=20 , __UpperCamelCase=2 , __UpperCamelCase=1 , __UpperCamelCase=0 , __UpperCamelCase=4 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = eos_token_id snake_case_ = pad_token_id snake_case_ = bos_token_id snake_case_ = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after snake_case_ = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests snake_case_ = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) snake_case_ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) snake_case_ = tf.concat([input_ids, eos_tensor] , axis=1 ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) snake_case_ = prepare_led_inputs_dict(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = tf.concat( [tf.zeros_like(__UpperCamelCase )[:, :-1], tf.ones_like(__UpperCamelCase )[:, -1:]] , axis=-1 , ) snake_case_ = global_attention_mask return config, inputs_dict def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = TFLEDModel(config=__UpperCamelCase ).get_decoder() snake_case_ = inputs_dict['input_ids'] snake_case_ = input_ids[:1, :] snake_case_ = inputs_dict['attention_mask'][:1, :] snake_case_ = 1 # first forward pass snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , use_cache=__UpperCamelCase ) snake_case_ , snake_case_ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case_ = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and snake_case_ = tf.concat([input_ids, next_tokens] , axis=-1 ) snake_case_ = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase )[0] snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , past_key_values=__UpperCamelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice snake_case_ = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) snake_case_ = output_from_no_past[:, -3:, random_slice_idx] snake_case_ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__UpperCamelCase , __UpperCamelCase , rtol=1E-3 ) def a(lowercase__ , lowercase__ , lowercase__ , lowercase__=None , lowercase__=None , lowercase__=None , lowercase__=None , ): '''simple docstring''' if attention_mask is None: snake_case_ = tf.cast(tf.math.not_equal(lowercase__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: snake_case_ = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: snake_case_ = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: snake_case_ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () __A = (TFLEDForConditionalGeneration,) if is_tf_available() else () __A = ( { """conversational""": TFLEDForConditionalGeneration, """feature-extraction""": TFLEDModel, """summarization""": TFLEDForConditionalGeneration, """text2text-generation""": TFLEDForConditionalGeneration, """translation""": TFLEDForConditionalGeneration, } if is_tf_available() else {} ) __A = True __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFLEDModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = tf.zeros_like(inputs_dict['attention_mask'] ) snake_case_ = 2 snake_case_ = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict['global_attention_mask'] , ) snake_case_ = True snake_case_ = self.model_tester.seq_length snake_case_ = self.model_tester.encoder_seq_length def check_decoder_attentions_output(__UpperCamelCase ): snake_case_ = outputs.decoder_attentions self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(__UpperCamelCase ): snake_case_ = [t.numpy() for t in outputs.encoder_attentions] snake_case_ = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: snake_case_ = True snake_case_ = False snake_case_ = False snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = len(__UpperCamelCase ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) if self.is_encoder_decoder: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_decoder_attentions_output(__UpperCamelCase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) # Check attention is always last and order is fine snake_case_ = True snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(__UpperCamelCase ) ) self.assertEqual(model.config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) @unittest.skip('LED keeps using potentially symbolic tensors in conditionals and breaks tracing.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" pass def a(lowercase__ ): '''simple docstring''' return tf.constant(lowercase__ , dtype=tf.intaa ) A = 1e-4 @slow @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ).led # change to intended input here snake_case_ = _long_tensor([5_12 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = _long_tensor([1_28 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = prepare_led_inputs_dict(model.config , __UpperCamelCase , __UpperCamelCase ) snake_case_ = model(**__UpperCamelCase )[0] snake_case_ = (1, 10_24, 7_68) self.assertEqual(output.shape , __UpperCamelCase ) # change to expected output here snake_case_ = tf.convert_to_tensor( [[2.3050, 2.8279, 0.6531], [-1.8457, -0.1455, -3.5661], [-1.0186, 0.4586, -2.2043]] , ) tf.debugging.assert_near(output[:, :3, :3] , __UpperCamelCase , atol=1E-3 ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ) # change to intended input here snake_case_ = _long_tensor([5_12 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = _long_tensor([1_28 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = prepare_led_inputs_dict(model.config , __UpperCamelCase , __UpperCamelCase ) snake_case_ = model(**__UpperCamelCase )[0] snake_case_ = (1, 10_24, model.config.vocab_size) self.assertEqual(output.shape , __UpperCamelCase ) # change to expected output here snake_case_ = tf.convert_to_tensor( [[33.6507, 6.4572, 16.8089], [5.8739, -2.4238, 11.2902], [-3.2139, -4.3149, 4.2783]] , ) tf.debugging.assert_near(output[:, :3, :3] , __UpperCamelCase , atol=1E-3 , rtol=1E-3 )
46
1
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=3 , __UpperCamelCase=16 , __UpperCamelCase=[1, 2, 1] , __UpperCamelCase=[2, 2, 4] , __UpperCamelCase=2 , __UpperCamelCase=2.0 , __UpperCamelCase=True , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.1 , __UpperCamelCase="gelu" , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=0.02 , __UpperCamelCase=1E-5 , __UpperCamelCase=True , __UpperCamelCase=None , __UpperCamelCase=True , __UpperCamelCase=10 , __UpperCamelCase=8 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = embed_dim snake_case_ = depths snake_case_ = num_heads snake_case_ = window_size snake_case_ = mlp_ratio snake_case_ = qkv_bias snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = drop_path_rate snake_case_ = hidden_act snake_case_ = use_absolute_embeddings snake_case_ = patch_norm snake_case_ = layer_norm_eps snake_case_ = initializer_range snake_case_ = is_training snake_case_ = scope snake_case_ = use_labels snake_case_ = type_sequence_label_size snake_case_ = encoder_stride def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = self.get_config() return config, pixel_values, labels def __lowerCAmelCase ( self ): """simple docstring""" return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = SwinvaModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) snake_case_ = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) snake_case_ = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = SwinvaForMaskedImageModeling(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images snake_case_ = 1 snake_case_ = SwinvaForMaskedImageModeling(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.type_sequence_label_size snake_case_ = SwinvaForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) __A = ( {"""feature-extraction""": SwinvaModel, """image-classification""": SwinvaForImageClassification} if is_torch_available() else {} ) __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = SwinvaModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , embed_dim=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase , nn.Linear ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = True for model_class in self.all_model_classes: snake_case_ = True snake_case_ = False snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.attentions snake_case_ = len(self.model_tester.depths ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # check that output_attentions also work using config del inputs_dict["output_attentions"] snake_case_ = True snake_case_ = config.window_size**2 snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) snake_case_ = len(__UpperCamelCase ) # Check attention is always last and order is fine snake_case_ = True snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): snake_case_ = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states snake_case_ = 2 self.assertEqual(out_len + added_hidden_states , len(__UpperCamelCase ) ) snake_case_ = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.hidden_states snake_case_ = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # Swinv2 has a different seq_length snake_case_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) snake_case_ = outputs.reshaped_hidden_states self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) snake_case_ , snake_case_ , snake_case_ , snake_case_ = reshaped_hidden_states[0].shape snake_case_ = ( reshaped_hidden_states[0].view(__UpperCamelCase , __UpperCamelCase , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) snake_case_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) snake_case_ = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = SwinvaModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = _config_zero_init(__UpperCamelCase ) for model_class in self.all_model_classes: snake_case_ = model_class(config=__UpperCamelCase ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @require_vision @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( __UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([-0.3947, -0.4306, 0.0026] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) )
46
from collections import defaultdict def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = first_str.lower().strip() snake_case_ = second_str.lower().strip() # Remove whitespace snake_case_ = first_str.replace(' ' , '' ) snake_case_ = second_str.replace(' ' , '' ) # Strings of different lengths are not anagrams if len(lowercase__ ) != len(lowercase__ ): return False # Default values for count should be 0 snake_case_ = defaultdict(lowercase__ ) # For each character in input strings, # increment count in the corresponding for i in range(len(lowercase__ ) ): count[first_str[i]] += 1 count[second_str[i]] -= 1 return all(_count == 0 for _count in count.values() ) if __name__ == "__main__": from doctest import testmod testmod() A = input('Enter the first string ').strip() A = input('Enter the second string ').strip() A = check_anagrams(input_a, input_b) print(f"""{input_a} and {input_b} are {"" if status else "not "}anagrams.""")
46
1
import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_MAPPING, LevitForImageClassification, LevitForImageClassificationWithTeacher, LevitModel, ) from transformers.models.levit.modeling_levit import LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'hidden_sizes' ) ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'num_attention_heads' ) ) class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=64 , __UpperCamelCase=3 , __UpperCamelCase=3 , __UpperCamelCase=2 , __UpperCamelCase=1 , __UpperCamelCase=16 , __UpperCamelCase=[1_28, 2_56, 3_84] , __UpperCamelCase=[4, 6, 8] , __UpperCamelCase=[2, 3, 4] , __UpperCamelCase=[16, 16, 16] , __UpperCamelCase=0 , __UpperCamelCase=[2, 2, 2] , __UpperCamelCase=[2, 2, 2] , __UpperCamelCase=0.02 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=2 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = num_channels snake_case_ = kernel_size snake_case_ = stride snake_case_ = padding snake_case_ = hidden_sizes snake_case_ = num_attention_heads snake_case_ = depths snake_case_ = key_dim snake_case_ = drop_path_rate snake_case_ = patch_size snake_case_ = attention_ratio snake_case_ = mlp_ratio snake_case_ = initializer_range snake_case_ = [ ['Subsample', key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ['Subsample', key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] snake_case_ = is_training snake_case_ = use_labels snake_case_ = num_labels snake_case_ = initializer_range def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.num_labels ) snake_case_ = self.get_config() return config, pixel_values, labels def __lowerCAmelCase ( self ): """simple docstring""" return LevitConfig( image_size=self.image_size , num_channels=self.num_channels , kernel_size=self.kernel_size , stride=self.stride , padding=self.padding , patch_size=self.patch_size , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , depths=self.depths , key_dim=self.key_dim , drop_path_rate=self.drop_path_rate , mlp_ratio=self.mlp_ratio , attention_ratio=self.attention_ratio , initializer_range=self.initializer_range , down_ops=self.down_ops , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = LevitModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) snake_case_ = (self.image_size, self.image_size) snake_case_ , snake_case_ = image_size[0], image_size[1] for _ in range(4 ): snake_case_ = floor(((height + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) snake_case_ = floor(((width + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, ceil(height / 4 ) * ceil(width / 4 ), self.hidden_sizes[-1]) , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = LevitForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( (LevitModel, LevitForImageClassification, LevitForImageClassificationWithTeacher) if is_torch_available() else () ) __A = ( { """feature-extraction""": LevitModel, """image-classification""": (LevitForImageClassification, LevitForImageClassificationWithTeacher), } if is_torch_available() else {} ) __A = False __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = LevitModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase , hidden_size=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowerCAmelCase ( self ): """simple docstring""" return @unittest.skip(reason='Levit does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='Levit does not support input and output embeddings' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='Levit does not output attentions' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" def check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.hidden_states snake_case_ = len(self.model_tester.depths ) + 1 self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) snake_case_ = (self.model_tester.image_size, self.model_tester.image_size) snake_case_ , snake_case_ = image_size[0], image_size[1] for _ in range(4 ): snake_case_ = floor( ( (height + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) snake_case_ = floor( ( (width + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [ height * width, self.model_tester.hidden_sizes[0], ] , ) snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=False ): """simple docstring""" snake_case_ = super()._prepare_for_class(__UpperCamelCase , __UpperCamelCase , return_labels=__UpperCamelCase ) if return_labels: if model_class.__name__ == "LevitForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" if not self.model_tester.is_training: return snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = True for model_class in self.all_model_classes: # LevitForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(__UpperCamelCase ) or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.train() snake_case_ = self._prepare_for_class(__UpperCamelCase , __UpperCamelCase , return_labels=__UpperCamelCase ) snake_case_ = model(**__UpperCamelCase ).loss loss.backward() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return snake_case_ = False snake_case_ = True for model_class in self.all_model_classes: if model_class in get_values(__UpperCamelCase ) or not model_class.supports_gradient_checkpointing: continue # LevitForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "LevitForImageClassificationWithTeacher": continue snake_case_ = model_class(__UpperCamelCase ) model.gradient_checkpointing_enable() model.to(__UpperCamelCase ) model.train() snake_case_ = self._prepare_for_class(__UpperCamelCase , __UpperCamelCase , return_labels=__UpperCamelCase ) snake_case_ = model(**__UpperCamelCase ).loss loss.backward() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = [ {'title': 'multi_label_classification', 'num_labels': 2, 'dtype': torch.float}, {'title': 'single_label_classification', 'num_labels': 1, 'dtype': torch.long}, {'title': 'regression', 'num_labels': 1, 'dtype': torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(__UpperCamelCase ), ] or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=f"""Testing {model_class} with {problem_type['title']}""" ): snake_case_ = problem_type['title'] snake_case_ = problem_type['num_labels'] snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.train() snake_case_ = self._prepare_for_class(__UpperCamelCase , __UpperCamelCase , return_labels=__UpperCamelCase ) if problem_type["num_labels"] > 1: snake_case_ = inputs['labels'].unsqueeze(1 ).repeat(1 , problem_type['num_labels'] ) snake_case_ = inputs['labels'].to(problem_type['dtype'] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=__UpperCamelCase ) as warning_list: snake_case_ = model(**__UpperCamelCase ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( f"""Something is going wrong in the regression problem: intercepted {w.message}""" ) loss.backward() @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = LevitModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def a(): '''simple docstring''' snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return LevitImageProcessor.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = LevitForImageClassificationWithTeacher.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( __UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([1.0448, -0.3745, -1.8317] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) )
46
import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.dummy_uncond_unet snake_case_ = ScoreSdeVeScheduler() snake_case_ = ScoreSdeVePipeline(unet=__UpperCamelCase , scheduler=__UpperCamelCase ) sde_ve.to(__UpperCamelCase ) sde_ve.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = sde_ve(num_inference_steps=2 , output_type='numpy' , generator=__UpperCamelCase ).images snake_case_ = torch.manual_seed(0 ) snake_case_ = sde_ve(num_inference_steps=2 , output_type='numpy' , generator=__UpperCamelCase , return_dict=__UpperCamelCase )[ 0 ] snake_case_ = image[0, -3:, -3:, -1] snake_case_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) snake_case_ = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = 'google/ncsnpp-church-256' snake_case_ = UNetaDModel.from_pretrained(__UpperCamelCase ) snake_case_ = ScoreSdeVeScheduler.from_pretrained(__UpperCamelCase ) snake_case_ = ScoreSdeVePipeline(unet=__UpperCamelCase , scheduler=__UpperCamelCase ) sde_ve.to(__UpperCamelCase ) sde_ve.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = sde_ve(num_inference_steps=10 , output_type='numpy' , generator=__UpperCamelCase ).images snake_case_ = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) snake_case_ = np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
46
1
import pyarrow.parquet as pq import pytest from datasets import Audio, Dataset, DatasetDict, Features, NamedSplit, Sequence, Value, config from datasets.features.image import Image from datasets.io.parquet import ParquetDatasetReader, ParquetDatasetWriter, get_writer_batch_size from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def a(lowercase__ , lowercase__ ): '''simple docstring''' assert isinstance(lowercase__ , lowercase__ ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('keep_in_memory' , [False, True] ) def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = tmp_path / 'cache' snake_case_ = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): snake_case_ = ParquetDatasetReader(lowercase__ , cache_dir=lowercase__ , keep_in_memory=lowercase__ ).read() _check_parquet_dataset(lowercase__ , lowercase__ ) @pytest.mark.parametrize( 'features' , [ None, {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'}, {'col_1': 'string', 'col_2': 'string', 'col_3': 'string'}, {'col_1': 'int32', 'col_2': 'int32', 'col_3': 'int32'}, {'col_1': 'float32', 'col_2': 'float32', 'col_3': 'float32'}, ] , ) def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = tmp_path / 'cache' snake_case_ = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} snake_case_ = features.copy() if features else default_expected_features snake_case_ = ( Features({feature: Value(lowercase__ ) for feature, dtype in features.items()} ) if features is not None else None ) snake_case_ = ParquetDatasetReader(lowercase__ , features=lowercase__ , cache_dir=lowercase__ ).read() _check_parquet_dataset(lowercase__ , lowercase__ ) @pytest.mark.parametrize('split' , [None, NamedSplit('train' ), 'train', 'test'] ) def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = tmp_path / 'cache' snake_case_ = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} snake_case_ = ParquetDatasetReader(lowercase__ , cache_dir=lowercase__ , split=lowercase__ ).read() _check_parquet_dataset(lowercase__ , lowercase__ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('path_type' , [str, list] ) def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' if issubclass(lowercase__ , lowercase__ ): snake_case_ = parquet_path elif issubclass(lowercase__ , lowercase__ ): snake_case_ = [parquet_path] snake_case_ = tmp_path / 'cache' snake_case_ = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} snake_case_ = ParquetDatasetReader(lowercase__ , cache_dir=lowercase__ ).read() _check_parquet_dataset(lowercase__ , lowercase__ ) def a(lowercase__ , lowercase__ , lowercase__=("train",) ): '''simple docstring''' assert isinstance(lowercase__ , lowercase__ ) for split in splits: snake_case_ = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('keep_in_memory' , [False, True] ) def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = tmp_path / 'cache' snake_case_ = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): snake_case_ = ParquetDatasetReader( {'train': parquet_path} , cache_dir=lowercase__ , keep_in_memory=lowercase__ ).read() _check_parquet_datasetdict(lowercase__ , lowercase__ ) @pytest.mark.parametrize( 'features' , [ None, {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'}, {'col_1': 'string', 'col_2': 'string', 'col_3': 'string'}, {'col_1': 'int32', 'col_2': 'int32', 'col_3': 'int32'}, {'col_1': 'float32', 'col_2': 'float32', 'col_3': 'float32'}, ] , ) def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = tmp_path / 'cache' snake_case_ = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} snake_case_ = features.copy() if features else default_expected_features snake_case_ = ( Features({feature: Value(lowercase__ ) for feature, dtype in features.items()} ) if features is not None else None ) snake_case_ = ParquetDatasetReader({'train': parquet_path} , features=lowercase__ , cache_dir=lowercase__ ).read() _check_parquet_datasetdict(lowercase__ , lowercase__ ) @pytest.mark.parametrize('split' , [None, NamedSplit('train' ), 'train', 'test'] ) def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' if split: snake_case_ = {split: parquet_path} else: snake_case_ = 'train' snake_case_ = {'train': parquet_path, 'test': parquet_path} snake_case_ = tmp_path / 'cache' snake_case_ = {'col_1': 'string', 'col_2': 'int64', 'col_3': 'float64'} snake_case_ = ParquetDatasetReader(lowercase__ , cache_dir=lowercase__ ).read() _check_parquet_datasetdict(lowercase__ , lowercase__ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = ParquetDatasetWriter(lowercase__ , tmp_path / 'foo.parquet' ) assert writer.write() > 0 snake_case_ = pq.ParquetFile(tmp_path / 'foo.parquet' ) snake_case_ = pf.read() assert dataset.data.table == output_table def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = str(shared_datadir / 'test_image_rgb.jpg' ) snake_case_ = {'image': [image_path]} snake_case_ = Features({'image': Image()} ) snake_case_ = Dataset.from_dict(lowercase__ , features=lowercase__ ) snake_case_ = ParquetDatasetWriter(lowercase__ , tmp_path / 'foo.parquet' ) assert writer.write() > 0 snake_case_ = Dataset.from_parquet(str(tmp_path / 'foo.parquet' ) ) assert dataset.features == reloaded_dataset.features snake_case_ = ParquetDatasetReader(str(tmp_path / 'foo.parquet' ) , streaming=lowercase__ ).read() assert dataset.features == reloaded_iterable_dataset.features @pytest.mark.parametrize( 'feature, expected' , [ (Features({'foo': Value('int32' )} ), None), (Features({'image': Image(), 'foo': Value('int32' )} ), config.PARQUET_ROW_GROUP_SIZE_FOR_IMAGE_DATASETS), (Features({'nested': Sequence(Audio() )} ), config.PARQUET_ROW_GROUP_SIZE_FOR_AUDIO_DATASETS), ] , ) def a(lowercase__ , lowercase__ ): '''simple docstring''' assert get_writer_batch_size(lowercase__ ) == expected
46
from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput A = logging.get_logger(__name__) # pylint: disable=invalid-name class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ): """simple docstring""" @register_to_config def __init__( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None ): """simple docstring""" super().__init__() snake_case_ = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" snake_case_ = torch.zeros(__UpperCamelCase , __UpperCamelCase ) else: snake_case_ = None snake_case_ = torch.nn.Parameter(__UpperCamelCase ) class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" super().__init__() self.register_modules( vqvae=__UpperCamelCase , transformer=__UpperCamelCase , text_encoder=__UpperCamelCase , tokenizer=__UpperCamelCase , scheduler=__UpperCamelCase , learned_classifier_free_sampling_embeddings=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = len(__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else 1 # get prompt text embeddings snake_case_ = self.tokenizer( __UpperCamelCase , padding='max_length' , max_length=self.tokenizer.model_max_length , return_tensors='pt' , ) snake_case_ = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: snake_case_ = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( 'The following part of your input was truncated because CLIP can only handle sequences up to' f""" {self.tokenizer.model_max_length} tokens: {removed_text}""" ) snake_case_ = text_input_ids[:, : self.tokenizer.model_max_length] snake_case_ = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 snake_case_ = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate text embeddings for each generation per prompt snake_case_ = prompt_embeds.repeat_interleave(__UpperCamelCase , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: snake_case_ = self.learned_classifier_free_sampling_embeddings.embeddings snake_case_ = negative_prompt_embeds.unsqueeze(0 ).repeat(__UpperCamelCase , 1 , 1 ) else: snake_case_ = [''] * batch_size snake_case_ = text_input_ids.shape[-1] snake_case_ = self.tokenizer( __UpperCamelCase , padding='max_length' , max_length=__UpperCamelCase , truncation=__UpperCamelCase , return_tensors='pt' , ) snake_case_ = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings snake_case_ = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method snake_case_ = negative_prompt_embeds.shape[1] snake_case_ = negative_prompt_embeds.repeat(1 , __UpperCamelCase , 1 ) snake_case_ = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __UpperCamelCase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes snake_case_ = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self , __UpperCamelCase , __UpperCamelCase = 1_00 , __UpperCamelCase = 5.0 , __UpperCamelCase = 1.0 , __UpperCamelCase = 1 , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = "pil" , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = 1 , ): """simple docstring""" if isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = 1 elif isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = len(__UpperCamelCase ) else: raise ValueError(f"""`prompt` has to be of type `str` or `list` but is {type(__UpperCamelCase )}""" ) snake_case_ = batch_size * num_images_per_prompt snake_case_ = guidance_scale > 1.0 snake_case_ = self._encode_prompt(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__UpperCamelCase , __UpperCamelCase ) or callback_steps <= 0) ): raise ValueError( f"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" f""" {type(__UpperCamelCase )}.""" ) # get the initial completely masked latents unless the user supplied it snake_case_ = (batch_size, self.transformer.num_latent_pixels) if latents is None: snake_case_ = self.transformer.num_vector_embeds - 1 snake_case_ = torch.full(__UpperCamelCase , __UpperCamelCase ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {latents_shape}""" ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( 'Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0,' f""" {self.transformer.num_vector_embeds - 1} (inclusive).""" ) snake_case_ = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(__UpperCamelCase , device=self.device ) snake_case_ = self.scheduler.timesteps.to(self.device ) snake_case_ = latents for i, t in enumerate(self.progress_bar(__UpperCamelCase ) ): # expand the sample if we are doing classifier free guidance snake_case_ = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` snake_case_ = self.transformer(__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , timestep=__UpperCamelCase ).sample if do_classifier_free_guidance: snake_case_ , snake_case_ = model_output.chunk(2 ) snake_case_ = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(__UpperCamelCase , dim=1 , keepdim=__UpperCamelCase ) snake_case_ = self.truncate(__UpperCamelCase , __UpperCamelCase ) # remove `log(0)`'s (`-inf`s) snake_case_ = model_output.clamp(-70 ) # compute the previous noisy sample x_t -> x_t-1 snake_case_ = self.scheduler.step(__UpperCamelCase , timestep=__UpperCamelCase , sample=__UpperCamelCase , generator=__UpperCamelCase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = self.vqvae.config.vq_embed_dim snake_case_ = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) snake_case_ = self.vqvae.quantize.get_codebook_entry(__UpperCamelCase , shape=__UpperCamelCase ) snake_case_ = self.vqvae.decode(__UpperCamelCase , force_not_quantize=__UpperCamelCase ).sample snake_case_ = (image / 2 + 0.5).clamp(0 , 1 ) snake_case_ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case_ = self.numpy_to_pil(__UpperCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ , snake_case_ = torch.sort(__UpperCamelCase , 1 , descending=__UpperCamelCase ) snake_case_ = torch.exp(__UpperCamelCase ) snake_case_ = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out snake_case_ = torch.full_like(keep_mask[:, 0:1, :] , __UpperCamelCase ) snake_case_ = torch.cat((all_true, keep_mask) , dim=1 ) snake_case_ = keep_mask[:, :-1, :] snake_case_ = keep_mask.gather(1 , indices.argsort(1 ) ) snake_case_ = log_p_x_0.clone() snake_case_ = -torch.inf # -inf = log(0) return rv
46
1
def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = '' for i in table: res += inp[i - 1] return res def a(lowercase__ ): '''simple docstring''' return data[1:] + data[0] def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = '' for i in range(len(lowercase__ ) ): if a[i] == b[i]: res += "0" else: res += "1" return res def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = int('0b' + data[0] + data[-1] , 2 ) snake_case_ = int('0b' + data[1:3] , 2 ) return bin(s[row][col] )[2:] def a(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = message[:4] snake_case_ = message[4:] snake_case_ = apply_table(lowercase__ , lowercase__ ) snake_case_ = xor(lowercase__ , lowercase__ ) snake_case_ = apply_sbox(lowercase__ , temp[:4] ) # noqa: E741 snake_case_ = apply_sbox(lowercase__ , temp[4:] ) snake_case_ = '0' * (2 - len(lowercase__ )) + l # noqa: E741 snake_case_ = '0' * (2 - len(lowercase__ )) + r snake_case_ = apply_table(l + r , lowercase__ ) snake_case_ = xor(lowercase__ , lowercase__ ) return temp + right if __name__ == "__main__": A = input('Enter 10 bit key: ') A = input('Enter 8 bit message: ') A = [6, 3, 7, 4, 8, 5, 10, 9] A = [3, 5, 2, 7, 4, 10, 1, 9, 8, 6] A = [2, 4, 3, 1] A = [2, 6, 3, 1, 4, 8, 5, 7] A = [4, 1, 3, 5, 7, 2, 8, 6] A = [4, 1, 2, 3, 2, 3, 4, 1] A = [[1, 0, 3, 2], [3, 2, 1, 0], [0, 2, 1, 3], [3, 1, 3, 2]] A = [[0, 1, 2, 3], [2, 0, 1, 3], [3, 0, 1, 0], [2, 1, 0, 3]] # key generation A = apply_table(key, paa_table) A = temp[:5] A = temp[5:] A = left_shift(left) A = left_shift(right) A = apply_table(left + right, pa_table) A = left_shift(left) A = left_shift(right) A = left_shift(left) A = left_shift(right) A = apply_table(left + right, pa_table) # encryption A = apply_table(message, IP) A = function(expansion, sa, sa, keya, temp) A = temp[4:] + temp[:4] A = function(expansion, sa, sa, keya, temp) A = apply_table(temp, IP_inv) print('Cipher text is:', CT) # decryption A = apply_table(CT, IP) A = function(expansion, sa, sa, keya, temp) A = temp[4:] + temp[:4] A = function(expansion, sa, sa, keya, temp) A = apply_table(temp, IP_inv) print('Plain text after decypting is:', PT)
46
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'hidden_sizes' ) ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'neck_hidden_sizes' ) ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'num_attention_heads' ) ) class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=3 , __UpperCamelCase=6_40 , __UpperCamelCase=4 , __UpperCamelCase="silu" , __UpperCamelCase=3 , __UpperCamelCase=32 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.02 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=10 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = last_hidden_size snake_case_ = num_attention_heads snake_case_ = hidden_act snake_case_ = conv_kernel_size snake_case_ = output_stride snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = classifier_dropout_prob snake_case_ = use_labels snake_case_ = is_training snake_case_ = num_labels snake_case_ = initializer_range snake_case_ = scope def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) snake_case_ = self.get_config() return config, pixel_values, labels, pixel_labels def __lowerCAmelCase ( self ): """simple docstring""" return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MobileViTModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MobileViTForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MobileViTForSemanticSegmentation(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __A = ( { """feature-extraction""": MobileViTModel, """image-classification""": MobileViTForImageClassification, """image-segmentation""": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTModelTester(self ) snake_case_ = MobileViTConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='MobileViT does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MobileViT does not support input and output embeddings' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MobileViT does not output attentions' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" def check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.hidden_states snake_case_ = 5 self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. snake_case_ = 2 for i in range(len(__UpperCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = MobileViTModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def a(): '''simple docstring''' snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return MobileViTImageProcessor.from_pretrained('apple/mobilevit-xx-small' ) if is_vision_available() else None @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForImageClassification.from_pretrained('apple/mobilevit-xx-small' ).to(__UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([-1.9364, -1.2327, -0.4653] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = model.to(__UpperCamelCase ) snake_case_ = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) snake_case_ = outputs.logits # verify the logits snake_case_ = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [ [[6.9713, 6.9786, 7.2422], [7.2893, 7.2825, 7.4446], [7.6580, 7.8797, 7.9420]], [[-10.6869, -10.3250, -10.3471], [-10.4228, -9.9868, -9.7132], [-11.0405, -11.0221, -10.7318]], [[-3.3089, -2.8539, -2.6740], [-3.2706, -2.5621, -2.5108], [-3.2534, -2.6615, -2.6651]], ] , device=__UpperCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = model.to(__UpperCamelCase ) snake_case_ = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) snake_case_ = outputs.logits.detach().cpu() snake_case_ = image_processor.post_process_semantic_segmentation(outputs=__UpperCamelCase , target_sizes=[(50, 60)] ) snake_case_ = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , __UpperCamelCase ) snake_case_ = image_processor.post_process_semantic_segmentation(outputs=__UpperCamelCase ) snake_case_ = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , __UpperCamelCase )
46
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_barthez import BarthezTokenizer else: A = None A = logging.get_logger(__name__) A = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} A = { 'vocab_file': { 'moussaKam/mbarthez': 'https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model', 'moussaKam/barthez': 'https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model', 'moussaKam/barthez-orangesum-title': ( 'https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model' ), }, 'tokenizer_file': { 'moussaKam/mbarthez': 'https://huggingface.co/moussaKam/mbarthez/resolve/main/tokenizer.json', 'moussaKam/barthez': 'https://huggingface.co/moussaKam/barthez/resolve/main/tokenizer.json', 'moussaKam/barthez-orangesum-title': ( 'https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/tokenizer.json' ), }, } A = { 'moussaKam/mbarthez': 1024, 'moussaKam/barthez': 1024, 'moussaKam/barthez-orangesum-title': 1024, } A = '▁' class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = VOCAB_FILES_NAMES __A = PRETRAINED_VOCAB_FILES_MAP __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = ["""input_ids""", """attention_mask"""] __A = BarthezTokenizer def __init__( self , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase="<s>" , __UpperCamelCase="</s>" , __UpperCamelCase="</s>" , __UpperCamelCase="<s>" , __UpperCamelCase="<unk>" , __UpperCamelCase="<pad>" , __UpperCamelCase="<mask>" , **__UpperCamelCase , ): """simple docstring""" snake_case_ = AddedToken(__UpperCamelCase , lstrip=__UpperCamelCase , rstrip=__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else mask_token super().__init__( __UpperCamelCase , tokenizer_file=__UpperCamelCase , bos_token=__UpperCamelCase , eos_token=__UpperCamelCase , unk_token=__UpperCamelCase , sep_token=__UpperCamelCase , cls_token=__UpperCamelCase , pad_token=__UpperCamelCase , mask_token=__UpperCamelCase , **__UpperCamelCase , ) snake_case_ = vocab_file snake_case_ = False if not self.vocab_file else True def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] snake_case_ = [self.cls_token_id] snake_case_ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None ): """simple docstring""" snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None ): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(__UpperCamelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return snake_case_ = os.path.join( __UpperCamelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCamelCase ): copyfile(self.vocab_file , __UpperCamelCase ) return (out_vocab_file,)
46
from ..utils import DummyObject, requires_backends class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] )
46
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_speech_available, is_torch_available A = { 'configuration_audio_spectrogram_transformer': [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ASTConfig', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'ASTForAudioClassification', 'ASTModel', 'ASTPreTrainedModel', ] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['ASTFeatureExtractor'] if TYPE_CHECKING: from .configuration_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ASTConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ASTForAudioClassification, ASTModel, ASTPreTrainedModel, ) try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_audio_spectrogram_transformer import ASTFeatureExtractor else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
from __future__ import annotations import os import tempfile import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import is_tensorflow_text_available, is_tf_available from transformers.testing_utils import require_tensorflow_text, require_tf, slow from ..test_modeling_tf_common import floats_tensor from .test_framework_agnostic import GenerationIntegrationTestsMixin if is_tf_available(): import tensorflow as tf from transformers import ( AutoTokenizer, TFAutoModelForCausalLM, TFAutoModelForSeqaSeqLM, TFAutoModelForSpeechSeqaSeq, TFAutoModelForVisionaSeq, TFBartForConditionalGeneration, TFLogitsProcessorList, TFMinLengthLogitsProcessor, tf_top_k_top_p_filtering, ) if is_tensorflow_text_available(): import tensorflow_text as text @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = tf.convert_to_tensor( [ [ 8.222_0991, # 3rd highest value; idx. 0 -0.562_0044, 5.2322_9752, 4.038_6393, -6.879_8378, -0.5478_5802, -3.201_2153, 2.9277_7176, 1.8817_1953, 7.3534_1276, # 5th highest value; idx. 9 8.4320_7833, # 2nd highest value; idx. 10 -9.8571_1836, -5.9620_9236, -1.1303_9161, -7.111_5294, -0.836_9633, -5.318_6408, 7.0642_7407, 0.8136_9344, -0.8202_3817, -5.917_9796, 0.5881_3443, -6.9977_8438, 4.7155_1189, -0.1877_1637, 7.4402_0759, # 4th highest value; idx. 25 9.3845_0987, # 1st highest value; idx. 26 2.1266_2941, -9.3256_2038, 2.3565_2522, ], # cummulative prob of 5 highest values <= 0.6 [ 0.5842_5518, 4.5313_9238, -5.5751_0464, -6.2803_0699, -7.1952_9503, -4.0212_2551, 1.3933_7037, -6.0670_7057, 1.5948_0517, -9.64_3119, 0.0390_7799, 0.6723_1762, -8.8820_6726, 6.2711_5922, # 4th highest value; idx. 13 2.2852_0723, 4.8276_7506, 4.3042_1368, 8.827_5313, # 2nd highest value; idx. 17 5.4402_9958, # 5th highest value; idx. 18 -4.473_5794, 7.3857_9536, # 3rd highest value; idx. 20 -2.9105_1663, 2.6194_6077, -2.567_4762, -9.4895_9302, -4.0292_2645, -1.3541_6918, 9.6770_2323, # 1st highest value; idx. 27 -5.8947_8553, 1.8537_0467, ], # cummulative prob of 5 highest values <= 0.6 ] , dtype=tf.floataa , ) snake_case_ = tf.convert_to_tensor( [[0, 0], [0, 9], [0, 10], [0, 25], [0, 26], [1, 13], [1, 17], [1, 18], [1, 20], [1, 27]] , dtype=tf.intaa , ) # expected non filtered idx as noted above snake_case_ = tf.convert_to_tensor( [8.22_2099, 7.353_4126, 8.43_2078, 7.440_2075, 9.3_8451, 6.27_1159, 8.82_7531, 5.440_2995, 7.385_7956, 9.67_7023] , dtype=tf.floataa , ) # expected non filtered values as noted above snake_case_ = tf_top_k_top_p_filtering(__UpperCamelCase , top_k=10 , top_p=0.6 , min_tokens_to_keep=4 ) snake_case_ = output[output != -float('inf' )] snake_case_ = tf.cast( tf.where(tf.not_equal(__UpperCamelCase , tf.constant(-float('inf' ) , dtype=tf.floataa ) ) ) , dtype=tf.intaa , ) tf.debugging.assert_near(__UpperCamelCase , __UpperCamelCase , rtol=1E-12 ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase , __snake_case ): """simple docstring""" if is_tf_available(): __A = { """AutoModelForCausalLM""": TFAutoModelForCausalLM, """AutoModelForSpeechSeq2Seq""": TFAutoModelForSpeechSeqaSeq, """AutoModelForSeq2SeqLM""": TFAutoModelForSeqaSeqLM, """AutoModelForVision2Seq""": TFAutoModelForVisionaSeq, """LogitsProcessorList""": TFLogitsProcessorList, """MinLengthLogitsProcessor""": TFMinLengthLogitsProcessor, """create_tensor_fn""": tf.convert_to_tensor, """floats_tensor""": floats_tensor, """return_tensors""": """tf""", } @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 2 snake_case_ = 2 class SCREAMING_SNAKE_CASE ( tf.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super(__UpperCamelCase , self ).__init__() snake_case_ = model @tf.function( input_signature=( tf.TensorSpec((None, input_length) , tf.intaa , name='input_ids' ), tf.TensorSpec((None, input_length) , tf.intaa , name='attention_mask' ), ) , jit_compile=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.model.generate( input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase , max_new_tokens=__UpperCamelCase , return_dict_in_generate=__UpperCamelCase , ) return {"sequences": outputs["sequences"]} snake_case_ = [[2, 0], [1_02, 1_03]] snake_case_ = [[1, 0], [1, 1]] snake_case_ = DummyModel(model=__UpperCamelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__UpperCamelCase , __UpperCamelCase , signatures={'serving_default': dummy_model.serving} ) snake_case_ = tf.saved_model.load(__UpperCamelCase ).signatures['serving_default'] for batch_size in range(1 , len(__UpperCamelCase ) + 1 ): snake_case_ = { 'input_ids': tf.constant(dummy_input_ids[:batch_size] ), 'attention_mask': tf.constant(dummy_attention_masks[:batch_size] ), } snake_case_ = serving_func(**__UpperCamelCase )['sequences'] snake_case_ = test_model.generate(**__UpperCamelCase , max_new_tokens=__UpperCamelCase ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 1 snake_case_ = 2 class SCREAMING_SNAKE_CASE ( tf.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super(__UpperCamelCase , self ).__init__() snake_case_ = model @tf.function( input_signature=( tf.TensorSpec((batch_size, None) , tf.intaa , name='input_ids' ), tf.TensorSpec((batch_size, None) , tf.intaa , name='attention_mask' ), ) , jit_compile=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.model.generate( input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase , max_new_tokens=__UpperCamelCase , return_dict_in_generate=__UpperCamelCase , ) return {"sequences": outputs["sequences"]} snake_case_ = [[2], [1_02, 1_03]] snake_case_ = [[1], [1, 1]] snake_case_ = DummyModel(model=__UpperCamelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__UpperCamelCase , __UpperCamelCase , signatures={'serving_default': dummy_model.serving} ) snake_case_ = tf.saved_model.load(__UpperCamelCase ).signatures['serving_default'] for input_row in range(len(__UpperCamelCase ) ): snake_case_ = { 'input_ids': tf.constant([dummy_input_ids[input_row]] ), 'attention_mask': tf.constant([dummy_attention_masks[input_row]] ), } snake_case_ = serving_func(**__UpperCamelCase )['sequences'] snake_case_ = test_model.generate(**__UpperCamelCase , max_new_tokens=__UpperCamelCase ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @slow @require_tensorflow_text def __lowerCAmelCase ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: # file needed to load the TF tokenizer hf_hub_download(repo_id='google/flan-t5-small' , filename='spiece.model' , local_dir=__UpperCamelCase ) class SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__() snake_case_ = text.SentencepieceTokenizer( model=tf.io.gfile.GFile(os.path.join(__UpperCamelCase , 'spiece.model' ) , 'rb' ).read() ) snake_case_ = TFAutoModelForSeqaSeqLM.from_pretrained('hf-internal-testing/tiny-random-t5' ) def __lowerCAmelCase ( self , __UpperCamelCase , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" snake_case_ = self.tokenizer.tokenize(__UpperCamelCase ) snake_case_ , snake_case_ = text.pad_model_inputs( __UpperCamelCase , max_seq_length=64 , pad_value=self.model.config.pad_token_id ) snake_case_ = self.model.generate(input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase ) return self.tokenizer.detokenize(__UpperCamelCase ) snake_case_ = CompleteSentenceTransformer() snake_case_ = tf.keras.layers.Input(shape=(1,) , dtype=tf.string , name='inputs' ) snake_case_ = complete_model(__UpperCamelCase ) snake_case_ = tf.keras.Model(__UpperCamelCase , __UpperCamelCase ) keras_model.save(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = { 'do_sample': True, 'num_beams': 1, 'top_p': 0.7, 'top_k': 10, 'temperature': 0.7, } snake_case_ = 14 snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 'Hello, my dog is cute and' snake_case_ = tokenizer(__UpperCamelCase , return_tensors='tf' ) snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 6_38 # forces the generation to happen on CPU, to avoid GPU-related quirks with tf.device(':/CPU:0' ): tf.random.set_seed(0 ) snake_case_ = model.generate(**__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) snake_case_ = [6_38, 1_98] with tf.device(':/CPU:0' ): tf.random.set_seed(0 ) snake_case_ = model.generate(**__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = 'Hugging Face is a technology company based in New York and Paris.' snake_case_ = bart_tokenizer(__UpperCamelCase , return_tensors='tf' ).input_ids snake_case_ = TFBartForConditionalGeneration.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = bart_model.generate(__UpperCamelCase ).numpy() class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase=None , **__UpperCamelCase ): """simple docstring""" return super().call(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = FakeBart.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = bart_model.generate(__UpperCamelCase , foo='bar' ).numpy() self.assertTrue(np.array_equal(__UpperCamelCase , __UpperCamelCase ) ) class SCREAMING_SNAKE_CASE ( bart_model.model.encoder.__class__ ): """simple docstring""" def __lowerCAmelCase ( self , __UpperCamelCase , **__UpperCamelCase ): """simple docstring""" return super().call(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = FakeEncoder(bart_model.config , bart_model.model.shared ) snake_case_ = fake_encoder # Normal generation still works (the output will be different because the encoder weights are different) snake_case_ = bart_model.generate(__UpperCamelCase ).numpy() with self.assertRaises(__UpperCamelCase ): # FakeEncoder.call() accepts **kwargs -> no filtering -> value error due to unexpected input "foo" bart_model.generate(__UpperCamelCase , foo='bar' )
46
1
from math import isqrt, loga def a(lowercase__ ): '''simple docstring''' snake_case_ = [True] * max_number for i in range(2 , isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2 , lowercase__ , lowercase__ ): snake_case_ = False return [i for i in range(2 , lowercase__ ) if is_prime[i]] def a(lowercase__ = 800800 , lowercase__ = 800800 ): '''simple docstring''' snake_case_ = degree * loga(lowercase__ ) snake_case_ = int(lowercase__ ) snake_case_ = calculate_prime_numbers(lowercase__ ) snake_case_ = 0 snake_case_ = 0 snake_case_ = len(lowercase__ ) - 1 while left < right: while ( prime_numbers[right] * loga(prime_numbers[left] ) + prime_numbers[left] * loga(prime_numbers[right] ) > upper_bound ): right -= 1 hybrid_integers_count += right - left left += 1 return hybrid_integers_count if __name__ == "__main__": print(f"""{solution() = }""")
46
import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=2 , __UpperCamelCase=8 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=99 , __UpperCamelCase=16 , __UpperCamelCase=5 , __UpperCamelCase=2 , __UpperCamelCase=36 , __UpperCamelCase="gelu" , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=5_12 , __UpperCamelCase=16 , __UpperCamelCase=2 , __UpperCamelCase=0.02 , __UpperCamelCase=3 , __UpperCamelCase=4 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_input_mask snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_labels snake_case_ = num_choices snake_case_ = scope def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None if self.use_token_type_ids: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowerCAmelCase ( self ): """simple docstring""" return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.get_config() snake_case_ = 3_00 return config def __lowerCAmelCase ( self ): """simple docstring""" ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = self.prepare_config_and_inputs() snake_case_ = True snake_case_ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase ) snake_case_ = model(__UpperCamelCase , token_type_ids=__UpperCamelCase ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" snake_case_ = True snake_case_ = MraModel(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , encoder_attention_mask=__UpperCamelCase , ) snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , ) snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraForMaskedLM(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraForQuestionAnswering(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , start_positions=__UpperCamelCase , end_positions=__UpperCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MraForSequenceClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MraForTokenClassification(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_choices snake_case_ = MraForMultipleChoice(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = config_and_inputs snake_case_ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , unittest.TestCase ): """simple docstring""" __A = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) __A = False __A = False __A = False __A = False __A = () def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , hidden_size=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: snake_case_ = type self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = MraModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) @unittest.skip(reason='MRA does not output attentions' ) def __lowerCAmelCase ( self ): """simple docstring""" return @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraModel.from_pretrained('uw-madison/mra-base-512-4' ) snake_case_ = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = torch.Size((1, 2_56, 7_68) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[-0.0140, 0.0830, -0.0381], [0.1546, 0.1402, 0.0220], [0.1162, 0.0851, 0.0165]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraForMaskedLM.from_pretrained('uw-madison/mra-base-512-4' ) snake_case_ = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = 5_02_65 snake_case_ = torch.Size((1, 2_56, vocab_size) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[9.2595, -3.6038, 11.8819], [9.3869, -3.2693, 11.0956], [11.8524, -3.4938, 13.1210]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraForMaskedLM.from_pretrained('uw-madison/mra-base-4096-8-d3' ) snake_case_ = torch.arange(40_96 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = 5_02_65 snake_case_ = torch.Size((1, 40_96, vocab_size) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[5.4789, -2.3564, 7.5064], [7.9067, -1.3369, 9.9668], [9.0712, -1.8106, 7.0380]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) )
46
1
import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer A = ['gpt2'] A = 'gpt2' if is_tf_available(): class SCREAMING_SNAKE_CASE ( tf.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super().__init__() snake_case_ = tokenizer snake_case_ = AutoConfig.from_pretrained(__UpperCamelCase ) snake_case_ = TFGPTaLMHeadModel.from_config(__UpperCamelCase ) @tf.function(input_signature=(tf.TensorSpec((None,) , tf.string , name='text' ),) ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = self.tokenizer(__UpperCamelCase ) snake_case_ = tokenized['input_ids'].to_tensor() snake_case_ = tf.cast(input_ids_dense > 0 , tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) snake_case_ = self.model(input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase )['logits'] return outputs @require_tf @require_keras_nlp class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" super().setUp() snake_case_ = [GPTaTokenizer.from_pretrained(__UpperCamelCase ) for checkpoint in (TOKENIZER_CHECKPOINTS)] snake_case_ = [TFGPTaTokenizer.from_pretrained(__UpperCamelCase ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) snake_case_ = [ 'This is a straightforward English test sentence.', 'This one has some weird characters\rto\nsee\r\nif those\u00E9break things.', 'Now we\'re going to add some Chinese: 一 二 三 一二三', 'And some much more rare Chinese: 齉 堃 齉堃', 'Je vais aussi écrire en français pour tester les accents', 'Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ', ] snake_case_ = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def __lowerCAmelCase ( self ): """simple docstring""" for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in self.test_sentences: snake_case_ = tokenizer([test_inputs] , return_tensors='tf' ) snake_case_ = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors snake_case_ = python_outputs[key].numpy() snake_case_ = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(__UpperCamelCase , tf.intaa ) == tf_outputs_values ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for tf_tokenizer in self.tf_tokenizers: snake_case_ = tf.function(__UpperCamelCase ) for test_inputs in self.test_sentences: snake_case_ = tf.constant(__UpperCamelCase ) snake_case_ = compiled_tokenizer(__UpperCamelCase ) snake_case_ = tf_tokenizer(__UpperCamelCase ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for tf_tokenizer in self.tf_tokenizers: snake_case_ = ModelToSave(tokenizer=__UpperCamelCase ) snake_case_ = tf.convert_to_tensor([self.test_sentences[0]] ) snake_case_ = model.serving(__UpperCamelCase ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: snake_case_ = Path(__UpperCamelCase ) / 'saved.model' tf.saved_model.save(__UpperCamelCase , __UpperCamelCase , signatures={'serving_default': model.serving} ) snake_case_ = tf.saved_model.load(__UpperCamelCase ) snake_case_ = loaded_model.signatures['serving_default'](__UpperCamelCase )['output_0'] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for tf_tokenizer in self.tf_tokenizers: snake_case_ = tf.convert_to_tensor([self.test_sentences[0]] ) snake_case_ = tf_tokenizer(__UpperCamelCase ) # Build model with some sample inputs snake_case_ = tf_tokenizer.get_config() snake_case_ = TFGPTaTokenizer.from_config(__UpperCamelCase ) snake_case_ = model_from_config(__UpperCamelCase ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for tf_tokenizer in self.tf_tokenizers: # for the test to run snake_case_ = 12_31_23 for max_length in [3, 5, 10_24]: snake_case_ = tf.convert_to_tensor([self.test_sentences[0]] ) snake_case_ = tf_tokenizer(__UpperCamelCase , max_length=__UpperCamelCase ) snake_case_ = out['input_ids'].numpy().shape[1] assert out_length == max_length
46
import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def a(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file snake_case_ = TapasConfig.from_json_file(lowercase__ ) # set absolute/relative position embeddings parameter snake_case_ = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "WTQ": # run_task_main.py hparams snake_case_ = 4 snake_case_ = True # hparam_utils.py hparams snake_case_ = 0.66_4694 snake_case_ = 0.20_7951 snake_case_ = 0.12_1194 snake_case_ = True snake_case_ = True snake_case_ = False snake_case_ = 0.035_2513 snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams snake_case_ = 4 snake_case_ = False # hparam_utils.py hparams snake_case_ = 36.4519 snake_case_ = 0.90_3421 snake_case_ = 222.088 snake_case_ = True snake_case_ = True snake_case_ = True snake_case_ = 0.76_3141 snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "TABFACT": snake_case_ = TapasForSequenceClassification(config=lowercase__ ) elif task == "MLM": snake_case_ = TapasForMaskedLM(config=lowercase__ ) elif task == "INTERMEDIATE_PRETRAINING": snake_case_ = TapasModel(config=lowercase__ ) else: raise ValueError(f"""Task {task} not supported.""" ) print(f"""Building PyTorch model from configuration: {config}""" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(lowercase__ , lowercase__ , lowercase__ ) # Save pytorch-model (weights and configuration) print(f"""Save PyTorch model to {pytorch_dump_path}""" ) model.save_pretrained(lowercase__ ) # Save tokenizer files print(f"""Save tokenizer files to {pytorch_dump_path}""" ) snake_case_ = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + 'vocab.txt' , model_max_length=512 ) tokenizer.save_pretrained(lowercase__ ) print('Used relative position embeddings:' , model.config.reset_position_index_per_cell ) if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( '--task', default='SQA', type=str, help='Model task for which to convert a checkpoint. Defaults to SQA.' ) parser.add_argument( '--reset_position_index_per_cell', default=False, action='store_true', help='Whether to use relative position embeddings or not. Defaults to True.', ) parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--tapas_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained TAPAS model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) A = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
46
1
# This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny model through reduction of a normal pre-trained model, but keeping the # full vocab, merges file, and thus also resulting in a larger model due to a large vocab size. # This gives ~3MB in total for all files. # # If you want a 50 times smaller than this see `fsmt-make-super-tiny-model.py`, which is slightly more complicated # # # It will be used then as "stas/tiny-wmt19-en-de" # Build from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration A = 'facebook/wmt19-en-de' A = FSMTTokenizer.from_pretrained(mname) # get the correct vocab sizes, etc. from the master model A = FSMTConfig.from_pretrained(mname) config.update( dict( d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) ) A = FSMTForConditionalGeneration(config) print(f"""num of params {tiny_model.num_parameters()}""") # Test A = tokenizer(['Making tiny model'], return_tensors='pt') A = tiny_model(**batch) print('test output:', len(outputs.logits[0])) # Save A = 'tiny-wmt19-en-de' tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(f"""Generated {mname_tiny}""") # Upload # transformers-cli upload tiny-wmt19-en-de
46
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=3 , __UpperCamelCase=16 , __UpperCamelCase=[1, 2, 1] , __UpperCamelCase=[2, 2, 4] , __UpperCamelCase=2 , __UpperCamelCase=2.0 , __UpperCamelCase=True , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.1 , __UpperCamelCase="gelu" , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=0.02 , __UpperCamelCase=1E-5 , __UpperCamelCase=True , __UpperCamelCase=None , __UpperCamelCase=True , __UpperCamelCase=10 , __UpperCamelCase=8 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = embed_dim snake_case_ = depths snake_case_ = num_heads snake_case_ = window_size snake_case_ = mlp_ratio snake_case_ = qkv_bias snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = drop_path_rate snake_case_ = hidden_act snake_case_ = use_absolute_embeddings snake_case_ = patch_norm snake_case_ = layer_norm_eps snake_case_ = initializer_range snake_case_ = is_training snake_case_ = scope snake_case_ = use_labels snake_case_ = type_sequence_label_size snake_case_ = encoder_stride def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = self.get_config() return config, pixel_values, labels def __lowerCAmelCase ( self ): """simple docstring""" return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = SwinvaModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) snake_case_ = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) snake_case_ = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = SwinvaForMaskedImageModeling(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images snake_case_ = 1 snake_case_ = SwinvaForMaskedImageModeling(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.type_sequence_label_size snake_case_ = SwinvaForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) __A = ( {"""feature-extraction""": SwinvaModel, """image-classification""": SwinvaForImageClassification} if is_torch_available() else {} ) __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = SwinvaModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , embed_dim=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase , nn.Linear ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = True for model_class in self.all_model_classes: snake_case_ = True snake_case_ = False snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.attentions snake_case_ = len(self.model_tester.depths ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # check that output_attentions also work using config del inputs_dict["output_attentions"] snake_case_ = True snake_case_ = config.window_size**2 snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) snake_case_ = len(__UpperCamelCase ) # Check attention is always last and order is fine snake_case_ = True snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): snake_case_ = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states snake_case_ = 2 self.assertEqual(out_len + added_hidden_states , len(__UpperCamelCase ) ) snake_case_ = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.hidden_states snake_case_ = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # Swinv2 has a different seq_length snake_case_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) snake_case_ = outputs.reshaped_hidden_states self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) snake_case_ , snake_case_ , snake_case_ , snake_case_ = reshaped_hidden_states[0].shape snake_case_ = ( reshaped_hidden_states[0].view(__UpperCamelCase , __UpperCamelCase , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) snake_case_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) snake_case_ = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = SwinvaModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = _config_zero_init(__UpperCamelCase ) for model_class in self.all_model_classes: snake_case_ = model_class(config=__UpperCamelCase ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @require_vision @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( __UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([-0.3947, -0.4306, 0.0026] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) )
46
1
from __future__ import annotations from random import random class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase = None ): """simple docstring""" snake_case_ = value snake_case_ = random() snake_case_ = None snake_case_ = None def __repr__( self ): """simple docstring""" from pprint import pformat if self.left is None and self.right is None: return f"""'{self.value}: {self.prior:.5}'""" else: return pformat( {f"""{self.value}: {self.prior:.5}""": (self.left, self.right)} , indent=1 ) def __str__( self ): """simple docstring""" snake_case_ = str(self.value ) + ' ' snake_case_ = str(self.left or '' ) snake_case_ = str(self.right or '' ) return value + left + right def a(lowercase__ , lowercase__ ): '''simple docstring''' if root is None: # None tree is split into 2 Nones return None, None elif root.value is None: return None, None else: if value < root.value: snake_case_ , snake_case_ = split(root.left , lowercase__ ) return left, root else: snake_case_ , snake_case_ = split(root.right , lowercase__ ) return root, right def a(lowercase__ , lowercase__ ): '''simple docstring''' if (not left) or (not right): # If one node is None, return the other return left or right elif left.prior < right.prior: snake_case_ = merge(left.right , lowercase__ ) return left else: snake_case_ = merge(lowercase__ , right.left ) return right def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = Node(lowercase__ ) snake_case_ , snake_case_ = split(lowercase__ , lowercase__ ) return merge(merge(lowercase__ , lowercase__ ) , lowercase__ ) def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ , snake_case_ = split(lowercase__ , value - 1 ) snake_case_ , snake_case_ = split(lowercase__ , lowercase__ ) return merge(lowercase__ , lowercase__ ) def a(lowercase__ ): '''simple docstring''' if not root: # None return else: inorder(root.left ) print(root.value , end=',' ) inorder(root.right ) def a(lowercase__ , lowercase__ ): '''simple docstring''' for arg in args.split(): if arg[0] == "+": snake_case_ = insert(lowercase__ , int(arg[1:] ) ) elif arg[0] == "-": snake_case_ = erase(lowercase__ , int(arg[1:] ) ) else: print('Unknown command' ) return root def a(): '''simple docstring''' snake_case_ = None print( 'enter numbers to create a tree, + value to add value into treap, ' '- value to erase all nodes with value. \'q\' to quit. ' ) snake_case_ = input() while args != "q": snake_case_ = interact_treap(lowercase__ , lowercase__ ) print(lowercase__ ) snake_case_ = input() print('good by!' ) if __name__ == "__main__": import doctest doctest.testmod() main()
46
import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline A = argparse.ArgumentParser('Stable Diffusion script with intel optimization', add_help=False) parser.add_argument('--dpm', action='store_true', help='Enable DPMSolver or not') parser.add_argument('--steps', default=None, type=int, help='Num inference steps') A = parser.parse_args() A = 'cpu' A = 'a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings' A = 'path-to-your-trained-model' A = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: A = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) A = pipe.to(device) # to channels last A = pipe.unet.to(memory_format=torch.channels_last) A = pipe.vae.to(memory_format=torch.channels_last) A = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: A = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex A = torch.randn(2, 4, 64, 64) A = torch.rand(1) * 999 A = torch.randn(2, 77, 768) A = (sample, timestep, encoder_hidden_status) try: A = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: A = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) A = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) A = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: A = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute A = 666 A = torch.Generator(device).manual_seed(seed) A = {'generator': generator} if args.steps is not None: A = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): A = pipe(prompt, **generate_kwargs).images[0] # save image image.save('generated.png')
46
1
def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def a(): '''simple docstring''' print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
46
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { 'microsoft/unispeech-sat-base-100h-libri-ft': ( 'https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json' ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = """unispeech-sat""" def __init__( self , __UpperCamelCase=32 , __UpperCamelCase=7_68 , __UpperCamelCase=12 , __UpperCamelCase=12 , __UpperCamelCase=30_72 , __UpperCamelCase="gelu" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.02 , __UpperCamelCase=1E-5 , __UpperCamelCase="group" , __UpperCamelCase="gelu" , __UpperCamelCase=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , __UpperCamelCase=(5, 2, 2, 2, 2, 2, 2) , __UpperCamelCase=(10, 3, 3, 3, 3, 2, 2) , __UpperCamelCase=False , __UpperCamelCase=1_28 , __UpperCamelCase=16 , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=0.05 , __UpperCamelCase=10 , __UpperCamelCase=2 , __UpperCamelCase=0.0 , __UpperCamelCase=10 , __UpperCamelCase=0 , __UpperCamelCase=3_20 , __UpperCamelCase=2 , __UpperCamelCase=0.1 , __UpperCamelCase=1_00 , __UpperCamelCase=2_56 , __UpperCamelCase=2_56 , __UpperCamelCase=0.1 , __UpperCamelCase="mean" , __UpperCamelCase=False , __UpperCamelCase=False , __UpperCamelCase=2_56 , __UpperCamelCase=(5_12, 5_12, 5_12, 5_12, 15_00) , __UpperCamelCase=(5, 3, 3, 1, 1) , __UpperCamelCase=(1, 2, 3, 1, 1) , __UpperCamelCase=5_12 , __UpperCamelCase=0 , __UpperCamelCase=1 , __UpperCamelCase=2 , __UpperCamelCase=5_04 , **__UpperCamelCase , ): """simple docstring""" super().__init__(**__UpperCamelCase , pad_token_id=__UpperCamelCase , bos_token_id=__UpperCamelCase , eos_token_id=__UpperCamelCase ) snake_case_ = hidden_size snake_case_ = feat_extract_norm snake_case_ = feat_extract_activation snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = conv_bias snake_case_ = num_conv_pos_embeddings snake_case_ = num_conv_pos_embedding_groups snake_case_ = len(self.conv_dim ) snake_case_ = num_hidden_layers snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = num_attention_heads snake_case_ = hidden_dropout snake_case_ = attention_dropout snake_case_ = activation_dropout snake_case_ = feat_proj_dropout snake_case_ = final_dropout snake_case_ = layerdrop snake_case_ = layer_norm_eps snake_case_ = initializer_range snake_case_ = vocab_size snake_case_ = num_clusters snake_case_ = do_stable_layer_norm snake_case_ = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( 'Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==' ' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =' f""" {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,""" f""" `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 snake_case_ = apply_spec_augment snake_case_ = mask_time_prob snake_case_ = mask_time_length snake_case_ = mask_time_min_masks snake_case_ = mask_feature_prob snake_case_ = mask_feature_length snake_case_ = mask_feature_min_masks # parameters for pretraining with codevector quantized representations snake_case_ = num_codevectors_per_group snake_case_ = num_codevector_groups snake_case_ = contrastive_logits_temperature snake_case_ = feat_quantizer_dropout snake_case_ = num_negatives snake_case_ = codevector_dim snake_case_ = proj_codevector_dim snake_case_ = diversity_loss_weight # ctc loss snake_case_ = ctc_loss_reduction snake_case_ = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. snake_case_ = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = xvector_output_dim @property def __lowerCAmelCase ( self ): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
46
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) A = { 'configuration_mobilebert': [ 'MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MobileBertConfig', 'MobileBertOnnxConfig', ], 'tokenization_mobilebert': ['MobileBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['MobileBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'MobileBertForMaskedLM', 'MobileBertForMultipleChoice', 'MobileBertForNextSentencePrediction', 'MobileBertForPreTraining', 'MobileBertForQuestionAnswering', 'MobileBertForSequenceClassification', 'MobileBertForTokenClassification', 'MobileBertLayer', 'MobileBertModel', 'MobileBertPreTrainedModel', 'load_tf_weights_in_mobilebert', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFMobileBertForMaskedLM', 'TFMobileBertForMultipleChoice', 'TFMobileBertForNextSentencePrediction', 'TFMobileBertForPreTraining', 'TFMobileBertForQuestionAnswering', 'TFMobileBertForSequenceClassification', 'TFMobileBertForTokenClassification', 'TFMobileBertMainLayer', 'TFMobileBertModel', 'TFMobileBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mobilebert import ( MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileBertConfig, MobileBertOnnxConfig, ) from .tokenization_mobilebert import MobileBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mobilebert_fast import MobileBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilebert import ( MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertLayer, MobileBertModel, MobileBertPreTrainedModel, load_tf_weights_in_mobilebert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilebert import ( TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertMainLayer, TFMobileBertModel, TFMobileBertPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = name snake_case_ = val def __str__( self ): """simple docstring""" return f"""{self.__class__.__name__}({self.name}, {self.val})""" def __lt__( self , __UpperCamelCase ): """simple docstring""" return self.val < other.val class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" snake_case_ = {} snake_case_ = {} snake_case_ = self.build_heap(__UpperCamelCase ) def __getitem__( self , __UpperCamelCase ): """simple docstring""" return self.get_value(__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return (idx - 1) // 2 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return idx * 2 + 1 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return idx * 2 + 2 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return self.heap_dict[key] def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = len(__UpperCamelCase ) - 1 snake_case_ = self.get_parent_idx(__UpperCamelCase ) for idx, i in enumerate(__UpperCamelCase ): snake_case_ = idx snake_case_ = i.val for i in range(__UpperCamelCase , -1 , -1 ): self.sift_down(__UpperCamelCase , __UpperCamelCase ) return array def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" while True: snake_case_ = self.get_left_child_idx(__UpperCamelCase ) # noqa: E741 snake_case_ = self.get_right_child_idx(__UpperCamelCase ) snake_case_ = idx if l < len(__UpperCamelCase ) and array[l] < array[idx]: snake_case_ = l if r < len(__UpperCamelCase ) and array[r] < array[smallest]: snake_case_ = r if smallest != idx: snake_case_ , snake_case_ = array[smallest], array[idx] ( ( snake_case_ ) , ( snake_case_ ) , ) = ( self.idx_of_element[array[smallest]], self.idx_of_element[array[idx]], ) snake_case_ = smallest else: break def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = self.get_parent_idx(__UpperCamelCase ) while p >= 0 and self.heap[p] > self.heap[idx]: snake_case_ , snake_case_ = self.heap[idx], self.heap[p] snake_case_ , snake_case_ = ( self.idx_of_element[self.heap[idx]], self.idx_of_element[self.heap[p]], ) snake_case_ = p snake_case_ = self.get_parent_idx(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" return self.heap[0] def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.heap[-1], self.heap[0] snake_case_ , snake_case_ = ( self.idx_of_element[self.heap[-1]], self.idx_of_element[self.heap[0]], ) snake_case_ = self.heap.pop() del self.idx_of_element[x] self.sift_down(0 , self.heap ) return x def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" self.heap.append(__UpperCamelCase ) snake_case_ = len(self.heap ) - 1 snake_case_ = node.val self.sift_up(len(self.heap ) - 1 ) def __lowerCAmelCase ( self ): """simple docstring""" return len(self.heap ) == 0 def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" assert ( self.heap[self.idx_of_element[node]].val > new_value ), "newValue must be less that current value" snake_case_ = new_value snake_case_ = new_value self.sift_up(self.idx_of_element[node] ) A = Node('R', -1) A = Node('B', 6) A = Node('A', 3) A = Node('X', 1) A = Node('E', 4) # Use one of these two ways to generate Min-Heap # Generating Min-Heap from array A = MinHeap([r, b, a, x, e]) # Generating Min-Heap by Insert method # myMinHeap.insert(a) # myMinHeap.insert(b) # myMinHeap.insert(x) # myMinHeap.insert(r) # myMinHeap.insert(e) # Before print('Min Heap - before decrease key') for i in my_min_heap.heap: print(i) print('Min Heap - After decrease key of node [B -> -17]') my_min_heap.decrease_key(b, -17) # After for i in my_min_heap.heap: print(i) if __name__ == "__main__": import doctest doctest.testmod()
46
1
from __future__ import annotations import os import tempfile import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import is_tensorflow_text_available, is_tf_available from transformers.testing_utils import require_tensorflow_text, require_tf, slow from ..test_modeling_tf_common import floats_tensor from .test_framework_agnostic import GenerationIntegrationTestsMixin if is_tf_available(): import tensorflow as tf from transformers import ( AutoTokenizer, TFAutoModelForCausalLM, TFAutoModelForSeqaSeqLM, TFAutoModelForSpeechSeqaSeq, TFAutoModelForVisionaSeq, TFBartForConditionalGeneration, TFLogitsProcessorList, TFMinLengthLogitsProcessor, tf_top_k_top_p_filtering, ) if is_tensorflow_text_available(): import tensorflow_text as text @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = tf.convert_to_tensor( [ [ 8.222_0991, # 3rd highest value; idx. 0 -0.562_0044, 5.2322_9752, 4.038_6393, -6.879_8378, -0.5478_5802, -3.201_2153, 2.9277_7176, 1.8817_1953, 7.3534_1276, # 5th highest value; idx. 9 8.4320_7833, # 2nd highest value; idx. 10 -9.8571_1836, -5.9620_9236, -1.1303_9161, -7.111_5294, -0.836_9633, -5.318_6408, 7.0642_7407, 0.8136_9344, -0.8202_3817, -5.917_9796, 0.5881_3443, -6.9977_8438, 4.7155_1189, -0.1877_1637, 7.4402_0759, # 4th highest value; idx. 25 9.3845_0987, # 1st highest value; idx. 26 2.1266_2941, -9.3256_2038, 2.3565_2522, ], # cummulative prob of 5 highest values <= 0.6 [ 0.5842_5518, 4.5313_9238, -5.5751_0464, -6.2803_0699, -7.1952_9503, -4.0212_2551, 1.3933_7037, -6.0670_7057, 1.5948_0517, -9.64_3119, 0.0390_7799, 0.6723_1762, -8.8820_6726, 6.2711_5922, # 4th highest value; idx. 13 2.2852_0723, 4.8276_7506, 4.3042_1368, 8.827_5313, # 2nd highest value; idx. 17 5.4402_9958, # 5th highest value; idx. 18 -4.473_5794, 7.3857_9536, # 3rd highest value; idx. 20 -2.9105_1663, 2.6194_6077, -2.567_4762, -9.4895_9302, -4.0292_2645, -1.3541_6918, 9.6770_2323, # 1st highest value; idx. 27 -5.8947_8553, 1.8537_0467, ], # cummulative prob of 5 highest values <= 0.6 ] , dtype=tf.floataa , ) snake_case_ = tf.convert_to_tensor( [[0, 0], [0, 9], [0, 10], [0, 25], [0, 26], [1, 13], [1, 17], [1, 18], [1, 20], [1, 27]] , dtype=tf.intaa , ) # expected non filtered idx as noted above snake_case_ = tf.convert_to_tensor( [8.22_2099, 7.353_4126, 8.43_2078, 7.440_2075, 9.3_8451, 6.27_1159, 8.82_7531, 5.440_2995, 7.385_7956, 9.67_7023] , dtype=tf.floataa , ) # expected non filtered values as noted above snake_case_ = tf_top_k_top_p_filtering(__UpperCamelCase , top_k=10 , top_p=0.6 , min_tokens_to_keep=4 ) snake_case_ = output[output != -float('inf' )] snake_case_ = tf.cast( tf.where(tf.not_equal(__UpperCamelCase , tf.constant(-float('inf' ) , dtype=tf.floataa ) ) ) , dtype=tf.intaa , ) tf.debugging.assert_near(__UpperCamelCase , __UpperCamelCase , rtol=1E-12 ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase , __snake_case ): """simple docstring""" if is_tf_available(): __A = { """AutoModelForCausalLM""": TFAutoModelForCausalLM, """AutoModelForSpeechSeq2Seq""": TFAutoModelForSpeechSeqaSeq, """AutoModelForSeq2SeqLM""": TFAutoModelForSeqaSeqLM, """AutoModelForVision2Seq""": TFAutoModelForVisionaSeq, """LogitsProcessorList""": TFLogitsProcessorList, """MinLengthLogitsProcessor""": TFMinLengthLogitsProcessor, """create_tensor_fn""": tf.convert_to_tensor, """floats_tensor""": floats_tensor, """return_tensors""": """tf""", } @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 2 snake_case_ = 2 class SCREAMING_SNAKE_CASE ( tf.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super(__UpperCamelCase , self ).__init__() snake_case_ = model @tf.function( input_signature=( tf.TensorSpec((None, input_length) , tf.intaa , name='input_ids' ), tf.TensorSpec((None, input_length) , tf.intaa , name='attention_mask' ), ) , jit_compile=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.model.generate( input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase , max_new_tokens=__UpperCamelCase , return_dict_in_generate=__UpperCamelCase , ) return {"sequences": outputs["sequences"]} snake_case_ = [[2, 0], [1_02, 1_03]] snake_case_ = [[1, 0], [1, 1]] snake_case_ = DummyModel(model=__UpperCamelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__UpperCamelCase , __UpperCamelCase , signatures={'serving_default': dummy_model.serving} ) snake_case_ = tf.saved_model.load(__UpperCamelCase ).signatures['serving_default'] for batch_size in range(1 , len(__UpperCamelCase ) + 1 ): snake_case_ = { 'input_ids': tf.constant(dummy_input_ids[:batch_size] ), 'attention_mask': tf.constant(dummy_attention_masks[:batch_size] ), } snake_case_ = serving_func(**__UpperCamelCase )['sequences'] snake_case_ = test_model.generate(**__UpperCamelCase , max_new_tokens=__UpperCamelCase ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 1 snake_case_ = 2 class SCREAMING_SNAKE_CASE ( tf.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super(__UpperCamelCase , self ).__init__() snake_case_ = model @tf.function( input_signature=( tf.TensorSpec((batch_size, None) , tf.intaa , name='input_ids' ), tf.TensorSpec((batch_size, None) , tf.intaa , name='attention_mask' ), ) , jit_compile=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.model.generate( input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase , max_new_tokens=__UpperCamelCase , return_dict_in_generate=__UpperCamelCase , ) return {"sequences": outputs["sequences"]} snake_case_ = [[2], [1_02, 1_03]] snake_case_ = [[1], [1, 1]] snake_case_ = DummyModel(model=__UpperCamelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__UpperCamelCase , __UpperCamelCase , signatures={'serving_default': dummy_model.serving} ) snake_case_ = tf.saved_model.load(__UpperCamelCase ).signatures['serving_default'] for input_row in range(len(__UpperCamelCase ) ): snake_case_ = { 'input_ids': tf.constant([dummy_input_ids[input_row]] ), 'attention_mask': tf.constant([dummy_attention_masks[input_row]] ), } snake_case_ = serving_func(**__UpperCamelCase )['sequences'] snake_case_ = test_model.generate(**__UpperCamelCase , max_new_tokens=__UpperCamelCase ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @slow @require_tensorflow_text def __lowerCAmelCase ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: # file needed to load the TF tokenizer hf_hub_download(repo_id='google/flan-t5-small' , filename='spiece.model' , local_dir=__UpperCamelCase ) class SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__() snake_case_ = text.SentencepieceTokenizer( model=tf.io.gfile.GFile(os.path.join(__UpperCamelCase , 'spiece.model' ) , 'rb' ).read() ) snake_case_ = TFAutoModelForSeqaSeqLM.from_pretrained('hf-internal-testing/tiny-random-t5' ) def __lowerCAmelCase ( self , __UpperCamelCase , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" snake_case_ = self.tokenizer.tokenize(__UpperCamelCase ) snake_case_ , snake_case_ = text.pad_model_inputs( __UpperCamelCase , max_seq_length=64 , pad_value=self.model.config.pad_token_id ) snake_case_ = self.model.generate(input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase ) return self.tokenizer.detokenize(__UpperCamelCase ) snake_case_ = CompleteSentenceTransformer() snake_case_ = tf.keras.layers.Input(shape=(1,) , dtype=tf.string , name='inputs' ) snake_case_ = complete_model(__UpperCamelCase ) snake_case_ = tf.keras.Model(__UpperCamelCase , __UpperCamelCase ) keras_model.save(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = { 'do_sample': True, 'num_beams': 1, 'top_p': 0.7, 'top_k': 10, 'temperature': 0.7, } snake_case_ = 14 snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 'Hello, my dog is cute and' snake_case_ = tokenizer(__UpperCamelCase , return_tensors='tf' ) snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 6_38 # forces the generation to happen on CPU, to avoid GPU-related quirks with tf.device(':/CPU:0' ): tf.random.set_seed(0 ) snake_case_ = model.generate(**__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) snake_case_ = [6_38, 1_98] with tf.device(':/CPU:0' ): tf.random.set_seed(0 ) snake_case_ = model.generate(**__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = 'Hugging Face is a technology company based in New York and Paris.' snake_case_ = bart_tokenizer(__UpperCamelCase , return_tensors='tf' ).input_ids snake_case_ = TFBartForConditionalGeneration.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = bart_model.generate(__UpperCamelCase ).numpy() class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase=None , **__UpperCamelCase ): """simple docstring""" return super().call(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = FakeBart.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = bart_model.generate(__UpperCamelCase , foo='bar' ).numpy() self.assertTrue(np.array_equal(__UpperCamelCase , __UpperCamelCase ) ) class SCREAMING_SNAKE_CASE ( bart_model.model.encoder.__class__ ): """simple docstring""" def __lowerCAmelCase ( self , __UpperCamelCase , **__UpperCamelCase ): """simple docstring""" return super().call(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = FakeEncoder(bart_model.config , bart_model.model.shared ) snake_case_ = fake_encoder # Normal generation still works (the output will be different because the encoder weights are different) snake_case_ = bart_model.generate(__UpperCamelCase ).numpy() with self.assertRaises(__UpperCamelCase ): # FakeEncoder.call() accepts **kwargs -> no filtering -> value error due to unexpected input "foo" bart_model.generate(__UpperCamelCase , foo='bar' )
46
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) A = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['PerceiverFeatureExtractor'] A = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
1
from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def a(lowercase__ ): '''simple docstring''' return {key.lstrip('-' ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def a(): '''simple docstring''' snake_case_ = ArgumentParser( 'HuggingFace Datasets CLI tool' , usage='datasets-cli <command> [<args>]' , allow_abbrev=lowercase__ ) snake_case_ = parser.add_subparsers(help='datasets-cli command helpers' ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(lowercase__ ) EnvironmentCommand.register_subcommand(lowercase__ ) TestCommand.register_subcommand(lowercase__ ) RunBeamCommand.register_subcommand(lowercase__ ) DummyDataCommand.register_subcommand(lowercase__ ) # Parse args snake_case_ , snake_case_ = parser.parse_known_args() if not hasattr(lowercase__ , 'func' ): parser.print_help() exit(1 ) snake_case_ = parse_unknown_args(lowercase__ ) # Run snake_case_ = args.func(lowercase__ , **lowercase__ ) service.run() if __name__ == "__main__": main()
46
def a(lowercase__ , lowercase__ ): '''simple docstring''' if not isinstance(lowercase__ , lowercase__ ): raise ValueError('iterations must be defined as integers' ) if not isinstance(lowercase__ , lowercase__ ) or not number >= 1: raise ValueError( 'starting number must be\n and integer and be more than 0' ) if not iterations >= 1: raise ValueError('Iterations must be done more than 0 times to play FizzBuzz' ) snake_case_ = '' while number <= iterations: if number % 3 == 0: out += "Fizz" if number % 5 == 0: out += "Buzz" if 0 not in (number % 3, number % 5): out += str(lowercase__ ) # print(out) number += 1 out += " " return out if __name__ == "__main__": import doctest doctest.testmod()
46
1
def a(lowercase__ = 50 ): '''simple docstring''' snake_case_ = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(f"""{solution() = }""")
46
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = 1.5 snake_case_ = int(factor * num_class_images ) snake_case_ = ClipClient( url='https://knn.laion.ai/knn-service' , indice_name='laion_400m' , num_images=lowercase__ , aesthetic_weight=0.1 ) os.makedirs(f"""{class_data_dir}/images""" , exist_ok=lowercase__ ) if len(list(Path(f"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: snake_case_ = client.query(text=lowercase__ ) if len(lowercase__ ) >= factor * num_class_images or num_images > 1e4: break else: snake_case_ = int(factor * num_images ) snake_case_ = ClipClient( url='https://knn.laion.ai/knn-service' , indice_name='laion_400m' , num_images=lowercase__ , aesthetic_weight=0.1 , ) snake_case_ = 0 snake_case_ = 0 snake_case_ = tqdm(desc='downloading real regularization images' , total=lowercase__ ) with open(f"""{class_data_dir}/caption.txt""" , 'w' ) as fa, open(f"""{class_data_dir}/urls.txt""" , 'w' ) as fa, open( f"""{class_data_dir}/images.txt""" , 'w' ) as fa: while total < num_class_images: snake_case_ = class_images[count] count += 1 try: snake_case_ = requests.get(images['url'] ) if img.status_code == 200: snake_case_ = Image.open(BytesIO(img.content ) ) with open(f"""{class_data_dir}/images/{total}.jpg""" , 'wb' ) as f: f.write(img.content ) fa.write(images['caption'] + '\n' ) fa.write(images['url'] + '\n' ) fa.write(f"""{class_data_dir}/images/{total}.jpg""" + '\n' ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def a(): '''simple docstring''' snake_case_ = argparse.ArgumentParser('' , add_help=lowercase__ ) parser.add_argument('--class_prompt' , help='text prompt to retrieve images' , required=lowercase__ , type=lowercase__ ) parser.add_argument('--class_data_dir' , help='path to save images' , required=lowercase__ , type=lowercase__ ) parser.add_argument('--num_class_images' , help='number of images to download' , default=200 , type=lowercase__ ) return parser.parse_args() if __name__ == "__main__": A = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
46
1
import operator as op def a(lowercase__ ): '''simple docstring''' snake_case_ = [] snake_case_ = lambda lowercase__ , lowercase__ : int(x / y ) # noqa: E731 integer division operation snake_case_ = { '^': op.pow, '*': op.mul, '/': div, '+': op.add, '-': op.sub, } # operators & their respective operation # print table header print('Symbol'.center(8 ) , 'Action'.center(12 ) , 'Stack' , sep=' | ' ) print('-' * (30 + len(lowercase__ )) ) for x in post_fix: if x.isdigit(): # if x in digit stack.append(lowercase__ ) # append x to stack # output in tabular format print(x.rjust(8 ) , ('push(' + x + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) else: snake_case_ = stack.pop() # pop stack # output in tabular format print(''.rjust(8 ) , ('pop(' + b + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) snake_case_ = stack.pop() # pop stack # output in tabular format print(''.rjust(8 ) , ('pop(' + a + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) stack.append( str(opr[x](int(lowercase__ ) , int(lowercase__ ) ) ) ) # evaluate the 2 values popped from stack & push result to stack # output in tabular format print( x.rjust(8 ) , ('push(' + a + x + b + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' , ) return int(stack[0] ) if __name__ == "__main__": A = input('\n\nEnter a Postfix Equation (space separated) = ').split(' ') print('\n\tResult = ', solve(Postfix))
46
# coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import platform import sys A = '3' print('Python version:', sys.version) print('OS platform:', platform.platform()) print('OS architecture:', platform.machine()) try: import torch print('Torch version:', torch.__version__) print('Cuda available:', torch.cuda.is_available()) print('Cuda version:', torch.version.cuda) print('CuDNN version:', torch.backends.cudnn.version()) print('Number of GPUs available:', torch.cuda.device_count()) except ImportError: print('Torch version:', None) try: import transformers print('transformers version:', transformers.__version__) except ImportError: print('transformers version:', None)
46
1
from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline A = logging.get_logger(__name__) # pylint: disable=invalid-name class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" super().__init__() self.register_modules(unet=__UpperCamelCase , scheduler=__UpperCamelCase ) @torch.no_grad() def __call__( self , __UpperCamelCase = 1 , __UpperCamelCase = 1_00 , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = True , ): """simple docstring""" if audio_length_in_s is None: snake_case_ = self.unet.config.sample_size / self.unet.config.sample_rate snake_case_ = audio_length_in_s * self.unet.config.sample_rate snake_case_ = 2 ** len(self.unet.up_blocks ) if sample_size < 3 * down_scale_factor: raise ValueError( f"""{audio_length_in_s} is too small. Make sure it's bigger or equal to""" f""" {3 * down_scale_factor / self.unet.config.sample_rate}.""" ) snake_case_ = int(__UpperCamelCase ) if sample_size % down_scale_factor != 0: snake_case_ = ( (audio_length_in_s * self.unet.config.sample_rate) // down_scale_factor + 1 ) * down_scale_factor logger.info( f"""{audio_length_in_s} is increased to {sample_size / self.unet.config.sample_rate} so that it can be handled""" f""" by the model. It will be cut to {original_sample_size / self.unet.config.sample_rate} after the denoising""" ' process.' ) snake_case_ = int(__UpperCamelCase ) snake_case_ = next(iter(self.unet.parameters() ) ).dtype snake_case_ = (batch_size, self.unet.config.in_channels, sample_size) if isinstance(__UpperCamelCase , __UpperCamelCase ) and len(__UpperCamelCase ) != batch_size: raise ValueError( f"""You have passed a list of generators of length {len(__UpperCamelCase )}, but requested an effective batch""" f""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) snake_case_ = randn_tensor(__UpperCamelCase , generator=__UpperCamelCase , device=self.device , dtype=__UpperCamelCase ) # set step values self.scheduler.set_timesteps(__UpperCamelCase , device=audio.device ) snake_case_ = self.scheduler.timesteps.to(__UpperCamelCase ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output snake_case_ = self.unet(__UpperCamelCase , __UpperCamelCase ).sample # 2. compute previous image: x_t -> t_t-1 snake_case_ = self.scheduler.step(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ).prev_sample snake_case_ = audio.clamp(-1 , 1 ).float().cpu().numpy() snake_case_ = audio[:, :, :original_sample_size] if not return_dict: return (audio,) return AudioPipelineOutput(audios=__UpperCamelCase )
46
import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') A = logging.getLogger(__name__) @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) __A = field( default=__snake_case , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) __A = field( default=__snake_case , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) __A = field( default=__snake_case , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) __A = field( default=__snake_case , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) __A = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) __A = field( default=__snake_case , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = field(default=__snake_case , metadata={"""help""": """The input training data file (a text file)."""} ) __A = field( default=__snake_case , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) __A = field( default=__snake_case , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) __A = field( default=__snake_case , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) __A = field( default=__snake_case , metadata={ """help""": ( """The maximum total input sequence length after tokenization. If passed, sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) __A = field( default=__snake_case , metadata={ """help""": ( """Whether to pad all samples to the maximum sentence length. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch. More """ """efficient on GPU but very bad for TPU.""" ) } , ) __A = field( default=__snake_case , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) __A = field( default=__snake_case , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def __lowerCAmelCase ( self ): """simple docstring""" if self.train_file is not None: snake_case_ = self.train_file.split('.' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: snake_case_ = self.validation_file.split('.' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = 42 __A = True __A = None __A = None def __call__( self , __UpperCamelCase ): """simple docstring""" snake_case_ = 'label' if 'label' in features[0].keys() else 'labels' snake_case_ = [feature.pop(__UpperCamelCase ) for feature in features] snake_case_ = len(__UpperCamelCase ) snake_case_ = len(features[0]['input_ids'] ) snake_case_ = [ [{k: v[i] for k, v in feature.items()} for i in range(__UpperCamelCase )] for feature in features ] snake_case_ = list(chain(*__UpperCamelCase ) ) snake_case_ = self.tokenizer.pad( __UpperCamelCase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='pt' , ) # Un-flatten snake_case_ = {k: v.view(__UpperCamelCase , __UpperCamelCase , -1 ) for k, v in batch.items()} # Add back labels snake_case_ = torch.tensor(__UpperCamelCase , dtype=torch.intaa ) return batch def a(): '''simple docstring''' # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. snake_case_ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. snake_case_ , snake_case_ , snake_case_ = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: snake_case_ , snake_case_ , snake_case_ = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_swag' , lowercase__ , lowercase__ ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() snake_case_ = training_args.get_process_log_level() logger.setLevel(lowercase__ ) datasets.utils.logging.set_verbosity(lowercase__ ) transformers.utils.logging.set_verbosity(lowercase__ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + f"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(f"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. snake_case_ = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: snake_case_ = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. """ 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: snake_case_ = {} if data_args.train_file is not None: snake_case_ = data_args.train_file if data_args.validation_file is not None: snake_case_ = data_args.validation_file snake_case_ = data_args.train_file.split('.' )[-1] snake_case_ = load_dataset( lowercase__ , data_files=lowercase__ , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. snake_case_ = load_dataset( 'swag' , 'regular' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case_ = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) snake_case_ = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) snake_case_ = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=lowercase__ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. snake_case_ = [f"""ending{i}""" for i in range(4 )] snake_case_ = 'sent1' snake_case_ = 'sent2' if data_args.max_seq_length is None: snake_case_ = tokenizer.model_max_length if max_seq_length > 1024: logger.warning( 'The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value' ' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can' ' override this default with `--block_size xxx`.' ) snake_case_ = 1024 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( f"""The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the""" f"""model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.""" ) snake_case_ = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(lowercase__ ): snake_case_ = [[context] * 4 for context in examples[context_name]] snake_case_ = examples[question_header_name] snake_case_ = [ [f"""{header} {examples[end][i]}""" for end in ending_names] for i, header in enumerate(lowercase__ ) ] # Flatten out snake_case_ = list(chain(*lowercase__ ) ) snake_case_ = list(chain(*lowercase__ ) ) # Tokenize snake_case_ = tokenizer( lowercase__ , lowercase__ , truncation=lowercase__ , max_length=lowercase__ , padding='max_length' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(lowercase__ ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('--do_train requires a train dataset' ) snake_case_ = raw_datasets['train'] if data_args.max_train_samples is not None: snake_case_ = min(len(lowercase__ ) , data_args.max_train_samples ) snake_case_ = train_dataset.select(range(lowercase__ ) ) with training_args.main_process_first(desc='train dataset map pre-processing' ): snake_case_ = train_dataset.map( lowercase__ , batched=lowercase__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('--do_eval requires a validation dataset' ) snake_case_ = raw_datasets['validation'] if data_args.max_eval_samples is not None: snake_case_ = min(len(lowercase__ ) , data_args.max_eval_samples ) snake_case_ = eval_dataset.select(range(lowercase__ ) ) with training_args.main_process_first(desc='validation dataset map pre-processing' ): snake_case_ = eval_dataset.map( lowercase__ , batched=lowercase__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator snake_case_ = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=lowercase__ , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(lowercase__ ): snake_case_ , snake_case_ = eval_predictions snake_case_ = np.argmax(lowercase__ , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer snake_case_ = Trainer( model=lowercase__ , args=lowercase__ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=lowercase__ , data_collator=lowercase__ , compute_metrics=lowercase__ , ) # Training if training_args.do_train: snake_case_ = None if training_args.resume_from_checkpoint is not None: snake_case_ = training_args.resume_from_checkpoint elif last_checkpoint is not None: snake_case_ = last_checkpoint snake_case_ = trainer.train(resume_from_checkpoint=lowercase__ ) trainer.save_model() # Saves the tokenizer too for easy upload snake_case_ = train_result.metrics snake_case_ = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowercase__ ) ) snake_case_ = min(lowercase__ , len(lowercase__ ) ) trainer.log_metrics('train' , lowercase__ ) trainer.save_metrics('train' , lowercase__ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) snake_case_ = trainer.evaluate() snake_case_ = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(lowercase__ ) snake_case_ = min(lowercase__ , len(lowercase__ ) ) trainer.log_metrics('eval' , lowercase__ ) trainer.save_metrics('eval' , lowercase__ ) snake_case_ = { 'finetuned_from': model_args.model_name_or_path, 'tasks': 'multiple-choice', 'dataset_tags': 'swag', 'dataset_args': 'regular', 'dataset': 'SWAG', 'language': 'en', } if training_args.push_to_hub: trainer.push_to_hub(**lowercase__ ) else: trainer.create_model_card(**lowercase__ ) def a(lowercase__ ): '''simple docstring''' # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
46
1
import argparse import os import re import packaging.version A = 'examples/' A = { 'examples': (re.compile(R'^check_min_version\("[^"]+"\)\s*$', re.MULTILINE), 'check_min_version("VERSION")\n'), 'init': (re.compile(R'^__version__\s+=\s+"([^"]+)"\s*$', re.MULTILINE), '__version__ = "VERSION"\n'), 'setup': (re.compile(R'^(\s*)version\s*=\s*"[^"]+",', re.MULTILINE), R'\1version="VERSION",'), 'doc': (re.compile(R'^(\s*)release\s*=\s*"[^"]+"$', re.MULTILINE), 'release = "VERSION"\n'), } A = { 'init': 'src/transformers/__init__.py', 'setup': 'setup.py', } A = 'README.md' def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' with open(lowercase__ , 'r' , encoding='utf-8' , newline='\n' ) as f: snake_case_ = f.read() snake_case_ , snake_case_ = REPLACE_PATTERNS[pattern] snake_case_ = replace.replace('VERSION' , lowercase__ ) snake_case_ = re_pattern.sub(lowercase__ , lowercase__ ) with open(lowercase__ , 'w' , encoding='utf-8' , newline='\n' ) as f: f.write(lowercase__ ) def a(lowercase__ ): '''simple docstring''' for folder, directories, fnames in os.walk(lowercase__ ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('research_projects' ) if "legacy" in directories: directories.remove('legacy' ) for fname in fnames: if fname.endswith('.py' ): update_version_in_file(os.path.join(lowercase__ , lowercase__ ) , lowercase__ , pattern='examples' ) def a(lowercase__ , lowercase__=False ): '''simple docstring''' for pattern, fname in REPLACE_FILES.items(): update_version_in_file(lowercase__ , lowercase__ , lowercase__ ) if not patch: update_version_in_examples(lowercase__ ) def a(): '''simple docstring''' snake_case_ = '🤗 Transformers currently provides the following architectures' snake_case_ = '1. Want to contribute a new model?' with open(lowercase__ , 'r' , encoding='utf-8' , newline='\n' ) as f: snake_case_ = f.readlines() # Find the start of the list. snake_case_ = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 snake_case_ = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('1.' ): snake_case_ = lines[index].replace( 'https://huggingface.co/docs/transformers/main/model_doc' , 'https://huggingface.co/docs/transformers/model_doc' , ) index += 1 with open(lowercase__ , 'w' , encoding='utf-8' , newline='\n' ) as f: f.writelines(lowercase__ ) def a(): '''simple docstring''' with open(REPLACE_FILES['init'] , 'r' ) as f: snake_case_ = f.read() snake_case_ = REPLACE_PATTERNS['init'][0].search(lowercase__ ).groups()[0] return packaging.version.parse(lowercase__ ) def a(lowercase__=False ): '''simple docstring''' snake_case_ = get_version() if patch and default_version.is_devrelease: raise ValueError('Can\'t create a patch version from the dev branch, checkout a released version!' ) if default_version.is_devrelease: snake_case_ = default_version.base_version elif patch: snake_case_ = f"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: snake_case_ = f"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. snake_case_ = input(f"""Which version are you releasing? [{default_version}]""" ) if len(lowercase__ ) == 0: snake_case_ = default_version print(f"""Updating version to {version}.""" ) global_version_update(lowercase__ , patch=lowercase__ ) if not patch: print('Cleaning main README, don\'t forget to run `make fix-copies`.' ) clean_main_ref_in_model_list() def a(): '''simple docstring''' snake_case_ = get_version() snake_case_ = f"""{current_version.major}.{current_version.minor + 1}.0.dev0""" snake_case_ = current_version.base_version # Check with the user we got that right. snake_case_ = input(f"""Which version are we developing now? [{dev_version}]""" ) if len(lowercase__ ) == 0: snake_case_ = dev_version print(f"""Updating version to {version}.""" ) global_version_update(lowercase__ ) print('Cleaning main README, don\'t forget to run `make fix-copies`.' ) clean_main_ref_in_model_list() if __name__ == "__main__": A = argparse.ArgumentParser() parser.add_argument('--post_release', action='store_true', help='Whether this is pre or post release.') parser.add_argument('--patch', action='store_true', help='Whether or not this is a patch release.') A = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('Nothing to do after a patch :-)') else: post_release_work()
46
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_speech_available, is_torch_available A = { 'configuration_audio_spectrogram_transformer': [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ASTConfig', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'ASTForAudioClassification', 'ASTModel', 'ASTPreTrainedModel', ] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['ASTFeatureExtractor'] if TYPE_CHECKING: from .configuration_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ASTConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ASTForAudioClassification, ASTModel, ASTPreTrainedModel, ) try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_audio_spectrogram_transformer import ASTFeatureExtractor else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
1
import math import sys def a(lowercase__ ): '''simple docstring''' if number != int(lowercase__ ): raise ValueError('the value of input must be a natural number' ) if number < 0: raise ValueError('the value of input must not be a negative number' ) if number == 0: return 1 snake_case_ = [-1] * (number + 1) snake_case_ = 0 for i in range(1 , number + 1 ): snake_case_ = sys.maxsize snake_case_ = int(math.sqrt(lowercase__ ) ) for j in range(1 , root + 1 ): snake_case_ = 1 + answers[i - (j**2)] snake_case_ = min(lowercase__ , lowercase__ ) snake_case_ = answer return answers[number] if __name__ == "__main__": import doctest doctest.testmod()
46
import operator as op def a(lowercase__ ): '''simple docstring''' snake_case_ = [] snake_case_ = lambda lowercase__ , lowercase__ : int(x / y ) # noqa: E731 integer division operation snake_case_ = { '^': op.pow, '*': op.mul, '/': div, '+': op.add, '-': op.sub, } # operators & their respective operation # print table header print('Symbol'.center(8 ) , 'Action'.center(12 ) , 'Stack' , sep=' | ' ) print('-' * (30 + len(lowercase__ )) ) for x in post_fix: if x.isdigit(): # if x in digit stack.append(lowercase__ ) # append x to stack # output in tabular format print(x.rjust(8 ) , ('push(' + x + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) else: snake_case_ = stack.pop() # pop stack # output in tabular format print(''.rjust(8 ) , ('pop(' + b + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) snake_case_ = stack.pop() # pop stack # output in tabular format print(''.rjust(8 ) , ('pop(' + a + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) stack.append( str(opr[x](int(lowercase__ ) , int(lowercase__ ) ) ) ) # evaluate the 2 values popped from stack & push result to stack # output in tabular format print( x.rjust(8 ) , ('push(' + a + x + b + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' , ) return int(stack[0] ) if __name__ == "__main__": A = input('\n\nEnter a Postfix Equation (space separated) = ').split(' ') print('\n\tResult = ', solve(Postfix))
46
1
import math class SCREAMING_SNAKE_CASE : """simple docstring""" def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = 0.0 snake_case_ = 0.0 for i in range(len(__UpperCamelCase ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" for i in range(len(__UpperCamelCase ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def a(): '''simple docstring''' # Training Examples ( m, n ) snake_case_ = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) snake_case_ = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training snake_case_ = SelfOrganizingMap() snake_case_ = 3 snake_case_ = 0.5 for _ in range(lowercase__ ): for j in range(len(lowercase__ ) ): # training sample snake_case_ = training_samples[j] # Compute the winning vector snake_case_ = self_organizing_map.get_winner(lowercase__ , lowercase__ ) # Update the winning vector snake_case_ = self_organizing_map.update(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) # classify test sample snake_case_ = [0, 0, 0, 1] snake_case_ = self_organizing_map.get_winner(lowercase__ , lowercase__ ) # results print(f"""Clusters that the test sample belongs to : {winner}""" ) print(f"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
46
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices A = logging.get_logger(__name__) A = { 'google/bit-50': 'https://huggingface.co/google/bit-50/resolve/main/config.json', } class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ): """simple docstring""" __A = """bit""" __A = ["""preactivation""", """bottleneck"""] __A = ["""SAME""", """VALID"""] def __init__( self , __UpperCamelCase=3 , __UpperCamelCase=64 , __UpperCamelCase=[2_56, 5_12, 10_24, 20_48] , __UpperCamelCase=[3, 4, 6, 3] , __UpperCamelCase="preactivation" , __UpperCamelCase="relu" , __UpperCamelCase=None , __UpperCamelCase=32 , __UpperCamelCase=0.0 , __UpperCamelCase=False , __UpperCamelCase=32 , __UpperCamelCase=1 , __UpperCamelCase=None , __UpperCamelCase=None , **__UpperCamelCase , ): """simple docstring""" super().__init__(**__UpperCamelCase ) if layer_type not in self.layer_types: raise ValueError(f"""layer_type={layer_type} is not one of {','.join(self.layer_types )}""" ) if global_padding is not None: if global_padding.upper() in self.supported_padding: snake_case_ = global_padding.upper() else: raise ValueError(f"""Padding strategy {global_padding} not supported""" ) snake_case_ = num_channels snake_case_ = embedding_size snake_case_ = hidden_sizes snake_case_ = depths snake_case_ = layer_type snake_case_ = hidden_act snake_case_ = global_padding snake_case_ = num_groups snake_case_ = drop_path_rate snake_case_ = embedding_dynamic_padding snake_case_ = output_stride snake_case_ = width_factor snake_case_ = ['stem'] + [f"""stage{idx}""" for idx in range(1 , len(__UpperCamelCase ) + 1 )] snake_case_ , snake_case_ = get_aligned_output_features_output_indices( out_features=__UpperCamelCase , out_indices=__UpperCamelCase , stage_names=self.stage_names )
46
1
import torch from transformers import CamembertForMaskedLM, CamembertTokenizer def a(lowercase__ , lowercase__ , lowercase__ , lowercase__=5 ): '''simple docstring''' # Adapted from https://github.com/pytorch/fairseq/blob/master/fairseq/models/roberta/hub_interface.py assert masked_input.count('<mask>' ) == 1 snake_case_ = torch.tensor(tokenizer.encode(lowercase__ , add_special_tokens=lowercase__ ) ).unsqueeze(0 ) # Batch size 1 snake_case_ = model(lowercase__ )[0] # The last hidden-state is the first element of the output tuple snake_case_ = (input_ids.squeeze() == tokenizer.mask_token_id).nonzero().item() snake_case_ = logits[0, masked_index, :] snake_case_ = logits.softmax(dim=0 ) snake_case_ , snake_case_ = prob.topk(k=lowercase__ , dim=0 ) snake_case_ = ' '.join( [tokenizer.convert_ids_to_tokens(indices[i].item() ) for i in range(len(lowercase__ ) )] ) snake_case_ = tokenizer.mask_token snake_case_ = [] for index, predicted_token_bpe in enumerate(topk_predicted_token_bpe.split(' ' ) ): snake_case_ = predicted_token_bpe.replace('\u2581' , ' ' ) if " {0}".format(lowercase__ ) in masked_input: topk_filled_outputs.append( ( masked_input.replace(' {0}'.format(lowercase__ ) , lowercase__ ), values[index].item(), predicted_token, ) ) else: topk_filled_outputs.append( ( masked_input.replace(lowercase__ , lowercase__ ), values[index].item(), predicted_token, ) ) return topk_filled_outputs A = CamembertTokenizer.from_pretrained('camembert-base') A = CamembertForMaskedLM.from_pretrained('camembert-base') model.eval() A = 'Le camembert est <mask> :)' print(fill_mask(masked_input, model, tokenizer, topk=3))
46
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = VQModel( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=3 , ) return model @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModel(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.dummy_uncond_unet snake_case_ = DDIMScheduler() snake_case_ = self.dummy_vq_model snake_case_ = LDMPipeline(unet=__UpperCamelCase , vqvae=__UpperCamelCase , scheduler=__UpperCamelCase ) ldm.to(__UpperCamelCase ) ldm.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=2 , output_type='numpy' ).images snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=2 , output_type='numpy' , return_dict=__UpperCamelCase )[0] snake_case_ = image[0, -3:, -3:, -1] snake_case_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) snake_case_ = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) snake_case_ = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(__UpperCamelCase ) ldm.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=5 , output_type='numpy' ).images snake_case_ = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) snake_case_ = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) snake_case_ = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
46
1
def a(lowercase__ = 1000 ): '''simple docstring''' snake_case_ , snake_case_ = 1, 1 snake_case_ = [] for i in range(1 , n + 1 ): snake_case_ = prev_numerator + 2 * prev_denominator snake_case_ = prev_numerator + prev_denominator if len(str(lowercase__ ) ) > len(str(lowercase__ ) ): result.append(lowercase__ ) snake_case_ = numerator snake_case_ = denominator return len(lowercase__ ) if __name__ == "__main__": print(f"""{solution() = }""")
46
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class SCREAMING_SNAKE_CASE : """simple docstring""" __A = LEDConfig __A = {} __A = """gelu""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=7 , __UpperCamelCase=True , __UpperCamelCase=False , __UpperCamelCase=99 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=4 , __UpperCamelCase=37 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=20 , __UpperCamelCase=2 , __UpperCamelCase=1 , __UpperCamelCase=0 , __UpperCamelCase=4 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = eos_token_id snake_case_ = pad_token_id snake_case_ = bos_token_id snake_case_ = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after snake_case_ = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests snake_case_ = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) snake_case_ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) snake_case_ = tf.concat([input_ids, eos_tensor] , axis=1 ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) snake_case_ = prepare_led_inputs_dict(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = tf.concat( [tf.zeros_like(__UpperCamelCase )[:, :-1], tf.ones_like(__UpperCamelCase )[:, -1:]] , axis=-1 , ) snake_case_ = global_attention_mask return config, inputs_dict def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = TFLEDModel(config=__UpperCamelCase ).get_decoder() snake_case_ = inputs_dict['input_ids'] snake_case_ = input_ids[:1, :] snake_case_ = inputs_dict['attention_mask'][:1, :] snake_case_ = 1 # first forward pass snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , use_cache=__UpperCamelCase ) snake_case_ , snake_case_ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case_ = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and snake_case_ = tf.concat([input_ids, next_tokens] , axis=-1 ) snake_case_ = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase )[0] snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , past_key_values=__UpperCamelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice snake_case_ = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) snake_case_ = output_from_no_past[:, -3:, random_slice_idx] snake_case_ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__UpperCamelCase , __UpperCamelCase , rtol=1E-3 ) def a(lowercase__ , lowercase__ , lowercase__ , lowercase__=None , lowercase__=None , lowercase__=None , lowercase__=None , ): '''simple docstring''' if attention_mask is None: snake_case_ = tf.cast(tf.math.not_equal(lowercase__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: snake_case_ = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: snake_case_ = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: snake_case_ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () __A = (TFLEDForConditionalGeneration,) if is_tf_available() else () __A = ( { """conversational""": TFLEDForConditionalGeneration, """feature-extraction""": TFLEDModel, """summarization""": TFLEDForConditionalGeneration, """text2text-generation""": TFLEDForConditionalGeneration, """translation""": TFLEDForConditionalGeneration, } if is_tf_available() else {} ) __A = True __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFLEDModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = tf.zeros_like(inputs_dict['attention_mask'] ) snake_case_ = 2 snake_case_ = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict['global_attention_mask'] , ) snake_case_ = True snake_case_ = self.model_tester.seq_length snake_case_ = self.model_tester.encoder_seq_length def check_decoder_attentions_output(__UpperCamelCase ): snake_case_ = outputs.decoder_attentions self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(__UpperCamelCase ): snake_case_ = [t.numpy() for t in outputs.encoder_attentions] snake_case_ = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: snake_case_ = True snake_case_ = False snake_case_ = False snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = len(__UpperCamelCase ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) if self.is_encoder_decoder: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_decoder_attentions_output(__UpperCamelCase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) # Check attention is always last and order is fine snake_case_ = True snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(__UpperCamelCase ) ) self.assertEqual(model.config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) @unittest.skip('LED keeps using potentially symbolic tensors in conditionals and breaks tracing.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" pass def a(lowercase__ ): '''simple docstring''' return tf.constant(lowercase__ , dtype=tf.intaa ) A = 1e-4 @slow @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ).led # change to intended input here snake_case_ = _long_tensor([5_12 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = _long_tensor([1_28 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = prepare_led_inputs_dict(model.config , __UpperCamelCase , __UpperCamelCase ) snake_case_ = model(**__UpperCamelCase )[0] snake_case_ = (1, 10_24, 7_68) self.assertEqual(output.shape , __UpperCamelCase ) # change to expected output here snake_case_ = tf.convert_to_tensor( [[2.3050, 2.8279, 0.6531], [-1.8457, -0.1455, -3.5661], [-1.0186, 0.4586, -2.2043]] , ) tf.debugging.assert_near(output[:, :3, :3] , __UpperCamelCase , atol=1E-3 ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ) # change to intended input here snake_case_ = _long_tensor([5_12 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = _long_tensor([1_28 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = prepare_led_inputs_dict(model.config , __UpperCamelCase , __UpperCamelCase ) snake_case_ = model(**__UpperCamelCase )[0] snake_case_ = (1, 10_24, model.config.vocab_size) self.assertEqual(output.shape , __UpperCamelCase ) # change to expected output here snake_case_ = tf.convert_to_tensor( [[33.6507, 6.4572, 16.8089], [5.8739, -2.4238, 11.2902], [-3.2139, -4.3149, 4.2783]] , ) tf.debugging.assert_near(output[:, :3, :3] , __UpperCamelCase , atol=1E-3 , rtol=1E-3 )
46
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging A = logging.get_logger(__name__) A = {'vocab_file': 'sentencepiece.bpe.model'} A = { 'vocab_file': { 'moussaKam/mbarthez': 'https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model', 'moussaKam/barthez': 'https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model', 'moussaKam/barthez-orangesum-title': ( 'https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model' ), }, } A = { 'moussaKam/mbarthez': 1024, 'moussaKam/barthez': 1024, 'moussaKam/barthez-orangesum-title': 1024, } A = '▁' class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = VOCAB_FILES_NAMES __A = PRETRAINED_VOCAB_FILES_MAP __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = ["""input_ids""", """attention_mask"""] def __init__( self , __UpperCamelCase , __UpperCamelCase="<s>" , __UpperCamelCase="</s>" , __UpperCamelCase="</s>" , __UpperCamelCase="<s>" , __UpperCamelCase="<unk>" , __UpperCamelCase="<pad>" , __UpperCamelCase="<mask>" , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" snake_case_ = AddedToken(__UpperCamelCase , lstrip=__UpperCamelCase , rstrip=__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else mask_token snake_case_ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__UpperCamelCase , eos_token=__UpperCamelCase , unk_token=__UpperCamelCase , sep_token=__UpperCamelCase , cls_token=__UpperCamelCase , pad_token=__UpperCamelCase , mask_token=__UpperCamelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCamelCase , ) snake_case_ = vocab_file snake_case_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__UpperCamelCase ) ) snake_case_ = {'<s>': 0, '<pad>': 1, '</s>': 2, '<unk>': 3} snake_case_ = len(self.sp_model ) - 1 snake_case_ = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] snake_case_ = [self.cls_token_id] snake_case_ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCamelCase , token_ids_a=__UpperCamelCase , already_has_special_tokens=__UpperCamelCase ) if token_ids_a is None: return [1] + ([0] * len(__UpperCamelCase )) + [1] return [1] + ([0] * len(__UpperCamelCase )) + [1, 1] + ([0] * len(__UpperCamelCase )) + [1] def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None ): """simple docstring""" snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def __lowerCAmelCase ( self ): """simple docstring""" return len(self.sp_model ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = {self.convert_ids_to_tokens(__UpperCamelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return self.sp_model.encode(__UpperCamelCase , out_type=__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] snake_case_ = self.sp_model.PieceToId(__UpperCamelCase ) return spm_id if spm_id else self.unk_token_id def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = [] snake_case_ = '' snake_case_ = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__UpperCamelCase ) + token snake_case_ = True snake_case_ = [] else: current_sub_tokens.append(__UpperCamelCase ) snake_case_ = False out_string += self.sp_model.decode(__UpperCamelCase ) return out_string.strip() def __getstate__( self ): """simple docstring""" snake_case_ = self.__dict__.copy() snake_case_ = None return state def __setstate__( self , __UpperCamelCase ): """simple docstring""" snake_case_ = d # for backward compatibility if not hasattr(self , 'sp_model_kwargs' ): snake_case_ = {} snake_case_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None ): """simple docstring""" if not os.path.isdir(__UpperCamelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return snake_case_ = os.path.join( __UpperCamelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCamelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCamelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCamelCase , 'wb' ) as fi: snake_case_ = self.sp_model.serialized_model_proto() fi.write(__UpperCamelCase ) return (out_vocab_file,)
46
from collections import defaultdict def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = first_str.lower().strip() snake_case_ = second_str.lower().strip() # Remove whitespace snake_case_ = first_str.replace(' ' , '' ) snake_case_ = second_str.replace(' ' , '' ) # Strings of different lengths are not anagrams if len(lowercase__ ) != len(lowercase__ ): return False # Default values for count should be 0 snake_case_ = defaultdict(lowercase__ ) # For each character in input strings, # increment count in the corresponding for i in range(len(lowercase__ ) ): count[first_str[i]] += 1 count[second_str[i]] -= 1 return all(_count == 0 for _count in count.values() ) if __name__ == "__main__": from doctest import testmod testmod() A = input('Enter the first string ').strip() A = input('Enter the second string ').strip() A = check_anagrams(input_a, input_b) print(f"""{input_a} and {input_b} are {"" if status else "not "}anagrams.""")
46
1
import sys def a(lowercase__ ): '''simple docstring''' snake_case_ = len(lowercase__ ) snake_case_ = [[0 for x in range(lowercase__ )] for x in range(lowercase__ )] snake_case_ = [[0 for x in range(lowercase__ )] for x in range(lowercase__ )] for chain_length in range(2 , lowercase__ ): for a in range(1 , n - chain_length + 1 ): snake_case_ = a + chain_length - 1 snake_case_ = sys.maxsize for c in range(lowercase__ , lowercase__ ): snake_case_ = ( matrix[a][c] + matrix[c + 1][b] + array[a - 1] * array[c] * array[b] ) if cost < matrix[a][b]: snake_case_ = cost snake_case_ = c return matrix, sol def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' if i == j: print('A' + str(lowercase__ ) , end=' ' ) else: print('(' , end=' ' ) print_optiomal_solution(lowercase__ , lowercase__ , optimal_solution[i][j] ) print_optiomal_solution(lowercase__ , optimal_solution[i][j] + 1 , lowercase__ ) print(')' , end=' ' ) def a(): '''simple docstring''' snake_case_ = [30, 35, 15, 5, 10, 20, 25] snake_case_ = len(lowercase__ ) # Size of matrix created from above array will be # 30*35 35*15 15*5 5*10 10*20 20*25 snake_case_ , snake_case_ = matrix_chain_order(lowercase__ ) print('No. of Operation required: ' + str(matrix[1][n - 1] ) ) print_optiomal_solution(lowercase__ , 1 , n - 1 ) if __name__ == "__main__": main()
46
import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.dummy_uncond_unet snake_case_ = ScoreSdeVeScheduler() snake_case_ = ScoreSdeVePipeline(unet=__UpperCamelCase , scheduler=__UpperCamelCase ) sde_ve.to(__UpperCamelCase ) sde_ve.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = sde_ve(num_inference_steps=2 , output_type='numpy' , generator=__UpperCamelCase ).images snake_case_ = torch.manual_seed(0 ) snake_case_ = sde_ve(num_inference_steps=2 , output_type='numpy' , generator=__UpperCamelCase , return_dict=__UpperCamelCase )[ 0 ] snake_case_ = image[0, -3:, -3:, -1] snake_case_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) snake_case_ = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = 'google/ncsnpp-church-256' snake_case_ = UNetaDModel.from_pretrained(__UpperCamelCase ) snake_case_ = ScoreSdeVeScheduler.from_pretrained(__UpperCamelCase ) snake_case_ = ScoreSdeVePipeline(unet=__UpperCamelCase , scheduler=__UpperCamelCase ) sde_ve.to(__UpperCamelCase ) sde_ve.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = sde_ve(num_inference_steps=10 , output_type='numpy' , generator=__UpperCamelCase ).images snake_case_ = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) snake_case_ = np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
46
1
import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, ftp_get, ftp_head, get_from_cache, http_get, http_head, ) A = '\\n Text data.\n Second line of data.' A = 'file' @pytest.fixture(scope='session' ) def a(lowercase__ ): '''simple docstring''' snake_case_ = tmp_path_factory.mktemp('data' ) / (FILE_PATH + '.zstd') snake_case_ = bytes(lowercase__ , 'utf-8' ) with zstd.open(lowercase__ , 'wb' ) as f: f.write(lowercase__ ) return path @pytest.fixture def a(lowercase__ ): '''simple docstring''' with open(os.path.join(tmpfs.local_root_dir , lowercase__ ) , 'w' ) as f: f.write(lowercase__ ) return FILE_PATH @pytest.mark.parametrize('compression_format' , ['gzip', 'xz', 'zstd'] ) def a(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = {'gzip': gz_file, 'xz': xz_file, 'zstd': zstd_path} snake_case_ = input_paths[compression_format] snake_case_ = tmp_path / 'cache' snake_case_ = DownloadConfig(cache_dir=lowercase__ , extract_compressed_file=lowercase__ ) snake_case_ = cached_path(lowercase__ , download_config=lowercase__ ) with open(lowercase__ ) as f: snake_case_ = f.read() with open(lowercase__ ) as f: snake_case_ = f.read() assert extracted_file_content == expected_file_content @pytest.mark.parametrize('default_extracted' , [True, False] ) @pytest.mark.parametrize('default_cache_dir' , [True, False] ) def a(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = 'custom_cache' snake_case_ = 'custom_extracted_dir' snake_case_ = tmp_path / 'custom_extracted_path' if default_extracted: snake_case_ = ('downloads' if default_cache_dir else custom_cache_dir, 'extracted') else: monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_DIR' , lowercase__ ) monkeypatch.setattr('datasets.config.EXTRACTED_DATASETS_PATH' , str(lowercase__ ) ) snake_case_ = custom_extracted_path.parts[-2:] if default_cache_dir else (custom_cache_dir, custom_extracted_dir) snake_case_ = xz_file snake_case_ = ( DownloadConfig(extract_compressed_file=lowercase__ ) if default_cache_dir else DownloadConfig(cache_dir=tmp_path / custom_cache_dir , extract_compressed_file=lowercase__ ) ) snake_case_ = cached_path(lowercase__ , download_config=lowercase__ ) assert Path(lowercase__ ).parent.parts[-2:] == expected def a(lowercase__ ): '''simple docstring''' # absolute path snake_case_ = str(Path(lowercase__ ).resolve() ) assert cached_path(lowercase__ ) == text_file # relative path snake_case_ = str(Path(lowercase__ ).resolve().relative_to(Path(os.getcwd() ) ) ) assert cached_path(lowercase__ ) == text_file def a(lowercase__ ): '''simple docstring''' # absolute path snake_case_ = str(tmp_path.resolve() / '__missing_file__.txt' ) with pytest.raises(lowercase__ ): cached_path(lowercase__ ) # relative path snake_case_ = './__missing_file__.txt' with pytest.raises(lowercase__ ): cached_path(lowercase__ ) def a(lowercase__ ): '''simple docstring''' snake_case_ = get_from_cache(f"""tmp://{tmpfs_file}""" ) with open(lowercase__ ) as f: snake_case_ = f.read() assert output_file_content == FILE_CONTENT @patch('datasets.config.HF_DATASETS_OFFLINE' , lowercase__ ) def a(): '''simple docstring''' with pytest.raises(lowercase__ ): cached_path('https://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE' , lowercase__ ) def a(lowercase__ ): '''simple docstring''' snake_case_ = tmp_path_factory.mktemp('data' ) / 'file.html' with pytest.raises(lowercase__ ): http_get('https://huggingface.co' , temp_file=lowercase__ ) with pytest.raises(lowercase__ ): http_head('https://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE' , lowercase__ ) def a(lowercase__ ): '''simple docstring''' snake_case_ = tmp_path_factory.mktemp('data' ) / 'file.html' with pytest.raises(lowercase__ ): ftp_get('ftp://huggingface.co' , temp_file=lowercase__ ) with pytest.raises(lowercase__ ): ftp_head('ftp://huggingface.co' ) @patch('datasets.config.HF_DATASETS_OFFLINE' , lowercase__ ) def a(lowercase__ ): '''simple docstring''' snake_case_ = tmp_path_factory.mktemp('data' ) / 'file.html' with pytest.raises(lowercase__ ): fsspec_get('s3://huggingface.co' , temp_file=lowercase__ ) with pytest.raises(lowercase__ ): fsspec_head('s3://huggingface.co' )
46
from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput A = logging.get_logger(__name__) # pylint: disable=invalid-name class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ): """simple docstring""" @register_to_config def __init__( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None ): """simple docstring""" super().__init__() snake_case_ = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" snake_case_ = torch.zeros(__UpperCamelCase , __UpperCamelCase ) else: snake_case_ = None snake_case_ = torch.nn.Parameter(__UpperCamelCase ) class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" super().__init__() self.register_modules( vqvae=__UpperCamelCase , transformer=__UpperCamelCase , text_encoder=__UpperCamelCase , tokenizer=__UpperCamelCase , scheduler=__UpperCamelCase , learned_classifier_free_sampling_embeddings=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = len(__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else 1 # get prompt text embeddings snake_case_ = self.tokenizer( __UpperCamelCase , padding='max_length' , max_length=self.tokenizer.model_max_length , return_tensors='pt' , ) snake_case_ = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: snake_case_ = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( 'The following part of your input was truncated because CLIP can only handle sequences up to' f""" {self.tokenizer.model_max_length} tokens: {removed_text}""" ) snake_case_ = text_input_ids[:, : self.tokenizer.model_max_length] snake_case_ = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 snake_case_ = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate text embeddings for each generation per prompt snake_case_ = prompt_embeds.repeat_interleave(__UpperCamelCase , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: snake_case_ = self.learned_classifier_free_sampling_embeddings.embeddings snake_case_ = negative_prompt_embeds.unsqueeze(0 ).repeat(__UpperCamelCase , 1 , 1 ) else: snake_case_ = [''] * batch_size snake_case_ = text_input_ids.shape[-1] snake_case_ = self.tokenizer( __UpperCamelCase , padding='max_length' , max_length=__UpperCamelCase , truncation=__UpperCamelCase , return_tensors='pt' , ) snake_case_ = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings snake_case_ = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method snake_case_ = negative_prompt_embeds.shape[1] snake_case_ = negative_prompt_embeds.repeat(1 , __UpperCamelCase , 1 ) snake_case_ = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __UpperCamelCase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes snake_case_ = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self , __UpperCamelCase , __UpperCamelCase = 1_00 , __UpperCamelCase = 5.0 , __UpperCamelCase = 1.0 , __UpperCamelCase = 1 , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = "pil" , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = 1 , ): """simple docstring""" if isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = 1 elif isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = len(__UpperCamelCase ) else: raise ValueError(f"""`prompt` has to be of type `str` or `list` but is {type(__UpperCamelCase )}""" ) snake_case_ = batch_size * num_images_per_prompt snake_case_ = guidance_scale > 1.0 snake_case_ = self._encode_prompt(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__UpperCamelCase , __UpperCamelCase ) or callback_steps <= 0) ): raise ValueError( f"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" f""" {type(__UpperCamelCase )}.""" ) # get the initial completely masked latents unless the user supplied it snake_case_ = (batch_size, self.transformer.num_latent_pixels) if latents is None: snake_case_ = self.transformer.num_vector_embeds - 1 snake_case_ = torch.full(__UpperCamelCase , __UpperCamelCase ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {latents_shape}""" ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( 'Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0,' f""" {self.transformer.num_vector_embeds - 1} (inclusive).""" ) snake_case_ = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(__UpperCamelCase , device=self.device ) snake_case_ = self.scheduler.timesteps.to(self.device ) snake_case_ = latents for i, t in enumerate(self.progress_bar(__UpperCamelCase ) ): # expand the sample if we are doing classifier free guidance snake_case_ = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` snake_case_ = self.transformer(__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , timestep=__UpperCamelCase ).sample if do_classifier_free_guidance: snake_case_ , snake_case_ = model_output.chunk(2 ) snake_case_ = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(__UpperCamelCase , dim=1 , keepdim=__UpperCamelCase ) snake_case_ = self.truncate(__UpperCamelCase , __UpperCamelCase ) # remove `log(0)`'s (`-inf`s) snake_case_ = model_output.clamp(-70 ) # compute the previous noisy sample x_t -> x_t-1 snake_case_ = self.scheduler.step(__UpperCamelCase , timestep=__UpperCamelCase , sample=__UpperCamelCase , generator=__UpperCamelCase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = self.vqvae.config.vq_embed_dim snake_case_ = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) snake_case_ = self.vqvae.quantize.get_codebook_entry(__UpperCamelCase , shape=__UpperCamelCase ) snake_case_ = self.vqvae.decode(__UpperCamelCase , force_not_quantize=__UpperCamelCase ).sample snake_case_ = (image / 2 + 0.5).clamp(0 , 1 ) snake_case_ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case_ = self.numpy_to_pil(__UpperCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ , snake_case_ = torch.sort(__UpperCamelCase , 1 , descending=__UpperCamelCase ) snake_case_ = torch.exp(__UpperCamelCase ) snake_case_ = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out snake_case_ = torch.full_like(keep_mask[:, 0:1, :] , __UpperCamelCase ) snake_case_ = torch.cat((all_true, keep_mask) , dim=1 ) snake_case_ = keep_mask[:, :-1, :] snake_case_ = keep_mask.gather(1 , indices.argsort(1 ) ) snake_case_ = log_p_x_0.clone() snake_case_ = -torch.inf # -inf = log(0) return rv
46
1
import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = 'ylacombe/bark-small' snake_case_ = tempfile.mkdtemp() snake_case_ = 'en_speaker_1' snake_case_ = 'This is a test string' snake_case_ = 'speaker_embeddings_path.json' snake_case_ = 'speaker_embeddings' def __lowerCAmelCase ( self , **__UpperCamelCase ): """simple docstring""" return AutoTokenizer.from_pretrained(self.checkpoint , **__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.get_tokenizer() snake_case_ = BarkProcessor(tokenizer=__UpperCamelCase ) processor.save_pretrained(self.tmpdirname ) snake_case_ = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) snake_case_ = self.get_tokenizer(bos_token='(BOS)' , eos_token='(EOS)' ) snake_case_ = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token='(BOS)' , eos_token='(EOS)' , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) snake_case_ = 35 snake_case_ = 2 snake_case_ = 8 snake_case_ = { 'semantic_prompt': np.ones(__UpperCamelCase ), 'coarse_prompt': np.ones((nb_codebooks_coarse, seq_len) ), 'fine_prompt': np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset snake_case_ = processor(text=self.input_string , voice_preset=__UpperCamelCase ) snake_case_ = inputs['history_prompt'] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(__UpperCamelCase , np.array([] ) ).tolist() ) # test loading voice preset from npz file snake_case_ = os.path.join(self.tmpdirname , 'file.npz' ) np.savez(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = processor(text=self.input_string , voice_preset=__UpperCamelCase ) snake_case_ = inputs['history_prompt'] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(__UpperCamelCase , np.array([] ) ).tolist() ) # test loading voice preset from the hub snake_case_ = processor(text=self.input_string , voice_preset=self.voice_preset ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.get_tokenizer() snake_case_ = BarkProcessor(tokenizer=__UpperCamelCase ) snake_case_ = processor(text=self.input_string ) snake_case_ = tokenizer( self.input_string , padding='max_length' , max_length=2_56 , add_special_tokens=__UpperCamelCase , return_attention_mask=__UpperCamelCase , return_token_type_ids=__UpperCamelCase , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
46
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'hidden_sizes' ) ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'neck_hidden_sizes' ) ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'num_attention_heads' ) ) class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=3 , __UpperCamelCase=6_40 , __UpperCamelCase=4 , __UpperCamelCase="silu" , __UpperCamelCase=3 , __UpperCamelCase=32 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.02 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=10 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = last_hidden_size snake_case_ = num_attention_heads snake_case_ = hidden_act snake_case_ = conv_kernel_size snake_case_ = output_stride snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = classifier_dropout_prob snake_case_ = use_labels snake_case_ = is_training snake_case_ = num_labels snake_case_ = initializer_range snake_case_ = scope def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) snake_case_ = self.get_config() return config, pixel_values, labels, pixel_labels def __lowerCAmelCase ( self ): """simple docstring""" return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MobileViTModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MobileViTForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MobileViTForSemanticSegmentation(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __A = ( { """feature-extraction""": MobileViTModel, """image-classification""": MobileViTForImageClassification, """image-segmentation""": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTModelTester(self ) snake_case_ = MobileViTConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='MobileViT does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MobileViT does not support input and output embeddings' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MobileViT does not output attentions' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" def check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.hidden_states snake_case_ = 5 self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. snake_case_ = 2 for i in range(len(__UpperCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = MobileViTModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def a(): '''simple docstring''' snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return MobileViTImageProcessor.from_pretrained('apple/mobilevit-xx-small' ) if is_vision_available() else None @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForImageClassification.from_pretrained('apple/mobilevit-xx-small' ).to(__UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([-1.9364, -1.2327, -0.4653] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = model.to(__UpperCamelCase ) snake_case_ = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) snake_case_ = outputs.logits # verify the logits snake_case_ = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [ [[6.9713, 6.9786, 7.2422], [7.2893, 7.2825, 7.4446], [7.6580, 7.8797, 7.9420]], [[-10.6869, -10.3250, -10.3471], [-10.4228, -9.9868, -9.7132], [-11.0405, -11.0221, -10.7318]], [[-3.3089, -2.8539, -2.6740], [-3.2706, -2.5621, -2.5108], [-3.2534, -2.6615, -2.6651]], ] , device=__UpperCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = model.to(__UpperCamelCase ) snake_case_ = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) snake_case_ = outputs.logits.detach().cpu() snake_case_ = image_processor.post_process_semantic_segmentation(outputs=__UpperCamelCase , target_sizes=[(50, 60)] ) snake_case_ = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , __UpperCamelCase ) snake_case_ = image_processor.post_process_semantic_segmentation(outputs=__UpperCamelCase ) snake_case_ = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , __UpperCamelCase )
46
1
def a(lowercase__ = 10 , lowercase__ = 22 ): '''simple docstring''' snake_case_ = range(1 , lowercase__ ) snake_case_ = range(1 , lowercase__ ) return sum( 1 for power in powers for base in bases if len(str(base**power ) ) == power ) if __name__ == "__main__": print(f"""{solution(10, 22) = }""")
46
from ..utils import DummyObject, requires_backends class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] )
46
1
import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=7 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=99 , __UpperCamelCase=32 , __UpperCamelCase=5 , __UpperCamelCase=4 , __UpperCamelCase=37 , __UpperCamelCase="gelu" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=5_12 , __UpperCamelCase=16 , __UpperCamelCase=2 , __UpperCamelCase=0.02 , __UpperCamelCase=3 , __UpperCamelCase=4 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_input_mask snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_labels snake_case_ = num_choices snake_case_ = scope def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None if self.use_token_type_ids: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowerCAmelCase ( self ): """simple docstring""" return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = LlamaModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" snake_case_ = True snake_case_ = LlamaModel(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , encoder_attention_mask=__UpperCamelCase , ) snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , ) snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" snake_case_ = LlamaForCausalLM(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" snake_case_ = True snake_case_ = True snake_case_ = LlamaForCausalLM(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() # first forward pass snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , encoder_attention_mask=__UpperCamelCase , use_cache=__UpperCamelCase , ) snake_case_ = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids snake_case_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case_ = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and snake_case_ = torch.cat([input_ids, next_tokens] , dim=-1 ) snake_case_ = torch.cat([input_mask, next_mask] , dim=-1 ) snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , encoder_attention_mask=__UpperCamelCase , output_hidden_states=__UpperCamelCase , )['hidden_states'][0] snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , encoder_attention_mask=__UpperCamelCase , past_key_values=__UpperCamelCase , output_hidden_states=__UpperCamelCase , )['hidden_states'][0] # select random slice snake_case_ = ids_tensor((1,) , output_from_past.shape[-1] ).item() snake_case_ = output_from_no_past[:, -3:, random_slice_idx].detach() snake_case_ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__UpperCamelCase , __UpperCamelCase , atol=1E-3 ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = config_and_inputs snake_case_ = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () __A = (LlamaForCausalLM,) if is_torch_available() else () __A = ( { """feature-extraction""": LlamaModel, """text-classification""": LlamaForSequenceClassification, """text-generation""": LlamaForCausalLM, """zero-shot""": LlamaForSequenceClassification, } if is_torch_available() else {} ) __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = LlamaModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , hidden_size=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: snake_case_ = type self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = input_dict['input_ids'] snake_case_ = input_ids.ne(1 ).to(__UpperCamelCase ) snake_case_ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) snake_case_ = LlamaForSequenceClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , labels=__UpperCamelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = 'single_label_classification' snake_case_ = input_dict['input_ids'] snake_case_ = input_ids.ne(1 ).to(__UpperCamelCase ) snake_case_ = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) snake_case_ = LlamaForSequenceClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , labels=__UpperCamelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = 'multi_label_classification' snake_case_ = input_dict['input_ids'] snake_case_ = input_ids.ne(1 ).to(__UpperCamelCase ) snake_case_ = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) snake_case_ = LlamaForSequenceClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , labels=__UpperCamelCase ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip('LLaMA buffers include complex numbers, which breaks this test' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @parameterized.expand([('linear',), ('dynamic',)] ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = ids_tensor([1, 10] , config.vocab_size ) snake_case_ = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights snake_case_ = LlamaModel(__UpperCamelCase ) original_model.to(__UpperCamelCase ) original_model.eval() snake_case_ = original_model(__UpperCamelCase ).last_hidden_state snake_case_ = original_model(__UpperCamelCase ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights snake_case_ = {'type': scaling_type, 'factor': 10.0} snake_case_ = LlamaModel(__UpperCamelCase ) scaled_model.to(__UpperCamelCase ) scaled_model.eval() snake_case_ = scaled_model(__UpperCamelCase ).last_hidden_state snake_case_ = scaled_model(__UpperCamelCase ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(__UpperCamelCase , __UpperCamelCase , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(__UpperCamelCase , __UpperCamelCase , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(__UpperCamelCase , __UpperCamelCase , atol=1E-5 ) ) @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [1, 3_06, 46_58, 2_78, 65_93, 3_10, 28_34, 3_38] snake_case_ = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-7b-hf' , device_map='auto' ) snake_case_ = model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 snake_case_ = torch.tensor([[-6.6550, -4.1227, -4.9859, -3.2406, 0.8262, -3.0033, 1.2964, -3.3699]] ) torch.testing.assert_close(out.mean(-1 ) , __UpperCamelCase , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off snake_case_ = torch.tensor([-12.8281, -7.4453, -0.4639, -8.0625, -7.2500, -8.0000, -6.4883, -7.7695, -7.8438, -7.0312, -6.2188, -7.1328, -1.8496, 1.9961, -8.6250, -6.7227, -12.8281, -6.9492, -7.0742, -7.7852, -7.5820, -7.9062, -6.9375, -7.9805, -8.3438, -8.1562, -8.0469, -7.6250, -7.7422, -7.3398,] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , __UpperCamelCase , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [1, 3_06, 46_58, 2_78, 65_93, 3_10, 28_34, 3_38] snake_case_ = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-hf' , device_map='auto' ) snake_case_ = model(torch.tensor(__UpperCamelCase ) ) # Expected mean on dim = -1 snake_case_ = torch.tensor([[-2.0622, -1.2794, -1.1638, -0.9788, -1.4603, -1.0238, -1.7893, -1.4411]] ) torch.testing.assert_close(out.mean(-1 ) , __UpperCamelCase , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off snake_case_ = torch.tensor([-8.1406, -8.0547, 2.7461, -1.2344, -0.1448, -1.8262, -1.0020, -1.8154, -1.6895, -1.8516, -2.3574, -0.9277, 3.7598, 6.5742, -1.2998, -0.1177, -8.1406, -2.9688, -2.9199, -3.1699, -3.5254, -2.3555, -2.7988, -3.4141, -2.8262, -4.5195, -3.3379, -3.3164, -2.7832, -3.0273] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , __UpperCamelCase , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [1, 3_06, 46_58, 2_78, 65_93, 3_10, 28_34, 3_38] snake_case_ = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-chat-hf' , device_map='auto' ) snake_case_ = model(torch.tensor(__UpperCamelCase ) ) # Expected mean on dim = -1 snake_case_ = torch.tensor([[-0.8562, -1.8520, -0.7551, -0.4162, -1.5161, -1.2038, -2.4823, -2.3254]] ) torch.testing.assert_close(out.mean(-1 ) , __UpperCamelCase , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off snake_case_ = torch.tensor([-2.2227, 4.8828, 0.9023, -0.4578, -0.7871, -0.1033, -0.6221, -0.5786, -0.7803, -1.0674, -1.2920, -0.1570, 0.8008, 2.0723, -0.9497, 0.2771, -2.2227, -0.7612, -1.4346, -1.2061, -1.6426, -0.3000, -0.7139, -1.1934, -1.8691, -1.6973, -1.5947, -1.2705, -0.3523, -0.5513] ) # fmt: on torch.testing.assert_close(out.mean(-1 ) , __UpperCamelCase , atol=1E-2 , rtol=1E-2 ) @unittest.skip( 'Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test' ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [1, 3_06, 46_58, 2_78, 65_93, 3_10, 28_34, 3_38] snake_case_ = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-70b-hf' , device_map='auto' ) snake_case_ = model(torch.tensor(__UpperCamelCase ) ) snake_case_ = torch.tensor( [[-4.2327, -3.3360, -4.6665, -4.7631, -1.8180, -3.4170, -1.4211, -3.1810]] , dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ) , __UpperCamelCase , atol=1E-2 , rtol=1E-2 ) # fmt: off snake_case_ = torch.tensor([-9.4922, -3.9551, 1.7998, -5.6758, -5.1055, -5.8984, -4.8320, -6.8086, -6.5391, -5.6172, -5.5820, -5.5352, 1.7881, 3.6289, -6.5117, -3.4785, -9.5000, -6.0352, -6.8125, -6.0195, -6.6836, -5.4727, -6.2812, -6.0391, -7.3398, -7.4297, -7.4844, -6.5820, -5.8789, -5.5312] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , __UpperCamelCase , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Model is curently gated' ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = 'Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the "princi' snake_case_ = 'Simply put, the theory of relativity states that ' snake_case_ = LlamaTokenizer.from_pretrained('meta-llama/Llama-2-13b-chat-hf' ) snake_case_ = tokenizer.encode(__UpperCamelCase , return_tensors='pt' ) snake_case_ = LlamaForCausalLM.from_pretrained( 'meta-llama/Llama-2-13b-chat-hf' , device_map='sequential' , use_safetensors=__UpperCamelCase ) # greedy generation outputs snake_case_ = model.generate(__UpperCamelCase , max_new_tokens=64 , top_p=__UpperCamelCase , temperature=1 , do_sample=__UpperCamelCase ) snake_case_ = tokenizer.decode(generated_ids[0] , skip_special_tokens=__UpperCamelCase ) self.assertEqual(__UpperCamelCase , __UpperCamelCase )
46
from __future__ import annotations import os import tempfile import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import is_tensorflow_text_available, is_tf_available from transformers.testing_utils import require_tensorflow_text, require_tf, slow from ..test_modeling_tf_common import floats_tensor from .test_framework_agnostic import GenerationIntegrationTestsMixin if is_tf_available(): import tensorflow as tf from transformers import ( AutoTokenizer, TFAutoModelForCausalLM, TFAutoModelForSeqaSeqLM, TFAutoModelForSpeechSeqaSeq, TFAutoModelForVisionaSeq, TFBartForConditionalGeneration, TFLogitsProcessorList, TFMinLengthLogitsProcessor, tf_top_k_top_p_filtering, ) if is_tensorflow_text_available(): import tensorflow_text as text @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = tf.convert_to_tensor( [ [ 8.222_0991, # 3rd highest value; idx. 0 -0.562_0044, 5.2322_9752, 4.038_6393, -6.879_8378, -0.5478_5802, -3.201_2153, 2.9277_7176, 1.8817_1953, 7.3534_1276, # 5th highest value; idx. 9 8.4320_7833, # 2nd highest value; idx. 10 -9.8571_1836, -5.9620_9236, -1.1303_9161, -7.111_5294, -0.836_9633, -5.318_6408, 7.0642_7407, 0.8136_9344, -0.8202_3817, -5.917_9796, 0.5881_3443, -6.9977_8438, 4.7155_1189, -0.1877_1637, 7.4402_0759, # 4th highest value; idx. 25 9.3845_0987, # 1st highest value; idx. 26 2.1266_2941, -9.3256_2038, 2.3565_2522, ], # cummulative prob of 5 highest values <= 0.6 [ 0.5842_5518, 4.5313_9238, -5.5751_0464, -6.2803_0699, -7.1952_9503, -4.0212_2551, 1.3933_7037, -6.0670_7057, 1.5948_0517, -9.64_3119, 0.0390_7799, 0.6723_1762, -8.8820_6726, 6.2711_5922, # 4th highest value; idx. 13 2.2852_0723, 4.8276_7506, 4.3042_1368, 8.827_5313, # 2nd highest value; idx. 17 5.4402_9958, # 5th highest value; idx. 18 -4.473_5794, 7.3857_9536, # 3rd highest value; idx. 20 -2.9105_1663, 2.6194_6077, -2.567_4762, -9.4895_9302, -4.0292_2645, -1.3541_6918, 9.6770_2323, # 1st highest value; idx. 27 -5.8947_8553, 1.8537_0467, ], # cummulative prob of 5 highest values <= 0.6 ] , dtype=tf.floataa , ) snake_case_ = tf.convert_to_tensor( [[0, 0], [0, 9], [0, 10], [0, 25], [0, 26], [1, 13], [1, 17], [1, 18], [1, 20], [1, 27]] , dtype=tf.intaa , ) # expected non filtered idx as noted above snake_case_ = tf.convert_to_tensor( [8.22_2099, 7.353_4126, 8.43_2078, 7.440_2075, 9.3_8451, 6.27_1159, 8.82_7531, 5.440_2995, 7.385_7956, 9.67_7023] , dtype=tf.floataa , ) # expected non filtered values as noted above snake_case_ = tf_top_k_top_p_filtering(__UpperCamelCase , top_k=10 , top_p=0.6 , min_tokens_to_keep=4 ) snake_case_ = output[output != -float('inf' )] snake_case_ = tf.cast( tf.where(tf.not_equal(__UpperCamelCase , tf.constant(-float('inf' ) , dtype=tf.floataa ) ) ) , dtype=tf.intaa , ) tf.debugging.assert_near(__UpperCamelCase , __UpperCamelCase , rtol=1E-12 ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase , __snake_case ): """simple docstring""" if is_tf_available(): __A = { """AutoModelForCausalLM""": TFAutoModelForCausalLM, """AutoModelForSpeechSeq2Seq""": TFAutoModelForSpeechSeqaSeq, """AutoModelForSeq2SeqLM""": TFAutoModelForSeqaSeqLM, """AutoModelForVision2Seq""": TFAutoModelForVisionaSeq, """LogitsProcessorList""": TFLogitsProcessorList, """MinLengthLogitsProcessor""": TFMinLengthLogitsProcessor, """create_tensor_fn""": tf.convert_to_tensor, """floats_tensor""": floats_tensor, """return_tensors""": """tf""", } @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 2 snake_case_ = 2 class SCREAMING_SNAKE_CASE ( tf.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super(__UpperCamelCase , self ).__init__() snake_case_ = model @tf.function( input_signature=( tf.TensorSpec((None, input_length) , tf.intaa , name='input_ids' ), tf.TensorSpec((None, input_length) , tf.intaa , name='attention_mask' ), ) , jit_compile=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.model.generate( input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase , max_new_tokens=__UpperCamelCase , return_dict_in_generate=__UpperCamelCase , ) return {"sequences": outputs["sequences"]} snake_case_ = [[2, 0], [1_02, 1_03]] snake_case_ = [[1, 0], [1, 1]] snake_case_ = DummyModel(model=__UpperCamelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__UpperCamelCase , __UpperCamelCase , signatures={'serving_default': dummy_model.serving} ) snake_case_ = tf.saved_model.load(__UpperCamelCase ).signatures['serving_default'] for batch_size in range(1 , len(__UpperCamelCase ) + 1 ): snake_case_ = { 'input_ids': tf.constant(dummy_input_ids[:batch_size] ), 'attention_mask': tf.constant(dummy_attention_masks[:batch_size] ), } snake_case_ = serving_func(**__UpperCamelCase )['sequences'] snake_case_ = test_model.generate(**__UpperCamelCase , max_new_tokens=__UpperCamelCase ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 1 snake_case_ = 2 class SCREAMING_SNAKE_CASE ( tf.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super(__UpperCamelCase , self ).__init__() snake_case_ = model @tf.function( input_signature=( tf.TensorSpec((batch_size, None) , tf.intaa , name='input_ids' ), tf.TensorSpec((batch_size, None) , tf.intaa , name='attention_mask' ), ) , jit_compile=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.model.generate( input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase , max_new_tokens=__UpperCamelCase , return_dict_in_generate=__UpperCamelCase , ) return {"sequences": outputs["sequences"]} snake_case_ = [[2], [1_02, 1_03]] snake_case_ = [[1], [1, 1]] snake_case_ = DummyModel(model=__UpperCamelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__UpperCamelCase , __UpperCamelCase , signatures={'serving_default': dummy_model.serving} ) snake_case_ = tf.saved_model.load(__UpperCamelCase ).signatures['serving_default'] for input_row in range(len(__UpperCamelCase ) ): snake_case_ = { 'input_ids': tf.constant([dummy_input_ids[input_row]] ), 'attention_mask': tf.constant([dummy_attention_masks[input_row]] ), } snake_case_ = serving_func(**__UpperCamelCase )['sequences'] snake_case_ = test_model.generate(**__UpperCamelCase , max_new_tokens=__UpperCamelCase ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @slow @require_tensorflow_text def __lowerCAmelCase ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: # file needed to load the TF tokenizer hf_hub_download(repo_id='google/flan-t5-small' , filename='spiece.model' , local_dir=__UpperCamelCase ) class SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__() snake_case_ = text.SentencepieceTokenizer( model=tf.io.gfile.GFile(os.path.join(__UpperCamelCase , 'spiece.model' ) , 'rb' ).read() ) snake_case_ = TFAutoModelForSeqaSeqLM.from_pretrained('hf-internal-testing/tiny-random-t5' ) def __lowerCAmelCase ( self , __UpperCamelCase , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" snake_case_ = self.tokenizer.tokenize(__UpperCamelCase ) snake_case_ , snake_case_ = text.pad_model_inputs( __UpperCamelCase , max_seq_length=64 , pad_value=self.model.config.pad_token_id ) snake_case_ = self.model.generate(input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase ) return self.tokenizer.detokenize(__UpperCamelCase ) snake_case_ = CompleteSentenceTransformer() snake_case_ = tf.keras.layers.Input(shape=(1,) , dtype=tf.string , name='inputs' ) snake_case_ = complete_model(__UpperCamelCase ) snake_case_ = tf.keras.Model(__UpperCamelCase , __UpperCamelCase ) keras_model.save(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = { 'do_sample': True, 'num_beams': 1, 'top_p': 0.7, 'top_k': 10, 'temperature': 0.7, } snake_case_ = 14 snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 'Hello, my dog is cute and' snake_case_ = tokenizer(__UpperCamelCase , return_tensors='tf' ) snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 6_38 # forces the generation to happen on CPU, to avoid GPU-related quirks with tf.device(':/CPU:0' ): tf.random.set_seed(0 ) snake_case_ = model.generate(**__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) snake_case_ = [6_38, 1_98] with tf.device(':/CPU:0' ): tf.random.set_seed(0 ) snake_case_ = model.generate(**__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = 'Hugging Face is a technology company based in New York and Paris.' snake_case_ = bart_tokenizer(__UpperCamelCase , return_tensors='tf' ).input_ids snake_case_ = TFBartForConditionalGeneration.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = bart_model.generate(__UpperCamelCase ).numpy() class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase=None , **__UpperCamelCase ): """simple docstring""" return super().call(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = FakeBart.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = bart_model.generate(__UpperCamelCase , foo='bar' ).numpy() self.assertTrue(np.array_equal(__UpperCamelCase , __UpperCamelCase ) ) class SCREAMING_SNAKE_CASE ( bart_model.model.encoder.__class__ ): """simple docstring""" def __lowerCAmelCase ( self , __UpperCamelCase , **__UpperCamelCase ): """simple docstring""" return super().call(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = FakeEncoder(bart_model.config , bart_model.model.shared ) snake_case_ = fake_encoder # Normal generation still works (the output will be different because the encoder weights are different) snake_case_ = bart_model.generate(__UpperCamelCase ).numpy() with self.assertRaises(__UpperCamelCase ): # FakeEncoder.call() accepts **kwargs -> no filtering -> value error due to unexpected input "foo" bart_model.generate(__UpperCamelCase , foo='bar' )
46
1
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=2 , __UpperCamelCase=24 , __UpperCamelCase=16 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=32 , __UpperCamelCase=5 , __UpperCamelCase=4 , __UpperCamelCase=37 , __UpperCamelCase="gelu" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=10 , __UpperCamelCase=0.02 , __UpperCamelCase=None , __UpperCamelCase=2 , __UpperCamelCase=2 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = patch_size snake_case_ = max_length snake_case_ = num_mel_bins snake_case_ = is_training snake_case_ = use_labels snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = scope snake_case_ = frequency_stride snake_case_ = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) snake_case_ = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 snake_case_ = (self.max_length - self.patch_size) // self.time_stride + 1 snake_case_ = frequency_out_dimension * time_out_dimension snake_case_ = num_patches + 2 def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = self.get_config() return config, input_values, labels def __lowerCAmelCase ( self ): """simple docstring""" return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = ASTModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = config_and_inputs snake_case_ = {'input_values': input_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) __A = ( {"""audio-classification""": ASTForAudioClassification, """feature-extraction""": ASTModel} if is_torch_available() else {} ) __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ASTModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase , hidden_size=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='AST does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase , nn.Linear ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['input_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = ASTModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def a(): '''simple docstring''' snake_case_ = hf_hub_download( repo_id='nielsr/audio-spectogram-transformer-checkpoint' , filename='sample_audio.flac' , repo_type='dataset' ) snake_case_ , snake_case_ = torchaudio.load(lowercase__ ) return audio, sampling_rate @require_torch @require_torchaudio class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return ( ASTFeatureExtractor.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ) if is_torchaudio_available() else None ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.default_feature_extractor snake_case_ = ASTForAudioClassification.from_pretrained('MIT/ast-finetuned-audioset-10-10-0.4593' ).to(__UpperCamelCase ) snake_case_ = self.default_feature_extractor snake_case_ , snake_case_ = prepare_audio() snake_case_ = audio.squeeze().numpy() snake_case_ = feature_extractor(__UpperCamelCase , sampling_rate=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 5_27) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([-0.8760, -7.0042, -8.6602] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) )
46
import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=2 , __UpperCamelCase=8 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=99 , __UpperCamelCase=16 , __UpperCamelCase=5 , __UpperCamelCase=2 , __UpperCamelCase=36 , __UpperCamelCase="gelu" , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=5_12 , __UpperCamelCase=16 , __UpperCamelCase=2 , __UpperCamelCase=0.02 , __UpperCamelCase=3 , __UpperCamelCase=4 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_input_mask snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_labels snake_case_ = num_choices snake_case_ = scope def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None if self.use_token_type_ids: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowerCAmelCase ( self ): """simple docstring""" return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.get_config() snake_case_ = 3_00 return config def __lowerCAmelCase ( self ): """simple docstring""" ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = self.prepare_config_and_inputs() snake_case_ = True snake_case_ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase ) snake_case_ = model(__UpperCamelCase , token_type_ids=__UpperCamelCase ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" snake_case_ = True snake_case_ = MraModel(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , encoder_attention_mask=__UpperCamelCase , ) snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , ) snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraForMaskedLM(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraForQuestionAnswering(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , start_positions=__UpperCamelCase , end_positions=__UpperCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MraForSequenceClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MraForTokenClassification(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_choices snake_case_ = MraForMultipleChoice(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = config_and_inputs snake_case_ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , unittest.TestCase ): """simple docstring""" __A = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) __A = False __A = False __A = False __A = False __A = () def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , hidden_size=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: snake_case_ = type self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = MraModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) @unittest.skip(reason='MRA does not output attentions' ) def __lowerCAmelCase ( self ): """simple docstring""" return @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraModel.from_pretrained('uw-madison/mra-base-512-4' ) snake_case_ = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = torch.Size((1, 2_56, 7_68) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[-0.0140, 0.0830, -0.0381], [0.1546, 0.1402, 0.0220], [0.1162, 0.0851, 0.0165]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraForMaskedLM.from_pretrained('uw-madison/mra-base-512-4' ) snake_case_ = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = 5_02_65 snake_case_ = torch.Size((1, 2_56, vocab_size) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[9.2595, -3.6038, 11.8819], [9.3869, -3.2693, 11.0956], [11.8524, -3.4938, 13.1210]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraForMaskedLM.from_pretrained('uw-madison/mra-base-4096-8-d3' ) snake_case_ = torch.arange(40_96 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = 5_02_65 snake_case_ = torch.Size((1, 40_96, vocab_size) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[5.4789, -2.3564, 7.5064], [7.9067, -1.3369, 9.9668], [9.0712, -1.8106, 7.0380]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) )
46
1
from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { 's-JoL/Open-Llama-V1': 'https://huggingface.co/s-JoL/Open-Llama-V1/blob/main/config.json', } class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = """open-llama""" def __init__( self , __UpperCamelCase=10_00_00 , __UpperCamelCase=40_96 , __UpperCamelCase=1_10_08 , __UpperCamelCase=32 , __UpperCamelCase=32 , __UpperCamelCase="silu" , __UpperCamelCase=20_48 , __UpperCamelCase=0.02 , __UpperCamelCase=1E-6 , __UpperCamelCase=True , __UpperCamelCase=0 , __UpperCamelCase=1 , __UpperCamelCase=2 , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=None , **__UpperCamelCase , ): """simple docstring""" snake_case_ = vocab_size snake_case_ = max_position_embeddings snake_case_ = hidden_size snake_case_ = intermediate_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = hidden_act snake_case_ = initializer_range snake_case_ = rms_norm_eps snake_case_ = use_cache snake_case_ = kwargs.pop( 'use_memorry_efficient_attention' , __UpperCamelCase ) snake_case_ = hidden_dropout_prob snake_case_ = attention_dropout_prob snake_case_ = use_stable_embedding snake_case_ = shared_input_output_embedding snake_case_ = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=__UpperCamelCase , bos_token_id=__UpperCamelCase , eos_token_id=__UpperCamelCase , tie_word_embeddings=__UpperCamelCase , **__UpperCamelCase , ) def __lowerCAmelCase ( self ): """simple docstring""" if self.rope_scaling is None: return if not isinstance(self.rope_scaling , __UpperCamelCase ) or len(self.rope_scaling ) != 2: raise ValueError( '`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, ' f"""got {self.rope_scaling}""" ) snake_case_ = self.rope_scaling.get('type' , __UpperCamelCase ) snake_case_ = self.rope_scaling.get('factor' , __UpperCamelCase ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"""`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}""" ) if rope_scaling_factor is None or not isinstance(__UpperCamelCase , __UpperCamelCase ) or rope_scaling_factor <= 1.0: raise ValueError(f"""`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}""" )
46
import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def a(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file snake_case_ = TapasConfig.from_json_file(lowercase__ ) # set absolute/relative position embeddings parameter snake_case_ = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "WTQ": # run_task_main.py hparams snake_case_ = 4 snake_case_ = True # hparam_utils.py hparams snake_case_ = 0.66_4694 snake_case_ = 0.20_7951 snake_case_ = 0.12_1194 snake_case_ = True snake_case_ = True snake_case_ = False snake_case_ = 0.035_2513 snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams snake_case_ = 4 snake_case_ = False # hparam_utils.py hparams snake_case_ = 36.4519 snake_case_ = 0.90_3421 snake_case_ = 222.088 snake_case_ = True snake_case_ = True snake_case_ = True snake_case_ = 0.76_3141 snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "TABFACT": snake_case_ = TapasForSequenceClassification(config=lowercase__ ) elif task == "MLM": snake_case_ = TapasForMaskedLM(config=lowercase__ ) elif task == "INTERMEDIATE_PRETRAINING": snake_case_ = TapasModel(config=lowercase__ ) else: raise ValueError(f"""Task {task} not supported.""" ) print(f"""Building PyTorch model from configuration: {config}""" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(lowercase__ , lowercase__ , lowercase__ ) # Save pytorch-model (weights and configuration) print(f"""Save PyTorch model to {pytorch_dump_path}""" ) model.save_pretrained(lowercase__ ) # Save tokenizer files print(f"""Save tokenizer files to {pytorch_dump_path}""" ) snake_case_ = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + 'vocab.txt' , model_max_length=512 ) tokenizer.save_pretrained(lowercase__ ) print('Used relative position embeddings:' , model.config.reset_position_index_per_cell ) if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( '--task', default='SQA', type=str, help='Model task for which to convert a checkpoint. Defaults to SQA.' ) parser.add_argument( '--reset_position_index_per_cell', default=False, action='store_true', help='Whether to use relative position embeddings or not. Defaults to True.', ) parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--tapas_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained TAPAS model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) A = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
46
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) A = { 'configuration_trocr': ['TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TrOCRConfig'], 'processing_trocr': ['TrOCRProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'TROCR_PRETRAINED_MODEL_ARCHIVE_LIST', 'TrOCRForCausalLM', 'TrOCRPreTrainedModel', ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=3 , __UpperCamelCase=16 , __UpperCamelCase=[1, 2, 1] , __UpperCamelCase=[2, 2, 4] , __UpperCamelCase=2 , __UpperCamelCase=2.0 , __UpperCamelCase=True , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.1 , __UpperCamelCase="gelu" , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=0.02 , __UpperCamelCase=1E-5 , __UpperCamelCase=True , __UpperCamelCase=None , __UpperCamelCase=True , __UpperCamelCase=10 , __UpperCamelCase=8 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = embed_dim snake_case_ = depths snake_case_ = num_heads snake_case_ = window_size snake_case_ = mlp_ratio snake_case_ = qkv_bias snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = drop_path_rate snake_case_ = hidden_act snake_case_ = use_absolute_embeddings snake_case_ = patch_norm snake_case_ = layer_norm_eps snake_case_ = initializer_range snake_case_ = is_training snake_case_ = scope snake_case_ = use_labels snake_case_ = type_sequence_label_size snake_case_ = encoder_stride def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = self.get_config() return config, pixel_values, labels def __lowerCAmelCase ( self ): """simple docstring""" return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = SwinvaModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) snake_case_ = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) snake_case_ = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = SwinvaForMaskedImageModeling(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images snake_case_ = 1 snake_case_ = SwinvaForMaskedImageModeling(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.type_sequence_label_size snake_case_ = SwinvaForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) __A = ( {"""feature-extraction""": SwinvaModel, """image-classification""": SwinvaForImageClassification} if is_torch_available() else {} ) __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = SwinvaModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , embed_dim=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase , nn.Linear ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = True for model_class in self.all_model_classes: snake_case_ = True snake_case_ = False snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.attentions snake_case_ = len(self.model_tester.depths ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # check that output_attentions also work using config del inputs_dict["output_attentions"] snake_case_ = True snake_case_ = config.window_size**2 snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) snake_case_ = len(__UpperCamelCase ) # Check attention is always last and order is fine snake_case_ = True snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): snake_case_ = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states snake_case_ = 2 self.assertEqual(out_len + added_hidden_states , len(__UpperCamelCase ) ) snake_case_ = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.hidden_states snake_case_ = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # Swinv2 has a different seq_length snake_case_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) snake_case_ = outputs.reshaped_hidden_states self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) snake_case_ , snake_case_ , snake_case_ , snake_case_ = reshaped_hidden_states[0].shape snake_case_ = ( reshaped_hidden_states[0].view(__UpperCamelCase , __UpperCamelCase , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) snake_case_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) snake_case_ = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = SwinvaModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = _config_zero_init(__UpperCamelCase ) for model_class in self.all_model_classes: snake_case_ = model_class(config=__UpperCamelCase ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @require_vision @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( __UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([-0.3947, -0.4306, 0.0026] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) )
46
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow if is_torch_available(): import torch from transformers import XLMRobertaModel @require_sentencepiece @require_tokenizers @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = XLMRobertaModel.from_pretrained('xlm-roberta-base' ) snake_case_ = torch.tensor([[0, 5_81, 1_02_69, 83, 9_99_42, 1_36, 6_07_42, 23, 70, 8_05_83, 1_82_76, 2]] ) # The dog is cute and lives in the garden house snake_case_ = torch.Size((1, 12, 7_68) ) # batch_size, sequence_length, embedding_vector_dim snake_case_ = torch.tensor( [[-0.0101, 0.1218, -0.0803, 0.0801, 0.1327, 0.0776, -0.1215, 0.2383, 0.3338, 0.3106, 0.0300, 0.0252]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): snake_case_ = model(__UpperCamelCase )['last_hidden_state'].detach() self.assertEqual(output.shape , __UpperCamelCase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , __UpperCamelCase , atol=1E-3 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = XLMRobertaModel.from_pretrained('xlm-roberta-large' ) snake_case_ = torch.tensor([[0, 5_81, 1_02_69, 83, 9_99_42, 1_36, 6_07_42, 23, 70, 8_05_83, 1_82_76, 2]] ) # The dog is cute and lives in the garden house snake_case_ = torch.Size((1, 12, 10_24) ) # batch_size, sequence_length, embedding_vector_dim snake_case_ = torch.tensor( [[-0.0699, -0.0318, 0.0705, -0.1241, 0.0999, -0.0520, 0.1004, -0.1838, -0.4704, 0.1437, 0.0821, 0.0126]] ) # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.large') # xlmr.eval() # expected_output_values_last_dim = xlmr.extract_features(input_ids[0])[:, :, -1] with torch.no_grad(): snake_case_ = model(__UpperCamelCase )['last_hidden_state'].detach() self.assertEqual(output.shape , __UpperCamelCase ) # compare the actual values for a slice of last dim self.assertTrue(torch.allclose(output[:, :, -1] , __UpperCamelCase , atol=1E-3 ) )
46
import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline A = argparse.ArgumentParser('Stable Diffusion script with intel optimization', add_help=False) parser.add_argument('--dpm', action='store_true', help='Enable DPMSolver or not') parser.add_argument('--steps', default=None, type=int, help='Num inference steps') A = parser.parse_args() A = 'cpu' A = 'a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings' A = 'path-to-your-trained-model' A = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: A = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) A = pipe.to(device) # to channels last A = pipe.unet.to(memory_format=torch.channels_last) A = pipe.vae.to(memory_format=torch.channels_last) A = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: A = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex A = torch.randn(2, 4, 64, 64) A = torch.rand(1) * 999 A = torch.randn(2, 77, 768) A = (sample, timestep, encoder_hidden_status) try: A = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: A = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) A = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) A = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: A = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute A = 666 A = torch.Generator(device).manual_seed(seed) A = {'generator': generator} if args.steps is not None: A = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): A = pipe(prompt, **generate_kwargs).images[0] # save image image.save('generated.png')
46
1
import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) A = logging.getLogger() def a(): '''simple docstring''' snake_case_ = argparse.ArgumentParser() parser.add_argument('-f' ) snake_case_ = parser.parse_args() return args.f class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = logging.StreamHandler(sys.stdout ) logger.addHandler(__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , 'run_glue_deebert.py' ) with patch.object(__UpperCamelCase , 'argv' , __UpperCamelCase ): snake_case_ = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(__UpperCamelCase , 0.666 ) @slow @require_torch_non_multi_gpu def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = '\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n '.split() self.run_and_check(__UpperCamelCase ) snake_case_ = '\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n '.split() self.run_and_check(__UpperCamelCase ) snake_case_ = '\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n '.split() self.run_and_check(__UpperCamelCase )
46
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { 'microsoft/unispeech-sat-base-100h-libri-ft': ( 'https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json' ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = """unispeech-sat""" def __init__( self , __UpperCamelCase=32 , __UpperCamelCase=7_68 , __UpperCamelCase=12 , __UpperCamelCase=12 , __UpperCamelCase=30_72 , __UpperCamelCase="gelu" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.02 , __UpperCamelCase=1E-5 , __UpperCamelCase="group" , __UpperCamelCase="gelu" , __UpperCamelCase=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , __UpperCamelCase=(5, 2, 2, 2, 2, 2, 2) , __UpperCamelCase=(10, 3, 3, 3, 3, 2, 2) , __UpperCamelCase=False , __UpperCamelCase=1_28 , __UpperCamelCase=16 , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=0.05 , __UpperCamelCase=10 , __UpperCamelCase=2 , __UpperCamelCase=0.0 , __UpperCamelCase=10 , __UpperCamelCase=0 , __UpperCamelCase=3_20 , __UpperCamelCase=2 , __UpperCamelCase=0.1 , __UpperCamelCase=1_00 , __UpperCamelCase=2_56 , __UpperCamelCase=2_56 , __UpperCamelCase=0.1 , __UpperCamelCase="mean" , __UpperCamelCase=False , __UpperCamelCase=False , __UpperCamelCase=2_56 , __UpperCamelCase=(5_12, 5_12, 5_12, 5_12, 15_00) , __UpperCamelCase=(5, 3, 3, 1, 1) , __UpperCamelCase=(1, 2, 3, 1, 1) , __UpperCamelCase=5_12 , __UpperCamelCase=0 , __UpperCamelCase=1 , __UpperCamelCase=2 , __UpperCamelCase=5_04 , **__UpperCamelCase , ): """simple docstring""" super().__init__(**__UpperCamelCase , pad_token_id=__UpperCamelCase , bos_token_id=__UpperCamelCase , eos_token_id=__UpperCamelCase ) snake_case_ = hidden_size snake_case_ = feat_extract_norm snake_case_ = feat_extract_activation snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = conv_bias snake_case_ = num_conv_pos_embeddings snake_case_ = num_conv_pos_embedding_groups snake_case_ = len(self.conv_dim ) snake_case_ = num_hidden_layers snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = num_attention_heads snake_case_ = hidden_dropout snake_case_ = attention_dropout snake_case_ = activation_dropout snake_case_ = feat_proj_dropout snake_case_ = final_dropout snake_case_ = layerdrop snake_case_ = layer_norm_eps snake_case_ = initializer_range snake_case_ = vocab_size snake_case_ = num_clusters snake_case_ = do_stable_layer_norm snake_case_ = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( 'Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==' ' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =' f""" {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,""" f""" `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 snake_case_ = apply_spec_augment snake_case_ = mask_time_prob snake_case_ = mask_time_length snake_case_ = mask_time_min_masks snake_case_ = mask_feature_prob snake_case_ = mask_feature_length snake_case_ = mask_feature_min_masks # parameters for pretraining with codevector quantized representations snake_case_ = num_codevectors_per_group snake_case_ = num_codevector_groups snake_case_ = contrastive_logits_temperature snake_case_ = feat_quantizer_dropout snake_case_ = num_negatives snake_case_ = codevector_dim snake_case_ = proj_codevector_dim snake_case_ = diversity_loss_weight # ctc loss snake_case_ = ctc_loss_reduction snake_case_ = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. snake_case_ = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = xvector_output_dim @property def __lowerCAmelCase ( self ): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
46
1
import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # and perform gradient accumulation # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## A = 16 A = 32 def a(lowercase__ , lowercase__ = 16 ): '''simple docstring''' snake_case_ = AutoTokenizer.from_pretrained('bert-base-cased' ) snake_case_ = load_dataset('glue' , 'mrpc' ) def tokenize_function(lowercase__ ): # max_length=None => use the model max length (it's actually the default) snake_case_ = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=lowercase__ , max_length=lowercase__ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): snake_case_ = datasets.map( lowercase__ , batched=lowercase__ , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library snake_case_ = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(lowercase__ ): # On TPU it's best to pad everything to the same length or training will be very slow. snake_case_ = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": snake_case_ = 16 elif accelerator.mixed_precision != "no": snake_case_ = 8 else: snake_case_ = None return tokenizer.pad( lowercase__ , padding='longest' , max_length=lowercase__ , pad_to_multiple_of=lowercase__ , return_tensors='pt' , ) # Instantiate dataloaders. snake_case_ = DataLoader( tokenized_datasets['train'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) snake_case_ = DataLoader( tokenized_datasets['validation'] , shuffle=lowercase__ , collate_fn=lowercase__ , batch_size=lowercase__ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders A = mocked_dataloaders # noqa: F811 def a(lowercase__ , lowercase__ ): '''simple docstring''' # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS' , lowercase__ ) == "1": snake_case_ = 2 # New Code # snake_case_ = int(args.gradient_accumulation_steps ) # Initialize accelerator snake_case_ = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , gradient_accumulation_steps=lowercase__ ) if accelerator.distributed_type == DistributedType.TPU and gradient_accumulation_steps > 1: raise NotImplementedError( 'Gradient accumulation on TPUs is currently not supported. Pass `gradient_accumulation_steps=1`' ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs snake_case_ = config['lr'] snake_case_ = int(config['num_epochs'] ) snake_case_ = int(config['seed'] ) snake_case_ = int(config['batch_size'] ) snake_case_ = evaluate.load('glue' , 'mrpc' ) set_seed(lowercase__ ) snake_case_ , snake_case_ = get_dataloaders(lowercase__ , lowercase__ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) snake_case_ = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=lowercase__ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). snake_case_ = model.to(accelerator.device ) # Instantiate optimizer snake_case_ = AdamW(params=model.parameters() , lr=lowercase__ ) # Instantiate scheduler snake_case_ = get_linear_schedule_with_warmup( optimizer=lowercase__ , num_warmup_steps=100 , num_training_steps=(len(lowercase__ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ = accelerator.prepare( lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ) # Now we train the model for epoch in range(lowercase__ ): model.train() for step, batch in enumerate(lowercase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(lowercase__ ): snake_case_ = model(**lowercase__ ) snake_case_ = output.loss accelerator.backward(lowercase__ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(lowercase__ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): snake_case_ = model(**lowercase__ ) snake_case_ = outputs.logits.argmax(dim=-1 ) snake_case_ , snake_case_ = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=lowercase__ , references=lowercase__ , ) snake_case_ = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , lowercase__ ) def a(): '''simple docstring''' snake_case_ = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=lowercase__ , default=lowercase__ , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) # New Code # parser.add_argument( '--gradient_accumulation_steps' , type=lowercase__ , default=1 , help='The number of minibatches to be ran before gradients are accumulated.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) snake_case_ = parser.parse_args() snake_case_ = {'lr': 2e-5, 'num_epochs': 3, 'seed': 42, 'batch_size': 16} training_function(lowercase__ , lowercase__ ) if __name__ == "__main__": main()
46
class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = name snake_case_ = val def __str__( self ): """simple docstring""" return f"""{self.__class__.__name__}({self.name}, {self.val})""" def __lt__( self , __UpperCamelCase ): """simple docstring""" return self.val < other.val class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" snake_case_ = {} snake_case_ = {} snake_case_ = self.build_heap(__UpperCamelCase ) def __getitem__( self , __UpperCamelCase ): """simple docstring""" return self.get_value(__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return (idx - 1) // 2 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return idx * 2 + 1 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return idx * 2 + 2 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return self.heap_dict[key] def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = len(__UpperCamelCase ) - 1 snake_case_ = self.get_parent_idx(__UpperCamelCase ) for idx, i in enumerate(__UpperCamelCase ): snake_case_ = idx snake_case_ = i.val for i in range(__UpperCamelCase , -1 , -1 ): self.sift_down(__UpperCamelCase , __UpperCamelCase ) return array def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" while True: snake_case_ = self.get_left_child_idx(__UpperCamelCase ) # noqa: E741 snake_case_ = self.get_right_child_idx(__UpperCamelCase ) snake_case_ = idx if l < len(__UpperCamelCase ) and array[l] < array[idx]: snake_case_ = l if r < len(__UpperCamelCase ) and array[r] < array[smallest]: snake_case_ = r if smallest != idx: snake_case_ , snake_case_ = array[smallest], array[idx] ( ( snake_case_ ) , ( snake_case_ ) , ) = ( self.idx_of_element[array[smallest]], self.idx_of_element[array[idx]], ) snake_case_ = smallest else: break def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = self.get_parent_idx(__UpperCamelCase ) while p >= 0 and self.heap[p] > self.heap[idx]: snake_case_ , snake_case_ = self.heap[idx], self.heap[p] snake_case_ , snake_case_ = ( self.idx_of_element[self.heap[idx]], self.idx_of_element[self.heap[p]], ) snake_case_ = p snake_case_ = self.get_parent_idx(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" return self.heap[0] def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.heap[-1], self.heap[0] snake_case_ , snake_case_ = ( self.idx_of_element[self.heap[-1]], self.idx_of_element[self.heap[0]], ) snake_case_ = self.heap.pop() del self.idx_of_element[x] self.sift_down(0 , self.heap ) return x def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" self.heap.append(__UpperCamelCase ) snake_case_ = len(self.heap ) - 1 snake_case_ = node.val self.sift_up(len(self.heap ) - 1 ) def __lowerCAmelCase ( self ): """simple docstring""" return len(self.heap ) == 0 def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" assert ( self.heap[self.idx_of_element[node]].val > new_value ), "newValue must be less that current value" snake_case_ = new_value snake_case_ = new_value self.sift_up(self.idx_of_element[node] ) A = Node('R', -1) A = Node('B', 6) A = Node('A', 3) A = Node('X', 1) A = Node('E', 4) # Use one of these two ways to generate Min-Heap # Generating Min-Heap from array A = MinHeap([r, b, a, x, e]) # Generating Min-Heap by Insert method # myMinHeap.insert(a) # myMinHeap.insert(b) # myMinHeap.insert(x) # myMinHeap.insert(r) # myMinHeap.insert(e) # Before print('Min Heap - before decrease key') for i in my_min_heap.heap: print(i) print('Min Heap - After decrease key of node [B -> -17]') my_min_heap.decrease_key(b, -17) # After for i in my_min_heap.heap: print(i) if __name__ == "__main__": import doctest doctest.testmod()
46
1
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=7 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=99 , __UpperCamelCase=32 , __UpperCamelCase=5 , __UpperCamelCase=4 , __UpperCamelCase=37 , __UpperCamelCase="gelu" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=5_12 , __UpperCamelCase=16 , __UpperCamelCase=2 , __UpperCamelCase=0.02 , __UpperCamelCase=4 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_attention_mask snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_choices def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_attention_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None if self.use_token_type_ids: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ = RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': attention_mask} return config, inputs_dict def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = True snake_case_ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class SCREAMING_SNAKE_CASE ( __snake_case , unittest.TestCase ): """simple docstring""" __A = True __A = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = FlaxRobertaModelTester(self ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_class_name in self.all_model_classes: snake_case_ = model_class_name.from_pretrained('roberta-base' , from_pt=__UpperCamelCase ) snake_case_ = model(np.ones((1, 1) ) ) self.assertIsNotNone(__UpperCamelCase )
46
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) A = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['PerceiverFeatureExtractor'] A = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A = { 'configuration_whisper': ['WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'WhisperConfig', 'WhisperOnnxConfig'], 'feature_extraction_whisper': ['WhisperFeatureExtractor'], 'processing_whisper': ['WhisperProcessor'], 'tokenization_whisper': ['WhisperTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['WhisperTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'WhisperForConditionalGeneration', 'WhisperModel', 'WhisperPreTrainedModel', 'WhisperForAudioClassification', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFWhisperForConditionalGeneration', 'TFWhisperModel', 'TFWhisperPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'FlaxWhisperForConditionalGeneration', 'FlaxWhisperModel', 'FlaxWhisperPreTrainedModel', 'FlaxWhisperForAudioClassification', ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
def a(lowercase__ , lowercase__ ): '''simple docstring''' if not isinstance(lowercase__ , lowercase__ ): raise ValueError('iterations must be defined as integers' ) if not isinstance(lowercase__ , lowercase__ ) or not number >= 1: raise ValueError( 'starting number must be\n and integer and be more than 0' ) if not iterations >= 1: raise ValueError('Iterations must be done more than 0 times to play FizzBuzz' ) snake_case_ = '' while number <= iterations: if number % 3 == 0: out += "Fizz" if number % 5 == 0: out += "Buzz" if 0 not in (number % 3, number % 5): out += str(lowercase__ ) # print(out) number += 1 out += " " return out if __name__ == "__main__": import doctest doctest.testmod()
46
1
import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline A = argparse.ArgumentParser('Stable Diffusion script with intel optimization', add_help=False) parser.add_argument('--dpm', action='store_true', help='Enable DPMSolver or not') parser.add_argument('--steps', default=None, type=int, help='Num inference steps') A = parser.parse_args() A = 'cpu' A = 'a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings' A = 'path-to-your-trained-model' A = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: A = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) A = pipe.to(device) # to channels last A = pipe.unet.to(memory_format=torch.channels_last) A = pipe.vae.to(memory_format=torch.channels_last) A = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: A = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex A = torch.randn(2, 4, 64, 64) A = torch.rand(1) * 999 A = torch.randn(2, 77, 768) A = (sample, timestep, encoder_hidden_status) try: A = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: A = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) A = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) A = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: A = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute A = 666 A = torch.Generator(device).manual_seed(seed) A = {'generator': generator} if args.steps is not None: A = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): A = pipe(prompt, **generate_kwargs).images[0] # save image image.save('generated.png')
46
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = 1.5 snake_case_ = int(factor * num_class_images ) snake_case_ = ClipClient( url='https://knn.laion.ai/knn-service' , indice_name='laion_400m' , num_images=lowercase__ , aesthetic_weight=0.1 ) os.makedirs(f"""{class_data_dir}/images""" , exist_ok=lowercase__ ) if len(list(Path(f"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: snake_case_ = client.query(text=lowercase__ ) if len(lowercase__ ) >= factor * num_class_images or num_images > 1e4: break else: snake_case_ = int(factor * num_images ) snake_case_ = ClipClient( url='https://knn.laion.ai/knn-service' , indice_name='laion_400m' , num_images=lowercase__ , aesthetic_weight=0.1 , ) snake_case_ = 0 snake_case_ = 0 snake_case_ = tqdm(desc='downloading real regularization images' , total=lowercase__ ) with open(f"""{class_data_dir}/caption.txt""" , 'w' ) as fa, open(f"""{class_data_dir}/urls.txt""" , 'w' ) as fa, open( f"""{class_data_dir}/images.txt""" , 'w' ) as fa: while total < num_class_images: snake_case_ = class_images[count] count += 1 try: snake_case_ = requests.get(images['url'] ) if img.status_code == 200: snake_case_ = Image.open(BytesIO(img.content ) ) with open(f"""{class_data_dir}/images/{total}.jpg""" , 'wb' ) as f: f.write(img.content ) fa.write(images['caption'] + '\n' ) fa.write(images['url'] + '\n' ) fa.write(f"""{class_data_dir}/images/{total}.jpg""" + '\n' ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def a(): '''simple docstring''' snake_case_ = argparse.ArgumentParser('' , add_help=lowercase__ ) parser.add_argument('--class_prompt' , help='text prompt to retrieve images' , required=lowercase__ , type=lowercase__ ) parser.add_argument('--class_data_dir' , help='path to save images' , required=lowercase__ , type=lowercase__ ) parser.add_argument('--num_class_images' , help='number of images to download' , default=200 , type=lowercase__ ) return parser.parse_args() if __name__ == "__main__": A = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
46
1
from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers A = [ 'python', 'tqdm', 'regex', 'requests', 'packaging', 'filelock', 'numpy', 'tokenizers', 'huggingface-hub', 'safetensors', 'accelerate', 'pyyaml', ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f"""can't find {pkg} in {deps.keys()}, check dependency_versions_table.py""") def a(lowercase__ , lowercase__=None ): '''simple docstring''' require_version(deps[pkg] , lowercase__ )
46
# coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import platform import sys A = '3' print('Python version:', sys.version) print('OS platform:', platform.platform()) print('OS architecture:', platform.machine()) try: import torch print('Torch version:', torch.__version__) print('Cuda available:', torch.cuda.is_available()) print('Cuda version:', torch.version.cuda) print('CuDNN version:', torch.backends.cudnn.version()) print('Number of GPUs available:', torch.cuda.device_count()) except ImportError: print('Torch version:', None) try: import transformers print('transformers version:', transformers.__version__) except ImportError: print('transformers version:', None)
46
1
import json import os from collections import Counter import torch import torchvision import torchvision.transforms as transforms from PIL import Image from torch import nn from torch.utils.data import Dataset A = {1: (1, 1), 2: (2, 1), 3: (3, 1), 4: (2, 2), 5: (5, 1), 6: (3, 2), 7: (7, 1), 8: (4, 2), 9: (3, 3)} class SCREAMING_SNAKE_CASE ( nn.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super().__init__() snake_case_ = torchvision.models.resnetaaa(pretrained=__UpperCamelCase ) snake_case_ = list(model.children() )[:-2] snake_case_ = nn.Sequential(*__UpperCamelCase ) snake_case_ = nn.AdaptiveAvgPoolad(POOLING_BREAKDOWN[args.num_image_embeds] ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = self.pool(self.model(__UpperCamelCase ) ) snake_case_ = torch.flatten(__UpperCamelCase , start_dim=2 ) snake_case_ = out.transpose(1 , 2 ).contiguous() return out # BxNx2048 class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = [json.loads(__UpperCamelCase ) for l in open(__UpperCamelCase )] snake_case_ = os.path.dirname(__UpperCamelCase ) snake_case_ = tokenizer snake_case_ = labels snake_case_ = len(__UpperCamelCase ) snake_case_ = max_seq_length snake_case_ = transforms def __len__( self ): """simple docstring""" return len(self.data ) def __getitem__( self , __UpperCamelCase ): """simple docstring""" snake_case_ = torch.LongTensor(self.tokenizer.encode(self.data[index]['text'] , add_special_tokens=__UpperCamelCase ) ) snake_case_ , snake_case_ , snake_case_ = sentence[0], sentence[1:-1], sentence[-1] snake_case_ = sentence[: self.max_seq_length] snake_case_ = torch.zeros(self.n_classes ) snake_case_ = 1 snake_case_ = Image.open(os.path.join(self.data_dir , self.data[index]['img'] ) ).convert('RGB' ) snake_case_ = self.transforms(__UpperCamelCase ) return { "image_start_token": start_token, "image_end_token": end_token, "sentence": sentence, "image": image, "label": label, } def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = Counter() for row in self.data: label_freqs.update(row['label'] ) return label_freqs def a(lowercase__ ): '''simple docstring''' snake_case_ = [len(row['sentence'] ) for row in batch] snake_case_ , snake_case_ = len(lowercase__ ), max(lowercase__ ) snake_case_ = torch.zeros(lowercase__ , lowercase__ , dtype=torch.long ) snake_case_ = torch.zeros(lowercase__ , lowercase__ , dtype=torch.long ) for i_batch, (input_row, length) in enumerate(zip(lowercase__ , lowercase__ ) ): snake_case_ = input_row['sentence'] snake_case_ = 1 snake_case_ = torch.stack([row['image'] for row in batch] ) snake_case_ = torch.stack([row['label'] for row in batch] ) snake_case_ = torch.stack([row['image_start_token'] for row in batch] ) snake_case_ = torch.stack([row['image_end_token'] for row in batch] ) return text_tensor, mask_tensor, img_tensor, img_start_token, img_end_token, tgt_tensor def a(): '''simple docstring''' return [ "Crime", "Drama", "Thriller", "Action", "Comedy", "Romance", "Documentary", "Short", "Mystery", "History", "Family", "Adventure", "Fantasy", "Sci-Fi", "Western", "Horror", "Sport", "War", "Music", "Musical", "Animation", "Biography", "Film-Noir", ] def a(): '''simple docstring''' return transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize( mean=[0.4677_7044, 0.4453_1429, 0.4066_1017] , std=[0.1222_1994, 0.1214_5835, 0.1438_0469] , ), ] )
46
import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') A = logging.getLogger(__name__) @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) __A = field( default=__snake_case , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) __A = field( default=__snake_case , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) __A = field( default=__snake_case , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) __A = field( default=__snake_case , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) __A = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) __A = field( default=__snake_case , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = field(default=__snake_case , metadata={"""help""": """The input training data file (a text file)."""} ) __A = field( default=__snake_case , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) __A = field( default=__snake_case , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) __A = field( default=__snake_case , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) __A = field( default=__snake_case , metadata={ """help""": ( """The maximum total input sequence length after tokenization. If passed, sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) __A = field( default=__snake_case , metadata={ """help""": ( """Whether to pad all samples to the maximum sentence length. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch. More """ """efficient on GPU but very bad for TPU.""" ) } , ) __A = field( default=__snake_case , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) __A = field( default=__snake_case , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def __lowerCAmelCase ( self ): """simple docstring""" if self.train_file is not None: snake_case_ = self.train_file.split('.' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: snake_case_ = self.validation_file.split('.' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = 42 __A = True __A = None __A = None def __call__( self , __UpperCamelCase ): """simple docstring""" snake_case_ = 'label' if 'label' in features[0].keys() else 'labels' snake_case_ = [feature.pop(__UpperCamelCase ) for feature in features] snake_case_ = len(__UpperCamelCase ) snake_case_ = len(features[0]['input_ids'] ) snake_case_ = [ [{k: v[i] for k, v in feature.items()} for i in range(__UpperCamelCase )] for feature in features ] snake_case_ = list(chain(*__UpperCamelCase ) ) snake_case_ = self.tokenizer.pad( __UpperCamelCase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='pt' , ) # Un-flatten snake_case_ = {k: v.view(__UpperCamelCase , __UpperCamelCase , -1 ) for k, v in batch.items()} # Add back labels snake_case_ = torch.tensor(__UpperCamelCase , dtype=torch.intaa ) return batch def a(): '''simple docstring''' # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. snake_case_ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. snake_case_ , snake_case_ , snake_case_ = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: snake_case_ , snake_case_ , snake_case_ = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_swag' , lowercase__ , lowercase__ ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() snake_case_ = training_args.get_process_log_level() logger.setLevel(lowercase__ ) datasets.utils.logging.set_verbosity(lowercase__ ) transformers.utils.logging.set_verbosity(lowercase__ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + f"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(f"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. snake_case_ = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: snake_case_ = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. """ 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: snake_case_ = {} if data_args.train_file is not None: snake_case_ = data_args.train_file if data_args.validation_file is not None: snake_case_ = data_args.validation_file snake_case_ = data_args.train_file.split('.' )[-1] snake_case_ = load_dataset( lowercase__ , data_files=lowercase__ , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. snake_case_ = load_dataset( 'swag' , 'regular' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case_ = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) snake_case_ = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) snake_case_ = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=lowercase__ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. snake_case_ = [f"""ending{i}""" for i in range(4 )] snake_case_ = 'sent1' snake_case_ = 'sent2' if data_args.max_seq_length is None: snake_case_ = tokenizer.model_max_length if max_seq_length > 1024: logger.warning( 'The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value' ' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can' ' override this default with `--block_size xxx`.' ) snake_case_ = 1024 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( f"""The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the""" f"""model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.""" ) snake_case_ = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(lowercase__ ): snake_case_ = [[context] * 4 for context in examples[context_name]] snake_case_ = examples[question_header_name] snake_case_ = [ [f"""{header} {examples[end][i]}""" for end in ending_names] for i, header in enumerate(lowercase__ ) ] # Flatten out snake_case_ = list(chain(*lowercase__ ) ) snake_case_ = list(chain(*lowercase__ ) ) # Tokenize snake_case_ = tokenizer( lowercase__ , lowercase__ , truncation=lowercase__ , max_length=lowercase__ , padding='max_length' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(lowercase__ ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('--do_train requires a train dataset' ) snake_case_ = raw_datasets['train'] if data_args.max_train_samples is not None: snake_case_ = min(len(lowercase__ ) , data_args.max_train_samples ) snake_case_ = train_dataset.select(range(lowercase__ ) ) with training_args.main_process_first(desc='train dataset map pre-processing' ): snake_case_ = train_dataset.map( lowercase__ , batched=lowercase__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('--do_eval requires a validation dataset' ) snake_case_ = raw_datasets['validation'] if data_args.max_eval_samples is not None: snake_case_ = min(len(lowercase__ ) , data_args.max_eval_samples ) snake_case_ = eval_dataset.select(range(lowercase__ ) ) with training_args.main_process_first(desc='validation dataset map pre-processing' ): snake_case_ = eval_dataset.map( lowercase__ , batched=lowercase__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator snake_case_ = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=lowercase__ , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(lowercase__ ): snake_case_ , snake_case_ = eval_predictions snake_case_ = np.argmax(lowercase__ , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer snake_case_ = Trainer( model=lowercase__ , args=lowercase__ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=lowercase__ , data_collator=lowercase__ , compute_metrics=lowercase__ , ) # Training if training_args.do_train: snake_case_ = None if training_args.resume_from_checkpoint is not None: snake_case_ = training_args.resume_from_checkpoint elif last_checkpoint is not None: snake_case_ = last_checkpoint snake_case_ = trainer.train(resume_from_checkpoint=lowercase__ ) trainer.save_model() # Saves the tokenizer too for easy upload snake_case_ = train_result.metrics snake_case_ = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowercase__ ) ) snake_case_ = min(lowercase__ , len(lowercase__ ) ) trainer.log_metrics('train' , lowercase__ ) trainer.save_metrics('train' , lowercase__ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) snake_case_ = trainer.evaluate() snake_case_ = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(lowercase__ ) snake_case_ = min(lowercase__ , len(lowercase__ ) ) trainer.log_metrics('eval' , lowercase__ ) trainer.save_metrics('eval' , lowercase__ ) snake_case_ = { 'finetuned_from': model_args.model_name_or_path, 'tasks': 'multiple-choice', 'dataset_tags': 'swag', 'dataset_args': 'regular', 'dataset': 'SWAG', 'language': 'en', } if training_args.push_to_hub: trainer.push_to_hub(**lowercase__ ) else: trainer.create_model_card(**lowercase__ ) def a(lowercase__ ): '''simple docstring''' # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
46
1
from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) # TODO Update this A = { 'facebook/esm-1b': 'https://huggingface.co/facebook/esm-1b/resolve/main/config.json', # See all ESM models at https://huggingface.co/models?filter=esm } class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = """esm""" def __init__( self , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=7_68 , __UpperCamelCase=12 , __UpperCamelCase=12 , __UpperCamelCase=30_72 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=10_26 , __UpperCamelCase=0.02 , __UpperCamelCase=1E-12 , __UpperCamelCase="absolute" , __UpperCamelCase=True , __UpperCamelCase=None , __UpperCamelCase=False , __UpperCamelCase=False , __UpperCamelCase=None , __UpperCamelCase=None , **__UpperCamelCase , ): """simple docstring""" super().__init__(pad_token_id=__UpperCamelCase , mask_token_id=__UpperCamelCase , **__UpperCamelCase ) snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = initializer_range snake_case_ = layer_norm_eps snake_case_ = position_embedding_type snake_case_ = use_cache snake_case_ = emb_layer_norm_before snake_case_ = token_dropout snake_case_ = is_folding_model if is_folding_model: if esmfold_config is None: logger.info('No esmfold_config supplied for folding model, using default values.' ) snake_case_ = EsmFoldConfig() elif isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = EsmFoldConfig(**__UpperCamelCase ) snake_case_ = esmfold_config if vocab_list is None: logger.warning('No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!' ) snake_case_ = get_default_vocab_list() else: snake_case_ = vocab_list else: snake_case_ = None snake_case_ = None if self.esmfold_config is not None and getattr(self.esmfold_config , 'use_esm_attn_map' , __UpperCamelCase ): raise ValueError('The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!' ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = super().to_dict() if isinstance(self.esmfold_config , __UpperCamelCase ): snake_case_ = self.esmfold_config.to_dict() return output @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = None __A = True __A = False __A = False __A = False __A = 0 __A = True __A = False __A = 1_2_8 __A = None def __lowerCAmelCase ( self ): """simple docstring""" if self.trunk is None: snake_case_ = TrunkConfig() elif isinstance(self.trunk , __UpperCamelCase ): snake_case_ = TrunkConfig(**self.trunk ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = asdict(self ) snake_case_ = self.trunk.to_dict() return output @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = 4_8 __A = 1_0_2_4 __A = 1_2_8 __A = 3_2 __A = 3_2 __A = 3_2 __A = 0 __A = 0 __A = False __A = 4 __A = 1_2_8 __A = None def __lowerCAmelCase ( self ): """simple docstring""" if self.structure_module is None: snake_case_ = StructureModuleConfig() elif isinstance(self.structure_module , __UpperCamelCase ): snake_case_ = StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(f"""`max_recycles` should be positive, got {self.max_recycles}.""" ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( '`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got' f""" {self.sequence_state_dim} and {self.sequence_state_dim}.""" ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( '`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got' f""" {self.pairwise_state_dim} and {self.pairwise_state_dim}.""" ) snake_case_ = self.sequence_state_dim // self.sequence_head_width snake_case_ = self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( '`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got' f""" {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}.""" ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( '`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got' f""" {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}.""" ) if self.pairwise_state_dim % 2 != 0: raise ValueError(f"""`pairwise_state_dim` should be even, got {self.pairwise_state_dim}.""" ) if self.dropout >= 0.4: raise ValueError(f"""`dropout` should not be greater than 0.4, got {self.dropout}.""" ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = asdict(self ) snake_case_ = self.structure_module.to_dict() return output @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = 3_8_4 __A = 1_2_8 __A = 1_6 __A = 1_2_8 __A = 1_2 __A = 4 __A = 8 __A = 0.1 __A = 8 __A = 1 __A = 2 __A = 7 __A = 1_0 __A = 1E-8 __A = 1E5 def __lowerCAmelCase ( self ): """simple docstring""" return asdict(self ) def a(): '''simple docstring''' return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
46
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_speech_available, is_torch_available A = { 'configuration_audio_spectrogram_transformer': [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ASTConfig', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'ASTForAudioClassification', 'ASTModel', 'ASTPreTrainedModel', ] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['ASTFeatureExtractor'] if TYPE_CHECKING: from .configuration_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ASTConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ASTForAudioClassification, ASTModel, ASTPreTrainedModel, ) try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_audio_spectrogram_transformer import ASTFeatureExtractor else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
1
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(__snake_case ) , """Tatoeba directory does not exist.""" ) class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = tempfile.mkdtemp() return TatoebaConverter(save_dir=__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" self.resolver.convert_models(['heb-eng'] ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.resolver.write_model_card('opus-mt-he-en' , dry_run=__UpperCamelCase ) assert mmeta["long_pair"] == "heb-eng"
46
import operator as op def a(lowercase__ ): '''simple docstring''' snake_case_ = [] snake_case_ = lambda lowercase__ , lowercase__ : int(x / y ) # noqa: E731 integer division operation snake_case_ = { '^': op.pow, '*': op.mul, '/': div, '+': op.add, '-': op.sub, } # operators & their respective operation # print table header print('Symbol'.center(8 ) , 'Action'.center(12 ) , 'Stack' , sep=' | ' ) print('-' * (30 + len(lowercase__ )) ) for x in post_fix: if x.isdigit(): # if x in digit stack.append(lowercase__ ) # append x to stack # output in tabular format print(x.rjust(8 ) , ('push(' + x + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) else: snake_case_ = stack.pop() # pop stack # output in tabular format print(''.rjust(8 ) , ('pop(' + b + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) snake_case_ = stack.pop() # pop stack # output in tabular format print(''.rjust(8 ) , ('pop(' + a + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) stack.append( str(opr[x](int(lowercase__ ) , int(lowercase__ ) ) ) ) # evaluate the 2 values popped from stack & push result to stack # output in tabular format print( x.rjust(8 ) , ('push(' + a + x + b + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' , ) return int(stack[0] ) if __name__ == "__main__": A = input('\n\nEnter a Postfix Equation (space separated) = ').split(' ') print('\n\tResult = ', solve(Postfix))
46
1
import importlib.metadata from typing import Union from packaging.version import Version, parse from .constants import STR_OPERATION_TO_FUNC A = parse(importlib.metadata.version('torch')) def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' if operation not in STR_OPERATION_TO_FUNC.keys(): raise ValueError(f"""`operation` must be one of {list(STR_OPERATION_TO_FUNC.keys() )}, received {operation}""" ) snake_case_ = STR_OPERATION_TO_FUNC[operation] if isinstance(lowercase__ , lowercase__ ): snake_case_ = parse(importlib.metadata.version(lowercase__ ) ) return operation(lowercase__ , parse(lowercase__ ) ) def a(lowercase__ , lowercase__ ): '''simple docstring''' return compare_versions(lowercase__ , lowercase__ , lowercase__ )
46
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices A = logging.get_logger(__name__) A = { 'google/bit-50': 'https://huggingface.co/google/bit-50/resolve/main/config.json', } class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ): """simple docstring""" __A = """bit""" __A = ["""preactivation""", """bottleneck"""] __A = ["""SAME""", """VALID"""] def __init__( self , __UpperCamelCase=3 , __UpperCamelCase=64 , __UpperCamelCase=[2_56, 5_12, 10_24, 20_48] , __UpperCamelCase=[3, 4, 6, 3] , __UpperCamelCase="preactivation" , __UpperCamelCase="relu" , __UpperCamelCase=None , __UpperCamelCase=32 , __UpperCamelCase=0.0 , __UpperCamelCase=False , __UpperCamelCase=32 , __UpperCamelCase=1 , __UpperCamelCase=None , __UpperCamelCase=None , **__UpperCamelCase , ): """simple docstring""" super().__init__(**__UpperCamelCase ) if layer_type not in self.layer_types: raise ValueError(f"""layer_type={layer_type} is not one of {','.join(self.layer_types )}""" ) if global_padding is not None: if global_padding.upper() in self.supported_padding: snake_case_ = global_padding.upper() else: raise ValueError(f"""Padding strategy {global_padding} not supported""" ) snake_case_ = num_channels snake_case_ = embedding_size snake_case_ = hidden_sizes snake_case_ = depths snake_case_ = layer_type snake_case_ = hidden_act snake_case_ = global_padding snake_case_ = num_groups snake_case_ = drop_path_rate snake_case_ = embedding_dynamic_padding snake_case_ = output_stride snake_case_ = width_factor snake_case_ = ['stem'] + [f"""stage{idx}""" for idx in range(1 , len(__UpperCamelCase ) + 1 )] snake_case_ , snake_case_ = get_aligned_output_features_output_indices( out_features=__UpperCamelCase , out_indices=__UpperCamelCase , stage_names=self.stage_names )
46
1
def a(lowercase__ ): '''simple docstring''' for i in range(len(lowercase__ ) - 1 , 0 , -1 ): snake_case_ = False for j in range(lowercase__ , 0 , -1 ): if unsorted[j] < unsorted[j - 1]: snake_case_ , snake_case_ = unsorted[j - 1], unsorted[j] snake_case_ = True for j in range(lowercase__ ): if unsorted[j] > unsorted[j + 1]: snake_case_ , snake_case_ = unsorted[j + 1], unsorted[j] snake_case_ = True if not swapped: break return unsorted if __name__ == "__main__": import doctest doctest.testmod() A = input('Enter numbers separated by a comma:\n').strip() A = [int(item) for item in user_input.split(',')] print(f"""{cocktail_shaker_sort(unsorted) = }""")
46
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = VQModel( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=3 , ) return model @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModel(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.dummy_uncond_unet snake_case_ = DDIMScheduler() snake_case_ = self.dummy_vq_model snake_case_ = LDMPipeline(unet=__UpperCamelCase , vqvae=__UpperCamelCase , scheduler=__UpperCamelCase ) ldm.to(__UpperCamelCase ) ldm.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=2 , output_type='numpy' ).images snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=2 , output_type='numpy' , return_dict=__UpperCamelCase )[0] snake_case_ = image[0, -3:, -3:, -1] snake_case_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) snake_case_ = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) snake_case_ = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(__UpperCamelCase ) ldm.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=5 , output_type='numpy' ).images snake_case_ = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) snake_case_ = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) snake_case_ = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
46
1
from __future__ import annotations from collections.abc import Callable A = list[list[float | int]] def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = len(lowercase__ ) snake_case_ = [[0 for _ in range(size + 1 )] for _ in range(lowercase__ )] snake_case_ = 42 snake_case_ = 42 snake_case_ = 42 snake_case_ = 42 snake_case_ = 42 snake_case_ = 42 for row in range(lowercase__ ): for col in range(lowercase__ ): snake_case_ = matrix[row][col] snake_case_ = vector[row][0] snake_case_ = 0 snake_case_ = 0 while row < size and col < size: # pivoting snake_case_ = max((abs(augmented[rowa][col] ), rowa) for rowa in range(lowercase__ , lowercase__ ) )[ 1 ] if augmented[pivot_row][col] == 0: col += 1 continue else: snake_case_ , snake_case_ = augmented[pivot_row], augmented[row] for rowa in range(row + 1 , lowercase__ ): snake_case_ = augmented[rowa][col] / augmented[row][col] snake_case_ = 0 for cola in range(col + 1 , size + 1 ): augmented[rowa][cola] -= augmented[row][cola] * ratio row += 1 col += 1 # back substitution for col in range(1 , lowercase__ ): for row in range(lowercase__ ): snake_case_ = augmented[row][col] / augmented[col][col] for cola in range(lowercase__ , size + 1 ): augmented[row][cola] -= augmented[col][cola] * ratio # round to get rid of numbers like 2.000000000000004 return [ [round(augmented[row][size] / augmented[row][row] , 10 )] for row in range(lowercase__ ) ] def a(lowercase__ ): '''simple docstring''' snake_case_ = len(lowercase__ ) snake_case_ = [[0 for _ in range(lowercase__ )] for _ in range(lowercase__ )] snake_case_ = [[0] for _ in range(lowercase__ )] snake_case_ = 42 snake_case_ = 42 snake_case_ = 42 snake_case_ = 42 for x_val, y_val in enumerate(lowercase__ ): for col in range(lowercase__ ): snake_case_ = (x_val + 1) ** (size - col - 1) snake_case_ = y_val snake_case_ = solve(lowercase__ , lowercase__ ) def interpolated_func(lowercase__ ) -> int: return sum( round(coeffs[x_val][0] ) * (var ** (size - x_val - 1)) for x_val in range(lowercase__ ) ) return interpolated_func def a(lowercase__ ): '''simple docstring''' return ( 1 - variable + variable**2 - variable**3 + variable**4 - variable**5 + variable**6 - variable**7 + variable**8 - variable**9 + variable**10 ) def a(lowercase__ = question_function , lowercase__ = 10 ): '''simple docstring''' snake_case_ = [func(lowercase__ ) for x_val in range(1 , order + 1 )] snake_case_ = [ interpolate(data_points[:max_coeff] ) for max_coeff in range(1 , order + 1 ) ] snake_case_ = 0 snake_case_ = 42 snake_case_ = 42 for poly in polynomials: snake_case_ = 1 while func(lowercase__ ) == poly(lowercase__ ): x_val += 1 ret += poly(lowercase__ ) return ret if __name__ == "__main__": print(f"""{solution() = }""")
46
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class SCREAMING_SNAKE_CASE : """simple docstring""" __A = LEDConfig __A = {} __A = """gelu""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=7 , __UpperCamelCase=True , __UpperCamelCase=False , __UpperCamelCase=99 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=4 , __UpperCamelCase=37 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=20 , __UpperCamelCase=2 , __UpperCamelCase=1 , __UpperCamelCase=0 , __UpperCamelCase=4 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = eos_token_id snake_case_ = pad_token_id snake_case_ = bos_token_id snake_case_ = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after snake_case_ = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests snake_case_ = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) snake_case_ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) snake_case_ = tf.concat([input_ids, eos_tensor] , axis=1 ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) snake_case_ = prepare_led_inputs_dict(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = tf.concat( [tf.zeros_like(__UpperCamelCase )[:, :-1], tf.ones_like(__UpperCamelCase )[:, -1:]] , axis=-1 , ) snake_case_ = global_attention_mask return config, inputs_dict def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = TFLEDModel(config=__UpperCamelCase ).get_decoder() snake_case_ = inputs_dict['input_ids'] snake_case_ = input_ids[:1, :] snake_case_ = inputs_dict['attention_mask'][:1, :] snake_case_ = 1 # first forward pass snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , use_cache=__UpperCamelCase ) snake_case_ , snake_case_ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case_ = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and snake_case_ = tf.concat([input_ids, next_tokens] , axis=-1 ) snake_case_ = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase )[0] snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , past_key_values=__UpperCamelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice snake_case_ = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) snake_case_ = output_from_no_past[:, -3:, random_slice_idx] snake_case_ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__UpperCamelCase , __UpperCamelCase , rtol=1E-3 ) def a(lowercase__ , lowercase__ , lowercase__ , lowercase__=None , lowercase__=None , lowercase__=None , lowercase__=None , ): '''simple docstring''' if attention_mask is None: snake_case_ = tf.cast(tf.math.not_equal(lowercase__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: snake_case_ = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: snake_case_ = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: snake_case_ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () __A = (TFLEDForConditionalGeneration,) if is_tf_available() else () __A = ( { """conversational""": TFLEDForConditionalGeneration, """feature-extraction""": TFLEDModel, """summarization""": TFLEDForConditionalGeneration, """text2text-generation""": TFLEDForConditionalGeneration, """translation""": TFLEDForConditionalGeneration, } if is_tf_available() else {} ) __A = True __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFLEDModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = tf.zeros_like(inputs_dict['attention_mask'] ) snake_case_ = 2 snake_case_ = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict['global_attention_mask'] , ) snake_case_ = True snake_case_ = self.model_tester.seq_length snake_case_ = self.model_tester.encoder_seq_length def check_decoder_attentions_output(__UpperCamelCase ): snake_case_ = outputs.decoder_attentions self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(__UpperCamelCase ): snake_case_ = [t.numpy() for t in outputs.encoder_attentions] snake_case_ = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: snake_case_ = True snake_case_ = False snake_case_ = False snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = len(__UpperCamelCase ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) if self.is_encoder_decoder: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_decoder_attentions_output(__UpperCamelCase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) # Check attention is always last and order is fine snake_case_ = True snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(__UpperCamelCase ) ) self.assertEqual(model.config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) @unittest.skip('LED keeps using potentially symbolic tensors in conditionals and breaks tracing.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" pass def a(lowercase__ ): '''simple docstring''' return tf.constant(lowercase__ , dtype=tf.intaa ) A = 1e-4 @slow @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ).led # change to intended input here snake_case_ = _long_tensor([5_12 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = _long_tensor([1_28 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = prepare_led_inputs_dict(model.config , __UpperCamelCase , __UpperCamelCase ) snake_case_ = model(**__UpperCamelCase )[0] snake_case_ = (1, 10_24, 7_68) self.assertEqual(output.shape , __UpperCamelCase ) # change to expected output here snake_case_ = tf.convert_to_tensor( [[2.3050, 2.8279, 0.6531], [-1.8457, -0.1455, -3.5661], [-1.0186, 0.4586, -2.2043]] , ) tf.debugging.assert_near(output[:, :3, :3] , __UpperCamelCase , atol=1E-3 ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ) # change to intended input here snake_case_ = _long_tensor([5_12 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = _long_tensor([1_28 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = prepare_led_inputs_dict(model.config , __UpperCamelCase , __UpperCamelCase ) snake_case_ = model(**__UpperCamelCase )[0] snake_case_ = (1, 10_24, model.config.vocab_size) self.assertEqual(output.shape , __UpperCamelCase ) # change to expected output here snake_case_ = tf.convert_to_tensor( [[33.6507, 6.4572, 16.8089], [5.8739, -2.4238, 11.2902], [-3.2139, -4.3149, 4.2783]] , ) tf.debugging.assert_near(output[:, :3, :3] , __UpperCamelCase , atol=1E-3 , rtol=1E-3 )
46
1
import importlib import sys from argparse import REMAINDER, ArgumentParser from pathlib import Path import torch_xla.distributed.xla_multiprocessing as xmp def a(): '''simple docstring''' snake_case_ = ArgumentParser( description=( 'PyTorch TPU distributed training launch ' 'helper utility that will spawn up ' 'multiple distributed processes' ) ) # Optional arguments for the launch helper parser.add_argument('--num_cores' , type=lowercase__ , default=1 , help='Number of TPU cores to use (1 or 8).' ) # positional parser.add_argument( 'training_script' , type=lowercase__ , help=( 'The full path to the single TPU training ' 'program/script to be launched in parallel, ' 'followed by all the arguments for the ' 'training script' ) , ) # rest from the training program parser.add_argument('training_script_args' , nargs=lowercase__ ) return parser.parse_args() def a(): '''simple docstring''' snake_case_ = parse_args() # Import training_script as a module. snake_case_ = Path(args.training_script ) sys.path.append(str(script_fpath.parent.resolve() ) ) snake_case_ = script_fpath.stem snake_case_ = importlib.import_module(lowercase__ ) # Patch sys.argv snake_case_ = [args.training_script] + args.training_script_args + ['--tpu_num_cores', str(args.num_cores )] xmp.spawn(mod._mp_fn , args=() , nprocs=args.num_cores ) if __name__ == "__main__": main()
46
from collections import defaultdict def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = first_str.lower().strip() snake_case_ = second_str.lower().strip() # Remove whitespace snake_case_ = first_str.replace(' ' , '' ) snake_case_ = second_str.replace(' ' , '' ) # Strings of different lengths are not anagrams if len(lowercase__ ) != len(lowercase__ ): return False # Default values for count should be 0 snake_case_ = defaultdict(lowercase__ ) # For each character in input strings, # increment count in the corresponding for i in range(len(lowercase__ ) ): count[first_str[i]] += 1 count[second_str[i]] -= 1 return all(_count == 0 for _count in count.values() ) if __name__ == "__main__": from doctest import testmod testmod() A = input('Enter the first string ').strip() A = input('Enter the second string ').strip() A = check_anagrams(input_a, input_b) print(f"""{input_a} and {input_b} are {"" if status else "not "}anagrams.""")
46
1
import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def a(): '''simple docstring''' print('Making key files...' ) make_key_files('rsa' , 1024 ) print('Key files generation successful.' ) def a(lowercase__ ): '''simple docstring''' print('Generating prime p...' ) snake_case_ = rabinMiller.generate_large_prime(lowercase__ ) print('Generating prime q...' ) snake_case_ = rabinMiller.generate_large_prime(lowercase__ ) snake_case_ = p * q print('Generating e that is relatively prime to (p - 1) * (q - 1)...' ) while True: snake_case_ = random.randrange(2 ** (key_size - 1) , 2 ** (key_size) ) if cryptoMath.gcd(lowercase__ , (p - 1) * (q - 1) ) == 1: break print('Calculating d that is mod inverse of e...' ) snake_case_ = cryptoMath.find_mod_inverse(lowercase__ , (p - 1) * (q - 1) ) snake_case_ = (n, e) snake_case_ = (n, d) return (public_key, private_key) def a(lowercase__ , lowercase__ ): '''simple docstring''' if os.path.exists(f"""{name}_pubkey.txt""" ) or os.path.exists(f"""{name}_privkey.txt""" ): print('\nWARNING:' ) print( f"""\"{name}_pubkey.txt\" or \"{name}_privkey.txt\" already exists. \n""" 'Use a different name or delete these files and re-run this program.' ) sys.exit() snake_case_ , snake_case_ = generate_key(lowercase__ ) print(f"""\nWriting public key to file {name}_pubkey.txt...""" ) with open(f"""{name}_pubkey.txt""" , 'w' ) as out_file: out_file.write(f"""{key_size},{public_key[0]},{public_key[1]}""" ) print(f"""Writing private key to file {name}_privkey.txt...""" ) with open(f"""{name}_privkey.txt""" , 'w' ) as out_file: out_file.write(f"""{key_size},{private_key[0]},{private_key[1]}""" ) if __name__ == "__main__": main()
46
import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.dummy_uncond_unet snake_case_ = ScoreSdeVeScheduler() snake_case_ = ScoreSdeVePipeline(unet=__UpperCamelCase , scheduler=__UpperCamelCase ) sde_ve.to(__UpperCamelCase ) sde_ve.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = sde_ve(num_inference_steps=2 , output_type='numpy' , generator=__UpperCamelCase ).images snake_case_ = torch.manual_seed(0 ) snake_case_ = sde_ve(num_inference_steps=2 , output_type='numpy' , generator=__UpperCamelCase , return_dict=__UpperCamelCase )[ 0 ] snake_case_ = image[0, -3:, -3:, -1] snake_case_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) snake_case_ = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = 'google/ncsnpp-church-256' snake_case_ = UNetaDModel.from_pretrained(__UpperCamelCase ) snake_case_ = ScoreSdeVeScheduler.from_pretrained(__UpperCamelCase ) snake_case_ = ScoreSdeVePipeline(unet=__UpperCamelCase , scheduler=__UpperCamelCase ) sde_ve.to(__UpperCamelCase ) sde_ve.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = sde_ve(num_inference_steps=10 , output_type='numpy' , generator=__UpperCamelCase ).images snake_case_ = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) snake_case_ = np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
46
1
import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin A = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE ( __snake_case , unittest.TestCase ): """simple docstring""" __A = XLMRobertaTokenizer __A = XLMRobertaTokenizerFast __A = True __A = True def __lowerCAmelCase ( self ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing snake_case_ = XLMRobertaTokenizer(__UpperCamelCase , keep_accents=__UpperCamelCase ) tokenizer.save_pretrained(self.tmpdirname ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = '<pad>' snake_case_ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCamelCase ) , __UpperCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCamelCase ) , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '<s>' ) self.assertEqual(vocab_keys[1] , '<pad>' ) self.assertEqual(vocab_keys[-1] , '<mask>' ) self.assertEqual(len(__UpperCamelCase ) , 10_02 ) def __lowerCAmelCase ( self ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 10_02 ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = XLMRobertaTokenizer(__UpperCamelCase , keep_accents=__UpperCamelCase ) snake_case_ = tokenizer.tokenize('This is a test' ) self.assertListEqual(__UpperCamelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__UpperCamelCase ) , [value + tokenizer.fairseq_offset for value in [2_85, 46, 10, 1_70, 3_82]] , ) snake_case_ = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( __UpperCamelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] , ) snake_case_ = tokenizer.convert_tokens_to_ids(__UpperCamelCase ) self.assertListEqual( __UpperCamelCase , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) snake_case_ = tokenizer.convert_ids_to_tokens(__UpperCamelCase ) self.assertListEqual( __UpperCamelCase , [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] , ) def __lowerCAmelCase ( self ): """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return snake_case_ = (self.rust_tokenizer_class, 'hf-internal-testing/tiny-xlm-roberta', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): snake_case_ = self.rust_tokenizer_class.from_pretrained(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = self.tokenizer_class.from_pretrained(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = tempfile.mkdtemp() snake_case_ = tokenizer_r.save_pretrained(__UpperCamelCase ) snake_case_ = tokenizer_p.save_pretrained(__UpperCamelCase ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) snake_case_ = tuple(f for f in tokenizer_r_files if 'tokenizer.json' not in f ) self.assertSequenceEqual(__UpperCamelCase , __UpperCamelCase ) # Checks everything loads correctly in the same way snake_case_ = tokenizer_r.from_pretrained(__UpperCamelCase ) snake_case_ = tokenizer_p.from_pretrained(__UpperCamelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__UpperCamelCase , __UpperCamelCase ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(__UpperCamelCase ) # Save tokenizer rust, legacy_format=True snake_case_ = tempfile.mkdtemp() snake_case_ = tokenizer_r.save_pretrained(__UpperCamelCase , legacy_format=__UpperCamelCase ) snake_case_ = tokenizer_p.save_pretrained(__UpperCamelCase ) # Checks it save with the same files self.assertSequenceEqual(__UpperCamelCase , __UpperCamelCase ) # Checks everything loads correctly in the same way snake_case_ = tokenizer_r.from_pretrained(__UpperCamelCase ) snake_case_ = tokenizer_p.from_pretrained(__UpperCamelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__UpperCamelCase , __UpperCamelCase ) ) shutil.rmtree(__UpperCamelCase ) # Save tokenizer rust, legacy_format=False snake_case_ = tempfile.mkdtemp() snake_case_ = tokenizer_r.save_pretrained(__UpperCamelCase , legacy_format=__UpperCamelCase ) snake_case_ = tokenizer_p.save_pretrained(__UpperCamelCase ) # Checks it saved the tokenizer.json file self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way snake_case_ = tokenizer_r.from_pretrained(__UpperCamelCase ) snake_case_ = tokenizer_p.from_pretrained(__UpperCamelCase ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(__UpperCamelCase , __UpperCamelCase ) ) shutil.rmtree(__UpperCamelCase ) @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return XLMRobertaTokenizer.from_pretrained('xlm-roberta-base' ) def __lowerCAmelCase ( self ): """simple docstring""" with tempfile.NamedTemporaryFile() as f: shutil.copyfile(__UpperCamelCase , f.name ) snake_case_ = XLMRobertaTokenizer(f.name , keep_accents=__UpperCamelCase ) snake_case_ = pickle.dumps(__UpperCamelCase ) pickle.loads(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" if not self.test_rust_tokenizer: return snake_case_ = self.get_tokenizer() snake_case_ = self.get_rust_tokenizer() snake_case_ = 'I was born in 92000, and this is falsé.' snake_case_ = tokenizer.tokenize(__UpperCamelCase ) snake_case_ = rust_tokenizer.tokenize(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) snake_case_ = tokenizer.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) snake_case_ = rust_tokenizer.encode(__UpperCamelCase , add_special_tokens=__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) snake_case_ = self.get_rust_tokenizer() snake_case_ = tokenizer.encode(__UpperCamelCase ) snake_case_ = rust_tokenizer.encode(__UpperCamelCase ) self.assertListEqual(__UpperCamelCase , __UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = 'Hello World!' snake_case_ = [0, 3_53_78, 66_61, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__UpperCamelCase , self.big_tokenizer.encode(__UpperCamelCase ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) snake_case_ = [ 0, 32_93, 83, 10, 45_52, 49_89, 79_86, 6_78, 10, 59_15, 1_11, 17_94_59, 12_48_50, 4, 60_44, 2_37, 12, 6, 5, 6, 4, 67_80, 7_05, 15, 13_88, 44, 3_78, 1_01_14, 7_11, 1_52, 20, 6, 5, 2_23_76, 6_42, 12_21, 1_51_90, 3_41_53, 4_50, 56_08, 9_59, 11_19, 5_77_02, 1_36, 1_86, 47, 10_98, 2_93_67, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 60_44, 2_37, 62_84, 5_09_01, 5_28, 31, 90, 34, 9_27, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(__UpperCamelCase , self.big_tokenizer.encode(__UpperCamelCase ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = {'input_ids': [[0, 1_10_62, 8_27_72, 7, 15, 8_27_72, 5_38, 5_15_29, 2_37, 1_71_98, 12_90, 2_06, 9, 21_51_75, 13_14, 1_36, 1_71_98, 12_90, 2_06, 9, 5_63_59, 42, 12_20_09, 9, 1_64_66, 16, 8_73_44, 45_37, 9, 47_17, 7_83_81, 6, 15_99_58, 7, 15, 2_44_80, 6_18, 4, 5_27, 2_26_93, 54_28, 4, 27_77, 2_44_80, 98_74, 4, 4_35_23, 5_94, 4, 8_03, 1_83_92, 3_31_89, 18, 4, 4_35_23, 2_44_47, 1_23_99, 1_00, 2_49_55, 8_36_58, 96_26, 14_40_57, 15, 8_39, 2_23_35, 16, 1_36, 2_49_55, 8_36_58, 8_34_79, 15, 3_91_02, 7_24, 16, 6_78, 6_45, 27_89, 13_28, 45_89, 42, 12_20_09, 11_57_74, 23, 8_05, 13_28, 4_68_76, 7, 1_36, 5_38_94, 19_40, 4_22_27, 4_11_59, 1_77_21, 8_23, 4_25, 4, 2_75_12, 9_87_22, 2_06, 1_36, 55_31, 49_70, 9_19, 1_73_36, 5, 2], [0, 2_00_80, 6_18, 83, 8_27_75, 47, 4_79, 9, 15_17, 73, 5_38_94, 3_33, 8_05_81, 11_01_17, 1_88_11, 52_56, 12_95, 51, 15_25_26, 2_97, 79_86, 3_90, 12_44_16, 5_38, 3_54_31, 2_14, 98, 1_50_44, 2_57_37, 1_36, 71_08, 4_37_01, 23, 7_56, 13_53_55, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_81, 6_37_73, 11_94_55, 6, 14_77_97, 8_82_03, 7, 6_45, 70, 21, 32_85, 1_02_69, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__UpperCamelCase , model_name='xlm-roberta-base' , revision='d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3' , )
46
from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput A = logging.get_logger(__name__) # pylint: disable=invalid-name class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ): """simple docstring""" @register_to_config def __init__( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None ): """simple docstring""" super().__init__() snake_case_ = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" snake_case_ = torch.zeros(__UpperCamelCase , __UpperCamelCase ) else: snake_case_ = None snake_case_ = torch.nn.Parameter(__UpperCamelCase ) class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" super().__init__() self.register_modules( vqvae=__UpperCamelCase , transformer=__UpperCamelCase , text_encoder=__UpperCamelCase , tokenizer=__UpperCamelCase , scheduler=__UpperCamelCase , learned_classifier_free_sampling_embeddings=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = len(__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else 1 # get prompt text embeddings snake_case_ = self.tokenizer( __UpperCamelCase , padding='max_length' , max_length=self.tokenizer.model_max_length , return_tensors='pt' , ) snake_case_ = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: snake_case_ = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( 'The following part of your input was truncated because CLIP can only handle sequences up to' f""" {self.tokenizer.model_max_length} tokens: {removed_text}""" ) snake_case_ = text_input_ids[:, : self.tokenizer.model_max_length] snake_case_ = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 snake_case_ = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate text embeddings for each generation per prompt snake_case_ = prompt_embeds.repeat_interleave(__UpperCamelCase , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: snake_case_ = self.learned_classifier_free_sampling_embeddings.embeddings snake_case_ = negative_prompt_embeds.unsqueeze(0 ).repeat(__UpperCamelCase , 1 , 1 ) else: snake_case_ = [''] * batch_size snake_case_ = text_input_ids.shape[-1] snake_case_ = self.tokenizer( __UpperCamelCase , padding='max_length' , max_length=__UpperCamelCase , truncation=__UpperCamelCase , return_tensors='pt' , ) snake_case_ = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings snake_case_ = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method snake_case_ = negative_prompt_embeds.shape[1] snake_case_ = negative_prompt_embeds.repeat(1 , __UpperCamelCase , 1 ) snake_case_ = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __UpperCamelCase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes snake_case_ = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self , __UpperCamelCase , __UpperCamelCase = 1_00 , __UpperCamelCase = 5.0 , __UpperCamelCase = 1.0 , __UpperCamelCase = 1 , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = "pil" , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = 1 , ): """simple docstring""" if isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = 1 elif isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = len(__UpperCamelCase ) else: raise ValueError(f"""`prompt` has to be of type `str` or `list` but is {type(__UpperCamelCase )}""" ) snake_case_ = batch_size * num_images_per_prompt snake_case_ = guidance_scale > 1.0 snake_case_ = self._encode_prompt(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__UpperCamelCase , __UpperCamelCase ) or callback_steps <= 0) ): raise ValueError( f"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" f""" {type(__UpperCamelCase )}.""" ) # get the initial completely masked latents unless the user supplied it snake_case_ = (batch_size, self.transformer.num_latent_pixels) if latents is None: snake_case_ = self.transformer.num_vector_embeds - 1 snake_case_ = torch.full(__UpperCamelCase , __UpperCamelCase ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {latents_shape}""" ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( 'Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0,' f""" {self.transformer.num_vector_embeds - 1} (inclusive).""" ) snake_case_ = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(__UpperCamelCase , device=self.device ) snake_case_ = self.scheduler.timesteps.to(self.device ) snake_case_ = latents for i, t in enumerate(self.progress_bar(__UpperCamelCase ) ): # expand the sample if we are doing classifier free guidance snake_case_ = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` snake_case_ = self.transformer(__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , timestep=__UpperCamelCase ).sample if do_classifier_free_guidance: snake_case_ , snake_case_ = model_output.chunk(2 ) snake_case_ = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(__UpperCamelCase , dim=1 , keepdim=__UpperCamelCase ) snake_case_ = self.truncate(__UpperCamelCase , __UpperCamelCase ) # remove `log(0)`'s (`-inf`s) snake_case_ = model_output.clamp(-70 ) # compute the previous noisy sample x_t -> x_t-1 snake_case_ = self.scheduler.step(__UpperCamelCase , timestep=__UpperCamelCase , sample=__UpperCamelCase , generator=__UpperCamelCase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = self.vqvae.config.vq_embed_dim snake_case_ = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) snake_case_ = self.vqvae.quantize.get_codebook_entry(__UpperCamelCase , shape=__UpperCamelCase ) snake_case_ = self.vqvae.decode(__UpperCamelCase , force_not_quantize=__UpperCamelCase ).sample snake_case_ = (image / 2 + 0.5).clamp(0 , 1 ) snake_case_ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case_ = self.numpy_to_pil(__UpperCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ , snake_case_ = torch.sort(__UpperCamelCase , 1 , descending=__UpperCamelCase ) snake_case_ = torch.exp(__UpperCamelCase ) snake_case_ = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out snake_case_ = torch.full_like(keep_mask[:, 0:1, :] , __UpperCamelCase ) snake_case_ = torch.cat((all_true, keep_mask) , dim=1 ) snake_case_ = keep_mask[:, :-1, :] snake_case_ = keep_mask.gather(1 , indices.argsort(1 ) ) snake_case_ = log_p_x_0.clone() snake_case_ = -torch.inf # -inf = log(0) return rv
46
1
import json import os import tempfile from transformers.testing_utils import check_json_file_has_correct_format class SCREAMING_SNAKE_CASE : """simple docstring""" __A = None def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.feature_extraction_class(**self.feat_extract_dict ) snake_case_ = json.loads(feat_extract.to_json_string() ) for key, value in self.feat_extract_dict.items(): self.assertEqual(obj[key] , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: snake_case_ = os.path.join(__UpperCamelCase , 'feat_extract.json' ) feat_extract_first.to_json_file(__UpperCamelCase ) snake_case_ = self.feature_extraction_class.from_json_file(__UpperCamelCase ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: snake_case_ = feat_extract_first.save_pretrained(__UpperCamelCase )[0] check_json_file_has_correct_format(__UpperCamelCase ) snake_case_ = self.feature_extraction_class.from_pretrained(__UpperCamelCase ) self.assertEqual(feat_extract_second.to_dict() , feat_extract_first.to_dict() ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.feature_extraction_class() self.assertIsNotNone(__UpperCamelCase )
46
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'hidden_sizes' ) ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'neck_hidden_sizes' ) ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'num_attention_heads' ) ) class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=3 , __UpperCamelCase=6_40 , __UpperCamelCase=4 , __UpperCamelCase="silu" , __UpperCamelCase=3 , __UpperCamelCase=32 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.02 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=10 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = last_hidden_size snake_case_ = num_attention_heads snake_case_ = hidden_act snake_case_ = conv_kernel_size snake_case_ = output_stride snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = classifier_dropout_prob snake_case_ = use_labels snake_case_ = is_training snake_case_ = num_labels snake_case_ = initializer_range snake_case_ = scope def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) snake_case_ = self.get_config() return config, pixel_values, labels, pixel_labels def __lowerCAmelCase ( self ): """simple docstring""" return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MobileViTModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MobileViTForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MobileViTForSemanticSegmentation(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __A = ( { """feature-extraction""": MobileViTModel, """image-classification""": MobileViTForImageClassification, """image-segmentation""": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTModelTester(self ) snake_case_ = MobileViTConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='MobileViT does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MobileViT does not support input and output embeddings' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MobileViT does not output attentions' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" def check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.hidden_states snake_case_ = 5 self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. snake_case_ = 2 for i in range(len(__UpperCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = MobileViTModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def a(): '''simple docstring''' snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return MobileViTImageProcessor.from_pretrained('apple/mobilevit-xx-small' ) if is_vision_available() else None @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForImageClassification.from_pretrained('apple/mobilevit-xx-small' ).to(__UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([-1.9364, -1.2327, -0.4653] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = model.to(__UpperCamelCase ) snake_case_ = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) snake_case_ = outputs.logits # verify the logits snake_case_ = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [ [[6.9713, 6.9786, 7.2422], [7.2893, 7.2825, 7.4446], [7.6580, 7.8797, 7.9420]], [[-10.6869, -10.3250, -10.3471], [-10.4228, -9.9868, -9.7132], [-11.0405, -11.0221, -10.7318]], [[-3.3089, -2.8539, -2.6740], [-3.2706, -2.5621, -2.5108], [-3.2534, -2.6615, -2.6651]], ] , device=__UpperCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = model.to(__UpperCamelCase ) snake_case_ = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) snake_case_ = outputs.logits.detach().cpu() snake_case_ = image_processor.post_process_semantic_segmentation(outputs=__UpperCamelCase , target_sizes=[(50, 60)] ) snake_case_ = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , __UpperCamelCase ) snake_case_ = image_processor.post_process_semantic_segmentation(outputs=__UpperCamelCase ) snake_case_ = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , __UpperCamelCase )
46
1
import unittest from transformers import AutoTokenizer, NystromformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, NystromformerModel, ) from transformers.models.nystromformer.modeling_nystromformer import NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=7 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=99 , __UpperCamelCase=32 , __UpperCamelCase=5 , __UpperCamelCase=4 , __UpperCamelCase=37 , __UpperCamelCase="gelu" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=5_12 , __UpperCamelCase=16 , __UpperCamelCase=2 , __UpperCamelCase=0.02 , __UpperCamelCase=3 , __UpperCamelCase=4 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_input_mask snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_labels snake_case_ = num_choices snake_case_ = scope def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None if self.use_token_type_ids: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowerCAmelCase ( self ): """simple docstring""" return NystromformerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = NystromformerModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase ) snake_case_ = model(__UpperCamelCase , token_type_ids=__UpperCamelCase ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = NystromformerForMaskedLM(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = NystromformerForQuestionAnswering(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , start_positions=__UpperCamelCase , end_positions=__UpperCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = NystromformerForSequenceClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = NystromformerForTokenClassification(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_choices snake_case_ = NystromformerForMultipleChoice(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = config_and_inputs snake_case_ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( ( NystromformerModel, NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, ) if is_torch_available() else () ) __A = ( { """feature-extraction""": NystromformerModel, """fill-mask""": NystromformerForMaskedLM, """question-answering""": NystromformerForQuestionAnswering, """text-classification""": NystromformerForSequenceClassification, """token-classification""": NystromformerForTokenClassification, """zero-shot""": NystromformerForSequenceClassification, } if is_torch_available() else {} ) __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = NystromformerModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , hidden_size=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: snake_case_ = type self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = NystromformerModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = NystromformerModel.from_pretrained('uw-madison/nystromformer-512' ) snake_case_ = torch.tensor([[0, 1, 2, 3, 4, 5]] ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = torch.Size((1, 6, 7_68) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[-0.4532, -0.0936, 0.5137], [-0.2676, 0.0628, 0.6186], [-0.3629, -0.1726, 0.4716]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = 'the [MASK] of Belgium is Brussels' snake_case_ = AutoTokenizer.from_pretrained('uw-madison/nystromformer-512' ) snake_case_ = NystromformerForMaskedLM.from_pretrained('uw-madison/nystromformer-512' ) snake_case_ = tokenizer(__UpperCamelCase , return_tensors='pt' ) with torch.no_grad(): snake_case_ = model(encoding.input_ids ).logits snake_case_ = token_logits[:, 2, :].argmax(-1 )[0] self.assertEqual(tokenizer.decode(__UpperCamelCase ) , 'capital' )
46
from ..utils import DummyObject, requires_backends class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] )
46
1
from __future__ import annotations from collections.abc import Sequence from typing import Literal def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = list(lowercase__ ) snake_case_ = list(lowercase__ ) snake_case_ = 0 for i in range(len(lowercase__ ) ): if lista[i] != lista[i]: count += 1 snake_case_ = '_' if count > 1: return False else: return "".join(lowercase__ ) def a(lowercase__ ): '''simple docstring''' snake_case_ = [] while True: snake_case_ = ['$'] * len(lowercase__ ) snake_case_ = [] for i in range(len(lowercase__ ) ): for j in range(i + 1 , len(lowercase__ ) ): snake_case_ = compare_string(binary[i] , binary[j] ) if k is False: snake_case_ = '*' snake_case_ = '*' temp.append('X' ) for i in range(len(lowercase__ ) ): if checka[i] == "$": pi.append(binary[i] ) if len(lowercase__ ) == 0: return pi snake_case_ = list(set(lowercase__ ) ) def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = [] for minterm in minterms: snake_case_ = '' for _ in range(lowercase__ ): snake_case_ = str(minterm % 2 ) + string minterm //= 2 temp.append(lowercase__ ) return temp def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = list(lowercase__ ) snake_case_ = list(lowercase__ ) snake_case_ = 0 for i in range(len(lowercase__ ) ): if lista[i] != lista[i]: count_n += 1 return count_n == count def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = [] snake_case_ = [0] * len(lowercase__ ) for i in range(len(chart[0] ) ): snake_case_ = 0 snake_case_ = -1 for j in range(len(lowercase__ ) ): if chart[j][i] == 1: count += 1 snake_case_ = j if count == 1: snake_case_ = 1 for i in range(len(lowercase__ ) ): if select[i] == 1: for j in range(len(chart[0] ) ): if chart[i][j] == 1: for k in range(len(lowercase__ ) ): snake_case_ = 0 temp.append(prime_implicants[i] ) while True: snake_case_ = 0 snake_case_ = -1 snake_case_ = 0 for i in range(len(lowercase__ ) ): snake_case_ = chart[i].count(1 ) if count_n > max_n: snake_case_ = count_n snake_case_ = i if max_n == 0: return temp temp.append(prime_implicants[rem] ) for i in range(len(chart[0] ) ): if chart[rem][i] == 1: for j in range(len(lowercase__ ) ): snake_case_ = 0 def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = [[0 for x in range(len(lowercase__ ) )] for x in range(len(lowercase__ ) )] for i in range(len(lowercase__ ) ): snake_case_ = prime_implicants[i].count('_' ) for j in range(len(lowercase__ ) ): if is_for_table(prime_implicants[i] , binary[j] , lowercase__ ): snake_case_ = 1 return chart def a(): '''simple docstring''' snake_case_ = int(input('Enter the no. of variables\n' ) ) snake_case_ = [ float(lowercase__ ) for x in input( 'Enter the decimal representation of Minterms \'Spaces Separated\'\n' ).split() ] snake_case_ = decimal_to_binary(lowercase__ , lowercase__ ) snake_case_ = check(lowercase__ ) print('Prime Implicants are:' ) print(lowercase__ ) snake_case_ = prime_implicant_chart(lowercase__ , lowercase__ ) snake_case_ = selection(lowercase__ , lowercase__ ) print('Essential Prime Implicants are:' ) print(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod() main()
46
from __future__ import annotations import os import tempfile import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import is_tensorflow_text_available, is_tf_available from transformers.testing_utils import require_tensorflow_text, require_tf, slow from ..test_modeling_tf_common import floats_tensor from .test_framework_agnostic import GenerationIntegrationTestsMixin if is_tf_available(): import tensorflow as tf from transformers import ( AutoTokenizer, TFAutoModelForCausalLM, TFAutoModelForSeqaSeqLM, TFAutoModelForSpeechSeqaSeq, TFAutoModelForVisionaSeq, TFBartForConditionalGeneration, TFLogitsProcessorList, TFMinLengthLogitsProcessor, tf_top_k_top_p_filtering, ) if is_tensorflow_text_available(): import tensorflow_text as text @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = tf.convert_to_tensor( [ [ 8.222_0991, # 3rd highest value; idx. 0 -0.562_0044, 5.2322_9752, 4.038_6393, -6.879_8378, -0.5478_5802, -3.201_2153, 2.9277_7176, 1.8817_1953, 7.3534_1276, # 5th highest value; idx. 9 8.4320_7833, # 2nd highest value; idx. 10 -9.8571_1836, -5.9620_9236, -1.1303_9161, -7.111_5294, -0.836_9633, -5.318_6408, 7.0642_7407, 0.8136_9344, -0.8202_3817, -5.917_9796, 0.5881_3443, -6.9977_8438, 4.7155_1189, -0.1877_1637, 7.4402_0759, # 4th highest value; idx. 25 9.3845_0987, # 1st highest value; idx. 26 2.1266_2941, -9.3256_2038, 2.3565_2522, ], # cummulative prob of 5 highest values <= 0.6 [ 0.5842_5518, 4.5313_9238, -5.5751_0464, -6.2803_0699, -7.1952_9503, -4.0212_2551, 1.3933_7037, -6.0670_7057, 1.5948_0517, -9.64_3119, 0.0390_7799, 0.6723_1762, -8.8820_6726, 6.2711_5922, # 4th highest value; idx. 13 2.2852_0723, 4.8276_7506, 4.3042_1368, 8.827_5313, # 2nd highest value; idx. 17 5.4402_9958, # 5th highest value; idx. 18 -4.473_5794, 7.3857_9536, # 3rd highest value; idx. 20 -2.9105_1663, 2.6194_6077, -2.567_4762, -9.4895_9302, -4.0292_2645, -1.3541_6918, 9.6770_2323, # 1st highest value; idx. 27 -5.8947_8553, 1.8537_0467, ], # cummulative prob of 5 highest values <= 0.6 ] , dtype=tf.floataa , ) snake_case_ = tf.convert_to_tensor( [[0, 0], [0, 9], [0, 10], [0, 25], [0, 26], [1, 13], [1, 17], [1, 18], [1, 20], [1, 27]] , dtype=tf.intaa , ) # expected non filtered idx as noted above snake_case_ = tf.convert_to_tensor( [8.22_2099, 7.353_4126, 8.43_2078, 7.440_2075, 9.3_8451, 6.27_1159, 8.82_7531, 5.440_2995, 7.385_7956, 9.67_7023] , dtype=tf.floataa , ) # expected non filtered values as noted above snake_case_ = tf_top_k_top_p_filtering(__UpperCamelCase , top_k=10 , top_p=0.6 , min_tokens_to_keep=4 ) snake_case_ = output[output != -float('inf' )] snake_case_ = tf.cast( tf.where(tf.not_equal(__UpperCamelCase , tf.constant(-float('inf' ) , dtype=tf.floataa ) ) ) , dtype=tf.intaa , ) tf.debugging.assert_near(__UpperCamelCase , __UpperCamelCase , rtol=1E-12 ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase , __snake_case ): """simple docstring""" if is_tf_available(): __A = { """AutoModelForCausalLM""": TFAutoModelForCausalLM, """AutoModelForSpeechSeq2Seq""": TFAutoModelForSpeechSeqaSeq, """AutoModelForSeq2SeqLM""": TFAutoModelForSeqaSeqLM, """AutoModelForVision2Seq""": TFAutoModelForVisionaSeq, """LogitsProcessorList""": TFLogitsProcessorList, """MinLengthLogitsProcessor""": TFMinLengthLogitsProcessor, """create_tensor_fn""": tf.convert_to_tensor, """floats_tensor""": floats_tensor, """return_tensors""": """tf""", } @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 2 snake_case_ = 2 class SCREAMING_SNAKE_CASE ( tf.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super(__UpperCamelCase , self ).__init__() snake_case_ = model @tf.function( input_signature=( tf.TensorSpec((None, input_length) , tf.intaa , name='input_ids' ), tf.TensorSpec((None, input_length) , tf.intaa , name='attention_mask' ), ) , jit_compile=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.model.generate( input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase , max_new_tokens=__UpperCamelCase , return_dict_in_generate=__UpperCamelCase , ) return {"sequences": outputs["sequences"]} snake_case_ = [[2, 0], [1_02, 1_03]] snake_case_ = [[1, 0], [1, 1]] snake_case_ = DummyModel(model=__UpperCamelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__UpperCamelCase , __UpperCamelCase , signatures={'serving_default': dummy_model.serving} ) snake_case_ = tf.saved_model.load(__UpperCamelCase ).signatures['serving_default'] for batch_size in range(1 , len(__UpperCamelCase ) + 1 ): snake_case_ = { 'input_ids': tf.constant(dummy_input_ids[:batch_size] ), 'attention_mask': tf.constant(dummy_attention_masks[:batch_size] ), } snake_case_ = serving_func(**__UpperCamelCase )['sequences'] snake_case_ = test_model.generate(**__UpperCamelCase , max_new_tokens=__UpperCamelCase ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 1 snake_case_ = 2 class SCREAMING_SNAKE_CASE ( tf.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super(__UpperCamelCase , self ).__init__() snake_case_ = model @tf.function( input_signature=( tf.TensorSpec((batch_size, None) , tf.intaa , name='input_ids' ), tf.TensorSpec((batch_size, None) , tf.intaa , name='attention_mask' ), ) , jit_compile=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.model.generate( input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase , max_new_tokens=__UpperCamelCase , return_dict_in_generate=__UpperCamelCase , ) return {"sequences": outputs["sequences"]} snake_case_ = [[2], [1_02, 1_03]] snake_case_ = [[1], [1, 1]] snake_case_ = DummyModel(model=__UpperCamelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__UpperCamelCase , __UpperCamelCase , signatures={'serving_default': dummy_model.serving} ) snake_case_ = tf.saved_model.load(__UpperCamelCase ).signatures['serving_default'] for input_row in range(len(__UpperCamelCase ) ): snake_case_ = { 'input_ids': tf.constant([dummy_input_ids[input_row]] ), 'attention_mask': tf.constant([dummy_attention_masks[input_row]] ), } snake_case_ = serving_func(**__UpperCamelCase )['sequences'] snake_case_ = test_model.generate(**__UpperCamelCase , max_new_tokens=__UpperCamelCase ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @slow @require_tensorflow_text def __lowerCAmelCase ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: # file needed to load the TF tokenizer hf_hub_download(repo_id='google/flan-t5-small' , filename='spiece.model' , local_dir=__UpperCamelCase ) class SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__() snake_case_ = text.SentencepieceTokenizer( model=tf.io.gfile.GFile(os.path.join(__UpperCamelCase , 'spiece.model' ) , 'rb' ).read() ) snake_case_ = TFAutoModelForSeqaSeqLM.from_pretrained('hf-internal-testing/tiny-random-t5' ) def __lowerCAmelCase ( self , __UpperCamelCase , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" snake_case_ = self.tokenizer.tokenize(__UpperCamelCase ) snake_case_ , snake_case_ = text.pad_model_inputs( __UpperCamelCase , max_seq_length=64 , pad_value=self.model.config.pad_token_id ) snake_case_ = self.model.generate(input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase ) return self.tokenizer.detokenize(__UpperCamelCase ) snake_case_ = CompleteSentenceTransformer() snake_case_ = tf.keras.layers.Input(shape=(1,) , dtype=tf.string , name='inputs' ) snake_case_ = complete_model(__UpperCamelCase ) snake_case_ = tf.keras.Model(__UpperCamelCase , __UpperCamelCase ) keras_model.save(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = { 'do_sample': True, 'num_beams': 1, 'top_p': 0.7, 'top_k': 10, 'temperature': 0.7, } snake_case_ = 14 snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 'Hello, my dog is cute and' snake_case_ = tokenizer(__UpperCamelCase , return_tensors='tf' ) snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 6_38 # forces the generation to happen on CPU, to avoid GPU-related quirks with tf.device(':/CPU:0' ): tf.random.set_seed(0 ) snake_case_ = model.generate(**__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) snake_case_ = [6_38, 1_98] with tf.device(':/CPU:0' ): tf.random.set_seed(0 ) snake_case_ = model.generate(**__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = 'Hugging Face is a technology company based in New York and Paris.' snake_case_ = bart_tokenizer(__UpperCamelCase , return_tensors='tf' ).input_ids snake_case_ = TFBartForConditionalGeneration.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = bart_model.generate(__UpperCamelCase ).numpy() class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase=None , **__UpperCamelCase ): """simple docstring""" return super().call(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = FakeBart.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = bart_model.generate(__UpperCamelCase , foo='bar' ).numpy() self.assertTrue(np.array_equal(__UpperCamelCase , __UpperCamelCase ) ) class SCREAMING_SNAKE_CASE ( bart_model.model.encoder.__class__ ): """simple docstring""" def __lowerCAmelCase ( self , __UpperCamelCase , **__UpperCamelCase ): """simple docstring""" return super().call(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = FakeEncoder(bart_model.config , bart_model.model.shared ) snake_case_ = fake_encoder # Normal generation still works (the output will be different because the encoder weights are different) snake_case_ = bart_model.generate(__UpperCamelCase ).numpy() with self.assertRaises(__UpperCamelCase ): # FakeEncoder.call() accepts **kwargs -> no filtering -> value error due to unexpected input "foo" bart_model.generate(__UpperCamelCase , foo='bar' )
46
1
from math import sqrt def a(lowercase__ = 1000000 ): '''simple docstring''' snake_case_ = 0 snake_case_ = 0 snake_case_ = 42 while num_cuboids <= limit: max_cuboid_size += 1 for sum_shortest_sides in range(2 , 2 * max_cuboid_size + 1 ): if sqrt(sum_shortest_sides**2 + max_cuboid_size**2 ).is_integer(): num_cuboids += ( min(lowercase__ , sum_shortest_sides // 2 ) - max(1 , sum_shortest_sides - max_cuboid_size ) + 1 ) return max_cuboid_size if __name__ == "__main__": print(f"""{solution() = }""")
46
import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=2 , __UpperCamelCase=8 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=99 , __UpperCamelCase=16 , __UpperCamelCase=5 , __UpperCamelCase=2 , __UpperCamelCase=36 , __UpperCamelCase="gelu" , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=5_12 , __UpperCamelCase=16 , __UpperCamelCase=2 , __UpperCamelCase=0.02 , __UpperCamelCase=3 , __UpperCamelCase=4 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_input_mask snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_labels snake_case_ = num_choices snake_case_ = scope def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None if self.use_token_type_ids: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowerCAmelCase ( self ): """simple docstring""" return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.get_config() snake_case_ = 3_00 return config def __lowerCAmelCase ( self ): """simple docstring""" ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = self.prepare_config_and_inputs() snake_case_ = True snake_case_ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase ) snake_case_ = model(__UpperCamelCase , token_type_ids=__UpperCamelCase ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" snake_case_ = True snake_case_ = MraModel(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , encoder_attention_mask=__UpperCamelCase , ) snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , ) snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraForMaskedLM(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraForQuestionAnswering(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , start_positions=__UpperCamelCase , end_positions=__UpperCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MraForSequenceClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MraForTokenClassification(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_choices snake_case_ = MraForMultipleChoice(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = config_and_inputs snake_case_ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , unittest.TestCase ): """simple docstring""" __A = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) __A = False __A = False __A = False __A = False __A = () def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , hidden_size=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: snake_case_ = type self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = MraModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) @unittest.skip(reason='MRA does not output attentions' ) def __lowerCAmelCase ( self ): """simple docstring""" return @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraModel.from_pretrained('uw-madison/mra-base-512-4' ) snake_case_ = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = torch.Size((1, 2_56, 7_68) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[-0.0140, 0.0830, -0.0381], [0.1546, 0.1402, 0.0220], [0.1162, 0.0851, 0.0165]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraForMaskedLM.from_pretrained('uw-madison/mra-base-512-4' ) snake_case_ = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = 5_02_65 snake_case_ = torch.Size((1, 2_56, vocab_size) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[9.2595, -3.6038, 11.8819], [9.3869, -3.2693, 11.0956], [11.8524, -3.4938, 13.1210]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraForMaskedLM.from_pretrained('uw-madison/mra-base-4096-8-d3' ) snake_case_ = torch.arange(40_96 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = 5_02_65 snake_case_ = torch.Size((1, 40_96, vocab_size) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[5.4789, -2.3564, 7.5064], [7.9067, -1.3369, 9.9668], [9.0712, -1.8106, 7.0380]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) )
46
1
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: A = None A = logging.get_logger(__name__) A = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} A = { 'vocab_file': { 'facebook/mbart-large-en-ro': ( 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model' ), 'facebook/mbart-large-cc25': ( 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model' ), }, 'tokenizer_file': { 'facebook/mbart-large-en-ro': 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json', 'facebook/mbart-large-cc25': 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json', }, } A = { 'facebook/mbart-large-en-ro': 1024, 'facebook/mbart-large-cc25': 1024, } # fmt: off A = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN'] class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = VOCAB_FILES_NAMES __A = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __A = PRETRAINED_VOCAB_FILES_MAP __A = ["""input_ids""", """attention_mask"""] __A = MBartTokenizer __A = [] __A = [] def __init__( self , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase="<s>" , __UpperCamelCase="</s>" , __UpperCamelCase="</s>" , __UpperCamelCase="<s>" , __UpperCamelCase="<unk>" , __UpperCamelCase="<pad>" , __UpperCamelCase="<mask>" , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=None , **__UpperCamelCase , ): """simple docstring""" snake_case_ = AddedToken(__UpperCamelCase , lstrip=__UpperCamelCase , rstrip=__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else mask_token super().__init__( vocab_file=__UpperCamelCase , tokenizer_file=__UpperCamelCase , bos_token=__UpperCamelCase , eos_token=__UpperCamelCase , sep_token=__UpperCamelCase , cls_token=__UpperCamelCase , unk_token=__UpperCamelCase , pad_token=__UpperCamelCase , mask_token=__UpperCamelCase , src_lang=__UpperCamelCase , tgt_lang=__UpperCamelCase , additional_special_tokens=__UpperCamelCase , **__UpperCamelCase , ) snake_case_ = vocab_file snake_case_ = False if not self.vocab_file else True snake_case_ = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({'additional_special_tokens': _additional_special_tokens} ) snake_case_ = { lang_code: self.convert_tokens_to_ids(__UpperCamelCase ) for lang_code in FAIRSEQ_LANGUAGE_CODES } snake_case_ = src_lang if src_lang is not None else 'en_XX' snake_case_ = self.convert_tokens_to_ids(self._src_lang ) snake_case_ = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __lowerCAmelCase ( self ): """simple docstring""" return self._src_lang @src_lang.setter def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None ): """simple docstring""" if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None ): """simple docstring""" snake_case_ = [self.sep_token_id] snake_case_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , **__UpperCamelCase ): """simple docstring""" if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) snake_case_ = src_lang snake_case_ = self(__UpperCamelCase , add_special_tokens=__UpperCamelCase , return_tensors=__UpperCamelCase , **__UpperCamelCase ) snake_case_ = self.convert_tokens_to_ids(__UpperCamelCase ) snake_case_ = tgt_lang_id return inputs def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = "en_XX" , __UpperCamelCase = None , __UpperCamelCase = "ro_RO" , **__UpperCamelCase , ): """simple docstring""" snake_case_ = src_lang snake_case_ = tgt_lang return super().prepare_seqaseq_batch(__UpperCamelCase , __UpperCamelCase , **__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" return self.set_src_lang_special_tokens(self.src_lang ) def __lowerCAmelCase ( self ): """simple docstring""" return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = self.convert_tokens_to_ids(__UpperCamelCase ) snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] snake_case_ = self.convert_ids_to_tokens(self.prefix_tokens ) snake_case_ = self.convert_ids_to_tokens(self.suffix_tokens ) snake_case_ = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = self.convert_tokens_to_ids(__UpperCamelCase ) snake_case_ = [] snake_case_ = [self.eos_token_id, self.cur_lang_code] snake_case_ = self.convert_ids_to_tokens(self.prefix_tokens ) snake_case_ = self.convert_ids_to_tokens(self.suffix_tokens ) snake_case_ = processors.TemplateProcessing( single=prefix_tokens_str + ['$A'] + suffix_tokens_str , pair=prefix_tokens_str + ['$A', '$B'] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None ): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(__UpperCamelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory.""" ) return snake_case_ = os.path.join( __UpperCamelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCamelCase ): copyfile(self.vocab_file , __UpperCamelCase ) return (out_vocab_file,)
46
import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def a(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file snake_case_ = TapasConfig.from_json_file(lowercase__ ) # set absolute/relative position embeddings parameter snake_case_ = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "WTQ": # run_task_main.py hparams snake_case_ = 4 snake_case_ = True # hparam_utils.py hparams snake_case_ = 0.66_4694 snake_case_ = 0.20_7951 snake_case_ = 0.12_1194 snake_case_ = True snake_case_ = True snake_case_ = False snake_case_ = 0.035_2513 snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams snake_case_ = 4 snake_case_ = False # hparam_utils.py hparams snake_case_ = 36.4519 snake_case_ = 0.90_3421 snake_case_ = 222.088 snake_case_ = True snake_case_ = True snake_case_ = True snake_case_ = 0.76_3141 snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "TABFACT": snake_case_ = TapasForSequenceClassification(config=lowercase__ ) elif task == "MLM": snake_case_ = TapasForMaskedLM(config=lowercase__ ) elif task == "INTERMEDIATE_PRETRAINING": snake_case_ = TapasModel(config=lowercase__ ) else: raise ValueError(f"""Task {task} not supported.""" ) print(f"""Building PyTorch model from configuration: {config}""" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(lowercase__ , lowercase__ , lowercase__ ) # Save pytorch-model (weights and configuration) print(f"""Save PyTorch model to {pytorch_dump_path}""" ) model.save_pretrained(lowercase__ ) # Save tokenizer files print(f"""Save tokenizer files to {pytorch_dump_path}""" ) snake_case_ = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + 'vocab.txt' , model_max_length=512 ) tokenizer.save_pretrained(lowercase__ ) print('Used relative position embeddings:' , model.config.reset_position_index_per_cell ) if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( '--task', default='SQA', type=str, help='Model task for which to convert a checkpoint. Defaults to SQA.' ) parser.add_argument( '--reset_position_index_per_cell', default=False, action='store_true', help='Whether to use relative position embeddings or not. Defaults to True.', ) parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--tapas_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained TAPAS model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) A = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
46
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL A = logging.get_logger(__name__) def a(lowercase__ ): '''simple docstring''' if isinstance(lowercase__ , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(lowercase__ , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(lowercase__ ): return [[videos]] raise ValueError(f"""Could not make batched video from {videos}""" ) class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = ["""pixel_values"""] def __init__( self , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = PILImageResampling.BILINEAR , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = True , __UpperCamelCase = 1 / 2_55 , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" super().__init__(**__UpperCamelCase ) snake_case_ = size if size is not None else {'shortest_edge': 2_24} snake_case_ = get_size_dict(__UpperCamelCase , default_to_square=__UpperCamelCase ) snake_case_ = crop_size if crop_size is not None else {'height': 2_24, 'width': 2_24} snake_case_ = get_size_dict(__UpperCamelCase , param_name='crop_size' ) snake_case_ = do_resize snake_case_ = size snake_case_ = do_center_crop snake_case_ = crop_size snake_case_ = resample snake_case_ = do_rescale snake_case_ = rescale_factor snake_case_ = do_normalize snake_case_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN snake_case_ = image_std if image_std is not None else IMAGENET_STANDARD_STD def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = PILImageResampling.BILINEAR , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" snake_case_ = get_size_dict(__UpperCamelCase , default_to_square=__UpperCamelCase ) if "shortest_edge" in size: snake_case_ = get_resize_output_image_size(__UpperCamelCase , size['shortest_edge'] , default_to_square=__UpperCamelCase ) elif "height" in size and "width" in size: snake_case_ = (size['height'], size['width']) else: raise ValueError(f"""Size must have 'height' and 'width' or 'shortest_edge' as keys. Got {size.keys()}""" ) return resize(__UpperCamelCase , size=__UpperCamelCase , resample=__UpperCamelCase , data_format=__UpperCamelCase , **__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" snake_case_ = get_size_dict(__UpperCamelCase ) if "height" not in size or "width" not in size: raise ValueError(f"""Size must have 'height' and 'width' as keys. Got {size.keys()}""" ) return center_crop(__UpperCamelCase , size=(size['height'], size['width']) , data_format=__UpperCamelCase , **__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" return rescale(__UpperCamelCase , scale=__UpperCamelCase , data_format=__UpperCamelCase , **__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" return normalize(__UpperCamelCase , mean=__UpperCamelCase , std=__UpperCamelCase , data_format=__UpperCamelCase , **__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = ChannelDimension.FIRST , ): """simple docstring""" if do_resize and size is None or resample is None: raise ValueError('Size and resample must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. snake_case_ = to_numpy_array(__UpperCamelCase ) if do_resize: snake_case_ = self.resize(image=__UpperCamelCase , size=__UpperCamelCase , resample=__UpperCamelCase ) if do_center_crop: snake_case_ = self.center_crop(__UpperCamelCase , size=__UpperCamelCase ) if do_rescale: snake_case_ = self.rescale(image=__UpperCamelCase , scale=__UpperCamelCase ) if do_normalize: snake_case_ = self.normalize(image=__UpperCamelCase , mean=__UpperCamelCase , std=__UpperCamelCase ) snake_case_ = to_channel_dimension_format(__UpperCamelCase , __UpperCamelCase ) return image def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = ChannelDimension.FIRST , **__UpperCamelCase , ): """simple docstring""" snake_case_ = do_resize if do_resize is not None else self.do_resize snake_case_ = resample if resample is not None else self.resample snake_case_ = do_center_crop if do_center_crop is not None else self.do_center_crop snake_case_ = do_rescale if do_rescale is not None else self.do_rescale snake_case_ = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case_ = do_normalize if do_normalize is not None else self.do_normalize snake_case_ = image_mean if image_mean is not None else self.image_mean snake_case_ = image_std if image_std is not None else self.image_std snake_case_ = size if size is not None else self.size snake_case_ = get_size_dict(__UpperCamelCase , default_to_square=__UpperCamelCase ) snake_case_ = crop_size if crop_size is not None else self.crop_size snake_case_ = get_size_dict(__UpperCamelCase , param_name='crop_size' ) if not valid_images(__UpperCamelCase ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) snake_case_ = make_batched(__UpperCamelCase ) snake_case_ = [ [ self._preprocess_image( image=__UpperCamelCase , do_resize=__UpperCamelCase , size=__UpperCamelCase , resample=__UpperCamelCase , do_center_crop=__UpperCamelCase , crop_size=__UpperCamelCase , do_rescale=__UpperCamelCase , rescale_factor=__UpperCamelCase , do_normalize=__UpperCamelCase , image_mean=__UpperCamelCase , image_std=__UpperCamelCase , data_format=__UpperCamelCase , ) for img in video ] for video in videos ] snake_case_ = {'pixel_values': videos} return BatchFeature(data=__UpperCamelCase , tensor_type=__UpperCamelCase )
46
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=3 , __UpperCamelCase=16 , __UpperCamelCase=[1, 2, 1] , __UpperCamelCase=[2, 2, 4] , __UpperCamelCase=2 , __UpperCamelCase=2.0 , __UpperCamelCase=True , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.1 , __UpperCamelCase="gelu" , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=0.02 , __UpperCamelCase=1E-5 , __UpperCamelCase=True , __UpperCamelCase=None , __UpperCamelCase=True , __UpperCamelCase=10 , __UpperCamelCase=8 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = embed_dim snake_case_ = depths snake_case_ = num_heads snake_case_ = window_size snake_case_ = mlp_ratio snake_case_ = qkv_bias snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = drop_path_rate snake_case_ = hidden_act snake_case_ = use_absolute_embeddings snake_case_ = patch_norm snake_case_ = layer_norm_eps snake_case_ = initializer_range snake_case_ = is_training snake_case_ = scope snake_case_ = use_labels snake_case_ = type_sequence_label_size snake_case_ = encoder_stride def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = self.get_config() return config, pixel_values, labels def __lowerCAmelCase ( self ): """simple docstring""" return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = SwinvaModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) snake_case_ = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) snake_case_ = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = SwinvaForMaskedImageModeling(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images snake_case_ = 1 snake_case_ = SwinvaForMaskedImageModeling(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.type_sequence_label_size snake_case_ = SwinvaForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) __A = ( {"""feature-extraction""": SwinvaModel, """image-classification""": SwinvaForImageClassification} if is_torch_available() else {} ) __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = SwinvaModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , embed_dim=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase , nn.Linear ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = True for model_class in self.all_model_classes: snake_case_ = True snake_case_ = False snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.attentions snake_case_ = len(self.model_tester.depths ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # check that output_attentions also work using config del inputs_dict["output_attentions"] snake_case_ = True snake_case_ = config.window_size**2 snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) snake_case_ = len(__UpperCamelCase ) # Check attention is always last and order is fine snake_case_ = True snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): snake_case_ = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states snake_case_ = 2 self.assertEqual(out_len + added_hidden_states , len(__UpperCamelCase ) ) snake_case_ = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.hidden_states snake_case_ = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # Swinv2 has a different seq_length snake_case_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) snake_case_ = outputs.reshaped_hidden_states self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) snake_case_ , snake_case_ , snake_case_ , snake_case_ = reshaped_hidden_states[0].shape snake_case_ = ( reshaped_hidden_states[0].view(__UpperCamelCase , __UpperCamelCase , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) snake_case_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) snake_case_ = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = SwinvaModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = _config_zero_init(__UpperCamelCase ) for model_class in self.all_model_classes: snake_case_ = model_class(config=__UpperCamelCase ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @require_vision @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( __UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([-0.3947, -0.4306, 0.0026] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) )
46
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A = { 'configuration_xmod': [ 'XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XmodConfig', 'XmodOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'XMOD_PRETRAINED_MODEL_ARCHIVE_LIST', 'XmodForCausalLM', 'XmodForMaskedLM', 'XmodForMultipleChoice', 'XmodForQuestionAnswering', 'XmodForSequenceClassification', 'XmodForTokenClassification', 'XmodModel', 'XmodPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline A = argparse.ArgumentParser('Stable Diffusion script with intel optimization', add_help=False) parser.add_argument('--dpm', action='store_true', help='Enable DPMSolver or not') parser.add_argument('--steps', default=None, type=int, help='Num inference steps') A = parser.parse_args() A = 'cpu' A = 'a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings' A = 'path-to-your-trained-model' A = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: A = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) A = pipe.to(device) # to channels last A = pipe.unet.to(memory_format=torch.channels_last) A = pipe.vae.to(memory_format=torch.channels_last) A = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: A = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex A = torch.randn(2, 4, 64, 64) A = torch.rand(1) * 999 A = torch.randn(2, 77, 768) A = (sample, timestep, encoder_hidden_status) try: A = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: A = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) A = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) A = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: A = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute A = 666 A = torch.Generator(device).manual_seed(seed) A = {'generator': generator} if args.steps is not None: A = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): A = pipe(prompt, **generate_kwargs).images[0] # save image image.save('generated.png')
46
1
import argparse import json import os import torch from torch import nn from transformers import NllbMoeConfig, NllbMoeModel from transformers.modeling_utils import dtype_byte_size from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME def a(lowercase__ ): '''simple docstring''' snake_case_ = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', 'decoder.output_projection.weight', '_float_tensor', 'encoder.embed_positions._float_tensor', 'decoder.embed_positions._float_tensor', ] for k in ignore_keys: state_dict.pop(lowercase__ , lowercase__ ) def a(lowercase__ ): '''simple docstring''' snake_case_ , snake_case_ = emb.weight.shape snake_case_ = nn.Linear(lowercase__ , lowercase__ , bias=lowercase__ ) snake_case_ = emb.weight.data return lin_layer def a(lowercase__ , lowercase__=None ): '''simple docstring''' snake_case_ = {} for old_key in state_dict.keys(): snake_case_ = old_key if "moe_layer.experts." in key: if expert_idx is not None: snake_case_ = key.replace('moe_layer.experts.0' , f"""ffn.experts.expert_{expert_idx}""" ) else: snake_case_ = key.replace('moe_layer.experts.' , 'ffn.experts.expert_' ) if "gate" in key: snake_case_ = key.replace('.moe_layer.gate.wg' , '.ffn.router.classifier' ) if "fc2" and "experts" not in key: snake_case_ = key.replace('.fc2.' , '.ffn.fc2.' ) if "fc1" and "experts" not in key: snake_case_ = key.replace('.fc1.' , '.ffn.fc1.' ) if ".encoder_attn." in key: snake_case_ = key.replace('.encoder_attn.' , '.cross_attention.' ) if "encoder_attn_layer_norm" in key: snake_case_ = key.replace('encoder_attn_layer_norm' , 'cross_attention_layer_norm' ) if "final_layer_norm" in key: snake_case_ = key.replace('final_layer_norm' , 'ff_layer_norm' ) snake_case_ = state_dict[old_key] return new_dict def a(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ = WEIGHTS_NAME ): '''simple docstring''' snake_case_ = [] snake_case_ = 0 os.makedirs(lowercase__ , exist_ok=lowercase__ ) for expert in range(lowercase__ ): snake_case_ = switch_checkpoint_path + f"""-rank-{expert}.pt""" if os.path.isfile(lowercase__ ): snake_case_ = torch.load(lowercase__ )['model'] remove_ignore_keys_(lowercase__ ) snake_case_ = rename_fairseq_keys(lowercase__ , lowercase__ ) snake_case_ = os.path.join( lowercase__ , weights_name.replace('.bin' , f"""-{len(lowercase__ )+1:05d}-of-???.bin""" ) ) torch.save(lowercase__ , lowercase__ ) sharded_state_dicts.append(expert_state.keys() ) total_size += sum([value.numel() for key, value in expert_state.items()] ) * dtype_byte_size( expert_state[list(lowercase__ )[0]].dtype ) # Add the last block snake_case_ = os.path.join(lowercase__ , weights_name.replace('.bin' , f"""-{len(lowercase__ )+1:05d}-of-???.bin""" ) ) snake_case_ = torch.load(switch_checkpoint_path + '-shared.pt' )['model'] remove_ignore_keys_(lowercase__ ) snake_case_ = rename_fairseq_keys(lowercase__ , lowercase__ ) snake_case_ = shared_weights['decoder.embed_tokens.weight'] sharded_state_dicts.append(shared_weights.keys() ) # If we only have the shared weights (dummy model/experts saved on the same file) if len(lowercase__ ) == 1: snake_case_ = os.path.join(lowercase__ , lowercase__ ) torch.save(lowercase__ , lowercase__ ) return {weights_name: sharded_state_dicts[0]}, None else: torch.save(lowercase__ , lowercase__ ) # Otherwise, let's build the index snake_case_ = {} for idx, shard in enumerate(lowercase__ ): snake_case_ = weights_name.replace('.bin' , f"""-{idx+1:05d}-of-{len(lowercase__ ):05d}.bin""" ) snake_case_ = os.path.join(lowercase__ , weights_name.replace('.bin' , f"""-{idx+1:05d}-of-???.bin""" ) ) os.rename(lowercase__ , os.path.join(lowercase__ , lowercase__ ) ) for key in shard: snake_case_ = shard_file # Add the metadata snake_case_ = {'total_size': total_size} snake_case_ = {'metadata': metadata, 'weight_map': weight_map} with open(os.path.join(lowercase__ , lowercase__ ) , 'w' , encoding='utf-8' ) as f: snake_case_ = json.dumps(lowercase__ , indent=2 , sort_keys=lowercase__ ) + '\n' f.write(lowercase__ ) return metadata, index if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( '--nllb_moe_checkpoint_path', default='/home/arthur_huggingface_co/fairseq/weights/checkpoints/model_moe_54b/checkpoint_2_300000', type=str, required=False, help='Path to a directory containing a folder per layer. Follows the original Google format.', ) parser.add_argument('--dtype', default='float32', type=str, required=False, help='dtype of the saved model') parser.add_argument( '--pytorch_dump_folder_path', default='/home/arthur_huggingface_co/fairseq/weights/checkpoints/hf-converted-moe-54b', type=str, required=False, help='Path to the output pytorch model.', ) A = parser.parse_args() A , A = shard_on_the_fly( args.nllb_moe_checkpoint_path, args.pytorch_dump_folder_path, 128, args.dtype, ) A = NllbMoeConfig.from_pretrained( 'facebook/nllb-200-3.3B', encoder_sparse_step=4, decoder_sparse_step=4, num_experts=128 ) config.save_pretrained(args.pytorch_dump_folder_path) A = NllbMoeModel.from_pretrained(args.pytorch_dump_folder_path) print('Done') model.save_pretrained(args.pytorch_dump_folder_path)
46
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { 'microsoft/unispeech-sat-base-100h-libri-ft': ( 'https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json' ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = """unispeech-sat""" def __init__( self , __UpperCamelCase=32 , __UpperCamelCase=7_68 , __UpperCamelCase=12 , __UpperCamelCase=12 , __UpperCamelCase=30_72 , __UpperCamelCase="gelu" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.02 , __UpperCamelCase=1E-5 , __UpperCamelCase="group" , __UpperCamelCase="gelu" , __UpperCamelCase=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , __UpperCamelCase=(5, 2, 2, 2, 2, 2, 2) , __UpperCamelCase=(10, 3, 3, 3, 3, 2, 2) , __UpperCamelCase=False , __UpperCamelCase=1_28 , __UpperCamelCase=16 , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=0.05 , __UpperCamelCase=10 , __UpperCamelCase=2 , __UpperCamelCase=0.0 , __UpperCamelCase=10 , __UpperCamelCase=0 , __UpperCamelCase=3_20 , __UpperCamelCase=2 , __UpperCamelCase=0.1 , __UpperCamelCase=1_00 , __UpperCamelCase=2_56 , __UpperCamelCase=2_56 , __UpperCamelCase=0.1 , __UpperCamelCase="mean" , __UpperCamelCase=False , __UpperCamelCase=False , __UpperCamelCase=2_56 , __UpperCamelCase=(5_12, 5_12, 5_12, 5_12, 15_00) , __UpperCamelCase=(5, 3, 3, 1, 1) , __UpperCamelCase=(1, 2, 3, 1, 1) , __UpperCamelCase=5_12 , __UpperCamelCase=0 , __UpperCamelCase=1 , __UpperCamelCase=2 , __UpperCamelCase=5_04 , **__UpperCamelCase , ): """simple docstring""" super().__init__(**__UpperCamelCase , pad_token_id=__UpperCamelCase , bos_token_id=__UpperCamelCase , eos_token_id=__UpperCamelCase ) snake_case_ = hidden_size snake_case_ = feat_extract_norm snake_case_ = feat_extract_activation snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = conv_bias snake_case_ = num_conv_pos_embeddings snake_case_ = num_conv_pos_embedding_groups snake_case_ = len(self.conv_dim ) snake_case_ = num_hidden_layers snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = num_attention_heads snake_case_ = hidden_dropout snake_case_ = attention_dropout snake_case_ = activation_dropout snake_case_ = feat_proj_dropout snake_case_ = final_dropout snake_case_ = layerdrop snake_case_ = layer_norm_eps snake_case_ = initializer_range snake_case_ = vocab_size snake_case_ = num_clusters snake_case_ = do_stable_layer_norm snake_case_ = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( 'Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==' ' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =' f""" {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,""" f""" `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 snake_case_ = apply_spec_augment snake_case_ = mask_time_prob snake_case_ = mask_time_length snake_case_ = mask_time_min_masks snake_case_ = mask_feature_prob snake_case_ = mask_feature_length snake_case_ = mask_feature_min_masks # parameters for pretraining with codevector quantized representations snake_case_ = num_codevectors_per_group snake_case_ = num_codevector_groups snake_case_ = contrastive_logits_temperature snake_case_ = feat_quantizer_dropout snake_case_ = num_negatives snake_case_ = codevector_dim snake_case_ = proj_codevector_dim snake_case_ = diversity_loss_weight # ctc loss snake_case_ = ctc_loss_reduction snake_case_ = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. snake_case_ = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = xvector_output_dim @property def __lowerCAmelCase ( self ): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
46
1
def a(lowercase__ ): '''simple docstring''' snake_case_ = [[0 for _ in range(lowercase__ )] for _ in range(m + 1 )] for i in range(m + 1 ): snake_case_ = 1 for n in range(m + 1 ): for k in range(1 , lowercase__ ): memo[n][k] += memo[n][k - 1] if n - k > 0: memo[n][k] += memo[n - k - 1][k] return memo[m][m - 1] if __name__ == "__main__": import sys if len(sys.argv) == 1: try: A = int(input('Enter a number: ').strip()) print(partition(n)) except ValueError: print('Please enter a number.') else: try: A = int(sys.argv[1]) print(partition(n)) except ValueError: print('Please pass a number.')
46
class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = name snake_case_ = val def __str__( self ): """simple docstring""" return f"""{self.__class__.__name__}({self.name}, {self.val})""" def __lt__( self , __UpperCamelCase ): """simple docstring""" return self.val < other.val class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" snake_case_ = {} snake_case_ = {} snake_case_ = self.build_heap(__UpperCamelCase ) def __getitem__( self , __UpperCamelCase ): """simple docstring""" return self.get_value(__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return (idx - 1) // 2 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return idx * 2 + 1 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return idx * 2 + 2 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return self.heap_dict[key] def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = len(__UpperCamelCase ) - 1 snake_case_ = self.get_parent_idx(__UpperCamelCase ) for idx, i in enumerate(__UpperCamelCase ): snake_case_ = idx snake_case_ = i.val for i in range(__UpperCamelCase , -1 , -1 ): self.sift_down(__UpperCamelCase , __UpperCamelCase ) return array def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" while True: snake_case_ = self.get_left_child_idx(__UpperCamelCase ) # noqa: E741 snake_case_ = self.get_right_child_idx(__UpperCamelCase ) snake_case_ = idx if l < len(__UpperCamelCase ) and array[l] < array[idx]: snake_case_ = l if r < len(__UpperCamelCase ) and array[r] < array[smallest]: snake_case_ = r if smallest != idx: snake_case_ , snake_case_ = array[smallest], array[idx] ( ( snake_case_ ) , ( snake_case_ ) , ) = ( self.idx_of_element[array[smallest]], self.idx_of_element[array[idx]], ) snake_case_ = smallest else: break def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = self.get_parent_idx(__UpperCamelCase ) while p >= 0 and self.heap[p] > self.heap[idx]: snake_case_ , snake_case_ = self.heap[idx], self.heap[p] snake_case_ , snake_case_ = ( self.idx_of_element[self.heap[idx]], self.idx_of_element[self.heap[p]], ) snake_case_ = p snake_case_ = self.get_parent_idx(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" return self.heap[0] def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.heap[-1], self.heap[0] snake_case_ , snake_case_ = ( self.idx_of_element[self.heap[-1]], self.idx_of_element[self.heap[0]], ) snake_case_ = self.heap.pop() del self.idx_of_element[x] self.sift_down(0 , self.heap ) return x def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" self.heap.append(__UpperCamelCase ) snake_case_ = len(self.heap ) - 1 snake_case_ = node.val self.sift_up(len(self.heap ) - 1 ) def __lowerCAmelCase ( self ): """simple docstring""" return len(self.heap ) == 0 def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" assert ( self.heap[self.idx_of_element[node]].val > new_value ), "newValue must be less that current value" snake_case_ = new_value snake_case_ = new_value self.sift_up(self.idx_of_element[node] ) A = Node('R', -1) A = Node('B', 6) A = Node('A', 3) A = Node('X', 1) A = Node('E', 4) # Use one of these two ways to generate Min-Heap # Generating Min-Heap from array A = MinHeap([r, b, a, x, e]) # Generating Min-Heap by Insert method # myMinHeap.insert(a) # myMinHeap.insert(b) # myMinHeap.insert(x) # myMinHeap.insert(r) # myMinHeap.insert(e) # Before print('Min Heap - before decrease key') for i in my_min_heap.heap: print(i) print('Min Heap - After decrease key of node [B -> -17]') my_min_heap.decrease_key(b, -17) # After for i in my_min_heap.heap: print(i) if __name__ == "__main__": import doctest doctest.testmod()
46
1
from __future__ import annotations def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ , snake_case_ = set(lowercase__ ), [start] while stack: snake_case_ = stack.pop() explored.add(lowercase__ ) # Differences from BFS: # 1) pop last element instead of first one # 2) add adjacent elements to stack without exploring them for adj in reversed(graph[v] ): if adj not in explored: stack.append(lowercase__ ) return explored A = { 'A': ['B', 'C', 'D'], 'B': ['A', 'D', 'E'], 'C': ['A', 'F'], 'D': ['B', 'D'], 'E': ['B', 'F'], 'F': ['C', 'E', 'G'], 'G': ['F'], } if __name__ == "__main__": import doctest doctest.testmod() print(depth_first_search(G, 'A'))
46
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) A = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['PerceiverFeatureExtractor'] A = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
1
import argparse from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import BigBirdPegasusConfig, BigBirdPegasusForConditionalGeneration A = [ # tf -> hf ('/', '.'), ('layer_', 'layers.'), ('kernel', 'weight'), ('beta', 'bias'), ('gamma', 'weight'), ('pegasus', 'model'), ] A = [ ('.output.dense', '.fc2'), ('intermediate.LayerNorm', 'final_layer_norm'), ('intermediate.dense', 'fc1'), ] A = ( INIT_COMMON + [ ('attention.self.LayerNorm', 'self_attn_layer_norm'), ('attention.output.dense', 'self_attn.out_proj'), ('attention.self', 'self_attn'), ('attention.encdec.LayerNorm', 'encoder_attn_layer_norm'), ('attention.encdec_output.dense', 'encoder_attn.out_proj'), ('attention.encdec', 'encoder_attn'), ('key', 'k_proj'), ('value', 'v_proj'), ('query', 'q_proj'), ('decoder.LayerNorm', 'decoder.layernorm_embedding'), ] + END_COMMON ) A = ( INIT_COMMON + [ ('embeddings.word_embeddings', 'shared.weight'), ('embeddings.position_embeddings', 'embed_positions.weight'), ('attention.self.LayerNorm', 'self_attn_layer_norm'), ('attention.output.dense', 'self_attn.output'), ('attention.self', 'self_attn.self'), ('encoder.LayerNorm', 'encoder.layernorm_embedding'), ] + END_COMMON ) A = [ 'encdec/key/bias', 'encdec/query/bias', 'encdec/value/bias', 'self/key/bias', 'self/query/bias', 'self/value/bias', 'encdec_output/dense/bias', 'attention/output/dense/bias', ] def a(lowercase__ , lowercase__ ): '''simple docstring''' for tf_name, hf_name in patterns: snake_case_ = k.replace(lowercase__ , lowercase__ ) return k def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = BigBirdPegasusConfig(**lowercase__ ) snake_case_ = BigBirdPegasusForConditionalGeneration(lowercase__ ) snake_case_ = torch_model.state_dict() snake_case_ = {} # separating decoder weights snake_case_ = {k: tf_weights[k] for k in tf_weights if k.startswith('pegasus/decoder' )} snake_case_ = {k: tf_weights[k] for k in tf_weights if not k.startswith('pegasus/decoder' )} for k, v in tqdm(decoder_weights.items() , 'tf -> hf conversion' ): snake_case_ = [k.endswith(lowercase__ ) for ending in KEYS_TO_IGNORE] if any(lowercase__ ): continue snake_case_ = DECODER_PATTERNS snake_case_ = rename_state_dict_key(lowercase__ , lowercase__ ) if new_k not in state_dict: raise ValueError(f"""could not find new key {new_k} in state dict. (converted from {k})""" ) if any(True if i in k else False for i in ['dense', 'query', 'key', 'value'] ): snake_case_ = v.T snake_case_ = torch.from_numpy(lowercase__ ) assert v.shape == state_dict[new_k].shape, f"""{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}""" for k, v in tqdm(remaining_weights.items() , 'tf -> hf conversion' ): snake_case_ = [k.endswith(lowercase__ ) for ending in KEYS_TO_IGNORE] if any(lowercase__ ): continue snake_case_ = REMAINING_PATTERNS snake_case_ = rename_state_dict_key(lowercase__ , lowercase__ ) if new_k not in state_dict and k != "pegasus/embeddings/position_embeddings": raise ValueError(f"""could not find new key {new_k} in state dict. (converted from {k})""" ) if any(True if i in k else False for i in ['dense', 'query', 'key', 'value'] ): snake_case_ = v.T snake_case_ = torch.from_numpy(lowercase__ ) if k != "pegasus/embeddings/position_embeddings": assert v.shape == state_dict[new_k].shape, f"""{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}""" snake_case_ = mapping['model.embed_positions.weight'] snake_case_ = mapping.pop('model.embed_positions.weight' ) snake_case_ , snake_case_ = torch_model.load_state_dict(lowercase__ , strict=lowercase__ ) snake_case_ = [ k for k in missing if k not in [ 'final_logits_bias', 'model.encoder.embed_tokens.weight', 'model.decoder.embed_tokens.weight', 'lm_head.weight', ] ] assert unexpected_missing == [], f"""no matches found for the following torch keys {unexpected_missing}""" assert extra == [], f"""no matches found for the following tf keys {extra}""" return torch_model def a(lowercase__ ): '''simple docstring''' snake_case_ = tf.train.list_variables(lowercase__ ) snake_case_ = {} snake_case_ = ['global_step'] for name, shape in tqdm(lowercase__ , desc='converting tf checkpoint to dict' ): snake_case_ = any(pat in name for pat in ignore_name ) if skip_key: continue snake_case_ = tf.train.load_variable(lowercase__ , lowercase__ ) snake_case_ = array return tf_weights def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = get_tf_weights_as_numpy(lowercase__ ) snake_case_ = convert_bigbird_pegasus(lowercase__ , lowercase__ ) torch_model.save_pretrained(lowercase__ ) if __name__ == "__main__": A = argparse.ArgumentParser() parser.add_argument('--tf_ckpt_path', type=str, help='passed to tf.train.list_variables') parser.add_argument('--save_dir', default=None, type=str, help='Path to the output PyTorch model.') A = parser.parse_args() A = {} convert_bigbird_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir, config_update=config_update)
46
def a(lowercase__ , lowercase__ ): '''simple docstring''' if not isinstance(lowercase__ , lowercase__ ): raise ValueError('iterations must be defined as integers' ) if not isinstance(lowercase__ , lowercase__ ) or not number >= 1: raise ValueError( 'starting number must be\n and integer and be more than 0' ) if not iterations >= 1: raise ValueError('Iterations must be done more than 0 times to play FizzBuzz' ) snake_case_ = '' while number <= iterations: if number % 3 == 0: out += "Fizz" if number % 5 == 0: out += "Buzz" if 0 not in (number % 3, number % 5): out += str(lowercase__ ) # print(out) number += 1 out += " " return out if __name__ == "__main__": import doctest doctest.testmod()
46
1
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to have a nice missing dependency error message early on import tqdm # noqa: F401 # Here to have a nice missing dependency error message early on from mauve import compute_mauve # From: mauve-text import datasets A = '\\n@inproceedings{pillutla-etal:mauve:neurips2021,\n title={MAUVE: Measuring the Gap Between Neural Text and Human Text using Divergence Frontiers},\n author={Pillutla, Krishna and Swayamdipta, Swabha and Zellers, Rowan and Thickstun, John and Welleck, Sean and Choi, Yejin and Harchaoui, Zaid},\n booktitle = {NeurIPS},\n year = {2021}\n}\n\n' A = '\\nMAUVE is a library built on PyTorch and HuggingFace Transformers to measure the gap between neural text and human text with the eponymous MAUVE measure.\n\nMAUVE summarizes both Type I and Type II errors measured softly using Kullback–Leibler (KL) divergences.\n\nFor details, see the MAUVE paper: https://arxiv.org/abs/2102.01454 (Neurips, 2021).\n\nThis metrics is a wrapper around the official implementation of MAUVE:\nhttps://github.com/krishnap25/mauve\n' A = '\nCalculates MAUVE scores between two lists of generated text and reference text.\nArgs:\n predictions: list of generated text to score. Each predictions\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\nOptional Args:\n num_buckets: the size of the histogram to quantize P and Q. Options: \'auto\' (default) or an integer\n pca_max_data: the number data points to use for PCA dimensionality reduction prior to clustering. If -1, use all the data. Default -1\n kmeans_explained_var: amount of variance of the data to keep in dimensionality reduction by PCA. Default 0.9\n kmeans_num_redo: number of times to redo k-means clustering (the best objective is kept). Default 5\n kmeans_max_iter: maximum number of k-means iterations. Default 500\n featurize_model_name: name of the model from which features are obtained. Default \'gpt2-large\' Use one of [\'gpt2\', \'gpt2-medium\', \'gpt2-large\', \'gpt2-xl\'].\n device_id: Device for featurization. Supply a GPU id (e.g. 0 or 3) to use GPU. If no GPU with this id is found, use CPU\n max_text_length: maximum number of tokens to consider. Default 1024\n divergence_curve_discretization_size: Number of points to consider on the divergence curve. Default 25\n mauve_scaling_factor: "c" from the paper. Default 5.\n verbose: If True (default), print running time updates\n seed: random seed to initialize k-means cluster assignments.\nReturns:\n mauve: MAUVE score, a number between 0 and 1. Larger values indicate that P and Q are closer,\n frontier_integral: Frontier Integral, a number between 0 and 1. Smaller values indicate that P and Q are closer,\n divergence_curve: a numpy.ndarray of shape (m, 2); plot it with matplotlib to view the divergence curve,\n p_hist: a discrete distribution, which is a quantized version of the text distribution p_text,\n q_hist: same as above, but with q_text.\nExamples:\n\n >>> # faiss segfaults in doctest for some reason, so the .compute call is not tested with doctest\n >>> import datasets\n >>> mauve = datasets.load_metric(\'mauve\')\n >>> predictions = ["hello there", "general kenobi"]\n >>> references = ["hello there", "general kenobi"]\n >>> out = mauve.compute(predictions=predictions, references=references) # doctest: +SKIP\n >>> print(out.mauve) # doctest: +SKIP\n 1.0\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE ( datasets.Metric ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='https://github.com/krishnap25/mauve' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/krishnap25/mauve'] , reference_urls=[ 'https://arxiv.org/abs/2102.01454', 'https://github.com/krishnap25/mauve', ] , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase="auto" , __UpperCamelCase=-1 , __UpperCamelCase=0.9 , __UpperCamelCase=5 , __UpperCamelCase=5_00 , __UpperCamelCase="gpt2-large" , __UpperCamelCase=-1 , __UpperCamelCase=10_24 , __UpperCamelCase=25 , __UpperCamelCase=5 , __UpperCamelCase=True , __UpperCamelCase=25 , ): """simple docstring""" snake_case_ = compute_mauve( p_text=__UpperCamelCase , q_text=__UpperCamelCase , p_features=__UpperCamelCase , q_features=__UpperCamelCase , p_tokens=__UpperCamelCase , q_tokens=__UpperCamelCase , num_buckets=__UpperCamelCase , pca_max_data=__UpperCamelCase , kmeans_explained_var=__UpperCamelCase , kmeans_num_redo=__UpperCamelCase , kmeans_max_iter=__UpperCamelCase , featurize_model_name=__UpperCamelCase , device_id=__UpperCamelCase , max_text_length=__UpperCamelCase , divergence_curve_discretization_size=__UpperCamelCase , mauve_scaling_factor=__UpperCamelCase , verbose=__UpperCamelCase , seed=__UpperCamelCase , ) return out
46
import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def a(lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = 1.5 snake_case_ = int(factor * num_class_images ) snake_case_ = ClipClient( url='https://knn.laion.ai/knn-service' , indice_name='laion_400m' , num_images=lowercase__ , aesthetic_weight=0.1 ) os.makedirs(f"""{class_data_dir}/images""" , exist_ok=lowercase__ ) if len(list(Path(f"""{class_data_dir}/images""" ).iterdir() ) ) >= num_class_images: return while True: snake_case_ = client.query(text=lowercase__ ) if len(lowercase__ ) >= factor * num_class_images or num_images > 1e4: break else: snake_case_ = int(factor * num_images ) snake_case_ = ClipClient( url='https://knn.laion.ai/knn-service' , indice_name='laion_400m' , num_images=lowercase__ , aesthetic_weight=0.1 , ) snake_case_ = 0 snake_case_ = 0 snake_case_ = tqdm(desc='downloading real regularization images' , total=lowercase__ ) with open(f"""{class_data_dir}/caption.txt""" , 'w' ) as fa, open(f"""{class_data_dir}/urls.txt""" , 'w' ) as fa, open( f"""{class_data_dir}/images.txt""" , 'w' ) as fa: while total < num_class_images: snake_case_ = class_images[count] count += 1 try: snake_case_ = requests.get(images['url'] ) if img.status_code == 200: snake_case_ = Image.open(BytesIO(img.content ) ) with open(f"""{class_data_dir}/images/{total}.jpg""" , 'wb' ) as f: f.write(img.content ) fa.write(images['caption'] + '\n' ) fa.write(images['url'] + '\n' ) fa.write(f"""{class_data_dir}/images/{total}.jpg""" + '\n' ) total += 1 pbar.update(1 ) else: continue except Exception: continue return def a(): '''simple docstring''' snake_case_ = argparse.ArgumentParser('' , add_help=lowercase__ ) parser.add_argument('--class_prompt' , help='text prompt to retrieve images' , required=lowercase__ , type=lowercase__ ) parser.add_argument('--class_data_dir' , help='path to save images' , required=lowercase__ , type=lowercase__ ) parser.add_argument('--num_class_images' , help='number of images to download' , default=200 , type=lowercase__ ) return parser.parse_args() if __name__ == "__main__": A = parse_args() retrieve(args.class_prompt, args.class_data_dir, args.num_class_images)
46
1
import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class SCREAMING_SNAKE_CASE : """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = TaEncoderModel.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) snake_case_ = UNetaDConditionModel( sample_size=32 , layers_per_block=1 , block_out_channels=[32, 64] , down_block_types=[ 'ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D', ] , mid_block_type='UNetMidBlock2DSimpleCrossAttn' , up_block_types=['SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'] , in_channels=3 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type='text' , addition_embed_type_num_heads=2 , cross_attention_norm='group_norm' , resnet_time_scale_shift='scale_shift' , act_fn='gelu' , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) snake_case_ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule='squaredcos_cap_v2' , beta_start=0.0001 , beta_end=0.02 , thresholding=__UpperCamelCase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type='epsilon' , variance_type='learned_range' , ) torch.manual_seed(0 ) snake_case_ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = TaEncoderModel.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-t5' ) torch.manual_seed(0 ) snake_case_ = UNetaDConditionModel( sample_size=32 , layers_per_block=[1, 2] , block_out_channels=[32, 64] , down_block_types=[ 'ResnetDownsampleBlock2D', 'SimpleCrossAttnDownBlock2D', ] , mid_block_type='UNetMidBlock2DSimpleCrossAttn' , up_block_types=['SimpleCrossAttnUpBlock2D', 'ResnetUpsampleBlock2D'] , in_channels=6 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type='text' , addition_embed_type_num_heads=2 , cross_attention_norm='group_norm' , resnet_time_scale_shift='scale_shift' , act_fn='gelu' , class_embed_type='timestep' , mid_block_scale_factor=1.414 , time_embedding_act_fn='gelu' , time_embedding_dim=32 , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) snake_case_ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule='squaredcos_cap_v2' , beta_start=0.0001 , beta_end=0.02 , thresholding=__UpperCamelCase , dynamic_thresholding_ratio=0.95 , sample_max_value=1.0 , prediction_type='epsilon' , variance_type='learned_range' , ) torch.manual_seed(0 ) snake_case_ = DDPMScheduler( num_train_timesteps=10_00 , beta_schedule='squaredcos_cap_v2' , beta_start=0.0001 , beta_end=0.02 , ) torch.manual_seed(0 ) snake_case_ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.get_dummy_components() snake_case_ = self.pipeline_class(**__UpperCamelCase ) pipe.to(__UpperCamelCase ) pipe.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = self.get_dummy_inputs(__UpperCamelCase ) snake_case_ = inputs['prompt'] snake_case_ = inputs['generator'] snake_case_ = inputs['num_inference_steps'] snake_case_ = inputs['output_type'] if "image" in inputs: snake_case_ = inputs['image'] else: snake_case_ = None if "mask_image" in inputs: snake_case_ = inputs['mask_image'] else: snake_case_ = None if "original_image" in inputs: snake_case_ = inputs['original_image'] else: snake_case_ = None snake_case_ , snake_case_ = pipe.encode_prompt(__UpperCamelCase ) # inputs with prompt converted to embeddings snake_case_ = { 'prompt_embeds': prompt_embeds, 'negative_prompt_embeds': negative_prompt_embeds, 'generator': generator, 'num_inference_steps': num_inference_steps, 'output_type': output_type, } if image is not None: snake_case_ = image if mask_image is not None: snake_case_ = mask_image if original_image is not None: snake_case_ = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = pipe(**__UpperCamelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__UpperCamelCase ) snake_case_ = self.pipeline_class.from_pretrained(__UpperCamelCase ) pipe_loaded.to(__UpperCamelCase ) pipe_loaded.set_progress_bar_config(disable=__UpperCamelCase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(__UpperCamelCase , __UpperCamelCase ) is None , f"""`{optional_component}` did not stay set to None after loading.""" , ) snake_case_ = self.get_dummy_inputs(__UpperCamelCase ) snake_case_ = inputs['generator'] snake_case_ = inputs['num_inference_steps'] snake_case_ = inputs['output_type'] # inputs with prompt converted to embeddings snake_case_ = { 'prompt_embeds': prompt_embeds, 'negative_prompt_embeds': negative_prompt_embeds, 'generator': generator, 'num_inference_steps': num_inference_steps, 'output_type': output_type, } if image is not None: snake_case_ = image if mask_image is not None: snake_case_ = mask_image if original_image is not None: snake_case_ = original_image snake_case_ = pipe_loaded(**__UpperCamelCase )[0] snake_case_ = np.abs(to_np(__UpperCamelCase ) - to_np(__UpperCamelCase ) ).max() self.assertLess(__UpperCamelCase , 1E-4 ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.get_dummy_components() snake_case_ = self.pipeline_class(**__UpperCamelCase ) pipe.to(__UpperCamelCase ) pipe.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = self.get_dummy_inputs(__UpperCamelCase ) snake_case_ = pipe(**__UpperCamelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__UpperCamelCase ) snake_case_ = self.pipeline_class.from_pretrained(__UpperCamelCase ) pipe_loaded.to(__UpperCamelCase ) pipe_loaded.set_progress_bar_config(disable=__UpperCamelCase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests snake_case_ = self.get_dummy_inputs(__UpperCamelCase ) snake_case_ = pipe_loaded(**__UpperCamelCase )[0] snake_case_ = np.abs(to_np(__UpperCamelCase ) - to_np(__UpperCamelCase ) ).max() self.assertLess(__UpperCamelCase , 1E-4 )
46
# coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # this script dumps information about the environment import os import platform import sys A = '3' print('Python version:', sys.version) print('OS platform:', platform.platform()) print('OS architecture:', platform.machine()) try: import torch print('Torch version:', torch.__version__) print('Cuda available:', torch.cuda.is_available()) print('Cuda version:', torch.version.cuda) print('CuDNN version:', torch.backends.cudnn.version()) print('Number of GPUs available:', torch.cuda.device_count()) except ImportError: print('Torch version:', None) try: import transformers print('transformers version:', transformers.__version__) except ImportError: print('transformers version:', None)
46
1
from __future__ import annotations def a(lowercase__ , lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = [] snake_case_ , snake_case_ = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) snake_case_ = result + left + right return input_list def a(lowercase__ ): '''simple docstring''' if len(lowercase__ ) <= 1: return input_list snake_case_ = list(lowercase__ ) # iteration for two-way merging snake_case_ = 2 while p <= len(lowercase__ ): # getting low, high and middle value for merge-sort of single list for i in range(0 , len(lowercase__ ) , lowercase__ ): snake_case_ = i snake_case_ = i + p - 1 snake_case_ = (low + high + 1) // 2 snake_case_ = merge(lowercase__ , lowercase__ , lowercase__ , lowercase__ ) # final merge of last two parts if p * 2 >= len(lowercase__ ): snake_case_ = i snake_case_ = merge(lowercase__ , 0 , lowercase__ , len(lowercase__ ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": A = input('Enter numbers separated by a comma:\n').strip() if user_input == "": A = [] else: A = [int(item.strip()) for item in user_input.split(',')] print(iter_merge_sort(unsorted))
46
import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, AutoTokenizer, HfArgumentParser, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.tokenization_utils_base import PreTrainedTokenizerBase from transformers.trainer_utils import get_last_checkpoint from transformers.utils import PaddingStrategy, check_min_version, send_example_telemetry # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.31.0') A = logging.getLogger(__name__) @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) __A = field( default=__snake_case , metadata={"""help""": """Pretrained config name or path if not the same as model_name"""} ) __A = field( default=__snake_case , metadata={"""help""": """Pretrained tokenizer name or path if not the same as model_name"""} ) __A = field( default=__snake_case , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) __A = field( default=__snake_case , metadata={"""help""": """Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."""} , ) __A = field( default="""main""" , metadata={"""help""": """The specific model version to use (can be a branch name, tag name or commit id)."""} , ) __A = field( default=__snake_case , metadata={ """help""": ( """Will use the token generated when running `huggingface-cli login` (necessary to use this script """ """with private models).""" ) } , ) @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = field(default=__snake_case , metadata={"""help""": """The input training data file (a text file)."""} ) __A = field( default=__snake_case , metadata={"""help""": """An optional input evaluation data file to evaluate the perplexity on (a text file)."""} , ) __A = field( default=__snake_case , metadata={"""help""": """Overwrite the cached training and evaluation sets"""} ) __A = field( default=__snake_case , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) __A = field( default=__snake_case , metadata={ """help""": ( """The maximum total input sequence length after tokenization. If passed, sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) } , ) __A = field( default=__snake_case , metadata={ """help""": ( """Whether to pad all samples to the maximum sentence length. """ """If False, will pad the samples dynamically when batching to the maximum length in the batch. More """ """efficient on GPU but very bad for TPU.""" ) } , ) __A = field( default=__snake_case , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of training examples to this """ """value if set.""" ) } , ) __A = field( default=__snake_case , metadata={ """help""": ( """For debugging purposes or quicker training, truncate the number of evaluation examples to this """ """value if set.""" ) } , ) def __lowerCAmelCase ( self ): """simple docstring""" if self.train_file is not None: snake_case_ = self.train_file.split('.' )[-1] assert extension in ["csv", "json"], "`train_file` should be a csv or a json file." if self.validation_file is not None: snake_case_ = self.validation_file.split('.' )[-1] assert extension in ["csv", "json"], "`validation_file` should be a csv or a json file." @dataclass class SCREAMING_SNAKE_CASE : """simple docstring""" __A = 42 __A = True __A = None __A = None def __call__( self , __UpperCamelCase ): """simple docstring""" snake_case_ = 'label' if 'label' in features[0].keys() else 'labels' snake_case_ = [feature.pop(__UpperCamelCase ) for feature in features] snake_case_ = len(__UpperCamelCase ) snake_case_ = len(features[0]['input_ids'] ) snake_case_ = [ [{k: v[i] for k, v in feature.items()} for i in range(__UpperCamelCase )] for feature in features ] snake_case_ = list(chain(*__UpperCamelCase ) ) snake_case_ = self.tokenizer.pad( __UpperCamelCase , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='pt' , ) # Un-flatten snake_case_ = {k: v.view(__UpperCamelCase , __UpperCamelCase , -1 ) for k, v in batch.items()} # Add back labels snake_case_ = torch.tensor(__UpperCamelCase , dtype=torch.intaa ) return batch def a(): '''simple docstring''' # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. snake_case_ = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. snake_case_ , snake_case_ , snake_case_ = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: snake_case_ , snake_case_ , snake_case_ = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_swag' , lowercase__ , lowercase__ ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() snake_case_ = training_args.get_process_log_level() logger.setLevel(lowercase__ ) datasets.utils.logging.set_verbosity(lowercase__ ) transformers.utils.logging.set_verbosity(lowercase__ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( f"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + f"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) logger.info(f"""Training/evaluation parameters {training_args}""" ) # Detecting last checkpoint. snake_case_ = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: snake_case_ = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. """ 'Use --overwrite_output_dir to overcome.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( f"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.train_file is not None or data_args.validation_file is not None: snake_case_ = {} if data_args.train_file is not None: snake_case_ = data_args.train_file if data_args.validation_file is not None: snake_case_ = data_args.validation_file snake_case_ = data_args.train_file.split('.' )[-1] snake_case_ = load_dataset( lowercase__ , data_files=lowercase__ , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) else: # Downloading and loading the swag dataset from the hub. snake_case_ = load_dataset( 'swag' , 'regular' , cache_dir=model_args.cache_dir , use_auth_token=True if model_args.use_auth_token else None , ) # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Load pretrained model and tokenizer # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case_ = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) snake_case_ = AutoTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) snake_case_ = AutoModelForMultipleChoice.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=lowercase__ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # When using your own dataset or a different dataset from swag, you will probably need to change this. snake_case_ = [f"""ending{i}""" for i in range(4 )] snake_case_ = 'sent1' snake_case_ = 'sent2' if data_args.max_seq_length is None: snake_case_ = tokenizer.model_max_length if max_seq_length > 1024: logger.warning( 'The chosen tokenizer supports a `model_max_length` that is longer than the default `block_size` value' ' of 1024. If you would like to use a longer `block_size` up to `tokenizer.model_max_length` you can' ' override this default with `--block_size xxx`.' ) snake_case_ = 1024 else: if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( f"""The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the""" f"""model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.""" ) snake_case_ = min(data_args.max_seq_length , tokenizer.model_max_length ) # Preprocessing the datasets. def preprocess_function(lowercase__ ): snake_case_ = [[context] * 4 for context in examples[context_name]] snake_case_ = examples[question_header_name] snake_case_ = [ [f"""{header} {examples[end][i]}""" for end in ending_names] for i, header in enumerate(lowercase__ ) ] # Flatten out snake_case_ = list(chain(*lowercase__ ) ) snake_case_ = list(chain(*lowercase__ ) ) # Tokenize snake_case_ = tokenizer( lowercase__ , lowercase__ , truncation=lowercase__ , max_length=lowercase__ , padding='max_length' if data_args.pad_to_max_length else False , ) # Un-flatten return {k: [v[i : i + 4] for i in range(0 , len(lowercase__ ) , 4 )] for k, v in tokenized_examples.items()} if training_args.do_train: if "train" not in raw_datasets: raise ValueError('--do_train requires a train dataset' ) snake_case_ = raw_datasets['train'] if data_args.max_train_samples is not None: snake_case_ = min(len(lowercase__ ) , data_args.max_train_samples ) snake_case_ = train_dataset.select(range(lowercase__ ) ) with training_args.main_process_first(desc='train dataset map pre-processing' ): snake_case_ = train_dataset.map( lowercase__ , batched=lowercase__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) if training_args.do_eval: if "validation" not in raw_datasets: raise ValueError('--do_eval requires a validation dataset' ) snake_case_ = raw_datasets['validation'] if data_args.max_eval_samples is not None: snake_case_ = min(len(lowercase__ ) , data_args.max_eval_samples ) snake_case_ = eval_dataset.select(range(lowercase__ ) ) with training_args.main_process_first(desc='validation dataset map pre-processing' ): snake_case_ = eval_dataset.map( lowercase__ , batched=lowercase__ , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , ) # Data collator snake_case_ = ( default_data_collator if data_args.pad_to_max_length else DataCollatorForMultipleChoice(tokenizer=lowercase__ , pad_to_multiple_of=8 if training_args.fpaa else None ) ) # Metric def compute_metrics(lowercase__ ): snake_case_ , snake_case_ = eval_predictions snake_case_ = np.argmax(lowercase__ , axis=1 ) return {"accuracy": (preds == label_ids).astype(np.floataa ).mean().item()} # Initialize our Trainer snake_case_ = Trainer( model=lowercase__ , args=lowercase__ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=lowercase__ , data_collator=lowercase__ , compute_metrics=lowercase__ , ) # Training if training_args.do_train: snake_case_ = None if training_args.resume_from_checkpoint is not None: snake_case_ = training_args.resume_from_checkpoint elif last_checkpoint is not None: snake_case_ = last_checkpoint snake_case_ = trainer.train(resume_from_checkpoint=lowercase__ ) trainer.save_model() # Saves the tokenizer too for easy upload snake_case_ = train_result.metrics snake_case_ = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowercase__ ) ) snake_case_ = min(lowercase__ , len(lowercase__ ) ) trainer.log_metrics('train' , lowercase__ ) trainer.save_metrics('train' , lowercase__ ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info('*** Evaluate ***' ) snake_case_ = trainer.evaluate() snake_case_ = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(lowercase__ ) snake_case_ = min(lowercase__ , len(lowercase__ ) ) trainer.log_metrics('eval' , lowercase__ ) trainer.save_metrics('eval' , lowercase__ ) snake_case_ = { 'finetuned_from': model_args.model_name_or_path, 'tasks': 'multiple-choice', 'dataset_tags': 'swag', 'dataset_args': 'regular', 'dataset': 'SWAG', 'language': 'en', } if training_args.push_to_hub: trainer.push_to_hub(**lowercase__ ) else: trainer.create_model_card(**lowercase__ ) def a(lowercase__ ): '''simple docstring''' # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
46
1
from __future__ import annotations def a(lowercase__ ): '''simple docstring''' return len(set(lowercase__ ) ) == len(lowercase__ ) if __name__ == "__main__": import doctest doctest.testmod()
46
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_speech_available, is_torch_available A = { 'configuration_audio_spectrogram_transformer': [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ASTConfig', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'ASTForAudioClassification', 'ASTModel', 'ASTPreTrainedModel', ] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['ASTFeatureExtractor'] if TYPE_CHECKING: from .configuration_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, ASTConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ASTForAudioClassification, ASTModel, ASTPreTrainedModel, ) try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_audio_spectrogram_transformer import ASTFeatureExtractor else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
1
class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = name snake_case_ = val def __str__( self ): """simple docstring""" return f"""{self.__class__.__name__}({self.name}, {self.val})""" def __lt__( self , __UpperCamelCase ): """simple docstring""" return self.val < other.val class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" snake_case_ = {} snake_case_ = {} snake_case_ = self.build_heap(__UpperCamelCase ) def __getitem__( self , __UpperCamelCase ): """simple docstring""" return self.get_value(__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return (idx - 1) // 2 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return idx * 2 + 1 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return idx * 2 + 2 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return self.heap_dict[key] def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = len(__UpperCamelCase ) - 1 snake_case_ = self.get_parent_idx(__UpperCamelCase ) for idx, i in enumerate(__UpperCamelCase ): snake_case_ = idx snake_case_ = i.val for i in range(__UpperCamelCase , -1 , -1 ): self.sift_down(__UpperCamelCase , __UpperCamelCase ) return array def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" while True: snake_case_ = self.get_left_child_idx(__UpperCamelCase ) # noqa: E741 snake_case_ = self.get_right_child_idx(__UpperCamelCase ) snake_case_ = idx if l < len(__UpperCamelCase ) and array[l] < array[idx]: snake_case_ = l if r < len(__UpperCamelCase ) and array[r] < array[smallest]: snake_case_ = r if smallest != idx: snake_case_ , snake_case_ = array[smallest], array[idx] ( ( snake_case_ ) , ( snake_case_ ) , ) = ( self.idx_of_element[array[smallest]], self.idx_of_element[array[idx]], ) snake_case_ = smallest else: break def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = self.get_parent_idx(__UpperCamelCase ) while p >= 0 and self.heap[p] > self.heap[idx]: snake_case_ , snake_case_ = self.heap[idx], self.heap[p] snake_case_ , snake_case_ = ( self.idx_of_element[self.heap[idx]], self.idx_of_element[self.heap[p]], ) snake_case_ = p snake_case_ = self.get_parent_idx(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" return self.heap[0] def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.heap[-1], self.heap[0] snake_case_ , snake_case_ = ( self.idx_of_element[self.heap[-1]], self.idx_of_element[self.heap[0]], ) snake_case_ = self.heap.pop() del self.idx_of_element[x] self.sift_down(0 , self.heap ) return x def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" self.heap.append(__UpperCamelCase ) snake_case_ = len(self.heap ) - 1 snake_case_ = node.val self.sift_up(len(self.heap ) - 1 ) def __lowerCAmelCase ( self ): """simple docstring""" return len(self.heap ) == 0 def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" assert ( self.heap[self.idx_of_element[node]].val > new_value ), "newValue must be less that current value" snake_case_ = new_value snake_case_ = new_value self.sift_up(self.idx_of_element[node] ) A = Node('R', -1) A = Node('B', 6) A = Node('A', 3) A = Node('X', 1) A = Node('E', 4) # Use one of these two ways to generate Min-Heap # Generating Min-Heap from array A = MinHeap([r, b, a, x, e]) # Generating Min-Heap by Insert method # myMinHeap.insert(a) # myMinHeap.insert(b) # myMinHeap.insert(x) # myMinHeap.insert(r) # myMinHeap.insert(e) # Before print('Min Heap - before decrease key') for i in my_min_heap.heap: print(i) print('Min Heap - After decrease key of node [B -> -17]') my_min_heap.decrease_key(b, -17) # After for i in my_min_heap.heap: print(i) if __name__ == "__main__": import doctest doctest.testmod()
46
import operator as op def a(lowercase__ ): '''simple docstring''' snake_case_ = [] snake_case_ = lambda lowercase__ , lowercase__ : int(x / y ) # noqa: E731 integer division operation snake_case_ = { '^': op.pow, '*': op.mul, '/': div, '+': op.add, '-': op.sub, } # operators & their respective operation # print table header print('Symbol'.center(8 ) , 'Action'.center(12 ) , 'Stack' , sep=' | ' ) print('-' * (30 + len(lowercase__ )) ) for x in post_fix: if x.isdigit(): # if x in digit stack.append(lowercase__ ) # append x to stack # output in tabular format print(x.rjust(8 ) , ('push(' + x + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) else: snake_case_ = stack.pop() # pop stack # output in tabular format print(''.rjust(8 ) , ('pop(' + b + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) snake_case_ = stack.pop() # pop stack # output in tabular format print(''.rjust(8 ) , ('pop(' + a + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' ) stack.append( str(opr[x](int(lowercase__ ) , int(lowercase__ ) ) ) ) # evaluate the 2 values popped from stack & push result to stack # output in tabular format print( x.rjust(8 ) , ('push(' + a + x + b + ')').ljust(12 ) , ','.join(lowercase__ ) , sep=' | ' , ) return int(stack[0] ) if __name__ == "__main__": A = input('\n\nEnter a Postfix Equation (space separated) = ').split(' ') print('\n\tResult = ', solve(Postfix))
46
1
import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [ 'safety_checker/pytorch_model.bin', 'safety_checker/model.safetensors', 'vae/diffusion_pytorch_model.bin', 'vae/diffusion_pytorch_model.safetensors', 'text_encoder/pytorch_model.bin', 'text_encoder/model.safetensors', 'unet/diffusion_pytorch_model.bin', 'unet/diffusion_pytorch_model.safetensors', ] self.assertTrue(is_safetensors_compatible(__UpperCamelCase ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [ 'unet/diffusion_pytorch_model.bin', 'unet/diffusion_pytorch_model.safetensors', ] self.assertTrue(is_safetensors_compatible(__UpperCamelCase ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [ 'safety_checker/pytorch_model.bin', 'safety_checker/model.safetensors', 'vae/diffusion_pytorch_model.bin', 'vae/diffusion_pytorch_model.safetensors', 'text_encoder/pytorch_model.bin', 'text_encoder/model.safetensors', 'unet/diffusion_pytorch_model.bin', # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(__UpperCamelCase ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [ 'text_encoder/pytorch_model.bin', 'text_encoder/model.safetensors', ] self.assertTrue(is_safetensors_compatible(__UpperCamelCase ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [ 'safety_checker/pytorch_model.bin', 'safety_checker/model.safetensors', 'vae/diffusion_pytorch_model.bin', 'vae/diffusion_pytorch_model.safetensors', 'text_encoder/pytorch_model.bin', # Removed: 'text_encoder/model.safetensors', 'unet/diffusion_pytorch_model.bin', 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(__UpperCamelCase ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [ 'safety_checker/pytorch_model.fp16.bin', 'safety_checker/model.fp16.safetensors', 'vae/diffusion_pytorch_model.fp16.bin', 'vae/diffusion_pytorch_model.fp16.safetensors', 'text_encoder/pytorch_model.fp16.bin', 'text_encoder/model.fp16.safetensors', 'unet/diffusion_pytorch_model.fp16.bin', 'unet/diffusion_pytorch_model.fp16.safetensors', ] snake_case_ = 'fp16' self.assertTrue(is_safetensors_compatible(__UpperCamelCase , variant=__UpperCamelCase ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [ 'unet/diffusion_pytorch_model.fp16.bin', 'unet/diffusion_pytorch_model.fp16.safetensors', ] snake_case_ = 'fp16' self.assertTrue(is_safetensors_compatible(__UpperCamelCase , variant=__UpperCamelCase ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [ 'unet/diffusion_pytorch_model.bin', 'unet/diffusion_pytorch_model.safetensors', ] snake_case_ = 'fp16' self.assertTrue(is_safetensors_compatible(__UpperCamelCase , variant=__UpperCamelCase ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [ 'safety_checker/pytorch_model.fp16.bin', 'safety_checker/model.fp16.safetensors', 'vae/diffusion_pytorch_model.fp16.bin', 'vae/diffusion_pytorch_model.fp16.safetensors', 'text_encoder/pytorch_model.fp16.bin', 'text_encoder/model.fp16.safetensors', 'unet/diffusion_pytorch_model.fp16.bin', # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] snake_case_ = 'fp16' self.assertFalse(is_safetensors_compatible(__UpperCamelCase , variant=__UpperCamelCase ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [ 'text_encoder/pytorch_model.fp16.bin', 'text_encoder/model.fp16.safetensors', ] snake_case_ = 'fp16' self.assertTrue(is_safetensors_compatible(__UpperCamelCase , variant=__UpperCamelCase ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [ 'text_encoder/pytorch_model.bin', 'text_encoder/model.safetensors', ] snake_case_ = 'fp16' self.assertTrue(is_safetensors_compatible(__UpperCamelCase , variant=__UpperCamelCase ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = [ 'safety_checker/pytorch_model.fp16.bin', 'safety_checker/model.fp16.safetensors', 'vae/diffusion_pytorch_model.fp16.bin', 'vae/diffusion_pytorch_model.fp16.safetensors', 'text_encoder/pytorch_model.fp16.bin', # 'text_encoder/model.fp16.safetensors', 'unet/diffusion_pytorch_model.fp16.bin', 'unet/diffusion_pytorch_model.fp16.safetensors', ] snake_case_ = 'fp16' self.assertFalse(is_safetensors_compatible(__UpperCamelCase , variant=__UpperCamelCase ) )
46
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices A = logging.get_logger(__name__) A = { 'google/bit-50': 'https://huggingface.co/google/bit-50/resolve/main/config.json', } class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ): """simple docstring""" __A = """bit""" __A = ["""preactivation""", """bottleneck"""] __A = ["""SAME""", """VALID"""] def __init__( self , __UpperCamelCase=3 , __UpperCamelCase=64 , __UpperCamelCase=[2_56, 5_12, 10_24, 20_48] , __UpperCamelCase=[3, 4, 6, 3] , __UpperCamelCase="preactivation" , __UpperCamelCase="relu" , __UpperCamelCase=None , __UpperCamelCase=32 , __UpperCamelCase=0.0 , __UpperCamelCase=False , __UpperCamelCase=32 , __UpperCamelCase=1 , __UpperCamelCase=None , __UpperCamelCase=None , **__UpperCamelCase , ): """simple docstring""" super().__init__(**__UpperCamelCase ) if layer_type not in self.layer_types: raise ValueError(f"""layer_type={layer_type} is not one of {','.join(self.layer_types )}""" ) if global_padding is not None: if global_padding.upper() in self.supported_padding: snake_case_ = global_padding.upper() else: raise ValueError(f"""Padding strategy {global_padding} not supported""" ) snake_case_ = num_channels snake_case_ = embedding_size snake_case_ = hidden_sizes snake_case_ = depths snake_case_ = layer_type snake_case_ = hidden_act snake_case_ = global_padding snake_case_ = num_groups snake_case_ = drop_path_rate snake_case_ = embedding_dynamic_padding snake_case_ = output_stride snake_case_ = width_factor snake_case_ = ['stem'] + [f"""stage{idx}""" for idx in range(1 , len(__UpperCamelCase ) + 1 )] snake_case_ , snake_case_ = get_aligned_output_features_output_indices( out_features=__UpperCamelCase , out_indices=__UpperCamelCase , stage_names=self.stage_names )
46
1
from __future__ import annotations def a(lowercase__ , lowercase__ ): '''simple docstring''' # Checks if the entire collection has been sorted if len(lowercase__ ) <= 1 or n <= 1: return insert_next(lowercase__ , n - 1 ) rec_insertion_sort(lowercase__ , n - 1 ) def a(lowercase__ , lowercase__ ): '''simple docstring''' # Checks order between adjacent elements if index >= len(lowercase__ ) or collection[index - 1] <= collection[index]: return # Swaps adjacent elements since they are not in ascending order snake_case_ , snake_case_ = ( collection[index], collection[index - 1], ) insert_next(lowercase__ , index + 1 ) if __name__ == "__main__": A = input('Enter integers separated by spaces: ') A = [int(num) for num in numbers.split()] rec_insertion_sort(number_list, len(number_list)) print(number_list)
46
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = VQModel( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=3 , ) return model @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModel(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.dummy_uncond_unet snake_case_ = DDIMScheduler() snake_case_ = self.dummy_vq_model snake_case_ = LDMPipeline(unet=__UpperCamelCase , vqvae=__UpperCamelCase , scheduler=__UpperCamelCase ) ldm.to(__UpperCamelCase ) ldm.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=2 , output_type='numpy' ).images snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=2 , output_type='numpy' , return_dict=__UpperCamelCase )[0] snake_case_ = image[0, -3:, -3:, -1] snake_case_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) snake_case_ = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) snake_case_ = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(__UpperCamelCase ) ldm.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=5 , output_type='numpy' ).images snake_case_ = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) snake_case_ = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) snake_case_ = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
46
1
from ..utils import DummyObject, requires_backends class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""transformers""", """torch""", """note_seq"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['transformers', 'torch', 'note_seq'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['transformers', 'torch', 'note_seq'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['transformers', 'torch', 'note_seq'] )
46
from __future__ import annotations import unittest from transformers import LEDConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFLEDForConditionalGeneration, TFLEDModel @require_tf class SCREAMING_SNAKE_CASE : """simple docstring""" __A = LEDConfig __A = {} __A = """gelu""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=7 , __UpperCamelCase=True , __UpperCamelCase=False , __UpperCamelCase=99 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=4 , __UpperCamelCase=37 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=20 , __UpperCamelCase=2 , __UpperCamelCase=1 , __UpperCamelCase=0 , __UpperCamelCase=4 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = eos_token_id snake_case_ = pad_token_id snake_case_ = bos_token_id snake_case_ = attention_window # `ModelTesterMixin.test_attention_outputs` is expecting attention tensors to be of size # [num_attention_heads, encoder_seq_length, encoder_key_length], but TFLongformerSelfAttention # returns attention of shape [num_attention_heads, encoder_seq_length, self.attention_window + 1] # because its local attention only attends to `self.attention_window` and one before and one after snake_case_ = self.attention_window + 2 # because of padding `encoder_seq_length`, is different from `seq_length`. Relevant for # the `test_attention_outputs` and `test_hidden_states_output` tests snake_case_ = ( self.seq_length + (self.attention_window - self.seq_length % self.attention_window) % self.attention_window ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) snake_case_ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) snake_case_ = tf.concat([input_ids, eos_tensor] , axis=1 ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , attention_window=self.attention_window , **self.config_updates , ) snake_case_ = prepare_led_inputs_dict(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = tf.concat( [tf.zeros_like(__UpperCamelCase )[:, :-1], tf.ones_like(__UpperCamelCase )[:, -1:]] , axis=-1 , ) snake_case_ = global_attention_mask return config, inputs_dict def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = TFLEDModel(config=__UpperCamelCase ).get_decoder() snake_case_ = inputs_dict['input_ids'] snake_case_ = input_ids[:1, :] snake_case_ = inputs_dict['attention_mask'][:1, :] snake_case_ = 1 # first forward pass snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , use_cache=__UpperCamelCase ) snake_case_ , snake_case_ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case_ = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case_ = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and snake_case_ = tf.concat([input_ids, next_tokens] , axis=-1 ) snake_case_ = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase )[0] snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , past_key_values=__UpperCamelCase )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice snake_case_ = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) snake_case_ = output_from_no_past[:, -3:, random_slice_idx] snake_case_ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(__UpperCamelCase , __UpperCamelCase , rtol=1E-3 ) def a(lowercase__ , lowercase__ , lowercase__ , lowercase__=None , lowercase__=None , lowercase__=None , lowercase__=None , ): '''simple docstring''' if attention_mask is None: snake_case_ = tf.cast(tf.math.not_equal(lowercase__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: snake_case_ = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: snake_case_ = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: snake_case_ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "attention_mask": attention_mask, "decoder_input_ids": decoder_input_ids, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, } @require_tf class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = (TFLEDForConditionalGeneration, TFLEDModel) if is_tf_available() else () __A = (TFLEDForConditionalGeneration,) if is_tf_available() else () __A = ( { """conversational""": TFLEDForConditionalGeneration, """feature-extraction""": TFLEDModel, """summarization""": TFLEDForConditionalGeneration, """text2text-generation""": TFLEDForConditionalGeneration, """translation""": TFLEDForConditionalGeneration, } if is_tf_available() else {} ) __A = True __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFLEDModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = tf.zeros_like(inputs_dict['attention_mask'] ) snake_case_ = 2 snake_case_ = tf.where( tf.range(self.model_tester.seq_length )[None, :] < num_global_attn_indices , 1 , inputs_dict['global_attention_mask'] , ) snake_case_ = True snake_case_ = self.model_tester.seq_length snake_case_ = self.model_tester.encoder_seq_length def check_decoder_attentions_output(__UpperCamelCase ): snake_case_ = outputs.decoder_attentions self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) def check_encoder_attentions_output(__UpperCamelCase ): snake_case_ = [t.numpy() for t in outputs.encoder_attentions] snake_case_ = [t.numpy() for t in outputs.encoder_global_attentions] self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertEqual(len(__UpperCamelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_length, seq_length] , ) self.assertListEqual( list(global_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, num_global_attn_indices] , ) for model_class in self.all_model_classes: snake_case_ = True snake_case_ = False snake_case_ = False snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = len(__UpperCamelCase ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) if self.is_encoder_decoder: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_decoder_attentions_output(__UpperCamelCase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) # Check attention is always last and order is fine snake_case_ = True snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) snake_case_ = model(self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(__UpperCamelCase ) ) self.assertEqual(model.config.output_hidden_states , __UpperCamelCase ) check_encoder_attentions_output(__UpperCamelCase ) @unittest.skip('LED keeps using potentially symbolic tensors in conditionals and breaks tracing.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" pass def a(lowercase__ ): '''simple docstring''' return tf.constant(lowercase__ , dtype=tf.intaa ) A = 1e-4 @slow @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ).led # change to intended input here snake_case_ = _long_tensor([5_12 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = _long_tensor([1_28 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = prepare_led_inputs_dict(model.config , __UpperCamelCase , __UpperCamelCase ) snake_case_ = model(**__UpperCamelCase )[0] snake_case_ = (1, 10_24, 7_68) self.assertEqual(output.shape , __UpperCamelCase ) # change to expected output here snake_case_ = tf.convert_to_tensor( [[2.3050, 2.8279, 0.6531], [-1.8457, -0.1455, -3.5661], [-1.0186, 0.4586, -2.2043]] , ) tf.debugging.assert_near(output[:, :3, :3] , __UpperCamelCase , atol=1E-3 ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFLEDForConditionalGeneration.from_pretrained('allenai/led-base-16384' ) # change to intended input here snake_case_ = _long_tensor([5_12 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = _long_tensor([1_28 * [0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69]] ) snake_case_ = prepare_led_inputs_dict(model.config , __UpperCamelCase , __UpperCamelCase ) snake_case_ = model(**__UpperCamelCase )[0] snake_case_ = (1, 10_24, model.config.vocab_size) self.assertEqual(output.shape , __UpperCamelCase ) # change to expected output here snake_case_ = tf.convert_to_tensor( [[33.6507, 6.4572, 16.8089], [5.8739, -2.4238, 11.2902], [-3.2139, -4.3149, 4.2783]] , ) tf.debugging.assert_near(output[:, :3, :3] , __UpperCamelCase , atol=1E-3 , rtol=1E-3 )
46
1
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import PoolFormerImageProcessor class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=7 , __UpperCamelCase=3 , __UpperCamelCase=30 , __UpperCamelCase=4_00 , __UpperCamelCase=True , __UpperCamelCase=None , __UpperCamelCase=0.9 , __UpperCamelCase=None , __UpperCamelCase=True , __UpperCamelCase=[0.5, 0.5, 0.5] , __UpperCamelCase=[0.5, 0.5, 0.5] , ): """simple docstring""" snake_case_ = size if size is not None else {'shortest_edge': 30} snake_case_ = crop_size if crop_size is not None else {'height': 30, 'width': 30} snake_case_ = parent snake_case_ = batch_size snake_case_ = num_channels snake_case_ = min_resolution snake_case_ = max_resolution snake_case_ = do_resize_and_center_crop snake_case_ = size snake_case_ = crop_pct snake_case_ = crop_size snake_case_ = do_normalize snake_case_ = image_mean snake_case_ = image_std def __lowerCAmelCase ( self ): """simple docstring""" return { "size": self.size, "do_resize_and_center_crop": self.do_resize_and_center_crop, "crop_pct": self.crop_pct, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, } @require_torch @require_vision class SCREAMING_SNAKE_CASE ( __snake_case , unittest.TestCase ): """simple docstring""" __A = PoolFormerImageProcessor if is_vision_available() else None def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = PoolFormerImageProcessingTester(self ) @property def __lowerCAmelCase ( self ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__UpperCamelCase , 'do_resize_and_center_crop' ) ) self.assertTrue(hasattr(__UpperCamelCase , 'size' ) ) self.assertTrue(hasattr(__UpperCamelCase , 'crop_pct' ) ) self.assertTrue(hasattr(__UpperCamelCase , 'do_normalize' ) ) self.assertTrue(hasattr(__UpperCamelCase , 'image_mean' ) ) self.assertTrue(hasattr(__UpperCamelCase , 'image_std' ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 30} ) self.assertEqual(image_processor.crop_size , {'height': 30, 'width': 30} ) snake_case_ = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {'shortest_edge': 42} ) self.assertEqual(image_processor.crop_size , {'height': 84, 'width': 84} ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images snake_case_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=__UpperCamelCase ) for image in image_inputs: self.assertIsInstance(__UpperCamelCase , Image.Image ) # Test not batched input snake_case_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched snake_case_ = image_processing(__UpperCamelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors snake_case_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=__UpperCamelCase , numpify=__UpperCamelCase ) for image in image_inputs: self.assertIsInstance(__UpperCamelCase , np.ndarray ) # Test not batched input snake_case_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched snake_case_ = image_processing(__UpperCamelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors snake_case_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=__UpperCamelCase , torchify=__UpperCamelCase ) for image in image_inputs: self.assertIsInstance(__UpperCamelCase , torch.Tensor ) # Test not batched input snake_case_ = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched snake_case_ = image_processing(__UpperCamelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , )
46
from collections import defaultdict def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = first_str.lower().strip() snake_case_ = second_str.lower().strip() # Remove whitespace snake_case_ = first_str.replace(' ' , '' ) snake_case_ = second_str.replace(' ' , '' ) # Strings of different lengths are not anagrams if len(lowercase__ ) != len(lowercase__ ): return False # Default values for count should be 0 snake_case_ = defaultdict(lowercase__ ) # For each character in input strings, # increment count in the corresponding for i in range(len(lowercase__ ) ): count[first_str[i]] += 1 count[second_str[i]] -= 1 return all(_count == 0 for _count in count.values() ) if __name__ == "__main__": from doctest import testmod testmod() A = input('Enter the first string ').strip() A = input('Enter the second string ').strip() A = check_anagrams(input_a, input_b) print(f"""{input_a} and {input_b} are {"" if status else "not "}anagrams.""")
46
1
import re import tempfile from pathlib import Path import pytest import yaml from datasets.utils.readme import ReadMe # @pytest.fixture # def example_yaml_structure(): A = yaml.safe_load( '\\nname: ""\nallow_empty: false\nallow_empty_text: true\nsubsections:\n - name: "Dataset Card for X" # First-level markdown heading\n allow_empty: false\n allow_empty_text: true\n subsections:\n - name: "Table of Contents"\n allow_empty: false\n allow_empty_text: false\n subsections: null\n - name: "Dataset Description"\n allow_empty: false\n allow_empty_text: false\n subsections:\n - name: "Dataset Summary"\n allow_empty: false\n allow_empty_text: false\n subsections: null\n - name: "Supported Tasks and Leaderboards"\n allow_empty: true\n allow_empty_text: true\n subsections: null\n - name: Languages\n allow_empty: false\n allow_empty_text: true\n subsections: null\n' ) A = { 'name': 'root', 'text': '', 'is_empty_text': True, 'subsections': [ { 'name': 'Dataset Card for My Dataset', 'text': '', 'is_empty_text': True, 'subsections': [ {'name': 'Table of Contents', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': []}, { 'name': 'Dataset Description', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': [ { 'name': 'Dataset Summary', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': [], }, { 'name': 'Supported Tasks and Leaderboards', 'text': '', 'is_empty_text': True, 'subsections': [], }, {'name': 'Languages', 'text': 'Language Text', 'is_empty_text': False, 'subsections': []}, ], }, ], } ], } A = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' A = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n#### Extra Ignored Subsection\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' A = { 'name': 'root', 'text': '', 'is_empty_text': True, 'subsections': [ { 'name': 'Dataset Card for My Dataset', 'text': '', 'is_empty_text': True, 'subsections': [ {'name': 'Table of Contents', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': []}, { 'name': 'Dataset Description', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': [ { 'name': 'Dataset Summary', 'text': 'Some text here.', 'is_empty_text': False, 'subsections': [ { 'name': 'Extra Ignored Subsection', 'text': '', 'is_empty_text': True, 'subsections': [], } ], }, { 'name': 'Supported Tasks and Leaderboards', 'text': '', 'is_empty_text': True, 'subsections': [], }, {'name': 'Languages', 'text': 'Language Text', 'is_empty_text': False, 'subsections': []}, ], }, ], } ], } A = '\\n---\n---\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' A = ( 'The following issues were found for the README at `{path}`:\n-\tEmpty YAML markers are present in the README.' ) A = '\\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' A = ( 'The following issues were found for the README at `{path}`:\n-\tNo YAML markers are present in the README.' ) A = '\\n---\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' A = 'The following issues were found for the README at `{path}`:\n-\tOnly the start of YAML tags present in the README.' A = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' A = 'The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Summary` but it is empty.\n-\tExpected some text in section `Dataset Summary` but it is empty (text in subsections are ignored).' A = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n' A = 'The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Dataset Card for My Dataset` but it is empty.\n-\tSection `Dataset Card for My Dataset` expected the following subsections: `Table of Contents`, `Dataset Description`. Found \'None\'.' A = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Languages\nLanguage Text\n' A = 'The following issues were found for the README at `{path}`:\n-\tSection `Dataset Description` is missing subsection: `Supported Tasks and Leaderboards`.' A = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\n' A = 'The following issues were found for the README at `{path}`:\n-\tExpected some content in section `Languages` but it is empty.' A = '\\n---\nlanguage:\n- zh\n- en\n---\n\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' A = 'The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README.' A = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n# Dataset Card My Dataset\n' A = 'The following issues were found for the README at `{path}`:\n-\tThe README has several first-level headings: `Dataset Card for My Dataset`, `Dataset Card My Dataset`. Only one heading is expected. Skipping further validation for this README.' A = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' A = 'The following issues were found for the README at `{path}`:\n-\tNo first-level heading starting with `Dataset Card for` found in README. Skipping further validation for this README.' A = '' A = 'The following issues were found for the README at `{path}`:\n-\tThe README has no first-level headings. One heading is expected. Skipping further validation for this README.\n-\tNo YAML markers are present in the README.' A = '\\n---\nlanguage:\n- zh\n- en\n---\n\n# Dataset Card for My Dataset\n# Dataset Card for My Dataset\n## Table of Contents\nSome text here.\n## Dataset Description\nSome text here.\n### Dataset Summary\nSome text here.\n### Supported Tasks and Leaderboards\n### Languages\nLanguage Text\n' A = 'The following issues were found while parsing the README at `{path}`:\n-\tMultiple sections with the same heading `Dataset Card for My Dataset` have been found. Please keep only one of these sections.' @pytest.mark.parametrize( 'readme_md, expected_dict' , [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ] , ) def a(lowercase__ , lowercase__ ): '''simple docstring''' assert ReadMe.from_string(lowercase__ , lowercase__ ).to_dict() == expected_dict @pytest.mark.parametrize( 'readme_md, expected_error' , [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ] , ) def a(lowercase__ , lowercase__ ): '''simple docstring''' with pytest.raises(lowercase__ , match=re.escape(expected_error.format(path='root' ) ) ): snake_case_ = ReadMe.from_string(lowercase__ , lowercase__ ) readme.validate() @pytest.mark.parametrize( 'readme_md, expected_error' , [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ] , ) def a(lowercase__ , lowercase__ ): '''simple docstring''' with pytest.raises(lowercase__ , match=re.escape(expected_error.format(path='root' ) ) ): ReadMe.from_string(lowercase__ , lowercase__ ) @pytest.mark.parametrize( 'readme_md,' , [ (README_MULTIPLE_SAME_HEADING_1), ] , ) def a(lowercase__ ): '''simple docstring''' ReadMe.from_string(lowercase__ , lowercase__ , suppress_parsing_errors=lowercase__ ) @pytest.mark.parametrize( 'readme_md, expected_dict' , [ (README_CORRECT, CORRECT_DICT), (README_CORRECT_FOUR_LEVEL, CORRECT_DICT_FOUR_LEVEL), ] , ) def a(lowercase__ , lowercase__ ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: snake_case_ = Path(lowercase__ ) / 'README.md' with open(lowercase__ , 'w+' ) as readme_file: readme_file.write(lowercase__ ) snake_case_ = ReadMe.from_readme(lowercase__ , lowercase__ ).to_dict() assert out["name"] == path assert out["text"] == "" assert out["is_empty_text"] assert out["subsections"] == expected_dict["subsections"] @pytest.mark.parametrize( 'readme_md, expected_error' , [ (README_NO_YAML, EXPECTED_ERROR_README_NO_YAML), (README_EMPTY_YAML, EXPECTED_ERROR_README_EMPTY_YAML), (README_INCORRECT_YAML, EXPECTED_ERROR_README_INCORRECT_YAML), (README_EMPTY, EXPECTED_ERROR_README_EMPTY), (README_NONE_SUBSECTION, EXPECTED_ERROR_README_NONE_SUBSECTION), (README_MISSING_FIRST_LEVEL, EXPECTED_ERROR_README_MISSING_FIRST_LEVEL), (README_MISSING_SUBSECTION, EXPECTED_ERROR_README_MISSING_SUBSECTION), (README_MISSING_TEXT, EXPECTED_ERROR_README_MISSING_TEXT), (README_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_WRONG_FIRST_LEVEL), (README_MULTIPLE_WRONG_FIRST_LEVEL, EXPECTED_ERROR_README_MULTIPLE_WRONG_FIRST_LEVEL), (README_MISSING_CONTENT, EXPECTED_ERROR_README_MISSING_CONTENT), ] , ) def a(lowercase__ , lowercase__ ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: snake_case_ = Path(lowercase__ ) / 'README.md' with open(lowercase__ , 'w+' ) as readme_file: readme_file.write(lowercase__ ) snake_case_ = expected_error.format(path=lowercase__ ) with pytest.raises(lowercase__ , match=re.escape(lowercase__ ) ): snake_case_ = ReadMe.from_readme(lowercase__ , lowercase__ ) readme.validate() @pytest.mark.parametrize( 'readme_md, expected_error' , [ (README_MULTIPLE_SAME_HEADING_1, EXPECTED_ERROR_README_MULTIPLE_SAME_HEADING_1), ] , ) def a(lowercase__ , lowercase__ ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: snake_case_ = Path(lowercase__ ) / 'README.md' with open(lowercase__ , 'w+' ) as readme_file: readme_file.write(lowercase__ ) snake_case_ = expected_error.format(path=lowercase__ ) with pytest.raises(lowercase__ , match=re.escape(lowercase__ ) ): ReadMe.from_readme(lowercase__ , lowercase__ ) @pytest.mark.parametrize( 'readme_md,' , [ (README_MULTIPLE_SAME_HEADING_1), ] , ) def a(lowercase__ ): '''simple docstring''' with tempfile.TemporaryDirectory() as tmp_dir: snake_case_ = Path(lowercase__ ) / 'README.md' with open(lowercase__ , 'w+' ) as readme_file: readme_file.write(lowercase__ ) ReadMe.from_readme(lowercase__ , lowercase__ , suppress_parsing_errors=lowercase__ )
46
import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.dummy_uncond_unet snake_case_ = ScoreSdeVeScheduler() snake_case_ = ScoreSdeVePipeline(unet=__UpperCamelCase , scheduler=__UpperCamelCase ) sde_ve.to(__UpperCamelCase ) sde_ve.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = sde_ve(num_inference_steps=2 , output_type='numpy' , generator=__UpperCamelCase ).images snake_case_ = torch.manual_seed(0 ) snake_case_ = sde_ve(num_inference_steps=2 , output_type='numpy' , generator=__UpperCamelCase , return_dict=__UpperCamelCase )[ 0 ] snake_case_ = image[0, -3:, -3:, -1] snake_case_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) snake_case_ = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = 'google/ncsnpp-church-256' snake_case_ = UNetaDModel.from_pretrained(__UpperCamelCase ) snake_case_ = ScoreSdeVeScheduler.from_pretrained(__UpperCamelCase ) snake_case_ = ScoreSdeVePipeline(unet=__UpperCamelCase , scheduler=__UpperCamelCase ) sde_ve.to(__UpperCamelCase ) sde_ve.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = sde_ve(num_inference_steps=10 , output_type='numpy' , generator=__UpperCamelCase ).images snake_case_ = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) snake_case_ = np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
46
1
from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { 'google/vivit-b-16x2-kinetics400': ( 'https://huggingface.co/google/vivit-b-16x2-kinetics400/resolve/main/config.json' ), # See all Vivit models at https://huggingface.co/models?filter=vivit } class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = """vivit""" def __init__( self , __UpperCamelCase=2_24 , __UpperCamelCase=32 , __UpperCamelCase=[2, 16, 16] , __UpperCamelCase=3 , __UpperCamelCase=7_68 , __UpperCamelCase=12 , __UpperCamelCase=12 , __UpperCamelCase=30_72 , __UpperCamelCase="gelu_fast" , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.02 , __UpperCamelCase=1E-06 , __UpperCamelCase=True , **__UpperCamelCase , ): """simple docstring""" snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = initializer_range snake_case_ = layer_norm_eps snake_case_ = image_size snake_case_ = num_frames snake_case_ = tubelet_size snake_case_ = num_channels snake_case_ = qkv_bias super().__init__(**__UpperCamelCase )
46
from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput A = logging.get_logger(__name__) # pylint: disable=invalid-name class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ): """simple docstring""" @register_to_config def __init__( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None ): """simple docstring""" super().__init__() snake_case_ = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" snake_case_ = torch.zeros(__UpperCamelCase , __UpperCamelCase ) else: snake_case_ = None snake_case_ = torch.nn.Parameter(__UpperCamelCase ) class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 __A = 42 def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" super().__init__() self.register_modules( vqvae=__UpperCamelCase , transformer=__UpperCamelCase , text_encoder=__UpperCamelCase , tokenizer=__UpperCamelCase , scheduler=__UpperCamelCase , learned_classifier_free_sampling_embeddings=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = len(__UpperCamelCase ) if isinstance(__UpperCamelCase , __UpperCamelCase ) else 1 # get prompt text embeddings snake_case_ = self.tokenizer( __UpperCamelCase , padding='max_length' , max_length=self.tokenizer.model_max_length , return_tensors='pt' , ) snake_case_ = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: snake_case_ = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( 'The following part of your input was truncated because CLIP can only handle sequences up to' f""" {self.tokenizer.model_max_length} tokens: {removed_text}""" ) snake_case_ = text_input_ids[:, : self.tokenizer.model_max_length] snake_case_ = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 snake_case_ = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate text embeddings for each generation per prompt snake_case_ = prompt_embeds.repeat_interleave(__UpperCamelCase , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: snake_case_ = self.learned_classifier_free_sampling_embeddings.embeddings snake_case_ = negative_prompt_embeds.unsqueeze(0 ).repeat(__UpperCamelCase , 1 , 1 ) else: snake_case_ = [''] * batch_size snake_case_ = text_input_ids.shape[-1] snake_case_ = self.tokenizer( __UpperCamelCase , padding='max_length' , max_length=__UpperCamelCase , truncation=__UpperCamelCase , return_tensors='pt' , ) snake_case_ = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings snake_case_ = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=__UpperCamelCase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method snake_case_ = negative_prompt_embeds.shape[1] snake_case_ = negative_prompt_embeds.repeat(1 , __UpperCamelCase , 1 ) snake_case_ = negative_prompt_embeds.view(batch_size * num_images_per_prompt , __UpperCamelCase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes snake_case_ = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self , __UpperCamelCase , __UpperCamelCase = 1_00 , __UpperCamelCase = 5.0 , __UpperCamelCase = 1.0 , __UpperCamelCase = 1 , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = "pil" , __UpperCamelCase = True , __UpperCamelCase = None , __UpperCamelCase = 1 , ): """simple docstring""" if isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = 1 elif isinstance(__UpperCamelCase , __UpperCamelCase ): snake_case_ = len(__UpperCamelCase ) else: raise ValueError(f"""`prompt` has to be of type `str` or `list` but is {type(__UpperCamelCase )}""" ) snake_case_ = batch_size * num_images_per_prompt snake_case_ = guidance_scale > 1.0 snake_case_ = self._encode_prompt(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(__UpperCamelCase , __UpperCamelCase ) or callback_steps <= 0) ): raise ValueError( f"""`callback_steps` has to be a positive integer but is {callback_steps} of type""" f""" {type(__UpperCamelCase )}.""" ) # get the initial completely masked latents unless the user supplied it snake_case_ = (batch_size, self.transformer.num_latent_pixels) if latents is None: snake_case_ = self.transformer.num_vector_embeds - 1 snake_case_ = torch.full(__UpperCamelCase , __UpperCamelCase ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(f"""Unexpected latents shape, got {latents.shape}, expected {latents_shape}""" ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( 'Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0,' f""" {self.transformer.num_vector_embeds - 1} (inclusive).""" ) snake_case_ = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(__UpperCamelCase , device=self.device ) snake_case_ = self.scheduler.timesteps.to(self.device ) snake_case_ = latents for i, t in enumerate(self.progress_bar(__UpperCamelCase ) ): # expand the sample if we are doing classifier free guidance snake_case_ = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` snake_case_ = self.transformer(__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , timestep=__UpperCamelCase ).sample if do_classifier_free_guidance: snake_case_ , snake_case_ = model_output.chunk(2 ) snake_case_ = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(__UpperCamelCase , dim=1 , keepdim=__UpperCamelCase ) snake_case_ = self.truncate(__UpperCamelCase , __UpperCamelCase ) # remove `log(0)`'s (`-inf`s) snake_case_ = model_output.clamp(-70 ) # compute the previous noisy sample x_t -> x_t-1 snake_case_ = self.scheduler.step(__UpperCamelCase , timestep=__UpperCamelCase , sample=__UpperCamelCase , generator=__UpperCamelCase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = self.vqvae.config.vq_embed_dim snake_case_ = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) snake_case_ = self.vqvae.quantize.get_codebook_entry(__UpperCamelCase , shape=__UpperCamelCase ) snake_case_ = self.vqvae.decode(__UpperCamelCase , force_not_quantize=__UpperCamelCase ).sample snake_case_ = (image / 2 + 0.5).clamp(0 , 1 ) snake_case_ = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": snake_case_ = self.numpy_to_pil(__UpperCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ , snake_case_ = torch.sort(__UpperCamelCase , 1 , descending=__UpperCamelCase ) snake_case_ = torch.exp(__UpperCamelCase ) snake_case_ = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out snake_case_ = torch.full_like(keep_mask[:, 0:1, :] , __UpperCamelCase ) snake_case_ = torch.cat((all_true, keep_mask) , dim=1 ) snake_case_ = keep_mask[:, :-1, :] snake_case_ = keep_mask.gather(1 , indices.argsort(1 ) ) snake_case_ = log_p_x_0.clone() snake_case_ = -torch.inf # -inf = log(0) return rv
46
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available A = {'configuration_yolos': ['YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'YolosConfig', 'YolosOnnxConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['YolosFeatureExtractor'] A = ['YolosImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST', 'YolosForObjectDetection', 'YolosModel', 'YolosPreTrainedModel', ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel from transformers.models.mobilevit.modeling_mobilevit import MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileViTImageProcessor class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'hidden_sizes' ) ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'neck_hidden_sizes' ) ) self.parent.assertTrue(hasattr(__UpperCamelCase , 'num_attention_heads' ) ) class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=3 , __UpperCamelCase=6_40 , __UpperCamelCase=4 , __UpperCamelCase="silu" , __UpperCamelCase=3 , __UpperCamelCase=32 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.02 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=10 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = last_hidden_size snake_case_ = num_attention_heads snake_case_ = hidden_act snake_case_ = conv_kernel_size snake_case_ = output_stride snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = classifier_dropout_prob snake_case_ = use_labels snake_case_ = is_training snake_case_ = num_labels snake_case_ = initializer_range snake_case_ = scope def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) snake_case_ = self.get_config() return config, pixel_values, labels, pixel_labels def __lowerCAmelCase ( self ): """simple docstring""" return MobileViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , num_attention_heads=self.num_attention_heads , hidden_act=self.hidden_act , conv_kernel_size=self.conv_kernel_size , output_stride=self.output_stride , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MobileViTModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MobileViTForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MobileViTForSemanticSegmentation(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( (MobileViTModel, MobileViTForImageClassification, MobileViTForSemanticSegmentation) if is_torch_available() else () ) __A = ( { """feature-extraction""": MobileViTModel, """image-classification""": MobileViTForImageClassification, """image-segmentation""": MobileViTForSemanticSegmentation, } if is_torch_available() else {} ) __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTModelTester(self ) snake_case_ = MobileViTConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='MobileViT does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MobileViT does not support input and output embeddings' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='MobileViT does not output attentions' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" def check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.hidden_states snake_case_ = 5 self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # MobileViT's feature maps are of shape (batch_size, num_channels, height, width) # with the width and height being successively divided by 2. snake_case_ = 2 for i in range(len(__UpperCamelCase ) ): self.assertListEqual( list(hidden_states[i].shape[-2:] ) , [self.model_tester.image_size // divisor, self.model_tester.image_size // divisor] , ) divisor *= 2 self.assertEqual(self.model_tester.output_stride , divisor // 2 ) snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = MobileViTModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def a(): '''simple docstring''' snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return MobileViTImageProcessor.from_pretrained('apple/mobilevit-xx-small' ) if is_vision_available() else None @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForImageClassification.from_pretrained('apple/mobilevit-xx-small' ).to(__UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([-1.9364, -1.2327, -0.4653] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = model.to(__UpperCamelCase ) snake_case_ = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) snake_case_ = outputs.logits # verify the logits snake_case_ = torch.Size((1, 21, 32, 32) ) self.assertEqual(logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [ [[6.9713, 6.9786, 7.2422], [7.2893, 7.2825, 7.4446], [7.6580, 7.8797, 7.9420]], [[-10.6869, -10.3250, -10.3471], [-10.4228, -9.9868, -9.7132], [-11.0405, -11.0221, -10.7318]], [[-3.3089, -2.8539, -2.6740], [-3.2706, -2.5621, -2.5108], [-3.2534, -2.6615, -2.6651]], ] , device=__UpperCamelCase , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MobileViTForSemanticSegmentation.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = model.to(__UpperCamelCase ) snake_case_ = MobileViTImageProcessor.from_pretrained('apple/deeplabv3-mobilevit-xx-small' ) snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) snake_case_ = outputs.logits.detach().cpu() snake_case_ = image_processor.post_process_semantic_segmentation(outputs=__UpperCamelCase , target_sizes=[(50, 60)] ) snake_case_ = torch.Size((50, 60) ) self.assertEqual(segmentation[0].shape , __UpperCamelCase ) snake_case_ = image_processor.post_process_semantic_segmentation(outputs=__UpperCamelCase ) snake_case_ = torch.Size((32, 32) ) self.assertEqual(segmentation[0].shape , __UpperCamelCase )
46
1
import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel from diffusers import DDIMScheduler, LDMPipeline, UNetaDModel, VQModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = VQModel( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=3 , ) return model @property def __lowerCAmelCase ( self ): """simple docstring""" torch.manual_seed(0 ) snake_case_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModel(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.dummy_uncond_unet snake_case_ = DDIMScheduler() snake_case_ = self.dummy_vq_model snake_case_ = LDMPipeline(unet=__UpperCamelCase , vqvae=__UpperCamelCase , scheduler=__UpperCamelCase ) ldm.to(__UpperCamelCase ) ldm.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=2 , output_type='numpy' ).images snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=2 , output_type='numpy' , return_dict=__UpperCamelCase )[0] snake_case_ = image[0, -3:, -3:, -1] snake_case_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) snake_case_ = np.array([0.8512, 0.818, 0.6411, 0.6808, 0.4465, 0.5618, 0.46, 0.6231, 0.5172] ) snake_case_ = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < tolerance @slow @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = LDMPipeline.from_pretrained('CompVis/ldm-celebahq-256' ) ldm.to(__UpperCamelCase ) ldm.set_progress_bar_config(disable=__UpperCamelCase ) snake_case_ = torch.manual_seed(0 ) snake_case_ = ldm(generator=__UpperCamelCase , num_inference_steps=5 , output_type='numpy' ).images snake_case_ = image[0, -3:, -3:, -1] assert image.shape == (1, 2_56, 2_56, 3) snake_case_ = np.array([0.4399, 0.4_4975, 0.4_6825, 0.474, 0.4359, 0.4581, 0.4_5095, 0.4341, 0.4447] ) snake_case_ = 1E-2 if torch_device != 'mps' else 3E-2 assert np.abs(image_slice.flatten() - expected_slice ).max() < tolerance
46
from ..utils import DummyObject, requires_backends class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) class SCREAMING_SNAKE_CASE ( metaclass=__snake_case ): """simple docstring""" __A = ["""torch""", """transformers""", """onnx"""] def __init__( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(self , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] ) @classmethod def __lowerCAmelCase ( cls , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" requires_backends(cls , ['torch', 'transformers', 'onnx'] )
46
1
import string from math import logaa def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = document.translate( str.maketrans('' , '' , string.punctuation ) ).replace('\n' , '' ) snake_case_ = document_without_punctuation.split(' ' ) # word tokenization return len([word for word in tokenize_document if word.lower() == term.lower()] ) def a(lowercase__ , lowercase__ ): '''simple docstring''' snake_case_ = corpus.lower().translate( str.maketrans('' , '' , string.punctuation ) ) # strip all punctuation and replace it with '' snake_case_ = corpus_without_punctuation.split('\n' ) snake_case_ = term.lower() return (len([doc for doc in docs if term in doc] ), len(lowercase__ )) def a(lowercase__ , lowercase__ , lowercase__=False ): '''simple docstring''' if smoothing: if n == 0: raise ValueError('log10(0) is undefined.' ) return round(1 + logaa(n / (1 + df) ) , 3 ) if df == 0: raise ZeroDivisionError('df must be > 0' ) elif n == 0: raise ValueError('log10(0) is undefined.' ) return round(logaa(n / df ) , 3 ) def a(lowercase__ , lowercase__ ): '''simple docstring''' return round(tf * idf , 3 )
46
from __future__ import annotations import os import tempfile import unittest import numpy as np from huggingface_hub import hf_hub_download from transformers import is_tensorflow_text_available, is_tf_available from transformers.testing_utils import require_tensorflow_text, require_tf, slow from ..test_modeling_tf_common import floats_tensor from .test_framework_agnostic import GenerationIntegrationTestsMixin if is_tf_available(): import tensorflow as tf from transformers import ( AutoTokenizer, TFAutoModelForCausalLM, TFAutoModelForSeqaSeqLM, TFAutoModelForSpeechSeqaSeq, TFAutoModelForVisionaSeq, TFBartForConditionalGeneration, TFLogitsProcessorList, TFMinLengthLogitsProcessor, tf_top_k_top_p_filtering, ) if is_tensorflow_text_available(): import tensorflow_text as text @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = tf.convert_to_tensor( [ [ 8.222_0991, # 3rd highest value; idx. 0 -0.562_0044, 5.2322_9752, 4.038_6393, -6.879_8378, -0.5478_5802, -3.201_2153, 2.9277_7176, 1.8817_1953, 7.3534_1276, # 5th highest value; idx. 9 8.4320_7833, # 2nd highest value; idx. 10 -9.8571_1836, -5.9620_9236, -1.1303_9161, -7.111_5294, -0.836_9633, -5.318_6408, 7.0642_7407, 0.8136_9344, -0.8202_3817, -5.917_9796, 0.5881_3443, -6.9977_8438, 4.7155_1189, -0.1877_1637, 7.4402_0759, # 4th highest value; idx. 25 9.3845_0987, # 1st highest value; idx. 26 2.1266_2941, -9.3256_2038, 2.3565_2522, ], # cummulative prob of 5 highest values <= 0.6 [ 0.5842_5518, 4.5313_9238, -5.5751_0464, -6.2803_0699, -7.1952_9503, -4.0212_2551, 1.3933_7037, -6.0670_7057, 1.5948_0517, -9.64_3119, 0.0390_7799, 0.6723_1762, -8.8820_6726, 6.2711_5922, # 4th highest value; idx. 13 2.2852_0723, 4.8276_7506, 4.3042_1368, 8.827_5313, # 2nd highest value; idx. 17 5.4402_9958, # 5th highest value; idx. 18 -4.473_5794, 7.3857_9536, # 3rd highest value; idx. 20 -2.9105_1663, 2.6194_6077, -2.567_4762, -9.4895_9302, -4.0292_2645, -1.3541_6918, 9.6770_2323, # 1st highest value; idx. 27 -5.8947_8553, 1.8537_0467, ], # cummulative prob of 5 highest values <= 0.6 ] , dtype=tf.floataa , ) snake_case_ = tf.convert_to_tensor( [[0, 0], [0, 9], [0, 10], [0, 25], [0, 26], [1, 13], [1, 17], [1, 18], [1, 20], [1, 27]] , dtype=tf.intaa , ) # expected non filtered idx as noted above snake_case_ = tf.convert_to_tensor( [8.22_2099, 7.353_4126, 8.43_2078, 7.440_2075, 9.3_8451, 6.27_1159, 8.82_7531, 5.440_2995, 7.385_7956, 9.67_7023] , dtype=tf.floataa , ) # expected non filtered values as noted above snake_case_ = tf_top_k_top_p_filtering(__UpperCamelCase , top_k=10 , top_p=0.6 , min_tokens_to_keep=4 ) snake_case_ = output[output != -float('inf' )] snake_case_ = tf.cast( tf.where(tf.not_equal(__UpperCamelCase , tf.constant(-float('inf' ) , dtype=tf.floataa ) ) ) , dtype=tf.intaa , ) tf.debugging.assert_near(__UpperCamelCase , __UpperCamelCase , rtol=1E-12 ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase , __snake_case ): """simple docstring""" if is_tf_available(): __A = { """AutoModelForCausalLM""": TFAutoModelForCausalLM, """AutoModelForSpeechSeq2Seq""": TFAutoModelForSpeechSeqaSeq, """AutoModelForSeq2SeqLM""": TFAutoModelForSeqaSeqLM, """AutoModelForVision2Seq""": TFAutoModelForVisionaSeq, """LogitsProcessorList""": TFLogitsProcessorList, """MinLengthLogitsProcessor""": TFMinLengthLogitsProcessor, """create_tensor_fn""": tf.convert_to_tensor, """floats_tensor""": floats_tensor, """return_tensors""": """tf""", } @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 2 snake_case_ = 2 class SCREAMING_SNAKE_CASE ( tf.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super(__UpperCamelCase , self ).__init__() snake_case_ = model @tf.function( input_signature=( tf.TensorSpec((None, input_length) , tf.intaa , name='input_ids' ), tf.TensorSpec((None, input_length) , tf.intaa , name='attention_mask' ), ) , jit_compile=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.model.generate( input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase , max_new_tokens=__UpperCamelCase , return_dict_in_generate=__UpperCamelCase , ) return {"sequences": outputs["sequences"]} snake_case_ = [[2, 0], [1_02, 1_03]] snake_case_ = [[1, 0], [1, 1]] snake_case_ = DummyModel(model=__UpperCamelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__UpperCamelCase , __UpperCamelCase , signatures={'serving_default': dummy_model.serving} ) snake_case_ = tf.saved_model.load(__UpperCamelCase ).signatures['serving_default'] for batch_size in range(1 , len(__UpperCamelCase ) + 1 ): snake_case_ = { 'input_ids': tf.constant(dummy_input_ids[:batch_size] ), 'attention_mask': tf.constant(dummy_attention_masks[:batch_size] ), } snake_case_ = serving_func(**__UpperCamelCase )['sequences'] snake_case_ = test_model.generate(**__UpperCamelCase , max_new_tokens=__UpperCamelCase ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 1 snake_case_ = 2 class SCREAMING_SNAKE_CASE ( tf.Module ): """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" super(__UpperCamelCase , self ).__init__() snake_case_ = model @tf.function( input_signature=( tf.TensorSpec((batch_size, None) , tf.intaa , name='input_ids' ), tf.TensorSpec((batch_size, None) , tf.intaa , name='attention_mask' ), ) , jit_compile=__UpperCamelCase , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.model.generate( input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase , max_new_tokens=__UpperCamelCase , return_dict_in_generate=__UpperCamelCase , ) return {"sequences": outputs["sequences"]} snake_case_ = [[2], [1_02, 1_03]] snake_case_ = [[1], [1, 1]] snake_case_ = DummyModel(model=__UpperCamelCase ) with tempfile.TemporaryDirectory() as tmp_dir: tf.saved_model.save(__UpperCamelCase , __UpperCamelCase , signatures={'serving_default': dummy_model.serving} ) snake_case_ = tf.saved_model.load(__UpperCamelCase ).signatures['serving_default'] for input_row in range(len(__UpperCamelCase ) ): snake_case_ = { 'input_ids': tf.constant([dummy_input_ids[input_row]] ), 'attention_mask': tf.constant([dummy_attention_masks[input_row]] ), } snake_case_ = serving_func(**__UpperCamelCase )['sequences'] snake_case_ = test_model.generate(**__UpperCamelCase , max_new_tokens=__UpperCamelCase ) tf.debugging.assert_equal(__UpperCamelCase , __UpperCamelCase ) @slow @require_tensorflow_text def __lowerCAmelCase ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: # file needed to load the TF tokenizer hf_hub_download(repo_id='google/flan-t5-small' , filename='spiece.model' , local_dir=__UpperCamelCase ) class SCREAMING_SNAKE_CASE ( tf.keras.layers.Layer ): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__() snake_case_ = text.SentencepieceTokenizer( model=tf.io.gfile.GFile(os.path.join(__UpperCamelCase , 'spiece.model' ) , 'rb' ).read() ) snake_case_ = TFAutoModelForSeqaSeqLM.from_pretrained('hf-internal-testing/tiny-random-t5' ) def __lowerCAmelCase ( self , __UpperCamelCase , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" snake_case_ = self.tokenizer.tokenize(__UpperCamelCase ) snake_case_ , snake_case_ = text.pad_model_inputs( __UpperCamelCase , max_seq_length=64 , pad_value=self.model.config.pad_token_id ) snake_case_ = self.model.generate(input_ids=__UpperCamelCase , attention_mask=__UpperCamelCase ) return self.tokenizer.detokenize(__UpperCamelCase ) snake_case_ = CompleteSentenceTransformer() snake_case_ = tf.keras.layers.Input(shape=(1,) , dtype=tf.string , name='inputs' ) snake_case_ = complete_model(__UpperCamelCase ) snake_case_ = tf.keras.Model(__UpperCamelCase , __UpperCamelCase ) keras_model.save(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = { 'do_sample': True, 'num_beams': 1, 'top_p': 0.7, 'top_k': 10, 'temperature': 0.7, } snake_case_ = 14 snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 'Hello, my dog is cute and' snake_case_ = tokenizer(__UpperCamelCase , return_tensors='tf' ) snake_case_ = TFAutoModelForCausalLM.from_pretrained('hf-internal-testing/tiny-random-gpt2' ) snake_case_ = 6_38 # forces the generation to happen on CPU, to avoid GPU-related quirks with tf.device(':/CPU:0' ): tf.random.set_seed(0 ) snake_case_ = model.generate(**__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) snake_case_ = [6_38, 1_98] with tf.device(':/CPU:0' ): tf.random.set_seed(0 ) snake_case_ = model.generate(**__UpperCamelCase , eos_token_id=__UpperCamelCase , **__UpperCamelCase ) self.assertTrue(expectation == len(generated_tokens[0] ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = AutoTokenizer.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = 'Hugging Face is a technology company based in New York and Paris.' snake_case_ = bart_tokenizer(__UpperCamelCase , return_tensors='tf' ).input_ids snake_case_ = TFBartForConditionalGeneration.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = bart_model.generate(__UpperCamelCase ).numpy() class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase=None , **__UpperCamelCase ): """simple docstring""" return super().call(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = FakeBart.from_pretrained('hf-internal-testing/tiny-random-bart' ) snake_case_ = bart_model.generate(__UpperCamelCase , foo='bar' ).numpy() self.assertTrue(np.array_equal(__UpperCamelCase , __UpperCamelCase ) ) class SCREAMING_SNAKE_CASE ( bart_model.model.encoder.__class__ ): """simple docstring""" def __lowerCAmelCase ( self , __UpperCamelCase , **__UpperCamelCase ): """simple docstring""" return super().call(__UpperCamelCase , **__UpperCamelCase ) snake_case_ = FakeEncoder(bart_model.config , bart_model.model.shared ) snake_case_ = fake_encoder # Normal generation still works (the output will be different because the encoder weights are different) snake_case_ = bart_model.generate(__UpperCamelCase ).numpy() with self.assertRaises(__UpperCamelCase ): # FakeEncoder.call() accepts **kwargs -> no filtering -> value error due to unexpected input "foo" bart_model.generate(__UpperCamelCase , foo='bar' )
46
1
import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger A = '<<<<<<< This should probably be modified because it mentions: ' A = '=======\n>>>>>>>\n' A = [ 'TextEncoderConfig', 'ByteTextEncoder', 'SubwordTextEncoder', 'encoder_config', 'maybe_build_from_corpus', 'manual_dir', ] A = [ # (pattern, replacement) # Order is important here for some replacements (R'tfds\.core', R'datasets'), (R'tf\.io\.gfile\.GFile', R'open'), (R'tf\.([\w\d]+)', R'datasets.Value(\'\1\')'), (R'tfds\.features\.Text\(\)', R'datasets.Value(\'string\')'), (R'tfds\.features\.Text\(', R'datasets.Value(\'string\'),'), (R'features\s*=\s*tfds.features.FeaturesDict\(', R'features=datasets.Features('), (R'tfds\.features\.FeaturesDict\(', R'dict('), (R'The TensorFlow Datasets Authors', R'The TensorFlow Datasets Authors and the HuggingFace Datasets Authors'), (R'tfds\.', R'datasets.'), (R'dl_manager\.manual_dir', R'self.config.data_dir'), (R'self\.builder_config', R'self.config'), ] def a(lowercase__ ): '''simple docstring''' return ConvertCommand(args.tfds_path , args.datasets_directory ) class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" @staticmethod def __lowerCAmelCase ( __UpperCamelCase ): """simple docstring""" snake_case_ = parser.add_parser( 'convert' , help='Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.' , ) train_parser.add_argument( '--tfds_path' , type=__UpperCamelCase , required=__UpperCamelCase , help='Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.' , ) train_parser.add_argument( '--datasets_directory' , type=__UpperCamelCase , required=__UpperCamelCase , help='Path to the HuggingFace Datasets folder.' ) train_parser.set_defaults(func=__UpperCamelCase ) def __init__( self , __UpperCamelCase , __UpperCamelCase , *__UpperCamelCase ): """simple docstring""" snake_case_ = get_logger('datasets-cli/converting' ) snake_case_ = tfds_path snake_case_ = datasets_directory def __lowerCAmelCase ( self ): """simple docstring""" if os.path.isdir(self._tfds_path ): snake_case_ = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): snake_case_ = os.path.dirname(self._tfds_path ) else: raise ValueError('--tfds_path is neither a directory nor a file. Please check path.' ) snake_case_ = os.path.abspath(self._datasets_directory ) self._logger.info(f"""Converting datasets from {abs_tfds_path} to {abs_datasets_path}""" ) snake_case_ = [] snake_case_ = [] snake_case_ = {} if os.path.isdir(self._tfds_path ): snake_case_ = os.listdir(__UpperCamelCase ) else: snake_case_ = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(f"""Looking at file {f_name}""" ) snake_case_ = os.path.join(__UpperCamelCase , __UpperCamelCase ) snake_case_ = os.path.join(__UpperCamelCase , __UpperCamelCase ) if not os.path.isfile(__UpperCamelCase ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info('Skipping file' ) continue with open(__UpperCamelCase , encoding='utf-8' ) as f: snake_case_ = f.readlines() snake_case_ = [] snake_case_ = False snake_case_ = False snake_case_ = [] for line in lines: snake_case_ = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: snake_case_ = 'import datasets\n' elif "import tensorflow" in out_line: # order is important here snake_case_ = '' continue elif "from absl import logging" in out_line: snake_case_ = 'from datasets import logging\n' elif "getLogger" in out_line: snake_case_ = out_line.replace('getLogger' , 'get_logger' ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): snake_case_ = True snake_case_ = list(filter(lambda __UpperCamelCase : e in out_line , __UpperCamelCase ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(__UpperCamelCase ) + '\n' ) out_lines.append(__UpperCamelCase ) out_lines.append(__UpperCamelCase ) continue else: for pattern, replacement in TO_CONVERT: snake_case_ = re.sub(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: snake_case_ = re.match(r'from\stensorflow_datasets.*import\s([^\.\r\n]+)' , __UpperCamelCase ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split(',' ) ) snake_case_ = 'from . import ' + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(f"""Error converting {out_line.strip()}""" ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: snake_case_ = True out_lines.append(__UpperCamelCase ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset snake_case_ = f_name.replace('.py' , '' ) snake_case_ = os.path.join(__UpperCamelCase , __UpperCamelCase ) snake_case_ = os.path.join(__UpperCamelCase , __UpperCamelCase ) os.makedirs(__UpperCamelCase , exist_ok=__UpperCamelCase ) self._logger.info(f"""Adding directory {output_dir}""" ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(__UpperCamelCase ) if needs_manual_update: with_manual_update.append(__UpperCamelCase ) with open(__UpperCamelCase , 'w' , encoding='utf-8' ) as f: f.writelines(__UpperCamelCase ) self._logger.info(f"""Converted in {output_file}""" ) for utils_file in utils_files: try: snake_case_ = os.path.basename(__UpperCamelCase ) snake_case_ = imports_to_builder_map[f_name.replace('.py' , '' )] self._logger.info(f"""Moving {dest_folder} to {utils_file}""" ) shutil.copy(__UpperCamelCase , __UpperCamelCase ) except KeyError: self._logger.error(f"""Cannot find destination folder for {utils_file}. Please copy manually.""" ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( f"""You need to manually update file {file_path} to remove configurations using 'TextEncoderConfig'.""" )
46
import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=2 , __UpperCamelCase=8 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=99 , __UpperCamelCase=16 , __UpperCamelCase=5 , __UpperCamelCase=2 , __UpperCamelCase=36 , __UpperCamelCase="gelu" , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=5_12 , __UpperCamelCase=16 , __UpperCamelCase=2 , __UpperCamelCase=0.02 , __UpperCamelCase=3 , __UpperCamelCase=4 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_input_mask snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_labels snake_case_ = num_choices snake_case_ = scope def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None if self.use_token_type_ids: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def __lowerCAmelCase ( self ): """simple docstring""" return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.get_config() snake_case_ = 3_00 return config def __lowerCAmelCase ( self ): """simple docstring""" ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = self.prepare_config_and_inputs() snake_case_ = True snake_case_ = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase ) snake_case_ = model(__UpperCamelCase , token_type_ids=__UpperCamelCase ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" snake_case_ = True snake_case_ = MraModel(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , encoder_attention_mask=__UpperCamelCase , ) snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , encoder_hidden_states=__UpperCamelCase , ) snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraForMaskedLM(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = MraForQuestionAnswering(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , start_positions=__UpperCamelCase , end_positions=__UpperCamelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MraForSequenceClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_labels snake_case_ = MraForTokenClassification(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.num_choices snake_case_ = MraForMultipleChoice(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case_ = model( __UpperCamelCase , attention_mask=__UpperCamelCase , token_type_ids=__UpperCamelCase , labels=__UpperCamelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ( snake_case_ ) , ) = config_and_inputs snake_case_ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , unittest.TestCase ): """simple docstring""" __A = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) __A = False __A = False __A = False __A = False __A = () def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , hidden_size=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: snake_case_ = type self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = MraModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) @unittest.skip(reason='MRA does not output attentions' ) def __lowerCAmelCase ( self ): """simple docstring""" return @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraModel.from_pretrained('uw-madison/mra-base-512-4' ) snake_case_ = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = torch.Size((1, 2_56, 7_68) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[-0.0140, 0.0830, -0.0381], [0.1546, 0.1402, 0.0220], [0.1162, 0.0851, 0.0165]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraForMaskedLM.from_pretrained('uw-madison/mra-base-512-4' ) snake_case_ = torch.arange(2_56 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = 5_02_65 snake_case_ = torch.Size((1, 2_56, vocab_size) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[9.2595, -3.6038, 11.8819], [9.3869, -3.2693, 11.0956], [11.8524, -3.4938, 13.1210]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = MraForMaskedLM.from_pretrained('uw-madison/mra-base-4096-8-d3' ) snake_case_ = torch.arange(40_96 ).unsqueeze(0 ) with torch.no_grad(): snake_case_ = model(__UpperCamelCase )[0] snake_case_ = 5_02_65 snake_case_ = torch.Size((1, 40_96, vocab_size) ) self.assertEqual(output.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[[5.4789, -2.3564, 7.5064], [7.9067, -1.3369, 9.9668], [9.0712, -1.8106, 7.0380]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , __UpperCamelCase , atol=1E-4 ) )
46
1
import re from typing import Callable, List, Optional, Union import tensorflow as tf try: from tensorflow.keras.optimizers.legacy import Adam except ImportError: from tensorflow.keras.optimizers import Adam class SCREAMING_SNAKE_CASE ( tf.keras.optimizers.schedules.LearningRateSchedule ): """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = 1.0 , __UpperCamelCase = None , ): """simple docstring""" super().__init__() snake_case_ = initial_learning_rate snake_case_ = warmup_steps snake_case_ = power snake_case_ = decay_schedule_fn snake_case_ = name def __call__( self , __UpperCamelCase ): """simple docstring""" with tf.name_scope(self.name or 'WarmUp' ) as name: # Implements polynomial warmup. i.e., if global_step < warmup_steps, the # learning rate will be `global_step/num_warmup_steps * init_lr`. snake_case_ = tf.cast(__UpperCamelCase , tf.floataa ) snake_case_ = tf.cast(self.warmup_steps , tf.floataa ) snake_case_ = global_step_float / warmup_steps_float snake_case_ = self.initial_learning_rate * tf.math.pow(__UpperCamelCase , self.power ) return tf.cond( global_step_float < warmup_steps_float , lambda: warmup_learning_rate , lambda: self.decay_schedule_fn(step - self.warmup_steps ) , name=__UpperCamelCase , ) def __lowerCAmelCase ( self ): """simple docstring""" return { "initial_learning_rate": self.initial_learning_rate, "decay_schedule_fn": self.decay_schedule_fn, "warmup_steps": self.warmup_steps, "power": self.power, "name": self.name, } def a(lowercase__ , lowercase__ , lowercase__ , lowercase__ = 0.0 , lowercase__ = 0.9 , lowercase__ = 0.999 , lowercase__ = 1e-8 , lowercase__ = None , lowercase__ = None , lowercase__ = 0.0 , lowercase__ = 1.0 , lowercase__ = None , ): '''simple docstring''' snake_case_ = tf.keras.optimizers.schedules.PolynomialDecay( initial_learning_rate=lowercase__ , decay_steps=num_train_steps - num_warmup_steps , end_learning_rate=init_lr * min_lr_ratio , power=lowercase__ , ) if num_warmup_steps: snake_case_ = WarmUp( initial_learning_rate=lowercase__ , decay_schedule_fn=lowercase__ , warmup_steps=lowercase__ , ) if weight_decay_rate > 0.0: snake_case_ = AdamWeightDecay( learning_rate=lowercase__ , weight_decay_rate=lowercase__ , beta_a=lowercase__ , beta_a=lowercase__ , epsilon=lowercase__ , clipnorm=lowercase__ , global_clipnorm=lowercase__ , exclude_from_weight_decay=['LayerNorm', 'layer_norm', 'bias'] , include_in_weight_decay=lowercase__ , ) else: snake_case_ = tf.keras.optimizers.Adam( learning_rate=lowercase__ , beta_a=lowercase__ , beta_a=lowercase__ , epsilon=lowercase__ , clipnorm=lowercase__ , global_clipnorm=lowercase__ , ) # We return the optimizer and the LR scheduler in order to better track the # evolution of the LR independently of the optimizer. return optimizer, lr_schedule class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __init__( self , __UpperCamelCase = 0.001 , __UpperCamelCase = 0.9 , __UpperCamelCase = 0.999 , __UpperCamelCase = 1E-7 , __UpperCamelCase = False , __UpperCamelCase = 0.0 , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = "AdamWeightDecay" , **__UpperCamelCase , ): """simple docstring""" super().__init__(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , **__UpperCamelCase ) snake_case_ = weight_decay_rate snake_case_ = include_in_weight_decay snake_case_ = exclude_from_weight_decay @classmethod def __lowerCAmelCase ( cls , __UpperCamelCase ): """simple docstring""" snake_case_ = {'WarmUp': WarmUp} return super(__UpperCamelCase , cls ).from_config(__UpperCamelCase , custom_objects=__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" super(__UpperCamelCase , self )._prepare_local(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) snake_case_ = tf.constant( self.weight_decay_rate , name='adam_weight_decay_rate' ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self._do_use_weight_decay(var.name ) if do_decay: return var.assign_sub( learning_rate * var * apply_state[(var.device, var.dtype.base_dtype)]['weight_decay_rate'] , use_locking=self._use_locking , ) return tf.no_op() def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase=None , **__UpperCamelCase ): """simple docstring""" snake_case_ , snake_case_ = list(zip(*__UpperCamelCase ) ) return super(__UpperCamelCase , self ).apply_gradients(zip(__UpperCamelCase , __UpperCamelCase ) , name=__UpperCamelCase , **__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" if apply_state is None: return self._decayed_lr_t[var_dtype], {} snake_case_ = apply_state or {} snake_case_ = apply_state.get((var_device, var_dtype) ) if coefficients is None: snake_case_ = self._fallback_apply_state(__UpperCamelCase , __UpperCamelCase ) snake_case_ = coefficients return coefficients["lr_t"], {"apply_state": apply_state} def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None ): """simple docstring""" snake_case_ , snake_case_ = self._get_lr(var.device , var.dtype.base_dtype , __UpperCamelCase ) snake_case_ = self._decay_weights_op(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) with tf.control_dependencies([decay] ): return super(__UpperCamelCase , self )._resource_apply_dense(__UpperCamelCase , __UpperCamelCase , **__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None ): """simple docstring""" snake_case_ , snake_case_ = self._get_lr(var.device , var.dtype.base_dtype , __UpperCamelCase ) snake_case_ = self._decay_weights_op(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) with tf.control_dependencies([decay] ): return super(__UpperCamelCase , self )._resource_apply_sparse(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , **__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = super().get_config() config.update({'weight_decay_rate': self.weight_decay_rate} ) return config def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" if self.weight_decay_rate == 0: return False if self._include_in_weight_decay: for r in self._include_in_weight_decay: if re.search(__UpperCamelCase , __UpperCamelCase ) is not None: return True if self._exclude_from_weight_decay: for r in self._exclude_from_weight_decay: if re.search(__UpperCamelCase , __UpperCamelCase ) is not None: return False return True class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" def __init__( self ): """simple docstring""" snake_case_ = [] snake_case_ = None @property def __lowerCAmelCase ( self ): """simple docstring""" if self._accum_steps is None: snake_case_ = tf.Variable( tf.constant(0 , dtype=tf.intaa ) , trainable=__UpperCamelCase , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) return self._accum_steps.value() @property def __lowerCAmelCase ( self ): """simple docstring""" if not self._gradients: raise ValueError('The accumulator should be called first to initialize the gradients' ) return [gradient.value() if gradient is not None else gradient for gradient in self._gradients] def __call__( self , __UpperCamelCase ): """simple docstring""" if not self._gradients: snake_case_ = self.step # Create the step variable. self._gradients.extend( [ tf.Variable( tf.zeros_like(__UpperCamelCase ) , trainable=__UpperCamelCase , synchronization=tf.VariableSynchronization.ON_READ , aggregation=tf.VariableAggregation.ONLY_FIRST_REPLICA , ) if gradient is not None else gradient for gradient in gradients ] ) if len(__UpperCamelCase ) != len(self._gradients ): raise ValueError(f"""Expected {len(self._gradients )} gradients, but got {len(__UpperCamelCase )}""" ) for accum_gradient, gradient in zip(self._gradients , __UpperCamelCase ): if accum_gradient is not None and gradient is not None: accum_gradient.assign_add(__UpperCamelCase ) self._accum_steps.assign_add(1 ) def __lowerCAmelCase ( self ): """simple docstring""" if not self._gradients: return self._accum_steps.assign(0 ) for gradient in self._gradients: if gradient is not None: gradient.assign(tf.zeros_like(__UpperCamelCase ) )
46
import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() def a(lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ ): '''simple docstring''' # Initialise PyTorch model. # If you want to convert a checkpoint that uses absolute position embeddings, make sure to set reset_position_index_per_cell of # TapasConfig to False. # initialize configuration from json file snake_case_ = TapasConfig.from_json_file(lowercase__ ) # set absolute/relative position embeddings parameter snake_case_ = reset_position_index_per_cell # set remaining parameters of TapasConfig as well as the model based on the task if task == "SQA": snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "WTQ": # run_task_main.py hparams snake_case_ = 4 snake_case_ = True # hparam_utils.py hparams snake_case_ = 0.66_4694 snake_case_ = 0.20_7951 snake_case_ = 0.12_1194 snake_case_ = True snake_case_ = True snake_case_ = False snake_case_ = 0.035_2513 snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "WIKISQL_SUPERVISED": # run_task_main.py hparams snake_case_ = 4 snake_case_ = False # hparam_utils.py hparams snake_case_ = 36.4519 snake_case_ = 0.90_3421 snake_case_ = 222.088 snake_case_ = True snake_case_ = True snake_case_ = True snake_case_ = 0.76_3141 snake_case_ = TapasForQuestionAnswering(config=lowercase__ ) elif task == "TABFACT": snake_case_ = TapasForSequenceClassification(config=lowercase__ ) elif task == "MLM": snake_case_ = TapasForMaskedLM(config=lowercase__ ) elif task == "INTERMEDIATE_PRETRAINING": snake_case_ = TapasModel(config=lowercase__ ) else: raise ValueError(f"""Task {task} not supported.""" ) print(f"""Building PyTorch model from configuration: {config}""" ) # Load weights from tf checkpoint load_tf_weights_in_tapas(lowercase__ , lowercase__ , lowercase__ ) # Save pytorch-model (weights and configuration) print(f"""Save PyTorch model to {pytorch_dump_path}""" ) model.save_pretrained(lowercase__ ) # Save tokenizer files print(f"""Save tokenizer files to {pytorch_dump_path}""" ) snake_case_ = TapasTokenizer(vocab_file=tf_checkpoint_path[:-10] + 'vocab.txt' , model_max_length=512 ) tokenizer.save_pretrained(lowercase__ ) print('Used relative position embeddings:' , model.config.reset_position_index_per_cell ) if __name__ == "__main__": A = argparse.ArgumentParser() # Required parameters parser.add_argument( '--task', default='SQA', type=str, help='Model task for which to convert a checkpoint. Defaults to SQA.' ) parser.add_argument( '--reset_position_index_per_cell', default=False, action='store_true', help='Whether to use relative position embeddings or not. Defaults to True.', ) parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--tapas_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained TAPAS model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) A = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.task, args.reset_position_index_per_cell, args.tf_checkpoint_path, args.tapas_config_file, args.pytorch_dump_path, )
46
1
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = ["""image_processor"""] __A = """SamImageProcessor""" def __init__( self , __UpperCamelCase ): """simple docstring""" super().__init__(__UpperCamelCase ) snake_case_ = self.image_processor snake_case_ = -10 snake_case_ = self.image_processor.size['longest_edge'] def __call__( self , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase = None , **__UpperCamelCase , ): """simple docstring""" snake_case_ = self.image_processor( __UpperCamelCase , return_tensors=__UpperCamelCase , **__UpperCamelCase , ) # pop arguments that are not used in the foward but used nevertheless snake_case_ = encoding_image_processor['original_sizes'] if hasattr(__UpperCamelCase , 'numpy' ): # Checks if Torch or TF tensor snake_case_ = original_sizes.numpy() snake_case_ , snake_case_ , snake_case_ = self._check_and_preprocess_points( input_points=__UpperCamelCase , input_labels=__UpperCamelCase , input_boxes=__UpperCamelCase , ) snake_case_ = self._normalize_and_convert( __UpperCamelCase , __UpperCamelCase , input_points=__UpperCamelCase , input_labels=__UpperCamelCase , input_boxes=__UpperCamelCase , return_tensors=__UpperCamelCase , ) return encoding_image_processor def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase="pt" , ): """simple docstring""" if input_points is not None: if len(__UpperCamelCase ) != len(__UpperCamelCase ): snake_case_ = [ self._normalize_coordinates(self.target_size , __UpperCamelCase , original_sizes[0] ) for point in input_points ] else: snake_case_ = [ self._normalize_coordinates(self.target_size , __UpperCamelCase , __UpperCamelCase ) for point, original_size in zip(__UpperCamelCase , __UpperCamelCase ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: snake_case_ , snake_case_ = self._pad_points_and_labels(__UpperCamelCase , __UpperCamelCase ) snake_case_ = np.array(__UpperCamelCase ) if input_labels is not None: snake_case_ = np.array(__UpperCamelCase ) if input_boxes is not None: if len(__UpperCamelCase ) != len(__UpperCamelCase ): snake_case_ = [ self._normalize_coordinates(self.target_size , __UpperCamelCase , original_sizes[0] , is_bounding_box=__UpperCamelCase ) for box in input_boxes ] else: snake_case_ = [ self._normalize_coordinates(self.target_size , __UpperCamelCase , __UpperCamelCase , is_bounding_box=__UpperCamelCase ) for box, original_size in zip(__UpperCamelCase , __UpperCamelCase ) ] snake_case_ = np.array(__UpperCamelCase ) if input_boxes is not None: if return_tensors == "pt": snake_case_ = torch.from_numpy(__UpperCamelCase ) # boxes batch size of 1 by default snake_case_ = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": snake_case_ = tf.convert_to_tensor(__UpperCamelCase ) # boxes batch size of 1 by default snake_case_ = tf.expand_dims(__UpperCamelCase , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({'input_boxes': input_boxes} ) if input_points is not None: if return_tensors == "pt": snake_case_ = torch.from_numpy(__UpperCamelCase ) # point batch size of 1 by default snake_case_ = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": snake_case_ = tf.convert_to_tensor(__UpperCamelCase ) # point batch size of 1 by default snake_case_ = tf.expand_dims(__UpperCamelCase , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({'input_points': input_points} ) if input_labels is not None: if return_tensors == "pt": snake_case_ = torch.from_numpy(__UpperCamelCase ) # point batch size of 1 by default snake_case_ = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": snake_case_ = tf.convert_to_tensor(__UpperCamelCase ) # point batch size of 1 by default snake_case_ = tf.expand_dims(__UpperCamelCase , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({'input_labels': input_labels} ) return encoding_image_processor def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = max([point.shape[0] for point in input_points] ) snake_case_ = [] for i, point in enumerate(__UpperCamelCase ): if point.shape[0] != expected_nb_points: snake_case_ = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) snake_case_ = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(__UpperCamelCase ) snake_case_ = processed_input_points return input_points, input_labels def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase=False ): """simple docstring""" snake_case_ , snake_case_ = original_size snake_case_ , snake_case_ = self.image_processor._get_preprocess_shape(__UpperCamelCase , longest_edge=__UpperCamelCase ) snake_case_ = deepcopy(__UpperCamelCase ).astype(__UpperCamelCase ) if is_bounding_box: snake_case_ = coords.reshape(-1 , 2 , 2 ) snake_case_ = coords[..., 0] * (new_w / old_w) snake_case_ = coords[..., 1] * (new_h / old_h) if is_bounding_box: snake_case_ = coords.reshape(-1 , 4 ) return coords def __lowerCAmelCase ( self , __UpperCamelCase=None , __UpperCamelCase=None , __UpperCamelCase=None , ): """simple docstring""" if input_points is not None: if hasattr(__UpperCamelCase , 'numpy' ): # Checks for TF or Torch tensor snake_case_ = input_points.numpy().tolist() if not isinstance(__UpperCamelCase , __UpperCamelCase ) or not isinstance(input_points[0] , __UpperCamelCase ): raise ValueError('Input points must be a list of list of floating points.' ) snake_case_ = [np.array(__UpperCamelCase ) for input_point in input_points] else: snake_case_ = None if input_labels is not None: if hasattr(__UpperCamelCase , 'numpy' ): snake_case_ = input_labels.numpy().tolist() if not isinstance(__UpperCamelCase , __UpperCamelCase ) or not isinstance(input_labels[0] , __UpperCamelCase ): raise ValueError('Input labels must be a list of list integers.' ) snake_case_ = [np.array(__UpperCamelCase ) for label in input_labels] else: snake_case_ = None if input_boxes is not None: if hasattr(__UpperCamelCase , 'numpy' ): snake_case_ = input_boxes.numpy().tolist() if ( not isinstance(__UpperCamelCase , __UpperCamelCase ) or not isinstance(input_boxes[0] , __UpperCamelCase ) or not isinstance(input_boxes[0][0] , __UpperCamelCase ) ): raise ValueError('Input boxes must be a list of list of list of floating points.' ) snake_case_ = [np.array(__UpperCamelCase ).astype(np.floataa ) for box in input_boxes] else: snake_case_ = None return input_points, input_labels, input_boxes @property def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.image_processor.model_input_names return list(dict.fromkeys(__UpperCamelCase ) ) def __lowerCAmelCase ( self , *__UpperCamelCase , **__UpperCamelCase ): """simple docstring""" return self.image_processor.post_process_masks(*__UpperCamelCase , **__UpperCamelCase )
46
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=32 , __UpperCamelCase=2 , __UpperCamelCase=3 , __UpperCamelCase=16 , __UpperCamelCase=[1, 2, 1] , __UpperCamelCase=[2, 2, 4] , __UpperCamelCase=2 , __UpperCamelCase=2.0 , __UpperCamelCase=True , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.1 , __UpperCamelCase="gelu" , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=0.02 , __UpperCamelCase=1E-5 , __UpperCamelCase=True , __UpperCamelCase=None , __UpperCamelCase=True , __UpperCamelCase=10 , __UpperCamelCase=8 , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = embed_dim snake_case_ = depths snake_case_ = num_heads snake_case_ = window_size snake_case_ = mlp_ratio snake_case_ = qkv_bias snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = drop_path_rate snake_case_ = hidden_act snake_case_ = use_absolute_embeddings snake_case_ = patch_norm snake_case_ = layer_norm_eps snake_case_ = initializer_range snake_case_ = is_training snake_case_ = scope snake_case_ = use_labels snake_case_ = type_sequence_label_size snake_case_ = encoder_stride def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = self.get_config() return config, pixel_values, labels def __lowerCAmelCase ( self ): """simple docstring""" return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = SwinvaModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) snake_case_ = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) snake_case_ = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = SwinvaForMaskedImageModeling(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images snake_case_ = 1 snake_case_ = SwinvaForMaskedImageModeling(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = self.type_sequence_label_size snake_case_ = SwinvaForImageClassification(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase , labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) __A = ( {"""feature-extraction""": SwinvaModel, """image-classification""": SwinvaForImageClassification} if is_torch_available() else {} ) __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = SwinvaModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , embed_dim=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) @unittest.skip(reason='Got `CUDA error: misaligned address` with PyTorch 2.0.0.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='Swinv2 does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase , nn.Linear ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = True for model_class in self.all_model_classes: snake_case_ = True snake_case_ = False snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.attentions snake_case_ = len(self.model_tester.depths ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # check that output_attentions also work using config del inputs_dict["output_attentions"] snake_case_ = True snake_case_ = config.window_size**2 snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) snake_case_ = len(__UpperCamelCase ) # Check attention is always last and order is fine snake_case_ = True snake_case_ = True snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) if hasattr(self.model_tester , 'num_hidden_states_types' ): snake_case_ = self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states snake_case_ = 2 self.assertEqual(out_len + added_hidden_states , len(__UpperCamelCase ) ) snake_case_ = outputs.attentions self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs.hidden_states snake_case_ = getattr( self.model_tester , 'expected_num_hidden_layers' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) # Swinv2 has a different seq_length snake_case_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) snake_case_ = outputs.reshaped_hidden_states self.assertEqual(len(__UpperCamelCase ) , __UpperCamelCase ) snake_case_ , snake_case_ , snake_case_ , snake_case_ = reshaped_hidden_states[0].shape snake_case_ = ( reshaped_hidden_states[0].view(__UpperCamelCase , __UpperCamelCase , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = 3 snake_case_ = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) snake_case_ = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case_ = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) snake_case_ = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case_ = True self.check_hidden_states_output(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , (padded_height, padded_width) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = SwinvaModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() snake_case_ = _config_zero_init(__UpperCamelCase ) for model_class in self.all_model_classes: snake_case_ = model_class(config=__UpperCamelCase ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @require_vision @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return ( AutoImageProcessor.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ) if is_vision_available() else None ) @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = SwinvaForImageClassification.from_pretrained('microsoft/swinv2-tiny-patch4-window8-256' ).to( __UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase ) # verify the logits snake_case_ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor([-0.3947, -0.4306, 0.0026] ).to(__UpperCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCamelCase , atol=1E-4 ) )
46
1
import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMAEForPreTraining, ViTMAEModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=13 , __UpperCamelCase=30 , __UpperCamelCase=2 , __UpperCamelCase=3 , __UpperCamelCase=True , __UpperCamelCase=True , __UpperCamelCase=32 , __UpperCamelCase=5 , __UpperCamelCase=4 , __UpperCamelCase=37 , __UpperCamelCase="gelu" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=10 , __UpperCamelCase=0.02 , __UpperCamelCase=3 , __UpperCamelCase=0.6 , __UpperCamelCase=None , ): """simple docstring""" snake_case_ = parent snake_case_ = batch_size snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = is_training snake_case_ = use_labels snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = mask_ratio snake_case_ = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) snake_case_ = (image_size // patch_size) ** 2 snake_case_ = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = self.get_config() return config, pixel_values, labels def __lowerCAmelCase ( self ): """simple docstring""" return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__UpperCamelCase , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = ViTMAEModel(config=__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = ViTMAEForPreTraining(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = model(__UpperCamelCase ) snake_case_ = (self.image_size // self.patch_size) ** 2 snake_case_ = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images snake_case_ = 1 snake_case_ = ViTMAEForPreTraining(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() snake_case_ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case_ = model(__UpperCamelCase ) snake_case_ = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.prepare_config_and_inputs() snake_case_ , snake_case_ , snake_case_ = config_and_inputs snake_case_ = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case , unittest.TestCase ): """simple docstring""" __A = (ViTMAEModel, ViTMAEForPreTraining) if is_torch_available() else () __A = {"""feature-extraction""": ViTMAEModel} if is_torch_available() else {} __A = False __A = False __A = False __A = False def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ViTMAEModelTester(self ) snake_case_ = ConfigTester(self , config_class=__UpperCamelCase , has_text_modality=__UpperCamelCase , hidden_size=37 ) def __lowerCAmelCase ( self ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='ViTMAE does not use inputs_embeds' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case_ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase , nn.Linear ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) snake_case_ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case_ = [*signature.parameters.keys()] snake_case_ = ['pixel_values'] self.assertListEqual(arg_names[:1] , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" np.random.seed(2 ) snake_case_ = int((pt_model.config.image_size // pt_model.config.patch_size) ** 2 ) snake_case_ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) snake_case_ = torch.from_numpy(__UpperCamelCase ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument snake_case_ = pt_noise super().check_pt_tf_models(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case_ = model_class(__UpperCamelCase ) model.to(__UpperCamelCase ) model.eval() # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) snake_case_ = outputs[0].cpu().numpy() snake_case_ = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__UpperCamelCase ) snake_case_ = model_class.from_pretrained(__UpperCamelCase ) model.to(__UpperCamelCase ) # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): snake_case_ = model(**self._prepare_for_class(__UpperCamelCase , __UpperCamelCase ) ) # Make sure we don't have nans snake_case_ = after_outputs[0].cpu().numpy() snake_case_ = 0 snake_case_ = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(__UpperCamelCase , 1E-5 ) @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip(reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def __lowerCAmelCase ( self ): """simple docstring""" pass @slow def __lowerCAmelCase ( self ): """simple docstring""" for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ = ViTMAEModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def a(): '''simple docstring''' snake_case_ = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @cached_property def __lowerCAmelCase ( self ): """simple docstring""" return ViTImageProcessor.from_pretrained('facebook/vit-mae-base' ) if is_vision_available() else None @slow def __lowerCAmelCase ( self ): """simple docstring""" np.random.seed(2 ) snake_case_ = ViTMAEForPreTraining.from_pretrained('facebook/vit-mae-base' ).to(__UpperCamelCase ) snake_case_ = self.default_image_processor snake_case_ = prepare_img() snake_case_ = image_processor(images=__UpperCamelCase , return_tensors='pt' ).to(__UpperCamelCase ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) snake_case_ = ViTMAEConfig() snake_case_ = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) snake_case_ = np.random.uniform(size=(1, num_patches) ) # forward pass with torch.no_grad(): snake_case_ = model(**__UpperCamelCase , noise=torch.from_numpy(__UpperCamelCase ).to(device=__UpperCamelCase ) ) # verify the logits snake_case_ = torch.Size((1, 1_96, 7_68) ) self.assertEqual(outputs.logits.shape , __UpperCamelCase ) snake_case_ = torch.tensor( [[-0.0548, -1.7023, -0.9325], [0.3721, -0.5670, -0.2233], [0.8235, -1.3878, -0.3524]] ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , expected_slice.to(__UpperCamelCase ) , atol=1E-4 ) )
46
import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline A = argparse.ArgumentParser('Stable Diffusion script with intel optimization', add_help=False) parser.add_argument('--dpm', action='store_true', help='Enable DPMSolver or not') parser.add_argument('--steps', default=None, type=int, help='Num inference steps') A = parser.parse_args() A = 'cpu' A = 'a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings' A = 'path-to-your-trained-model' A = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: A = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) A = pipe.to(device) # to channels last A = pipe.unet.to(memory_format=torch.channels_last) A = pipe.vae.to(memory_format=torch.channels_last) A = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: A = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex A = torch.randn(2, 4, 64, 64) A = torch.rand(1) * 999 A = torch.randn(2, 77, 768) A = (sample, timestep, encoder_hidden_status) try: A = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: A = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) A = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) A = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: A = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute A = 666 A = torch.Generator(device).manual_seed(seed) A = {'generator': generator} if args.steps is not None: A = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): A = pipe(prompt, **generate_kwargs).images[0] # save image image.save('generated.png')
46
1
def a(lowercase__ , lowercase__ ): '''simple docstring''' return x if y == 0 else greatest_common_divisor(lowercase__ , x % y ) def a(lowercase__ , lowercase__ ): '''simple docstring''' return (x * y) // greatest_common_divisor(lowercase__ , lowercase__ ) def a(lowercase__ = 20 ): '''simple docstring''' snake_case_ = 1 for i in range(1 , n + 1 ): snake_case_ = lcm(lowercase__ , lowercase__ ) return g if __name__ == "__main__": print(f"""{solution() = }""")
46
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging A = logging.get_logger(__name__) A = { 'microsoft/unispeech-sat-base-100h-libri-ft': ( 'https://huggingface.co/microsoft/unispeech-sat-base-100h-libri-ft/resolve/main/config.json' ), # See all UniSpeechSat models at https://huggingface.co/models?filter=unispeech_sat } class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = """unispeech-sat""" def __init__( self , __UpperCamelCase=32 , __UpperCamelCase=7_68 , __UpperCamelCase=12 , __UpperCamelCase=12 , __UpperCamelCase=30_72 , __UpperCamelCase="gelu" , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.0 , __UpperCamelCase=0.0 , __UpperCamelCase=0.1 , __UpperCamelCase=0.1 , __UpperCamelCase=0.02 , __UpperCamelCase=1E-5 , __UpperCamelCase="group" , __UpperCamelCase="gelu" , __UpperCamelCase=(5_12, 5_12, 5_12, 5_12, 5_12, 5_12, 5_12) , __UpperCamelCase=(5, 2, 2, 2, 2, 2, 2) , __UpperCamelCase=(10, 3, 3, 3, 3, 2, 2) , __UpperCamelCase=False , __UpperCamelCase=1_28 , __UpperCamelCase=16 , __UpperCamelCase=False , __UpperCamelCase=True , __UpperCamelCase=0.05 , __UpperCamelCase=10 , __UpperCamelCase=2 , __UpperCamelCase=0.0 , __UpperCamelCase=10 , __UpperCamelCase=0 , __UpperCamelCase=3_20 , __UpperCamelCase=2 , __UpperCamelCase=0.1 , __UpperCamelCase=1_00 , __UpperCamelCase=2_56 , __UpperCamelCase=2_56 , __UpperCamelCase=0.1 , __UpperCamelCase="mean" , __UpperCamelCase=False , __UpperCamelCase=False , __UpperCamelCase=2_56 , __UpperCamelCase=(5_12, 5_12, 5_12, 5_12, 15_00) , __UpperCamelCase=(5, 3, 3, 1, 1) , __UpperCamelCase=(1, 2, 3, 1, 1) , __UpperCamelCase=5_12 , __UpperCamelCase=0 , __UpperCamelCase=1 , __UpperCamelCase=2 , __UpperCamelCase=5_04 , **__UpperCamelCase , ): """simple docstring""" super().__init__(**__UpperCamelCase , pad_token_id=__UpperCamelCase , bos_token_id=__UpperCamelCase , eos_token_id=__UpperCamelCase ) snake_case_ = hidden_size snake_case_ = feat_extract_norm snake_case_ = feat_extract_activation snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = conv_bias snake_case_ = num_conv_pos_embeddings snake_case_ = num_conv_pos_embedding_groups snake_case_ = len(self.conv_dim ) snake_case_ = num_hidden_layers snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = num_attention_heads snake_case_ = hidden_dropout snake_case_ = attention_dropout snake_case_ = activation_dropout snake_case_ = feat_proj_dropout snake_case_ = final_dropout snake_case_ = layerdrop snake_case_ = layer_norm_eps snake_case_ = initializer_range snake_case_ = vocab_size snake_case_ = num_clusters snake_case_ = do_stable_layer_norm snake_case_ = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( 'Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==' ' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =' f""" {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,""" f""" `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 snake_case_ = apply_spec_augment snake_case_ = mask_time_prob snake_case_ = mask_time_length snake_case_ = mask_time_min_masks snake_case_ = mask_feature_prob snake_case_ = mask_feature_length snake_case_ = mask_feature_min_masks # parameters for pretraining with codevector quantized representations snake_case_ = num_codevectors_per_group snake_case_ = num_codevector_groups snake_case_ = contrastive_logits_temperature snake_case_ = feat_quantizer_dropout snake_case_ = num_negatives snake_case_ = codevector_dim snake_case_ = proj_codevector_dim snake_case_ = diversity_loss_weight # ctc loss snake_case_ = ctc_loss_reduction snake_case_ = ctc_zero_infinity # SequenceClassification-specific parameter. Feel free to ignore for other classes. snake_case_ = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = list(__UpperCamelCase ) snake_case_ = xvector_output_dim @property def __lowerCAmelCase ( self ): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
46
1
def a(lowercase__ , lowercase__ ): '''simple docstring''' while second != 0: snake_case_ = first & second first ^= second snake_case_ = c << 1 return first if __name__ == "__main__": import doctest doctest.testmod() A = int(input('Enter the first number: ').strip()) A = int(input('Enter the second number: ').strip()) print(f"""{add(first, second) = }""")
46
class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" snake_case_ = name snake_case_ = val def __str__( self ): """simple docstring""" return f"""{self.__class__.__name__}({self.name}, {self.val})""" def __lt__( self , __UpperCamelCase ): """simple docstring""" return self.val < other.val class SCREAMING_SNAKE_CASE : """simple docstring""" def __init__( self , __UpperCamelCase ): """simple docstring""" snake_case_ = {} snake_case_ = {} snake_case_ = self.build_heap(__UpperCamelCase ) def __getitem__( self , __UpperCamelCase ): """simple docstring""" return self.get_value(__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return (idx - 1) // 2 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return idx * 2 + 1 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return idx * 2 + 2 def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" return self.heap_dict[key] def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = len(__UpperCamelCase ) - 1 snake_case_ = self.get_parent_idx(__UpperCamelCase ) for idx, i in enumerate(__UpperCamelCase ): snake_case_ = idx snake_case_ = i.val for i in range(__UpperCamelCase , -1 , -1 ): self.sift_down(__UpperCamelCase , __UpperCamelCase ) return array def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" while True: snake_case_ = self.get_left_child_idx(__UpperCamelCase ) # noqa: E741 snake_case_ = self.get_right_child_idx(__UpperCamelCase ) snake_case_ = idx if l < len(__UpperCamelCase ) and array[l] < array[idx]: snake_case_ = l if r < len(__UpperCamelCase ) and array[r] < array[smallest]: snake_case_ = r if smallest != idx: snake_case_ , snake_case_ = array[smallest], array[idx] ( ( snake_case_ ) , ( snake_case_ ) , ) = ( self.idx_of_element[array[smallest]], self.idx_of_element[array[idx]], ) snake_case_ = smallest else: break def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" snake_case_ = self.get_parent_idx(__UpperCamelCase ) while p >= 0 and self.heap[p] > self.heap[idx]: snake_case_ , snake_case_ = self.heap[idx], self.heap[p] snake_case_ , snake_case_ = ( self.idx_of_element[self.heap[idx]], self.idx_of_element[self.heap[p]], ) snake_case_ = p snake_case_ = self.get_parent_idx(__UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" return self.heap[0] def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ , snake_case_ = self.heap[-1], self.heap[0] snake_case_ , snake_case_ = ( self.idx_of_element[self.heap[-1]], self.idx_of_element[self.heap[0]], ) snake_case_ = self.heap.pop() del self.idx_of_element[x] self.sift_down(0 , self.heap ) return x def __lowerCAmelCase ( self , __UpperCamelCase ): """simple docstring""" self.heap.append(__UpperCamelCase ) snake_case_ = len(self.heap ) - 1 snake_case_ = node.val self.sift_up(len(self.heap ) - 1 ) def __lowerCAmelCase ( self ): """simple docstring""" return len(self.heap ) == 0 def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" assert ( self.heap[self.idx_of_element[node]].val > new_value ), "newValue must be less that current value" snake_case_ = new_value snake_case_ = new_value self.sift_up(self.idx_of_element[node] ) A = Node('R', -1) A = Node('B', 6) A = Node('A', 3) A = Node('X', 1) A = Node('E', 4) # Use one of these two ways to generate Min-Heap # Generating Min-Heap from array A = MinHeap([r, b, a, x, e]) # Generating Min-Heap by Insert method # myMinHeap.insert(a) # myMinHeap.insert(b) # myMinHeap.insert(x) # myMinHeap.insert(r) # myMinHeap.insert(e) # Before print('Min Heap - before decrease key') for i in my_min_heap.heap: print(i) print('Min Heap - After decrease key of node [B -> -17]') my_min_heap.decrease_key(b, -17) # After for i in my_min_heap.heap: print(i) if __name__ == "__main__": import doctest doctest.testmod()
46
1
# DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from dataclasses import dataclass from typing import Optional, Tuple, Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin, SchedulerOutput @dataclass class SCREAMING_SNAKE_CASE ( __snake_case ): """simple docstring""" __A = 42 __A = 42 class SCREAMING_SNAKE_CASE ( __snake_case , __snake_case ): """simple docstring""" __A = 1 @register_to_config def __init__( self , __UpperCamelCase = 20_00 , __UpperCamelCase = 0.15 , __UpperCamelCase = 0.01 , __UpperCamelCase = 1348.0 , __UpperCamelCase = 1E-5 , __UpperCamelCase = 1 , ): """simple docstring""" snake_case_ = sigma_max # setable values snake_case_ = None self.set_sigmas(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None ): """simple docstring""" return sample def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None ): """simple docstring""" snake_case_ = sampling_eps if sampling_eps is not None else self.config.sampling_eps snake_case_ = torch.linspace(1 , __UpperCamelCase , __UpperCamelCase , device=__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = None , __UpperCamelCase = None ): """simple docstring""" snake_case_ = sigma_min if sigma_min is not None else self.config.sigma_min snake_case_ = sigma_max if sigma_max is not None else self.config.sigma_max snake_case_ = sampling_eps if sampling_eps is not None else self.config.sampling_eps if self.timesteps is None: self.set_timesteps(__UpperCamelCase , __UpperCamelCase ) snake_case_ = sigma_min * (sigma_max / sigma_min) ** (self.timesteps / sampling_eps) snake_case_ = torch.exp(torch.linspace(math.log(__UpperCamelCase ) , math.log(__UpperCamelCase ) , __UpperCamelCase ) ) snake_case_ = torch.tensor([sigma_min * (sigma_max / sigma_min) ** t for t in self.timesteps] ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase ): """simple docstring""" return torch.where( timesteps == 0 , torch.zeros_like(t.to(timesteps.device ) ) , self.discrete_sigmas[timesteps - 1].to(timesteps.device ) , ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = True , ): """simple docstring""" if self.timesteps is None: raise ValueError( '`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler' ) snake_case_ = timestep * torch.ones( sample.shape[0] , device=sample.device ) # torch.repeat_interleave(timestep, sample.shape[0]) snake_case_ = (timestep * (len(self.timesteps ) - 1)).long() # mps requires indices to be in the same device, so we use cpu as is the default with cuda snake_case_ = timesteps.to(self.discrete_sigmas.device ) snake_case_ = self.discrete_sigmas[timesteps].to(sample.device ) snake_case_ = self.get_adjacent_sigma(__UpperCamelCase , __UpperCamelCase ).to(sample.device ) snake_case_ = torch.zeros_like(__UpperCamelCase ) snake_case_ = (sigma**2 - adjacent_sigma**2) ** 0.5 # equation 6 in the paper: the model_output modeled by the network is grad_x log pt(x) # also equation 47 shows the analog from SDE models to ancestral sampling methods snake_case_ = diffusion.flatten() while len(diffusion.shape ) < len(sample.shape ): snake_case_ = diffusion.unsqueeze(-1 ) snake_case_ = drift - diffusion**2 * model_output # equation 6: sample noise for the diffusion term of snake_case_ = randn_tensor( sample.shape , layout=sample.layout , generator=__UpperCamelCase , device=sample.device , dtype=sample.dtype ) snake_case_ = sample - drift # subtract because `dt` is a small negative timestep # TODO is the variable diffusion the correct scaling term for the noise? snake_case_ = prev_sample_mean + diffusion * noise # add impact of diffusion field g if not return_dict: return (prev_sample, prev_sample_mean) return SdeVeOutput(prev_sample=__UpperCamelCase , prev_sample_mean=__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = None , __UpperCamelCase = True , ): """simple docstring""" if self.timesteps is None: raise ValueError( '`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler' ) # For small batch sizes, the paper "suggest replacing norm(z) with sqrt(d), where d is the dim. of z" # sample noise for correction snake_case_ = randn_tensor(sample.shape , layout=sample.layout , generator=__UpperCamelCase ).to(sample.device ) # compute step size from the model_output, the noise, and the snr snake_case_ = torch.norm(model_output.reshape(model_output.shape[0] , -1 ) , dim=-1 ).mean() snake_case_ = torch.norm(noise.reshape(noise.shape[0] , -1 ) , dim=-1 ).mean() snake_case_ = (self.config.snr * noise_norm / grad_norm) ** 2 * 2 snake_case_ = step_size * torch.ones(sample.shape[0] ).to(sample.device ) # self.repeat_scalar(step_size, sample.shape[0]) # compute corrected sample: model_output term and noise term snake_case_ = step_size.flatten() while len(step_size.shape ) < len(sample.shape ): snake_case_ = step_size.unsqueeze(-1 ) snake_case_ = sample + step_size * model_output snake_case_ = prev_sample_mean + ((step_size * 2) ** 0.5) * noise if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__UpperCamelCase ) def __lowerCAmelCase ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , ): """simple docstring""" snake_case_ = timesteps.to(original_samples.device ) snake_case_ = self.discrete_sigmas.to(original_samples.device )[timesteps] snake_case_ = ( noise * sigmas[:, None, None, None] if noise is not None else torch.randn_like(__UpperCamelCase ) * sigmas[:, None, None, None] ) snake_case_ = noise + original_samples return noisy_samples def __len__( self ): """simple docstring""" return self.config.num_train_timesteps
46
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) A = { 'configuration_perceiver': ['PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PerceiverConfig', 'PerceiverOnnxConfig'], 'tokenization_perceiver': ['PerceiverTokenizer'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = ['PerceiverFeatureExtractor'] A = ['PerceiverImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A = [ 'PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PerceiverForImageClassificationConvProcessing', 'PerceiverForImageClassificationFourier', 'PerceiverForImageClassificationLearned', 'PerceiverForMaskedLM', 'PerceiverForMultimodalAutoencoding', 'PerceiverForOpticalFlow', 'PerceiverForSequenceClassification', 'PerceiverLayer', 'PerceiverModel', 'PerceiverPreTrainedModel', ] if TYPE_CHECKING: from .configuration_perceiver import PERCEIVER_PRETRAINED_CONFIG_ARCHIVE_MAP, PerceiverConfig, PerceiverOnnxConfig from .tokenization_perceiver import PerceiverTokenizer try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_perceiver import PerceiverFeatureExtractor from .image_processing_perceiver import PerceiverImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_perceiver import ( PERCEIVER_PRETRAINED_MODEL_ARCHIVE_LIST, PerceiverForImageClassificationConvProcessing, PerceiverForImageClassificationFourier, PerceiverForImageClassificationLearned, PerceiverForMaskedLM, PerceiverForMultimodalAutoencoding, PerceiverForOpticalFlow, PerceiverForSequenceClassification, PerceiverLayer, PerceiverModel, PerceiverPreTrainedModel, ) else: import sys A = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
46
1
import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __init__( self , __UpperCamelCase , __UpperCamelCase=7 , __UpperCamelCase=3 , __UpperCamelCase=18 , __UpperCamelCase=30 , __UpperCamelCase=4_00 , __UpperCamelCase=True , __UpperCamelCase=None , __UpperCamelCase=True , ): """simple docstring""" snake_case_ = size if size is not None else {'height': 18, 'width': 18} snake_case_ = parent snake_case_ = batch_size snake_case_ = num_channels snake_case_ = image_size snake_case_ = min_resolution snake_case_ = max_resolution snake_case_ = do_resize snake_case_ = size snake_case_ = do_normalize def __lowerCAmelCase ( self ): """simple docstring""" return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.8866_4436_3403_3203, 0.6618_8293_6954_4983, 0.3891_7464_0178_6804], [-0.6042_5591_4688_1104, -0.0_2295_0088_6052_8469, 0.5423_7973_6900_3296], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class SCREAMING_SNAKE_CASE ( __snake_case , unittest.TestCase ): """simple docstring""" __A = ImageGPTImageProcessor if is_vision_available() else None def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ImageGPTImageProcessingTester(self ) @property def __lowerCAmelCase ( self ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__UpperCamelCase , 'clusters' ) ) self.assertTrue(hasattr(__UpperCamelCase , 'do_resize' ) ) self.assertTrue(hasattr(__UpperCamelCase , 'size' ) ) self.assertTrue(hasattr(__UpperCamelCase , 'do_normalize' ) ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'height': 18, 'width': 18} ) snake_case_ = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'height': 42, 'width': 42} ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.image_processing_class(**self.image_processor_dict ) snake_case_ = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(__UpperCamelCase , obj[key] ) ) else: self.assertEqual(obj[key] , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: snake_case_ = os.path.join(__UpperCamelCase , 'image_processor.json' ) image_processor_first.to_json_file(__UpperCamelCase ) snake_case_ = self.image_processing_class.from_json_file(__UpperCamelCase ).to_dict() snake_case_ = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(__UpperCamelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , __UpperCamelCase ) def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(__UpperCamelCase ) snake_case_ = self.image_processing_class.from_pretrained(__UpperCamelCase ).to_dict() snake_case_ = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(__UpperCamelCase , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , __UpperCamelCase ) @unittest.skip('ImageGPT requires clusters at initialization' ) def __lowerCAmelCase ( self ): """simple docstring""" pass def a(): '''simple docstring''' snake_case_ = load_dataset('hf-internal-testing/fixtures_image_utils' , split='test' ) snake_case_ = Image.open(dataset[4]['file'] ) snake_case_ = Image.open(dataset[5]['file'] ) snake_case_ = [imagea, imagea] return images @require_vision @require_torch class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @slow def __lowerCAmelCase ( self ): """simple docstring""" snake_case_ = ImageGPTImageProcessor.from_pretrained('openai/imagegpt-small' ) snake_case_ = prepare_images() # test non-batched snake_case_ = image_processing(images[0] , return_tensors='pt' ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 10_24) ) snake_case_ = [3_06, 1_91, 1_91] self.assertEqual(encoding.input_ids[0, :3].tolist() , __UpperCamelCase ) # test batched snake_case_ = image_processing(__UpperCamelCase , return_tensors='pt' ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 10_24) ) snake_case_ = [3_03, 13, 13] self.assertEqual(encoding.input_ids[1, -3:].tolist() , __UpperCamelCase )
46
def a(lowercase__ , lowercase__ ): '''simple docstring''' if not isinstance(lowercase__ , lowercase__ ): raise ValueError('iterations must be defined as integers' ) if not isinstance(lowercase__ , lowercase__ ) or not number >= 1: raise ValueError( 'starting number must be\n and integer and be more than 0' ) if not iterations >= 1: raise ValueError('Iterations must be done more than 0 times to play FizzBuzz' ) snake_case_ = '' while number <= iterations: if number % 3 == 0: out += "Fizz" if number % 5 == 0: out += "Buzz" if 0 not in (number % 3, number % 5): out += str(lowercase__ ) # print(out) number += 1 out += " " return out if __name__ == "__main__": import doctest doctest.testmod()
46
1