code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" from __future__ import annotations __SCREAMING_SNAKE_CASE : int = '''Muhammad Umer Farooq''' __SCREAMING_SNAKE_CASE : Optional[Any] = '''MIT''' __SCREAMING_SNAKE_CASE : Any = '''1.0.0''' __SCREAMING_SNAKE_CASE : Any = '''Muhammad Umer Farooq''' __SCREAMING_SNAKE_CASE : List[Any] = '''contact@muhammadumerfarooq.me''' __SCREAMING_SNAKE_CASE : List[str] = '''Alpha''' import re from html.parser import HTMLParser from urllib import parse import requests class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ ): super().__init__() _lowerCamelCase = [] _lowerCamelCase = domain def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): # Only parse the 'anchor' tag. if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: _lowerCamelCase = parse.urljoin(self.domain , lowerCamelCase__ ) self.urls.append(lowerCamelCase__ ) def lowerCAmelCase_( lowercase_ : str ) -> str: return ".".join(get_sub_domain_name(lowercase_ ).split('''.''' )[-2:] ) def lowerCAmelCase_( lowercase_ : str ) -> str: return parse.urlparse(lowercase_ ).netloc def lowerCAmelCase_( lowercase_ : str = "https://github.com" ) -> list[str]: _lowerCamelCase = get_domain_name(lowercase_ ) # Initialize the parser _lowerCamelCase = Parser(lowercase_ ) try: # Open URL _lowerCamelCase = requests.get(lowercase_ ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through _lowerCamelCase = set() for link in parser.urls: # open URL. # read = requests.get(link) try: _lowerCamelCase = requests.get(lowercase_ ) # Get the valid email. _lowerCamelCase = re.findall('''[a-zA-Z0-9]+@''' + domain , read.text ) # If not in list then append it. for email in emails: valid_emails.add(lowercase_ ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(lowercase_ ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : str = emails_from_url('''https://github.com''') print(F"""{len(emails)} emails found:""") print('''\n'''.join(sorted(emails)))
661
"""simple docstring""" import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=3 , lowerCamelCase__=3_2 , lowerCamelCase__=3 , lowerCamelCase__=1_0 , lowerCamelCase__=[8, 1_6, 3_2, 6_4] , lowerCamelCase__=[1, 1, 2, 1] , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__="relu" , lowerCamelCase__=3 , lowerCamelCase__=None , lowerCamelCase__=["stage2", "stage3", "stage4"] , lowerCamelCase__=[2, 3, 4] , lowerCamelCase__=1 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = num_channels _lowerCamelCase = embeddings_size _lowerCamelCase = hidden_sizes _lowerCamelCase = depths _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_act _lowerCamelCase = num_labels _lowerCamelCase = scope _lowerCamelCase = len(lowerCamelCase__ ) _lowerCamelCase = out_features _lowerCamelCase = out_indices _lowerCamelCase = num_groups def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.num_labels ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = BitModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = BitForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = BitBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None _lowerCamelCase = None _lowerCamelCase = BitBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Dict = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () lowercase__ : Any = ( {'feature-extraction': BitModel, 'image-classification': BitForImageClassification} if is_torch_available() else {} ) lowercase__ : Union[str, Any] = False lowercase__ : List[Any] = False lowercase__ : Any = False lowercase__ : List[str] = False lowercase__ : Any = False def snake_case__ ( self ): _lowerCamelCase = BitModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ ) def snake_case__ ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case__ ( self ): return @unittest.skip(reason='''Bit does not output attentions''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''Bit does not use inputs_embeds''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''Bit does not support input and output embeddings''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(config=lowerCamelCase__ ) for name, module in model.named_modules(): if isinstance(lowerCamelCase__ , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) def snake_case__ ( self ): def check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _lowerCamelCase = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase__ ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = ['''preactivation''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: _lowerCamelCase = layer_type _lowerCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) @unittest.skip(reason='''Bit does not use feedforward chunking''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = BitModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def snake_case__ ( self ): _lowerCamelCase = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowerCamelCase__ ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**lowerCamelCase__ ) # verify the logits _lowerCamelCase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([[-0.6_5_2_6, -0.5_2_6_3, -1.4_3_9_8]] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) ) @require_torch class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = (BitBackbone,) if is_torch_available() else () lowercase__ : Tuple = BitConfig lowercase__ : Any = False def snake_case__ ( self ): _lowerCamelCase = BitModelTester(self )
661
1
"""simple docstring""" from abc import ABC, abstractmethod from typing import List, Optional class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self ): # test for the above condition self.test() def snake_case__ ( self ): _lowerCamelCase = 0 _lowerCamelCase = False while not completed: if counter == 1: self.reset() _lowerCamelCase = self.advance() if not self.does_advance(lowerCamelCase__ ): raise Exception( '''Custom Constraint is not defined correctly. self.does_advance(self.advance()) must be true.''' ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.update(lowerCamelCase__ ) counter += 1 if counter > 1_0_0_0_0: raise Exception('''update() does not fulfill the constraint.''' ) if self.remaining() != 0: raise Exception('''Custom Constraint is not defined correctly.''' ) @abstractmethod def snake_case__ ( self ): raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def snake_case__ ( self , lowerCamelCase__ ): raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def snake_case__ ( self , lowerCamelCase__ ): raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def snake_case__ ( self ): raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def snake_case__ ( self ): raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def snake_case__ ( self , lowerCamelCase__=False ): raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ ): super(lowerCamelCase__ , self ).__init__() if not isinstance(lowerCamelCase__ , lowerCamelCase__ ) or len(lowerCamelCase__ ) == 0: raise ValueError(F"""`token_ids` has to be a non-empty list, but is {token_ids}.""" ) if any((not isinstance(lowerCamelCase__ , lowerCamelCase__ ) or token_id < 0) for token_id in token_ids ): raise ValueError(F"""Each list in `token_ids` has to be a list of positive integers, but is {token_ids}.""" ) _lowerCamelCase = token_ids _lowerCamelCase = len(self.token_ids ) _lowerCamelCase = -1 # the index of the currently fulfilled step _lowerCamelCase = False def snake_case__ ( self ): if self.completed: return None return self.token_ids[self.fulfilled_idx + 1] def snake_case__ ( self , lowerCamelCase__ ): if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): raise ValueError(F"""`token_id` has to be an `int`, but is {token_id} of type {type(lowerCamelCase__ )}""" ) if self.completed: return False return token_id == self.token_ids[self.fulfilled_idx + 1] def snake_case__ ( self , lowerCamelCase__ ): if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): raise ValueError(F"""`token_id` has to be an `int`, but is {token_id} of type {type(lowerCamelCase__ )}""" ) _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False if self.does_advance(lowerCamelCase__ ): self.fulfilled_idx += 1 _lowerCamelCase = True if self.fulfilled_idx == (self.seqlen - 1): _lowerCamelCase = True _lowerCamelCase = completed else: # failed to make progress. _lowerCamelCase = True self.reset() return stepped, completed, reset def snake_case__ ( self ): _lowerCamelCase = False _lowerCamelCase = 0 def snake_case__ ( self ): return self.seqlen - (self.fulfilled_idx + 1) def snake_case__ ( self , lowerCamelCase__=False ): _lowerCamelCase = PhrasalConstraint(self.token_ids ) if stateful: _lowerCamelCase = self.seqlen _lowerCamelCase = self.fulfilled_idx _lowerCamelCase = self.completed return new_constraint class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=True ): _lowerCamelCase = max([len(lowerCamelCase__ ) for one in nested_token_ids] ) _lowerCamelCase = {} for token_ids in nested_token_ids: _lowerCamelCase = root for tidx, token_id in enumerate(lowerCamelCase__ ): if token_id not in level: _lowerCamelCase = {} _lowerCamelCase = level[token_id] if no_subsets and self.has_subsets(lowerCamelCase__ , lowerCamelCase__ ): raise ValueError( '''Each list in `nested_token_ids` can\'t be a complete subset of another list, but is''' F""" {nested_token_ids}.""" ) _lowerCamelCase = root def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = self.trie for current_token in current_seq: _lowerCamelCase = start[current_token] _lowerCamelCase = list(start.keys() ) return next_tokens def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = self.next_tokens(lowerCamelCase__ ) return len(lowerCamelCase__ ) == 0 def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = list(root.values() ) if len(lowerCamelCase__ ) == 0: return 1 else: return sum([self.count_leaves(lowerCamelCase__ ) for nn in next_nodes] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.count_leaves(lowerCamelCase__ ) return len(lowerCamelCase__ ) != leaf_count class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , lowerCamelCase__ ): super(lowerCamelCase__ , self ).__init__() if not isinstance(lowerCamelCase__ , lowerCamelCase__ ) or len(lowerCamelCase__ ) == 0: raise ValueError(F"""`nested_token_ids` has to be a non-empty list, but is {nested_token_ids}.""" ) if any(not isinstance(lowerCamelCase__ , lowerCamelCase__ ) for token_ids in nested_token_ids ): raise ValueError(F"""`nested_token_ids` has to be a list of lists, but is {nested_token_ids}.""" ) if any( any((not isinstance(lowerCamelCase__ , lowerCamelCase__ ) or token_id < 0) for token_id in token_ids ) for token_ids in nested_token_ids ): raise ValueError( F"""Each list in `nested_token_ids` has to be a list of positive integers, but is {nested_token_ids}.""" ) _lowerCamelCase = DisjunctiveTrie(lowerCamelCase__ ) _lowerCamelCase = nested_token_ids _lowerCamelCase = self.trie.max_height _lowerCamelCase = [] _lowerCamelCase = False def snake_case__ ( self ): _lowerCamelCase = self.trie.next_tokens(self.current_seq ) if len(lowerCamelCase__ ) == 0: return None else: return token_list def snake_case__ ( self , lowerCamelCase__ ): if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): raise ValueError(F"""`token_id` is supposed to be type `int`, but is {token_id} of type {type(lowerCamelCase__ )}""" ) _lowerCamelCase = self.trie.next_tokens(self.current_seq ) return token_id in next_tokens def snake_case__ ( self , lowerCamelCase__ ): if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): raise ValueError(F"""`token_id` is supposed to be type `int`, but is {token_id} of type {type(lowerCamelCase__ )}""" ) _lowerCamelCase = False _lowerCamelCase = False _lowerCamelCase = False if self.does_advance(lowerCamelCase__ ): self.current_seq.append(lowerCamelCase__ ) _lowerCamelCase = True else: _lowerCamelCase = True self.reset() _lowerCamelCase = self.trie.reached_leaf(self.current_seq ) _lowerCamelCase = completed return stepped, completed, reset def snake_case__ ( self ): _lowerCamelCase = False _lowerCamelCase = [] def snake_case__ ( self ): if self.completed: # since this can be completed without reaching max height return 0 else: return self.seqlen - len(self.current_seq ) def snake_case__ ( self , lowerCamelCase__=False ): _lowerCamelCase = DisjunctiveConstraint(self.token_ids ) if stateful: _lowerCamelCase = self.seqlen _lowerCamelCase = self.current_seq _lowerCamelCase = self.completed return new_constraint class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ ): _lowerCamelCase = constraints # max # of steps required to fulfill a given constraint _lowerCamelCase = max([c.seqlen for c in constraints] ) _lowerCamelCase = len(lowerCamelCase__ ) _lowerCamelCase = False self.init_state() def snake_case__ ( self ): _lowerCamelCase = [] _lowerCamelCase = None _lowerCamelCase = [constraint.copy(stateful=lowerCamelCase__ ) for constraint in self.constraints] def snake_case__ ( self ): _lowerCamelCase = 0 if self.inprogress_constraint: # extra points for having a constraint mid-fulfilled add += self.max_seqlen - self.inprogress_constraint.remaining() return (len(self.complete_constraints ) * self.max_seqlen) + add def snake_case__ ( self ): _lowerCamelCase = [] if self.inprogress_constraint is None: for constraint in self.pending_constraints: # "pending" == "unfulfilled yet" _lowerCamelCase = constraint.advance() if isinstance(lowerCamelCase__ , lowerCamelCase__ ): token_list.append(lowerCamelCase__ ) elif isinstance(lowerCamelCase__ , lowerCamelCase__ ): token_list.extend(lowerCamelCase__ ) else: _lowerCamelCase = self.inprogress_constraint.advance() if isinstance(lowerCamelCase__ , lowerCamelCase__ ): token_list.append(lowerCamelCase__ ) elif isinstance(lowerCamelCase__ , lowerCamelCase__ ): token_list.extend(lowerCamelCase__ ) if len(lowerCamelCase__ ) == 0: return None else: return token_list def snake_case__ ( self , lowerCamelCase__ ): self.init_state() if token_ids is not None: for token in token_ids: # completes or steps **one** constraint _lowerCamelCase , _lowerCamelCase = self.add(lowerCamelCase__ ) # the entire list of constraints are fulfilled if self.completed: break def snake_case__ ( self , lowerCamelCase__ ): if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): raise ValueError(F"""`token_id` should be an `int`, but is `{token_id}`.""" ) _lowerCamelCase , _lowerCamelCase = False, False if self.completed: _lowerCamelCase = True _lowerCamelCase = False return complete, stepped if self.inprogress_constraint is not None: # In the middle of fulfilling a constraint. If the `token_id` *does* makes an incremental progress to current # job, simply update the state _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = self.inprogress_constraint.update(lowerCamelCase__ ) if reset: # 1. If the next token breaks the progress, then we must restart. # e.g. constraint = "I love pies" and sequence so far is "I love" but `token_id` == "books". # But that doesn't mean we self.init_state(), since we only reset the state for this particular # constraint, not the full list of constraints. self.pending_constraints.append(self.inprogress_constraint.copy(stateful=lowerCamelCase__ ) ) _lowerCamelCase = None if complete: # 2. If the next token completes the constraint, move it to completed list, set # inprogress to None. If there are no pending constraints either, then this full list of constraints # is complete. self.complete_constraints.append(self.inprogress_constraint ) _lowerCamelCase = None if len(self.pending_constraints ) == 0: # we're done! _lowerCamelCase = True else: # Not in the middle of fulfilling a constraint. So does this `token_id` helps us step towards any of our list # of constraints? for cidx, pending_constraint in enumerate(self.pending_constraints ): if pending_constraint.does_advance(lowerCamelCase__ ): _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = pending_constraint.update(lowerCamelCase__ ) if not stepped: raise Exception( '''`constraint.update(token_id)` is not yielding incremental progress, ''' '''even though `constraint.does_advance(token_id)` is true.''' ) if complete: self.complete_constraints.append(lowerCamelCase__ ) _lowerCamelCase = None if not complete and stepped: _lowerCamelCase = pending_constraint if complete or stepped: # If we made any progress at all, then it's at least not a "pending constraint". _lowerCamelCase = ( self.pending_constraints[:cidx] + self.pending_constraints[cidx + 1 :] ) if len(self.pending_constraints ) == 0 and self.inprogress_constraint is None: # If there's no longer any pending after this and no inprogress either, then we must be # complete. _lowerCamelCase = True break # prevent accidentally stepping through multiple constraints with just one token. return complete, stepped def snake_case__ ( self , lowerCamelCase__=True ): _lowerCamelCase = ConstraintListState(self.constraints ) # we actually never though self.constraints objects # throughout this process. So it's at initialization state. if stateful: _lowerCamelCase = [ constraint.copy(stateful=lowerCamelCase__ ) for constraint in self.complete_constraints ] if self.inprogress_constraint is not None: _lowerCamelCase = self.inprogress_constraint.copy(stateful=lowerCamelCase__ ) _lowerCamelCase = [constraint.copy() for constraint in self.pending_constraints] return new_state
661
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=2 , lowerCamelCase__=2_4 , lowerCamelCase__=1_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=3_2 , lowerCamelCase__=5 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1_0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=None , lowerCamelCase__=2 , lowerCamelCase__=2 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = patch_size _lowerCamelCase = max_length _lowerCamelCase = num_mel_bins _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = scope _lowerCamelCase = frequency_stride _lowerCamelCase = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _lowerCamelCase = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _lowerCamelCase = (self.max_length - self.patch_size) // self.time_stride + 1 _lowerCamelCase = frequency_out_dimension * time_out_dimension _lowerCamelCase = num_patches + 2 def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, input_values, labels def snake_case__ ( self ): return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = ASTModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = config_and_inputs _lowerCamelCase = {'''input_values''': input_values} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Any = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) lowercase__ : List[str] = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) lowercase__ : int = False lowercase__ : str = False lowercase__ : Union[str, Any] = False lowercase__ : List[str] = False def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def snake_case__ ( self ): _lowerCamelCase = ASTModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , nn.Linear ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''input_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = ASTModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> str: _lowerCamelCase = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) _lowerCamelCase , _lowerCamelCase = torchaudio.load(lowercase_ ) return audio, sampling_rate @require_torch @require_torchaudio class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def snake_case__ ( self ): _lowerCamelCase = self.default_feature_extractor _lowerCamelCase = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(lowerCamelCase__ ) _lowerCamelCase = self.default_feature_extractor _lowerCamelCase , _lowerCamelCase = prepare_audio() _lowerCamelCase = audio.squeeze().numpy() _lowerCamelCase = feature_extractor(lowerCamelCase__ , sampling_rate=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**lowerCamelCase__ ) # verify the logits _lowerCamelCase = torch.Size((1, 5_2_7) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([-0.8_7_6_0, -7.0_0_4_2, -8.6_6_0_2] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) )
661
1
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowerCamelCase_( A__, A__, A__ ): '''simple docstring''' lowercase__ : List[Any] = [r'h\.\d+\.attn\.bias', r'h\.\d+\.attn\.masked_bias'] @register_to_config def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = 5_0_2_5_7 , lowerCamelCase__ = 1_0_2_4 , lowerCamelCase__ = 7_6_8 , lowerCamelCase__ = 1_2 , lowerCamelCase__ = 1_2 , lowerCamelCase__ = None , lowerCamelCase__ = "gelu_new" , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 1e-5 , lowerCamelCase__ = 0.0_2 , lowerCamelCase__ = True , lowerCamelCase__ = True , lowerCamelCase__ = False , lowerCamelCase__ = False , ): super().__init__() _lowerCamelCase = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F"""`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and""" F""" `n_embd`: {n_embd} are not equal.""" ) _lowerCamelCase = prefix_inner_dim _lowerCamelCase = prefix_hidden_dim _lowerCamelCase = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCamelCase = ( nn.Linear(self.prefix_hidden_dim , lowerCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCamelCase = GPTaConfig( vocab_size=lowerCamelCase__ , n_positions=lowerCamelCase__ , n_embd=lowerCamelCase__ , n_layer=lowerCamelCase__ , n_head=lowerCamelCase__ , n_inner=lowerCamelCase__ , activation_function=lowerCamelCase__ , resid_pdrop=lowerCamelCase__ , embd_pdrop=lowerCamelCase__ , attn_pdrop=lowerCamelCase__ , layer_norm_epsilon=lowerCamelCase__ , initializer_range=lowerCamelCase__ , scale_attn_weights=lowerCamelCase__ , use_cache=lowerCamelCase__ , scale_attn_by_inverse_layer_idx=lowerCamelCase__ , reorder_and_upcast_attn=lowerCamelCase__ , ) _lowerCamelCase = GPTaLMHeadModel(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None , ): _lowerCamelCase = self.transformer.transformer.wte(lowerCamelCase__ ) _lowerCamelCase = self.encode_prefix(lowerCamelCase__ ) _lowerCamelCase = self.decode_prefix(lowerCamelCase__ ) _lowerCamelCase = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: _lowerCamelCase = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) _lowerCamelCase = torch.cat((dummy_token, input_ids) , dim=1 ) _lowerCamelCase = self.transformer(inputs_embeds=lowerCamelCase__ , labels=lowerCamelCase__ , attention_mask=lowerCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): return torch.zeros(lowerCamelCase__ , self.prefix_length , dtype=torch.intaa , device=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): return self.encode_prefix(lowerCamelCase__ ) @torch.no_grad() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = torch.split(lowerCamelCase__ , 1 , dim=0 ) _lowerCamelCase = [] _lowerCamelCase = [] for feature in features: _lowerCamelCase = self.decode_prefix(feature.to(lowerCamelCase__ ) ) # back to the clip feature # Only support beam search for now _lowerCamelCase , _lowerCamelCase = self.generate_beam( input_embeds=lowerCamelCase__ , device=lowerCamelCase__ , eos_token_id=lowerCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) _lowerCamelCase = torch.stack(lowerCamelCase__ ) _lowerCamelCase = torch.stack(lowerCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def snake_case__ ( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__ = 5 , lowerCamelCase__ = 6_7 , lowerCamelCase__ = 1.0 , lowerCamelCase__ = None , ): _lowerCamelCase = eos_token_id _lowerCamelCase = None _lowerCamelCase = None _lowerCamelCase = torch.ones(lowerCamelCase__ , device=lowerCamelCase__ , dtype=torch.int ) _lowerCamelCase = torch.zeros(lowerCamelCase__ , device=lowerCamelCase__ , dtype=torch.bool ) if input_embeds is not None: _lowerCamelCase = input_embeds else: _lowerCamelCase = self.transformer.transformer.wte(lowerCamelCase__ ) for i in range(lowerCamelCase__ ): _lowerCamelCase = self.transformer(inputs_embeds=lowerCamelCase__ ) _lowerCamelCase = outputs.logits _lowerCamelCase = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) _lowerCamelCase = logits.softmax(-1 ).log() if scores is None: _lowerCamelCase , _lowerCamelCase = logits.topk(lowerCamelCase__ , -1 ) _lowerCamelCase = generated.expand(lowerCamelCase__ , *generated.shape[1:] ) _lowerCamelCase , _lowerCamelCase = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: _lowerCamelCase = next_tokens else: _lowerCamelCase = tokens.expand(lowerCamelCase__ , *tokens.shape[1:] ) _lowerCamelCase = torch.cat((tokens, next_tokens) , dim=1 ) else: _lowerCamelCase = -float(np.inf ) _lowerCamelCase = 0 _lowerCamelCase = scores[:, None] + logits seq_lengths[~is_stopped] += 1 _lowerCamelCase = scores_sum / seq_lengths[:, None] _lowerCamelCase , _lowerCamelCase = scores_sum_average.view(-1 ).topk(lowerCamelCase__ , -1 ) _lowerCamelCase = next_tokens // scores_sum.shape[1] _lowerCamelCase = seq_lengths[next_tokens_source] _lowerCamelCase = next_tokens % scores_sum.shape[1] _lowerCamelCase = next_tokens.unsqueeze(1 ) _lowerCamelCase = tokens[next_tokens_source] _lowerCamelCase = torch.cat((tokens, next_tokens) , dim=1 ) _lowerCamelCase = generated[next_tokens_source] _lowerCamelCase = scores_sum_average * seq_lengths _lowerCamelCase = is_stopped[next_tokens_source] _lowerCamelCase = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) _lowerCamelCase = torch.cat((generated, next_token_embed) , dim=1 ) _lowerCamelCase = is_stopped + next_tokens.eq(lowerCamelCase__ ).squeeze() if is_stopped.all(): break _lowerCamelCase = scores / seq_lengths _lowerCamelCase = scores.argsort(descending=lowerCamelCase__ ) # tokens tensors are already padded to max_seq_length _lowerCamelCase = [tokens[i] for i in order] _lowerCamelCase = torch.stack(lowerCamelCase__ , dim=0 ) _lowerCamelCase = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
661
"""simple docstring""" import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class lowerCamelCase_: '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return None class lowerCamelCase_: '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return None class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = [ # (model_name, model_kwargs) ('bert-base-cased', {}), ('gpt2', {'use_cache': False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase__ , '''tf''' , 1_2 , **lowerCamelCase__ ) @require_torch @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , **lowerCamelCase__ ) @require_torch @slow def snake_case__ ( self ): from transformers import BertModel _lowerCamelCase = ['''[UNK]''', '''[SEP]''', '''[CLS]''', '''[PAD]''', '''[MASK]''', '''some''', '''other''', '''words'''] with NamedTemporaryFile(mode='''w+t''' ) as vocab_file: vocab_file.write('''\n'''.join(lowerCamelCase__ ) ) vocab_file.flush() _lowerCamelCase = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: _lowerCamelCase = BertModel(BertConfig(vocab_size=len(lowerCamelCase__ ) ) ) model.save_pretrained(lowerCamelCase__ ) self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , lowerCamelCase__ ) @require_tf @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCamelCase = self._test_export(lowerCamelCase__ , '''tf''' , 1_2 , **lowerCamelCase__ ) _lowerCamelCase = quantize(Path(lowerCamelCase__ ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) @require_torch @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCamelCase = self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , **lowerCamelCase__ ) _lowerCamelCase = quantize(lowerCamelCase__ ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): try: # Compute path with TemporaryDirectory() as tempdir: _lowerCamelCase = Path(lowerCamelCase__ ).joinpath('''model.onnx''' ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) return path except Exception as e: self.fail(lowerCamelCase__ ) @require_torch @require_tokenizers @slow def snake_case__ ( self ): from transformers import BertModel _lowerCamelCase = BertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) _lowerCamelCase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(lowerCamelCase__ , lowerCamelCase__ , '''pt''' ) @require_tf @require_tokenizers @slow def snake_case__ ( self ): from transformers import TFBertModel _lowerCamelCase = TFBertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) _lowerCamelCase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(lowerCamelCase__ , lowerCamelCase__ , '''tf''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = FeatureExtractionPipeline(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''output_0''', '''output_1'''] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = infer_shapes(lowerCamelCase__ , lowerCamelCase__ ) # Assert all variables are present self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] , lowerCamelCase__ ) self.assertSequenceEqual(variable_names[3:] , lowerCamelCase__ ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] , {0: '''batch''', 1: '''sequence'''} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes['''output_0'''] , {0: '''batch''', 1: '''sequence'''} ) self.assertDictEqual(shapes['''output_1'''] , {0: '''batch'''} ) def snake_case__ ( self ): _lowerCamelCase = ['''input_ids''', '''attention_mask''', '''token_type_ids'''] _lowerCamelCase = {'''input_ids''': [1, 2, 3, 4], '''attention_mask''': [0, 0, 0, 0], '''token_type_ids''': [1, 1, 1, 1]} _lowerCamelCase , _lowerCamelCase = ensure_valid_input(FuncContiguousArgs() , lowerCamelCase__ , lowerCamelCase__ ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(lowerCamelCase__ ) , 3 ) # Should have exactly the same input names self.assertEqual(set(lowerCamelCase__ ) , set(lowerCamelCase__ ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(lowerCamelCase__ , (tokens['''input_ids'''], tokens['''token_type_ids'''], tokens['''attention_mask''']) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) _lowerCamelCase , _lowerCamelCase = ensure_valid_input(FuncNonContiguousArgs() , lowerCamelCase__ , lowerCamelCase__ ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(lowerCamelCase__ ) , 1 ) self.assertEqual(len(lowerCamelCase__ ) , 1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] , tokens['''input_ids'''] ) self.assertEqual(ordered_input_names[0] , '''input_ids''' ) def snake_case__ ( self ): _lowerCamelCase = generate_identified_filename(Path('''/home/something/my_fake_model.onnx''' ) , '''-test''' ) self.assertEqual('''/home/something/my_fake_model-test.onnx''' , generated.as_posix() )
661
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int ) -> int: assert isinstance(lowercase_ , lowercase_ ), F"""The input value of [n={number}] is not an integer""" if number == 1: return 2 elif number < 1: _lowerCamelCase = F"""The input value of [n={number}] has to be > 0""" raise ValueError(lowercase_ ) else: _lowerCamelCase = sylvester(number - 1 ) _lowerCamelCase = num - 1 _lowerCamelCase = num return lower * upper + 1 if __name__ == "__main__": print(F"""The 8th number in Sylvester's sequence: {sylvester(8)}""")
661
"""simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = {str(digit): digit**5 for digit in range(1_0)} def lowerCAmelCase_( lowercase_ : int ) -> int: return sum(DIGITS_FIFTH_POWER[digit] for digit in str(lowercase_ ) ) def lowerCAmelCase_( ) -> int: return sum( number for number in range(10_00 , 1_00_00_00 ) if number == digits_fifth_powers_sum(lowercase_ ) ) if __name__ == "__main__": print(solution())
661
1
"""simple docstring""" import argparse import gc import json import os import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler __SCREAMING_SNAKE_CASE : Any = 1_6 __SCREAMING_SNAKE_CASE : str = 3_2 def lowerCAmelCase_( lowercase_ : Optional[Any] ) -> int: return int(x / 2**20 ) class lowerCamelCase_: '''simple docstring''' def __enter__( self ): gc.collect() torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() # reset the peak gauge to zero _lowerCamelCase = torch.cuda.memory_allocated() return self def __exit__( self , *lowerCamelCase__ ): gc.collect() torch.cuda.empty_cache() _lowerCamelCase = torch.cuda.memory_allocated() _lowerCamelCase = torch.cuda.max_memory_allocated() _lowerCamelCase = bamb(self.end - self.begin ) _lowerCamelCase = bamb(self.peak - self.begin ) # print(f"delta used/peak {self.used:4d}/{self.peaked:4d}") def lowerCAmelCase_( lowercase_ : Accelerator , lowercase_ : int = 16 , lowercase_ : str = "bert-base-cased" , lowercase_ : int = 3_20 , lowercase_ : int = 1_60 , ) -> Dict: _lowerCamelCase = AutoTokenizer.from_pretrained(lowercase_ ) _lowerCamelCase = load_dataset( '''glue''' , '''mrpc''' , split={'''train''': F"""train[:{n_train}]""", '''validation''': F"""validation[:{n_val}]"""} ) def tokenize_function(lowercase_ : Any ): # max_length=None => use the model max length (it's actually the default) _lowerCamelCase = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=lowercase_ , max_length=lowercase_ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset _lowerCamelCase = datasets.map( lowercase_ , batched=lowercase_ , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , load_from_cache_file=lowercase_ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _lowerCamelCase = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(lowercase_ : Dict ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(lowercase_ , padding='''max_length''' , max_length=1_28 , return_tensors='''pt''' ) return tokenizer.pad(lowercase_ , padding='''longest''' , return_tensors='''pt''' ) # Instantiate dataloaders. _lowerCamelCase = DataLoader( tokenized_datasets['''train'''] , shuffle=lowercase_ , collate_fn=lowercase_ , batch_size=lowercase_ ) _lowerCamelCase = DataLoader( tokenized_datasets['''validation'''] , shuffle=lowercase_ , collate_fn=lowercase_ , batch_size=lowercase_ ) return train_dataloader, eval_dataloader def lowerCAmelCase_( lowercase_ : List[Any] , lowercase_ : Tuple ) -> List[str]: # Initialize accelerator _lowerCamelCase = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _lowerCamelCase = config['''lr'''] _lowerCamelCase = int(config['''num_epochs'''] ) _lowerCamelCase = int(config['''seed'''] ) _lowerCamelCase = int(config['''batch_size'''] ) _lowerCamelCase = args.model_name_or_path set_seed(lowercase_ ) _lowerCamelCase , _lowerCamelCase = get_dataloaders(lowercase_ , lowercase_ , lowercase_ , args.n_train , args.n_val ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _lowerCamelCase = AutoModelForSequenceClassification.from_pretrained(lowercase_ , return_dict=lowercase_ ) # Instantiate optimizer _lowerCamelCase = ( AdamW if accelerator.state.deepspeed_plugin is None or '''optimizer''' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) _lowerCamelCase = optimizer_cls(params=model.parameters() , lr=lowercase_ ) if accelerator.state.deepspeed_plugin is not None: _lowerCamelCase = accelerator.state.deepspeed_plugin.deepspeed_config[ '''gradient_accumulation_steps''' ] else: _lowerCamelCase = 1 _lowerCamelCase = (len(lowercase_ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): _lowerCamelCase = get_linear_schedule_with_warmup( optimizer=lowercase_ , num_warmup_steps=0 , num_training_steps=lowercase_ , ) else: _lowerCamelCase = DummyScheduler(lowercase_ , total_num_steps=lowercase_ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = accelerator.prepare( lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) # We need to keep track of how many total steps we have iterated over _lowerCamelCase = 0 # We also need to keep track of the stating epoch so files are named properly _lowerCamelCase = 0 # Now we train the model _lowerCamelCase = {} for epoch in range(lowercase_ , lowercase_ ): with TorchTracemalloc() as tracemalloc: model.train() for step, batch in enumerate(lowercase_ ): _lowerCamelCase = model(**lowercase_ ) _lowerCamelCase = outputs.loss _lowerCamelCase = loss / gradient_accumulation_steps accelerator.backward(lowercase_ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 # Printing the GPU memory usage details such as allocated memory, peak memory, and total memory usage accelerator.print('''Memory before entering the train : {}'''.format(bamb(tracemalloc.begin ) ) ) accelerator.print('''Memory consumed at the end of the train (end-begin): {}'''.format(tracemalloc.used ) ) accelerator.print('''Peak Memory consumed during the train (max-begin): {}'''.format(tracemalloc.peaked ) ) accelerator.print( '''Total Peak Memory consumed during the train (max): {}'''.format( tracemalloc.peaked + bamb(tracemalloc.begin ) ) ) _lowerCamelCase = tracemalloc.peaked + bamb(tracemalloc.begin ) if args.peak_memory_upper_bound is not None: assert ( train_total_peak_memory[F"""epoch-{epoch}"""] <= args.peak_memory_upper_bound ), "Peak memory usage exceeded the upper bound" accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , '''peak_memory_utilization.json''' ) , '''w''' ) as f: json.dump(lowercase_ , lowercase_ ) def lowerCAmelCase_( ) -> str: _lowerCamelCase = argparse.ArgumentParser(description='''Simple example of training script tracking peak GPU memory usage.''' ) parser.add_argument( '''--model_name_or_path''' , type=lowercase_ , default='''bert-base-cased''' , help='''Path to pretrained model or model identifier from huggingface.co/models.''' , required=lowercase_ , ) parser.add_argument( '''--output_dir''' , type=lowercase_ , default='''.''' , help='''Optional save directory where all checkpoint folders will be stored. Default is the current working directory.''' , ) parser.add_argument( '''--peak_memory_upper_bound''' , type=lowercase_ , default=lowercase_ , help='''The upper bound of peak memory usage in MB. If set, the training will throw an error if the peak memory usage exceeds this value.''' , ) parser.add_argument( '''--n_train''' , type=lowercase_ , default=3_20 , help='''Number of training examples to use.''' , ) parser.add_argument( '''--n_val''' , type=lowercase_ , default=1_60 , help='''Number of validation examples to use.''' , ) parser.add_argument( '''--num_epochs''' , type=lowercase_ , default=1 , help='''Number of train epochs.''' , ) _lowerCamelCase = parser.parse_args() _lowerCamelCase = {'''lr''': 2e-5, '''num_epochs''': args.num_epochs, '''seed''': 42, '''batch_size''': 16} training_function(lowercase_ , lowercase_ ) if __name__ == "__main__": main()
661
"""simple docstring""" from __future__ import annotations import os from collections.abc import Mapping __SCREAMING_SNAKE_CASE : str = tuple[int, int] class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = vertices _lowerCamelCase = { (min(lowerCamelCase__ ), max(lowerCamelCase__ )): weight for edge, weight in edges.items() } def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): self.vertices.add(edge[0] ) self.vertices.add(edge[1] ) _lowerCamelCase = weight def snake_case__ ( self ): _lowerCamelCase = Graph({min(self.vertices )} , {} ) _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 while len(subgraph.vertices ) < len(self.vertices ): _lowerCamelCase = max(self.edges.values() ) + 1 for edge, weight in self.edges.items(): if (edge[0] in subgraph.vertices) ^ (edge[1] in subgraph.vertices): if weight < min_weight: _lowerCamelCase = edge _lowerCamelCase = weight subgraph.add_edge(lowerCamelCase__ , lowerCamelCase__ ) return subgraph def lowerCAmelCase_( lowercase_ : str = "p107_network.txt" ) -> int: _lowerCamelCase = os.path.abspath(os.path.dirname(lowercase_ ) ) _lowerCamelCase = os.path.join(lowercase_ , lowercase_ ) _lowerCamelCase = {} _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 with open(lowercase_ ) as f: _lowerCamelCase = f.read().strip().split('''\n''' ) _lowerCamelCase = [line.split(''',''' ) for line in data] for edgea in range(1 , len(lowercase_ ) ): for edgea in range(lowercase_ ): if adjaceny_matrix[edgea][edgea] != "-": _lowerCamelCase = int(adjaceny_matrix[edgea][edgea] ) _lowerCamelCase = Graph(set(range(len(lowercase_ ) ) ) , lowercase_ ) _lowerCamelCase = graph.prims_algorithm() _lowerCamelCase = sum(graph.edges.values() ) _lowerCamelCase = sum(subgraph.edges.values() ) return initial_total - optimal_total if __name__ == "__main__": print(F"""{solution() = }""")
661
1
"""simple docstring""" import uuid from typing import Any, Dict, List, Optional, Union from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch __SCREAMING_SNAKE_CASE : Optional[Any] = logging.get_logger(__name__) class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__=None , lowerCamelCase__=None ): if not conversation_id: _lowerCamelCase = uuid.uuida() if past_user_inputs is None: _lowerCamelCase = [] if generated_responses is None: _lowerCamelCase = [] _lowerCamelCase = conversation_id _lowerCamelCase = past_user_inputs _lowerCamelCase = generated_responses _lowerCamelCase = text def __eq__( self , lowerCamelCase__ ): if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): return False if self.uuid == other.uuid: return True return ( self.new_user_input == other.new_user_input and self.past_user_inputs == other.past_user_inputs and self.generated_responses == other.generated_responses ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = False ): if self.new_user_input: if overwrite: logger.warning( F"""User input added while unprocessed input was existing: \"{self.new_user_input}\" was overwritten """ F"""with: \"{text}\".""" ) _lowerCamelCase = text else: logger.warning( F"""User input added while unprocessed input was existing: \"{self.new_user_input}\" new input """ F"""ignored: \"{text}\". Set `overwrite` to True to overwrite unprocessed user input""" ) else: _lowerCamelCase = text def snake_case__ ( self ): if self.new_user_input: self.past_user_inputs.append(self.new_user_input ) _lowerCamelCase = None def snake_case__ ( self , lowerCamelCase__ ): self.generated_responses.append(lowerCamelCase__ ) def snake_case__ ( self ): for user_input, generated_response in zip(self.past_user_inputs , self.generated_responses ): yield True, user_input yield False, generated_response if self.new_user_input: yield True, self.new_user_input def __repr__( self ): _lowerCamelCase = F"""Conversation id: {self.uuid} \n""" for is_user, text in self.iter_texts(): _lowerCamelCase = '''user''' if is_user else '''bot''' output += F"""{name} >> {text} \n""" return output @add_end_docstrings( A__, r'\n min_length_for_response (`int`, *optional*, defaults to 32):\n The minimum length (in number of tokens) for a response.\n minimum_tokens (`int`, *optional*, defaults to 10):\n The minimum length of tokens to leave for a response.\n ', ) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , *lowerCamelCase__ , **lowerCamelCase__ ): super().__init__(*lowerCamelCase__ , **lowerCamelCase__ ) if self.tokenizer.pad_token_id is None: _lowerCamelCase = self.tokenizer.eos_token def snake_case__ ( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , **lowerCamelCase__ ): _lowerCamelCase = {} _lowerCamelCase = {} _lowerCamelCase = {} if min_length_for_response is not None: _lowerCamelCase = min_length_for_response if minimum_tokens is not None: _lowerCamelCase = minimum_tokens if "max_length" in generate_kwargs: _lowerCamelCase = generate_kwargs['''max_length'''] # self.max_length = generate_kwargs.get("max_length", self.model.config.max_length) if clean_up_tokenization_spaces is not None: _lowerCamelCase = clean_up_tokenization_spaces if generate_kwargs: forward_params.update(lowerCamelCase__ ) return preprocess_params, forward_params, postprocess_params def __call__( self , lowerCamelCase__ , lowerCamelCase__=0 , **lowerCamelCase__ ): _lowerCamelCase = super().__call__(lowerCamelCase__ , num_workers=lowerCamelCase__ , **lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and len(lowerCamelCase__ ) == 1: return outputs[0] return outputs def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=3_2 ): if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): raise ValueError('''ConversationalPipeline, expects Conversation as inputs''' ) if conversation.new_user_input is None: raise ValueError( F"""Conversation with UUID {type(conversation.uuid )} does not contain new user input to process. """ '''Add user inputs with the conversation\'s `add_user_input` method''' ) if hasattr(self.tokenizer , '''_build_conversation_input_ids''' ): _lowerCamelCase = self.tokenizer._build_conversation_input_ids(lowerCamelCase__ ) else: # If the tokenizer cannot handle conversations, we default to only the old version _lowerCamelCase = self._legacy_parse_and_tokenize(lowerCamelCase__ ) if self.framework == "pt": _lowerCamelCase = torch.LongTensor([input_ids] ) elif self.framework == "tf": _lowerCamelCase = tf.constant([input_ids] ) return {"input_ids": input_ids, "conversation": conversation} def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=1_0 , **lowerCamelCase__ ): _lowerCamelCase = generate_kwargs.get('''max_length''' , self.model.config.max_length ) _lowerCamelCase = model_inputs['''input_ids'''].shape[1] if max_length - minimum_tokens < n: logger.warning(F"""Conversation input is to long ({n}), trimming it to ({max_length} - {minimum_tokens})""" ) _lowerCamelCase = max_length - minimum_tokens _lowerCamelCase = model_inputs['''input_ids'''][:, -trim:] if "attention_mask" in model_inputs: _lowerCamelCase = model_inputs['''attention_mask'''][:, -trim:] _lowerCamelCase = model_inputs.pop('''conversation''' ) _lowerCamelCase = max_length _lowerCamelCase = self.model.generate(**lowerCamelCase__ , **lowerCamelCase__ ) if self.model.config.is_encoder_decoder: _lowerCamelCase = 1 else: _lowerCamelCase = n return {"output_ids": output_ids[:, start_position:], "conversation": conversation} def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=True ): _lowerCamelCase = model_outputs['''output_ids'''] _lowerCamelCase = self.tokenizer.decode( output_ids[0] , skip_special_tokens=lowerCamelCase__ , clean_up_tokenization_spaces=lowerCamelCase__ , ) _lowerCamelCase = model_outputs['''conversation'''] conversation.mark_processed() conversation.append_response(lowerCamelCase__ ) return conversation def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = self.tokenizer.eos_token_id _lowerCamelCase = [] for is_user, text in conversation.iter_texts(): if eos_token_id is not None: input_ids.extend(self.tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) + [eos_token_id] ) else: input_ids.extend(self.tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) ) if len(lowerCamelCase__ ) > self.tokenizer.model_max_length: _lowerCamelCase = input_ids[-self.tokenizer.model_max_length :] return input_ids
661
"""simple docstring""" from __future__ import annotations from math import ceil, floor, sqrt def lowerCAmelCase_( lowercase_ : int = 2_00_00_00 ) -> int: _lowerCamelCase = [0] _lowerCamelCase = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target _lowerCamelCase = 0 # the area corresponding to the grid that gives the product closest to target _lowerCamelCase = 0 # an estimate of b, using the quadratic formula _lowerCamelCase = 42 # the largest integer less than b_estimate _lowerCamelCase = 42 # the largest integer less than b_estimate _lowerCamelCase = 42 # the triangle number corresponding to b_floor _lowerCamelCase = 42 # the triangle number corresponding to b_ceil _lowerCamelCase = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): _lowerCamelCase = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 _lowerCamelCase = floor(lowercase_ ) _lowerCamelCase = ceil(lowercase_ ) _lowerCamelCase = triangle_numbers[b_floor] _lowerCamelCase = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): _lowerCamelCase = triangle_b_first_guess * triangle_a _lowerCamelCase = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): _lowerCamelCase = triangle_b_second_guess * triangle_a _lowerCamelCase = idx_a * b_ceil return area if __name__ == "__main__": print(F"""{solution() = }""")
661
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor __SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , *lowerCamelCase__ , **lowerCamelCase__ ): warnings.warn( '''The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use FlavaImageProcessor instead.''' , lowerCamelCase__ , ) super().__init__(*lowerCamelCase__ , **lowerCamelCase__ )
661
"""simple docstring""" import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , *lowerCamelCase__ , **lowerCamelCase__ ): warnings.warn( '''The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use DeiTImageProcessor instead.''' , lowerCamelCase__ , ) super().__init__(*lowerCamelCase__ , **lowerCamelCase__ )
661
1
"""simple docstring""" __SCREAMING_SNAKE_CASE : dict[tuple[int, int, int], int] = {} def lowerCAmelCase_( lowercase_ : int , lowercase_ : int , lowercase_ : int ) -> int: # if we are absent twice, or late 3 consecutive days, # no further prize strings are possible if late == 3 or absent == 2: return 0 # if we have no days left, and have not failed any other rules, # we have a prize string if days == 0: return 1 # No easy solution, so now we need to do the recursive calculation # First, check if the combination is already in the cache, and # if yes, return the stored value from there since we already # know the number of possible prize strings from this point on _lowerCamelCase = (days, absent, late) if key in cache: return cache[key] # now we calculate the three possible ways that can unfold from # this point on, depending on our attendance today # 1) if we are late (but not absent), the "absent" counter stays as # it is, but the "late" counter increases by one _lowerCamelCase = _calculate(days - 1 , lowercase_ , late + 1 ) # 2) if we are absent, the "absent" counter increases by 1, and the # "late" counter resets to 0 _lowerCamelCase = _calculate(days - 1 , absent + 1 , 0 ) # 3) if we are on time, this resets the "late" counter and keeps the # absent counter _lowerCamelCase = _calculate(days - 1 , lowercase_ , 0 ) _lowerCamelCase = state_late + state_absent + state_ontime _lowerCamelCase = prizestrings return prizestrings def lowerCAmelCase_( lowercase_ : int = 30 ) -> int: return _calculate(lowercase_ , absent=0 , late=0 ) if __name__ == "__main__": print(solution())
661
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Optional[Any] = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''vocab_file''': { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json''' ), '''distilbert-base-german-cased''': ( '''https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json''' ), '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json''' ), }, } __SCREAMING_SNAKE_CASE : str = { '''distilbert-base-uncased''': 5_1_2, '''distilbert-base-uncased-distilled-squad''': 5_1_2, '''distilbert-base-cased''': 5_1_2, '''distilbert-base-cased-distilled-squad''': 5_1_2, '''distilbert-base-german-cased''': 5_1_2, '''distilbert-base-multilingual-cased''': 5_1_2, } __SCREAMING_SNAKE_CASE : str = { '''distilbert-base-uncased''': {'''do_lower_case''': True}, '''distilbert-base-uncased-distilled-squad''': {'''do_lower_case''': True}, '''distilbert-base-cased''': {'''do_lower_case''': False}, '''distilbert-base-cased-distilled-squad''': {'''do_lower_case''': False}, '''distilbert-base-german-cased''': {'''do_lower_case''': False}, '''distilbert-base-multilingual-cased''': {'''do_lower_case''': False}, } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[Any] = VOCAB_FILES_NAMES lowercase__ : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP lowercase__ : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : str = PRETRAINED_INIT_CONFIGURATION lowercase__ : int = ['input_ids', 'attention_mask'] lowercase__ : Tuple = DistilBertTokenizer def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__="[UNK]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ): super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) _lowerCamelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , lowerCamelCase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' , lowerCamelCase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , lowerCamelCase__ ) != tokenize_chinese_chars ): _lowerCamelCase = getattr(lowerCamelCase__ , normalizer_state.pop('''type''' ) ) _lowerCamelCase = do_lower_case _lowerCamelCase = strip_accents _lowerCamelCase = tokenize_chinese_chars _lowerCamelCase = normalizer_class(**lowerCamelCase__ ) _lowerCamelCase = do_lower_case def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=None ): _lowerCamelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
661
1
"""simple docstring""" import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : Optional[int] ) -> List[str]: _lowerCamelCase = checkpoint _lowerCamelCase = {} _lowerCamelCase = vae_state_dict['''encoder.conv_in.weight'''] _lowerCamelCase = vae_state_dict['''encoder.conv_in.bias'''] _lowerCamelCase = vae_state_dict['''encoder.conv_out.weight'''] _lowerCamelCase = vae_state_dict['''encoder.conv_out.bias'''] _lowerCamelCase = vae_state_dict['''encoder.norm_out.weight'''] _lowerCamelCase = vae_state_dict['''encoder.norm_out.bias'''] _lowerCamelCase = vae_state_dict['''decoder.conv_in.weight'''] _lowerCamelCase = vae_state_dict['''decoder.conv_in.bias'''] _lowerCamelCase = vae_state_dict['''decoder.conv_out.weight'''] _lowerCamelCase = vae_state_dict['''decoder.conv_out.bias'''] _lowerCamelCase = vae_state_dict['''decoder.norm_out.weight'''] _lowerCamelCase = vae_state_dict['''decoder.norm_out.bias'''] _lowerCamelCase = vae_state_dict['''quant_conv.weight'''] _lowerCamelCase = vae_state_dict['''quant_conv.bias'''] _lowerCamelCase = vae_state_dict['''post_quant_conv.weight'''] _lowerCamelCase = vae_state_dict['''post_quant_conv.bias'''] # Retrieves the keys for the encoder down blocks only _lowerCamelCase = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''encoder.down''' in layer} ) _lowerCamelCase = { layer_id: [key for key in vae_state_dict if F"""down.{layer_id}""" in key] for layer_id in range(lowercase_ ) } # Retrieves the keys for the decoder up blocks only _lowerCamelCase = len({'''.'''.join(layer.split('''.''' )[:3] ) for layer in vae_state_dict if '''decoder.up''' in layer} ) _lowerCamelCase = { layer_id: [key for key in vae_state_dict if F"""up.{layer_id}""" in key] for layer_id in range(lowercase_ ) } for i in range(lowercase_ ): _lowerCamelCase = [key for key in down_blocks[i] if F"""down.{i}""" in key and F"""down.{i}.downsample""" not in key] if F"""encoder.down.{i}.downsample.conv.weight""" in vae_state_dict: _lowerCamelCase = vae_state_dict.pop( F"""encoder.down.{i}.downsample.conv.weight""" ) _lowerCamelCase = vae_state_dict.pop( F"""encoder.down.{i}.downsample.conv.bias""" ) _lowerCamelCase = renew_vae_resnet_paths(lowercase_ ) _lowerCamelCase = {'''old''': F"""down.{i}.block""", '''new''': F"""down_blocks.{i}.resnets"""} assign_to_checkpoint(lowercase_ , lowercase_ , lowercase_ , additional_replacements=[meta_path] , config=lowercase_ ) _lowerCamelCase = [key for key in vae_state_dict if '''encoder.mid.block''' in key] _lowerCamelCase = 2 for i in range(1 , num_mid_res_blocks + 1 ): _lowerCamelCase = [key for key in mid_resnets if F"""encoder.mid.block_{i}""" in key] _lowerCamelCase = renew_vae_resnet_paths(lowercase_ ) _lowerCamelCase = {'''old''': F"""mid.block_{i}""", '''new''': F"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(lowercase_ , lowercase_ , lowercase_ , additional_replacements=[meta_path] , config=lowercase_ ) _lowerCamelCase = [key for key in vae_state_dict if '''encoder.mid.attn''' in key] _lowerCamelCase = renew_vae_attention_paths(lowercase_ ) _lowerCamelCase = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(lowercase_ , lowercase_ , lowercase_ , additional_replacements=[meta_path] , config=lowercase_ ) conv_attn_to_linear(lowercase_ ) for i in range(lowercase_ ): _lowerCamelCase = num_up_blocks - 1 - i _lowerCamelCase = [ key for key in up_blocks[block_id] if F"""up.{block_id}""" in key and F"""up.{block_id}.upsample""" not in key ] if F"""decoder.up.{block_id}.upsample.conv.weight""" in vae_state_dict: _lowerCamelCase = vae_state_dict[ F"""decoder.up.{block_id}.upsample.conv.weight""" ] _lowerCamelCase = vae_state_dict[ F"""decoder.up.{block_id}.upsample.conv.bias""" ] _lowerCamelCase = renew_vae_resnet_paths(lowercase_ ) _lowerCamelCase = {'''old''': F"""up.{block_id}.block""", '''new''': F"""up_blocks.{i}.resnets"""} assign_to_checkpoint(lowercase_ , lowercase_ , lowercase_ , additional_replacements=[meta_path] , config=lowercase_ ) _lowerCamelCase = [key for key in vae_state_dict if '''decoder.mid.block''' in key] _lowerCamelCase = 2 for i in range(1 , num_mid_res_blocks + 1 ): _lowerCamelCase = [key for key in mid_resnets if F"""decoder.mid.block_{i}""" in key] _lowerCamelCase = renew_vae_resnet_paths(lowercase_ ) _lowerCamelCase = {'''old''': F"""mid.block_{i}""", '''new''': F"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(lowercase_ , lowercase_ , lowercase_ , additional_replacements=[meta_path] , config=lowercase_ ) _lowerCamelCase = [key for key in vae_state_dict if '''decoder.mid.attn''' in key] _lowerCamelCase = renew_vae_attention_paths(lowercase_ ) _lowerCamelCase = {'''old''': '''mid.attn_1''', '''new''': '''mid_block.attentions.0'''} assign_to_checkpoint(lowercase_ , lowercase_ , lowercase_ , additional_replacements=[meta_path] , config=lowercase_ ) conv_attn_to_linear(lowercase_ ) return new_checkpoint def lowerCAmelCase_( lowercase_ : str , lowercase_ : str , ) -> Union[str, Any]: # Only support V1 _lowerCamelCase = requests.get( ''' https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml''' ) _lowerCamelCase = io.BytesIO(r.content ) _lowerCamelCase = OmegaConf.load(lowercase_ ) _lowerCamelCase = 5_12 _lowerCamelCase = '''cuda''' if torch.cuda.is_available() else '''cpu''' if checkpoint_path.endswith('''safetensors''' ): from safetensors import safe_open _lowerCamelCase = {} with safe_open(lowercase_ , framework='''pt''' , device='''cpu''' ) as f: for key in f.keys(): _lowerCamelCase = f.get_tensor(lowercase_ ) else: _lowerCamelCase = torch.load(lowercase_ , map_location=lowercase_ )['''state_dict'''] # Convert the VAE model. _lowerCamelCase = create_vae_diffusers_config(lowercase_ , image_size=lowercase_ ) _lowerCamelCase = custom_convert_ldm_vae_checkpoint(lowercase_ , lowercase_ ) _lowerCamelCase = AutoencoderKL(**lowercase_ ) vae.load_state_dict(lowercase_ ) vae.save_pretrained(lowercase_ ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : int = argparse.ArgumentParser() parser.add_argument('''--vae_pt_path''', default=None, type=str, required=True, help='''Path to the VAE.pt to convert.''') parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the VAE.pt to convert.''') __SCREAMING_SNAKE_CASE : str = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
661
"""simple docstring""" from math import sqrt import numpy as np from sympy import symbols # Coefficient # Speed of light (m/s) __SCREAMING_SNAKE_CASE : str = 2_9_9_7_9_2_4_5_8 # Symbols __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Union[str, Any] = symbols('''ct x y z''') def lowerCAmelCase_( lowercase_ : float ) -> float: if velocity > c: raise ValueError('''Speed must not exceed light speed 299,792,458 [m/s]!''' ) elif velocity < 1: # Usually the speed should be much higher than 1 (c order of magnitude) raise ValueError('''Speed must be greater than or equal to 1!''' ) return velocity / c def lowerCAmelCase_( lowercase_ : float ) -> float: return 1 / sqrt(1 - beta(lowercase_ ) ** 2 ) def lowerCAmelCase_( lowercase_ : float ) -> np.ndarray: return np.array( [ [gamma(lowercase_ ), -gamma(lowercase_ ) * beta(lowercase_ ), 0, 0], [-gamma(lowercase_ ) * beta(lowercase_ ), gamma(lowercase_ ), 0, 0], [0, 0, 1, 0], [0, 0, 0, 1], ] ) def lowerCAmelCase_( lowercase_ : float , lowercase_ : np.ndarray | None = None ) -> np.ndarray: # Ensure event is not empty if event is None: _lowerCamelCase = np.array([ct, x, y, z] ) # Symbolic four vector else: event[0] *= c # x0 is ct (speed of light * time) return transformation_matrix(lowercase_ ) @ event if __name__ == "__main__": import doctest doctest.testmod() # Example of symbolic vector: __SCREAMING_SNAKE_CASE : List[str] = transform(2_9_9_7_9_2_4_5) print('''Example of four vector: ''') print(F"""ct' = {four_vector[0]}""") print(F"""x' = {four_vector[1]}""") print(F"""y' = {four_vector[2]}""") print(F"""z' = {four_vector[3]}""") # Substitute symbols with numerical values __SCREAMING_SNAKE_CASE : Tuple = {ct: c, x: 1, y: 1, z: 1} __SCREAMING_SNAKE_CASE : int = [four_vector[i].subs(sub_dict) for i in range(4)] print(F"""\n{numerical_vector}""")
661
1
"""simple docstring""" import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class lowerCamelCase_: '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return None class lowerCamelCase_: '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return None class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = [ # (model_name, model_kwargs) ('bert-base-cased', {}), ('gpt2', {'use_cache': False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase__ , '''tf''' , 1_2 , **lowerCamelCase__ ) @require_torch @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , **lowerCamelCase__ ) @require_torch @slow def snake_case__ ( self ): from transformers import BertModel _lowerCamelCase = ['''[UNK]''', '''[SEP]''', '''[CLS]''', '''[PAD]''', '''[MASK]''', '''some''', '''other''', '''words'''] with NamedTemporaryFile(mode='''w+t''' ) as vocab_file: vocab_file.write('''\n'''.join(lowerCamelCase__ ) ) vocab_file.flush() _lowerCamelCase = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: _lowerCamelCase = BertModel(BertConfig(vocab_size=len(lowerCamelCase__ ) ) ) model.save_pretrained(lowerCamelCase__ ) self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , lowerCamelCase__ ) @require_tf @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCamelCase = self._test_export(lowerCamelCase__ , '''tf''' , 1_2 , **lowerCamelCase__ ) _lowerCamelCase = quantize(Path(lowerCamelCase__ ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) @require_torch @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCamelCase = self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , **lowerCamelCase__ ) _lowerCamelCase = quantize(lowerCamelCase__ ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): try: # Compute path with TemporaryDirectory() as tempdir: _lowerCamelCase = Path(lowerCamelCase__ ).joinpath('''model.onnx''' ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) return path except Exception as e: self.fail(lowerCamelCase__ ) @require_torch @require_tokenizers @slow def snake_case__ ( self ): from transformers import BertModel _lowerCamelCase = BertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) _lowerCamelCase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(lowerCamelCase__ , lowerCamelCase__ , '''pt''' ) @require_tf @require_tokenizers @slow def snake_case__ ( self ): from transformers import TFBertModel _lowerCamelCase = TFBertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) _lowerCamelCase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(lowerCamelCase__ , lowerCamelCase__ , '''tf''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = FeatureExtractionPipeline(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''output_0''', '''output_1'''] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = infer_shapes(lowerCamelCase__ , lowerCamelCase__ ) # Assert all variables are present self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] , lowerCamelCase__ ) self.assertSequenceEqual(variable_names[3:] , lowerCamelCase__ ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] , {0: '''batch''', 1: '''sequence'''} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes['''output_0'''] , {0: '''batch''', 1: '''sequence'''} ) self.assertDictEqual(shapes['''output_1'''] , {0: '''batch'''} ) def snake_case__ ( self ): _lowerCamelCase = ['''input_ids''', '''attention_mask''', '''token_type_ids'''] _lowerCamelCase = {'''input_ids''': [1, 2, 3, 4], '''attention_mask''': [0, 0, 0, 0], '''token_type_ids''': [1, 1, 1, 1]} _lowerCamelCase , _lowerCamelCase = ensure_valid_input(FuncContiguousArgs() , lowerCamelCase__ , lowerCamelCase__ ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(lowerCamelCase__ ) , 3 ) # Should have exactly the same input names self.assertEqual(set(lowerCamelCase__ ) , set(lowerCamelCase__ ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(lowerCamelCase__ , (tokens['''input_ids'''], tokens['''token_type_ids'''], tokens['''attention_mask''']) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) _lowerCamelCase , _lowerCamelCase = ensure_valid_input(FuncNonContiguousArgs() , lowerCamelCase__ , lowerCamelCase__ ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(lowerCamelCase__ ) , 1 ) self.assertEqual(len(lowerCamelCase__ ) , 1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] , tokens['''input_ids'''] ) self.assertEqual(ordered_input_names[0] , '''input_ids''' ) def snake_case__ ( self ): _lowerCamelCase = generate_identified_filename(Path('''/home/something/my_fake_model.onnx''' ) , '''-test''' ) self.assertEqual('''/home/something/my_fake_model-test.onnx''' , generated.as_posix() )
661
"""simple docstring""" from __future__ import annotations __SCREAMING_SNAKE_CASE : Dict = 1.6_0_2_1e-1_9 # units = C def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float , ) -> tuple[str, float]: if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif conductivity < 0: raise ValueError('''Conductivity cannot be negative''' ) elif electron_conc < 0: raise ValueError('''Electron concentration cannot be negative''' ) elif mobility < 0: raise ValueError('''mobility cannot be negative''' ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
661
1
"""simple docstring""" import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Any = GPTSanJapaneseTokenizer lowercase__ : List[Any] = False lowercase__ : Union[str, Any] = {'do_clean_text': False, 'add_prefix_space': False} def snake_case__ ( self ): super().setUp() # fmt: off _lowerCamelCase = ['''こん''', '''こんに''', '''にちは''', '''ばんは''', '''世界,㔺界''', '''、''', '''。''', '''<BR>''', '''<SP>''', '''<TAB>''', '''<URL>''', '''<EMAIL>''', '''<TEL>''', '''<DATE>''', '''<PRICE>''', '''<BLOCK>''', '''<KIGOU>''', '''<U2000U2BFF>''', '''<|emoji1|>''', '''<unk>''', '''<|bagoftoken|>''', '''<|endoftext|>'''] # fmt: on _lowerCamelCase = {'''emoji''': {'''\ud83d\ude00''': '''<|emoji1|>'''}, '''emoji_inv''': {'''<|emoji1|>''': '''\ud83d\ude00'''}} # 😀 _lowerCamelCase = {'''unk_token''': '''<unk>'''} _lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''emoji_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) with open(self.emoji_file , '''w''' ) as emoji_writer: emoji_writer.write(json.dumps(lowerCamelCase__ ) ) def snake_case__ ( self , **lowerCamelCase__ ): kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = '''こんにちは、世界。 \nこんばんは、㔺界。😀''' _lowerCamelCase = '''こんにちは、世界。 \nこんばんは、世界。😀''' return input_text, output_text def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase , _lowerCamelCase = self.get_input_output_texts(lowerCamelCase__ ) _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) _lowerCamelCase = tokenizer.decode(lowerCamelCase__ , clean_up_tokenization_spaces=lowerCamelCase__ ) return text, ids def snake_case__ ( self ): pass # TODO add if relevant def snake_case__ ( self ): pass # TODO add if relevant def snake_case__ ( self ): pass # TODO add if relevant def snake_case__ ( self ): _lowerCamelCase = self.get_tokenizer() # Testing tokenization _lowerCamelCase = '''こんにちは、世界。 こんばんは、㔺界。''' _lowerCamelCase = ['''こん''', '''にちは''', '''、''', '''世界''', '''。''', '''<SP>''', '''こん''', '''ばんは''', '''、''', '''㔺界''', '''。'''] _lowerCamelCase = tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) # Testing conversion to ids without special tokens _lowerCamelCase = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] _lowerCamelCase = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) # Testing conversion to ids with special tokens _lowerCamelCase = tokens + [tokenizer.unk_token] _lowerCamelCase = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 1_9] _lowerCamelCase = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.get_tokenizer() # Testing tokenization _lowerCamelCase = '''こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。''' _lowerCamelCase = '''こんにちは、、、、世界。こんばんは、、、、世界。''' _lowerCamelCase = tokenizer.encode(lowerCamelCase__ ) _lowerCamelCase = tokenizer.decode(lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): _lowerCamelCase = self.tokenizer_class.from_pretrained('''Tanrei/GPTSAN-japanese''' ) # Testing tokenization _lowerCamelCase = '''こんにちは、世界。''' _lowerCamelCase = '''こんばんは、㔺界。😀''' _lowerCamelCase = '''こんにちは、世界。こんばんは、世界。😀''' _lowerCamelCase = tokenizer.encode(prefix_text + input_text ) _lowerCamelCase = tokenizer.encode('''''' , prefix_text=prefix_text + input_text ) _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , prefix_text=lowerCamelCase__ ) _lowerCamelCase = tokenizer.decode(lowerCamelCase__ ) _lowerCamelCase = tokenizer.decode(lowerCamelCase__ ) _lowerCamelCase = tokenizer.decode(lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): _lowerCamelCase = self.tokenizer_class.from_pretrained('''Tanrei/GPTSAN-japanese''' ) # Testing tokenization _lowerCamelCase = '''こんにちは、世界。''' _lowerCamelCase = '''こんばんは、㔺界。😀''' _lowerCamelCase = len(tokenizer.encode(lowerCamelCase__ ) ) - 2 _lowerCamelCase = len(tokenizer.encode(lowerCamelCase__ ) ) - 2 _lowerCamelCase = [1] + [0] * (len_prefix + len_text + 1) _lowerCamelCase = [1] * (len_prefix + len_text + 1) + [0] _lowerCamelCase = [1] + [1] * (len_prefix) + [0] * (len_text + 1) _lowerCamelCase = tokenizer(prefix_text + input_text ).token_type_ids _lowerCamelCase = tokenizer('''''' , prefix_text=prefix_text + input_text ).token_type_ids _lowerCamelCase = tokenizer(lowerCamelCase__ , prefix_text=lowerCamelCase__ ).token_type_ids self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): _lowerCamelCase = self.tokenizer_class.from_pretrained('''Tanrei/GPTSAN-japanese''' ) _lowerCamelCase = tokenizer.encode('''あンいワ''' ) _lowerCamelCase = tokenizer.encode('''''' , prefix_text='''あンいワ''' ) _lowerCamelCase = tokenizer.encode('''いワ''' , prefix_text='''あン''' ) self.assertEqual(tokenizer.decode(lowerCamelCase__ ) , tokenizer.decode(lowerCamelCase__ ) ) self.assertEqual(tokenizer.decode(lowerCamelCase__ ) , tokenizer.decode(lowerCamelCase__ ) ) self.assertNotEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertNotEqual(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def snake_case__ ( self ): _lowerCamelCase = self.tokenizer_class.from_pretrained('''Tanrei/GPTSAN-japanese''' ) _lowerCamelCase = [['''武田信玄''', '''は、'''], ['''織田信長''', '''の配下の、''']] _lowerCamelCase = tokenizer(lowerCamelCase__ , padding=lowerCamelCase__ ) _lowerCamelCase = tokenizer.batch_encode_plus(lowerCamelCase__ , padding=lowerCamelCase__ ) # fmt: off _lowerCamelCase = [[3_5_9_9_3, 8_6_4_0, 2_5_9_4_8, 3_5_9_9_8, 3_0_6_4_7, 3_5_6_7_5, 3_5_9_9_9, 3_5_9_9_9], [3_5_9_9_3, 1_0_3_8_2, 9_8_6_8, 3_5_9_9_8, 3_0_6_4_6, 9_4_5_9, 3_0_6_4_6, 3_5_6_7_5]] _lowerCamelCase = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] _lowerCamelCase = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , lowerCamelCase__ ) self.assertListEqual(x_token.token_type_ids , lowerCamelCase__ ) self.assertListEqual(x_token.attention_mask , lowerCamelCase__ ) self.assertListEqual(x_token_a.input_ids , lowerCamelCase__ ) self.assertListEqual(x_token_a.token_type_ids , lowerCamelCase__ ) self.assertListEqual(x_token_a.attention_mask , lowerCamelCase__ ) def snake_case__ ( self ): # Intentionally convert some words to accommodate character fluctuations unique to Japanese pass def snake_case__ ( self ): # tokenizer has no padding token pass
661
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): _lowerCamelCase = { '''task_specific_params''': { '''summarization''': {'''length_penalty''': 1.0, '''max_length''': 1_2_8, '''min_length''': 1_2, '''num_beams''': 4}, '''summarization_cnn''': {'''length_penalty''': 2.0, '''max_length''': 1_4_2, '''min_length''': 5_6, '''num_beams''': 4}, '''summarization_xsum''': {'''length_penalty''': 1.0, '''max_length''': 6_2, '''min_length''': 1_1, '''num_beams''': 6}, } } _lowerCamelCase = { '''task_specific_params.summarization.length_penalty''': 1.0, '''task_specific_params.summarization.max_length''': 1_2_8, '''task_specific_params.summarization.min_length''': 1_2, '''task_specific_params.summarization.num_beams''': 4, '''task_specific_params.summarization_cnn.length_penalty''': 2.0, '''task_specific_params.summarization_cnn.max_length''': 1_4_2, '''task_specific_params.summarization_cnn.min_length''': 5_6, '''task_specific_params.summarization_cnn.num_beams''': 4, '''task_specific_params.summarization_xsum.length_penalty''': 1.0, '''task_specific_params.summarization_xsum.max_length''': 6_2, '''task_specific_params.summarization_xsum.min_length''': 1_1, '''task_specific_params.summarization_xsum.num_beams''': 6, } self.assertEqual(flatten_dict(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , x.transpose() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , np.asarray(transpose(lowerCamelCase__ ) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , np.asarray(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.reshape(lowerCamelCase__ , (4, 3) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , np.reshape(lowerCamelCase__ , (1_2, 5) ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , reshape(lowerCamelCase__ , (1_2, 5) ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , reshape(lowerCamelCase__ , (1_2, 5) ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.asarray(reshape(lowerCamelCase__ , (4, 3) ) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , np.asarray(reshape(lowerCamelCase__ , (1_2, 5) ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.squeeze(lowerCamelCase__ ) ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.squeeze(lowerCamelCase__ , axis=2 ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.asarray(squeeze(lowerCamelCase__ ) ) ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.asarray(squeeze(lowerCamelCase__ , axis=2 ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.expand_dims(lowerCamelCase__ , axis=1 ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.asarray(expand_dims(lowerCamelCase__ , axis=1 ) ) ) )
661
1
"""simple docstring""" import unittest from transformers.testing_utils import CaptureStdout from transformers.tools.python_interpreter import evaluate def lowerCAmelCase_( lowercase_ : List[Any] ) -> Optional[int]: return x + 2 class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): _lowerCamelCase = '''x = 3''' _lowerCamelCase = {} _lowerCamelCase = evaluate(lowerCamelCase__ , {} , state=lowerCamelCase__ ) assert result == 3 self.assertDictEqual(lowerCamelCase__ , {'''x''': 3} ) _lowerCamelCase = '''x = y''' _lowerCamelCase = {'''y''': 5} _lowerCamelCase = evaluate(lowerCamelCase__ , {} , state=lowerCamelCase__ ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(lowerCamelCase__ , {'''x''': 5, '''y''': 5} ) def snake_case__ ( self ): _lowerCamelCase = '''y = add_two(x)''' _lowerCamelCase = {'''x''': 3} _lowerCamelCase = evaluate(lowerCamelCase__ , {'''add_two''': add_two} , state=lowerCamelCase__ ) assert result == 5 self.assertDictEqual(lowerCamelCase__ , {'''x''': 3, '''y''': 5} ) # Won't work without the tool with CaptureStdout() as out: _lowerCamelCase = evaluate(lowerCamelCase__ , {} , state=lowerCamelCase__ ) assert result is None assert "tried to execute add_two" in out.out def snake_case__ ( self ): _lowerCamelCase = '''x = 3''' _lowerCamelCase = {} _lowerCamelCase = evaluate(lowerCamelCase__ , {} , state=lowerCamelCase__ ) assert result == 3 self.assertDictEqual(lowerCamelCase__ , {'''x''': 3} ) def snake_case__ ( self ): _lowerCamelCase = '''test_dict = {\'x\': x, \'y\': add_two(x)}''' _lowerCamelCase = {'''x''': 3} _lowerCamelCase = evaluate(lowerCamelCase__ , {'''add_two''': add_two} , state=lowerCamelCase__ ) self.assertDictEqual(lowerCamelCase__ , {'''x''': 3, '''y''': 5} ) self.assertDictEqual(lowerCamelCase__ , {'''x''': 3, '''test_dict''': {'''x''': 3, '''y''': 5}} ) def snake_case__ ( self ): _lowerCamelCase = '''x = 3\ny = 5''' _lowerCamelCase = {} _lowerCamelCase = evaluate(lowerCamelCase__ , {} , state=lowerCamelCase__ ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(lowerCamelCase__ , {'''x''': 3, '''y''': 5} ) def snake_case__ ( self ): _lowerCamelCase = '''text = f\'This is x: {x}.\'''' _lowerCamelCase = {'''x''': 3} _lowerCamelCase = evaluate(lowerCamelCase__ , {} , state=lowerCamelCase__ ) # evaluate returns the value of the last assignment. assert result == "This is x: 3." self.assertDictEqual(lowerCamelCase__ , {'''x''': 3, '''text''': '''This is x: 3.'''} ) def snake_case__ ( self ): _lowerCamelCase = '''if x <= 3:\n y = 2\nelse:\n y = 5''' _lowerCamelCase = {'''x''': 3} _lowerCamelCase = evaluate(lowerCamelCase__ , {} , state=lowerCamelCase__ ) # evaluate returns the value of the last assignment. assert result == 2 self.assertDictEqual(lowerCamelCase__ , {'''x''': 3, '''y''': 2} ) _lowerCamelCase = {'''x''': 8} _lowerCamelCase = evaluate(lowerCamelCase__ , {} , state=lowerCamelCase__ ) # evaluate returns the value of the last assignment. assert result == 5 self.assertDictEqual(lowerCamelCase__ , {'''x''': 8, '''y''': 5} ) def snake_case__ ( self ): _lowerCamelCase = '''test_list = [x, add_two(x)]''' _lowerCamelCase = {'''x''': 3} _lowerCamelCase = evaluate(lowerCamelCase__ , {'''add_two''': add_two} , state=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , [3, 5] ) self.assertDictEqual(lowerCamelCase__ , {'''x''': 3, '''test_list''': [3, 5]} ) def snake_case__ ( self ): _lowerCamelCase = '''y = x''' _lowerCamelCase = {'''x''': 3} _lowerCamelCase = evaluate(lowerCamelCase__ , {} , state=lowerCamelCase__ ) assert result == 3 self.assertDictEqual(lowerCamelCase__ , {'''x''': 3, '''y''': 3} ) def snake_case__ ( self ): _lowerCamelCase = '''test_list = [x, add_two(x)]\ntest_list[1]''' _lowerCamelCase = {'''x''': 3} _lowerCamelCase = evaluate(lowerCamelCase__ , {'''add_two''': add_two} , state=lowerCamelCase__ ) assert result == 5 self.assertDictEqual(lowerCamelCase__ , {'''x''': 3, '''test_list''': [3, 5]} ) _lowerCamelCase = '''test_dict = {\'x\': x, \'y\': add_two(x)}\ntest_dict[\'y\']''' _lowerCamelCase = {'''x''': 3} _lowerCamelCase = evaluate(lowerCamelCase__ , {'''add_two''': add_two} , state=lowerCamelCase__ ) assert result == 5 self.assertDictEqual(lowerCamelCase__ , {'''x''': 3, '''test_dict''': {'''x''': 3, '''y''': 5}} ) def snake_case__ ( self ): _lowerCamelCase = '''x = 0\nfor i in range(3):\n x = i''' _lowerCamelCase = {} _lowerCamelCase = evaluate(lowerCamelCase__ , {'''range''': range} , state=lowerCamelCase__ ) assert result == 2 self.assertDictEqual(lowerCamelCase__ , {'''x''': 2, '''i''': 2} )
661
"""simple docstring""" import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } __SCREAMING_SNAKE_CASE : Tuple = { '''vocab_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json'''}, '''merges_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt'''}, '''tokenizer_config_file''': { '''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json''' }, } __SCREAMING_SNAKE_CASE : str = {'''facebook/blenderbot-3B''': 1_2_8} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def lowerCAmelCase_( ) -> str: _lowerCamelCase = ( list(range(ord('''!''' ) , ord('''~''' ) + 1 ) ) + list(range(ord('''¡''' ) , ord('''¬''' ) + 1 ) ) + list(range(ord('''®''' ) , ord('''ÿ''' ) + 1 ) ) ) _lowerCamelCase = bs[:] _lowerCamelCase = 0 for b in range(2**8 ): if b not in bs: bs.append(lowercase_ ) cs.append(2**8 + n ) n += 1 _lowerCamelCase = [chr(lowercase_ ) for n in cs] return dict(zip(lowercase_ , lowercase_ ) ) def lowerCAmelCase_( lowercase_ : str ) -> Dict: _lowerCamelCase = set() _lowerCamelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _lowerCamelCase = char return pairs class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Union[str, Any] = VOCAB_FILES_NAMES lowercase__ : int = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : Tuple = ['input_ids', 'attention_mask'] def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__="replace" , lowerCamelCase__="<s>" , lowerCamelCase__="</s>" , lowerCamelCase__="</s>" , lowerCamelCase__="<s>" , lowerCamelCase__="<unk>" , lowerCamelCase__="<pad>" , lowerCamelCase__="<mask>" , lowerCamelCase__=False , **lowerCamelCase__ , ): _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else bos_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else eos_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else sep_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else cls_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else unk_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else mask_token super().__init__( errors=lowerCamelCase__ , bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , add_prefix_space=lowerCamelCase__ , **lowerCamelCase__ , ) with open(lowerCamelCase__ , encoding='''utf-8''' ) as vocab_handle: _lowerCamelCase = json.load(lowerCamelCase__ ) _lowerCamelCase = {v: k for k, v in self.encoder.items()} _lowerCamelCase = errors # how to handle errors in decoding _lowerCamelCase = bytes_to_unicode() _lowerCamelCase = {v: k for k, v in self.byte_encoder.items()} with open(lowerCamelCase__ , encoding='''utf-8''' ) as merges_handle: _lowerCamelCase = merges_handle.read().split('''\n''' )[1:-1] _lowerCamelCase = [tuple(merge.split() ) for merge in bpe_merges] _lowerCamelCase = dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) _lowerCamelCase = {} _lowerCamelCase = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions _lowerCamelCase = re.compile(R'''\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+''' ) @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def snake_case__ ( self ): return len(self.encoder ) def snake_case__ ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def snake_case__ ( self , lowerCamelCase__ ): if token in self.cache: return self.cache[token] _lowerCamelCase = tuple(lowerCamelCase__ ) _lowerCamelCase = get_pairs(lowerCamelCase__ ) if not pairs: return token while True: _lowerCamelCase = min(lowerCamelCase__ , key=lambda lowerCamelCase__ : self.bpe_ranks.get(lowerCamelCase__ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break _lowerCamelCase , _lowerCamelCase = bigram _lowerCamelCase = [] _lowerCamelCase = 0 while i < len(lowerCamelCase__ ): try: _lowerCamelCase = word.index(lowerCamelCase__ , lowerCamelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _lowerCamelCase = j if word[i] == first and i < len(lowerCamelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _lowerCamelCase = tuple(lowerCamelCase__ ) _lowerCamelCase = new_word if len(lowerCamelCase__ ) == 1: break else: _lowerCamelCase = get_pairs(lowerCamelCase__ ) _lowerCamelCase = ''' '''.join(lowerCamelCase__ ) _lowerCamelCase = word return word def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] for token in re.findall(self.pat , lowerCamelCase__ ): _lowerCamelCase = ''''''.join( self.byte_encoder[b] for b in token.encode('''utf-8''' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCamelCase__ ).split(''' ''' ) ) return bpe_tokens def snake_case__ ( self , lowerCamelCase__ ): return self.encoder.get(lowerCamelCase__ , self.encoder.get(self.unk_token ) ) def snake_case__ ( self , lowerCamelCase__ ): return self.decoder.get(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = ''''''.join(lowerCamelCase__ ) _lowerCamelCase = bytearray([self.byte_decoder[c] for c in text] ).decode('''utf-8''' , errors=self.errors ) return text def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): if not os.path.isdir(lowerCamelCase__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(lowerCamelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCamelCase__ , ensure_ascii=lowerCamelCase__ ) + '''\n''' ) _lowerCamelCase = 0 with open(lowerCamelCase__ , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCamelCase__ : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ''' Please check that the tokenizer is not corrupted!''' ) _lowerCamelCase = token_index writer.write(''' '''.join(lowerCamelCase__ ) + '''\n''' ) index += 1 return vocab_file, merge_file def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase__ )) + [1] return [1] + ([0] * len(lowerCamelCase__ )) + [1, 1] + ([0] * len(lowerCamelCase__ )) + [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=False , **lowerCamelCase__ ): _lowerCamelCase = kwargs.pop('''add_prefix_space''' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCamelCase__ ) > 0 and not text[0].isspace()): _lowerCamelCase = ''' ''' + text return (text, kwargs) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): return token_ids_a + [self.eos_token_id] def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(''' ''' + text ) else: # Generated responses should contain them already. inputs.append(lowerCamelCase__ ) _lowerCamelCase = ''' '''.join(lowerCamelCase__ ) _lowerCamelCase = self.encode(lowerCamelCase__ ) if len(lowerCamelCase__ ) > self.model_max_length: _lowerCamelCase = input_ids[-self.model_max_length :] logger.warning(F"""Trimmed input from conversation as it was longer than {self.model_max_length} tokens.""" ) return input_ids
661
1
"""simple docstring""" from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_torch_available, is_torch_tpu_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_torch_available(): import torch if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) @dataclass class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Dict = [ 'no_inference', 'no_cuda', 'no_tpu', 'no_speed', 'no_memory', 'no_env_print', 'no_multi_process', ] def __init__( self , **lowerCamelCase__ ): for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: _lowerCamelCase = deprecated_arg[3:] setattr(self , lowerCamelCase__ , not kwargs.pop(lowerCamelCase__ ) ) logger.warning( F"""{deprecated_arg} is depreciated. Please use --no_{positive_arg} or""" F""" {positive_arg}={kwargs[positive_arg]}""" ) _lowerCamelCase = kwargs.pop('''torchscript''' , self.torchscript ) _lowerCamelCase = kwargs.pop('''torch_xla_tpu_print_metrics''' , self.torch_xla_tpu_print_metrics ) _lowerCamelCase = kwargs.pop('''fp16_opt_level''' , self.fpaa_opt_level ) super().__init__(**lowerCamelCase__ ) lowercase__ : bool = field(default=A__, metadata={'help': 'Trace the models using torchscript'} ) lowercase__ : bool = field(default=A__, metadata={'help': 'Print Xla/PyTorch tpu metrics'} ) lowercase__ : str = field( default='O1', metadata={ 'help': ( 'For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\']. ' 'See details at https://nvidia.github.io/apex/amp.html' ) }, ) @cached_property def snake_case__ ( self ): requires_backends(self , ['''torch'''] ) logger.info('''PyTorch: setting up devices''' ) if not self.cuda: _lowerCamelCase = torch.device('''cpu''' ) _lowerCamelCase = 0 elif is_torch_tpu_available(): _lowerCamelCase = xm.xla_device() _lowerCamelCase = 0 else: _lowerCamelCase = torch.device('''cuda''' if torch.cuda.is_available() else '''cpu''' ) _lowerCamelCase = torch.cuda.device_count() return device, n_gpu @property def snake_case__ ( self ): return is_torch_tpu_available() and self.tpu @property def snake_case__ ( self ): requires_backends(self , ['''torch'''] ) # TODO(PVP): currently only single GPU is supported return torch.cuda.current_device() @property def snake_case__ ( self ): requires_backends(self , ['''torch'''] ) return self._setup_devices[0] @property def snake_case__ ( self ): requires_backends(self , ['''torch'''] ) return self._setup_devices[1] @property def snake_case__ ( self ): return self.n_gpu > 0
661
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Dict = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Union[str, Any] = XLMRobertaTokenizer lowercase__ : Optional[int] = XLMRobertaTokenizerFast lowercase__ : List[str] = True lowercase__ : Union[str, Any] = True def snake_case__ ( self ): super().setUp() # We have a SentencePiece fixture for testing _lowerCamelCase = XLMRobertaTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case__ ( self ): _lowerCamelCase = '''<pad>''' _lowerCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(lowerCamelCase__ ) , 1_0_0_2 ) def snake_case__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_2 ) def snake_case__ ( self ): _lowerCamelCase = XLMRobertaTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) _lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(lowerCamelCase__ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) _lowerCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) _lowerCamelCase = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) _lowerCamelCase = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def snake_case__ ( self ): if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _lowerCamelCase = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-xlm-roberta''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _lowerCamelCase = self.rust_tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = self.tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) _lowerCamelCase = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(lowerCamelCase__ , lowerCamelCase__ ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(lowerCamelCase__ ) # Save tokenizer rust, legacy_format=True _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ , legacy_format=lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it save with the same files self.assertSequenceEqual(lowerCamelCase__ , lowerCamelCase__ ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) shutil.rmtree(lowerCamelCase__ ) # Save tokenizer rust, legacy_format=False _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ , legacy_format=lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) shutil.rmtree(lowerCamelCase__ ) @cached_property def snake_case__ ( self ): return XLMRobertaTokenizer.from_pretrained('''xlm-roberta-base''' ) def snake_case__ ( self ): with tempfile.NamedTemporaryFile() as f: shutil.copyfile(lowerCamelCase__ , f.name ) _lowerCamelCase = XLMRobertaTokenizer(f.name , keep_accents=lowerCamelCase__ ) _lowerCamelCase = pickle.dumps(lowerCamelCase__ ) pickle.loads(lowerCamelCase__ ) def snake_case__ ( self ): if not self.test_rust_tokenizer: return _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = '''I was born in 92000, and this is falsé.''' _lowerCamelCase = tokenizer.tokenize(lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = tokenizer.encode(lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): _lowerCamelCase = '''Hello World!''' _lowerCamelCase = [0, 3_5_3_7_8, 6_6_6_1, 3_8, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(lowerCamelCase__ , self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def snake_case__ ( self ): _lowerCamelCase = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) _lowerCamelCase = [ 0, 3_2_9_3, 8_3, 1_0, 4_5_5_2, 4_9_8_9, 7_9_8_6, 6_7_8, 1_0, 5_9_1_5, 1_1_1, 1_7_9_4_5_9, 1_2_4_8_5_0, 4, 6_0_4_4, 2_3_7, 1_2, 6, 5, 6, 4, 6_7_8_0, 7_0_5, 1_5, 1_3_8_8, 4_4, 3_7_8, 1_0_1_1_4, 7_1_1, 1_5_2, 2_0, 6, 5, 2_2_3_7_6, 6_4_2, 1_2_2_1, 1_5_1_9_0, 3_4_1_5_3, 4_5_0, 5_6_0_8, 9_5_9, 1_1_1_9, 5_7_7_0_2, 1_3_6, 1_8_6, 4_7, 1_0_9_8, 2_9_3_6_7, 4_7, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6_0_4_4, 2_3_7, 6_2_8_4, 5_0_9_0_1, 5_2_8, 3_1, 9_0, 3_4, 9_2_7, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(lowerCamelCase__ , self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def snake_case__ ( self ): # fmt: off _lowerCamelCase = {'''input_ids''': [[0, 1_1_0_6_2, 8_2_7_7_2, 7, 1_5, 8_2_7_7_2, 5_3_8, 5_1_5_2_9, 2_3_7, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 2_1_5_1_7_5, 1_3_1_4, 1_3_6, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 5_6_3_5_9, 4_2, 1_2_2_0_0_9, 9, 1_6_4_6_6, 1_6, 8_7_3_4_4, 4_5_3_7, 9, 4_7_1_7, 7_8_3_8_1, 6, 1_5_9_9_5_8, 7, 1_5, 2_4_4_8_0, 6_1_8, 4, 5_2_7, 2_2_6_9_3, 5_4_2_8, 4, 2_7_7_7, 2_4_4_8_0, 9_8_7_4, 4, 4_3_5_2_3, 5_9_4, 4, 8_0_3, 1_8_3_9_2, 3_3_1_8_9, 1_8, 4, 4_3_5_2_3, 2_4_4_4_7, 1_2_3_9_9, 1_0_0, 2_4_9_5_5, 8_3_6_5_8, 9_6_2_6, 1_4_4_0_5_7, 1_5, 8_3_9, 2_2_3_3_5, 1_6, 1_3_6, 2_4_9_5_5, 8_3_6_5_8, 8_3_4_7_9, 1_5, 3_9_1_0_2, 7_2_4, 1_6, 6_7_8, 6_4_5, 2_7_8_9, 1_3_2_8, 4_5_8_9, 4_2, 1_2_2_0_0_9, 1_1_5_7_7_4, 2_3, 8_0_5, 1_3_2_8, 4_6_8_7_6, 7, 1_3_6, 5_3_8_9_4, 1_9_4_0, 4_2_2_2_7, 4_1_1_5_9, 1_7_7_2_1, 8_2_3, 4_2_5, 4, 2_7_5_1_2, 9_8_7_2_2, 2_0_6, 1_3_6, 5_5_3_1, 4_9_7_0, 9_1_9, 1_7_3_3_6, 5, 2], [0, 2_0_0_8_0, 6_1_8, 8_3, 8_2_7_7_5, 4_7, 4_7_9, 9, 1_5_1_7, 7_3, 5_3_8_9_4, 3_3_3, 8_0_5_8_1, 1_1_0_1_1_7, 1_8_8_1_1, 5_2_5_6, 1_2_9_5, 5_1, 1_5_2_5_2_6, 2_9_7, 7_9_8_6, 3_9_0, 1_2_4_4_1_6, 5_3_8, 3_5_4_3_1, 2_1_4, 9_8, 1_5_0_4_4, 2_5_7_3_7, 1_3_6, 7_1_0_8, 4_3_7_0_1, 2_3, 7_5_6, 1_3_5_3_5_5, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_8_1, 6_3_7_7_3, 1_1_9_4_5_5, 6, 1_4_7_7_9_7, 8_8_2_0_3, 7, 6_4_5, 7_0, 2_1, 3_2_8_5, 1_0_2_6_9, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ , model_name='''xlm-roberta-base''' , revision='''d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3''' , )
661
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> Tuple: assert x is not None assert y is not None _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = len(lowercase_ ) # declaring the array for storing the dp values _lowerCamelCase = [[0] * (n + 1) for _ in range(m + 1 )] # noqa: E741 for i in range(1 , m + 1 ): for j in range(1 , n + 1 ): _lowerCamelCase = 1 if x[i - 1] == y[j - 1] else 0 _lowerCamelCase = max(l[i - 1][j] , l[i][j - 1] , l[i - 1][j - 1] + match ) _lowerCamelCase = '''''' _lowerCamelCase , _lowerCamelCase = m, n while i > 0 and j > 0: _lowerCamelCase = 1 if x[i - 1] == y[j - 1] else 0 if l[i][j] == l[i - 1][j - 1] + match: if match == 1: _lowerCamelCase = x[i - 1] + seq i -= 1 j -= 1 elif l[i][j] == l[i - 1][j]: i -= 1 else: j -= 1 return l[m][n], seq if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Tuple = '''AGGTAB''' __SCREAMING_SNAKE_CASE : List[str] = '''GXTXAYB''' __SCREAMING_SNAKE_CASE : Any = 4 __SCREAMING_SNAKE_CASE : Optional[Any] = '''GTAB''' __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : List[str] = longest_common_subsequence(a, b) print('''len =''', ln, ''', sub-sequence =''', subseq) import doctest doctest.testmod()
661
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int = 10 ) -> str: if not isinstance(lowercase_ , lowercase_ ) or n < 0: raise ValueError('''Invalid input''' ) _lowerCamelCase = 10**n _lowerCamelCase = 2_84_33 * (pow(2 , 7_83_04_57 , lowercase_ )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(F"""{solution(1_0) = }""")
661
1
"""simple docstring""" from ...processing_utils import ProcessorMixin class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[Any] = ['image_processor', 'feature_extractor'] lowercase__ : str = 'TvltImageProcessor' lowercase__ : List[str] = 'TvltFeatureExtractor' def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): super().__init__(image_processor=lowerCamelCase__ , feature_extractor=lowerCamelCase__ ) _lowerCamelCase = image_processor _lowerCamelCase = feature_extractor def __call__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=False , lowerCamelCase__=False , *lowerCamelCase__ , **lowerCamelCase__ , ): if images is None and audio is None: raise ValueError('''You need to specify either an `images` or `audio` input to process.''' ) _lowerCamelCase = None if images is not None: _lowerCamelCase = self.image_processor(lowerCamelCase__ , mask_pixel=lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) if images_mixed is not None: _lowerCamelCase = self.image_processor(lowerCamelCase__ , is_mixed=lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) if audio is not None: _lowerCamelCase = self.feature_extractor( lowerCamelCase__ , *lowerCamelCase__ , sampling_rate=lowerCamelCase__ , mask_audio=lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = {} if audio is not None: output_dict.update(lowerCamelCase__ ) if images is not None: output_dict.update(lowerCamelCase__ ) if images_mixed_dict is not None: output_dict.update(lowerCamelCase__ ) return output_dict @property def snake_case__ ( self ): _lowerCamelCase = self.image_processor.model_input_names _lowerCamelCase = self.feature_extractor.model_input_names return list(dict.fromkeys(image_processor_input_names + feature_extractor_input_names ) )
661
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __SCREAMING_SNAKE_CASE : Optional[int] = { '''configuration_informer''': [ '''INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ '''INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InformerForPrediction''', '''InformerModel''', '''InformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
661
1
"""simple docstring""" from __future__ import annotations __SCREAMING_SNAKE_CASE : Tuple = tuple[int, int, int] __SCREAMING_SNAKE_CASE : List[str] = tuple[str, str, str] # used alphabet -------------------------- # from string.ascii_uppercase __SCREAMING_SNAKE_CASE : List[Any] = '''ABCDEFGHIJKLMNOPQRSTUVWXYZ''' # -------------------------- default selection -------------------------- # rotors -------------------------- __SCREAMING_SNAKE_CASE : int = '''EGZWVONAHDCLFQMSIPJBYUKXTR''' __SCREAMING_SNAKE_CASE : List[Any] = '''FOBHMDKEXQNRAULPGSJVTYICZW''' __SCREAMING_SNAKE_CASE : List[str] = '''ZJXESIUQLHAVRMDOYGTNFWPBKC''' # reflector -------------------------- __SCREAMING_SNAKE_CASE : str = { '''A''': '''N''', '''N''': '''A''', '''B''': '''O''', '''O''': '''B''', '''C''': '''P''', '''P''': '''C''', '''D''': '''Q''', '''Q''': '''D''', '''E''': '''R''', '''R''': '''E''', '''F''': '''S''', '''S''': '''F''', '''G''': '''T''', '''T''': '''G''', '''H''': '''U''', '''U''': '''H''', '''I''': '''V''', '''V''': '''I''', '''J''': '''W''', '''W''': '''J''', '''K''': '''X''', '''X''': '''K''', '''L''': '''Y''', '''Y''': '''L''', '''M''': '''Z''', '''Z''': '''M''', } # -------------------------- extra rotors -------------------------- __SCREAMING_SNAKE_CASE : Any = '''RMDJXFUWGISLHVTCQNKYPBEZOA''' __SCREAMING_SNAKE_CASE : str = '''SGLCPQWZHKXAREONTFBVIYJUDM''' __SCREAMING_SNAKE_CASE : List[Any] = '''HVSICLTYKQUBXDWAJZOMFGPREN''' __SCREAMING_SNAKE_CASE : int = '''RZWQHFMVDBKICJLNTUXAGYPSOE''' __SCREAMING_SNAKE_CASE : Any = '''LFKIJODBEGAMQPXVUHYSTCZRWN''' __SCREAMING_SNAKE_CASE : Any = '''KOAEGVDHXPQZMLFTYWJNBRCIUS''' def lowerCAmelCase_( lowercase_ : RotorPositionT , lowercase_ : RotorSelectionT , lowercase_ : str ) -> tuple[RotorPositionT, RotorSelectionT, dict[str, str]]: # Checks if there are 3 unique rotors if (unique_rotsel := len(set(lowercase_ ) )) < 3: _lowerCamelCase = F"""Please use 3 unique rotors (not {unique_rotsel})""" raise Exception(lowercase_ ) # Checks if rotor positions are valid _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = rotpos if not 0 < rotorposa <= len(lowercase_ ): _lowerCamelCase = F"""First rotor position is not within range of 1..26 ({rotorposa}""" raise ValueError(lowercase_ ) if not 0 < rotorposa <= len(lowercase_ ): _lowerCamelCase = F"""Second rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(lowercase_ ) if not 0 < rotorposa <= len(lowercase_ ): _lowerCamelCase = F"""Third rotor position is not within range of 1..26 ({rotorposa})""" raise ValueError(lowercase_ ) # Validates string and returns dict _lowerCamelCase = _plugboard(lowercase_ ) return rotpos, rotsel, pbdict def lowerCAmelCase_( lowercase_ : str ) -> dict[str, str]: # tests the input string if it # a) is type string # b) has even length (so pairs can be made) if not isinstance(lowercase_ , lowercase_ ): _lowerCamelCase = F"""Plugboard setting isn't type string ({type(lowercase_ )})""" raise TypeError(lowercase_ ) elif len(lowercase_ ) % 2 != 0: _lowerCamelCase = F"""Odd number of symbols ({len(lowercase_ )})""" raise Exception(lowercase_ ) elif pbstring == "": return {} pbstring.replace(''' ''' , '''''' ) # Checks if all characters are unique _lowerCamelCase = set() for i in pbstring: if i not in abc: _lowerCamelCase = F"""'{i}' not in list of symbols""" raise Exception(lowercase_ ) elif i in tmppbl: _lowerCamelCase = F"""Duplicate symbol ({i})""" raise Exception(lowercase_ ) else: tmppbl.add(lowercase_ ) del tmppbl # Created the dictionary _lowerCamelCase = {} for j in range(0 , len(lowercase_ ) - 1 , 2 ): _lowerCamelCase = pbstring[j + 1] _lowerCamelCase = pbstring[j] return pb def lowerCAmelCase_( lowercase_ : str , lowercase_ : RotorPositionT , lowercase_ : RotorSelectionT = (rotora, rotora, rotora) , lowercase_ : str = "" , ) -> str: _lowerCamelCase = text.upper() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = _validator( lowercase_ , lowercase_ , plugb.upper() ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = rotor_position _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = rotor_selection rotorposa -= 1 rotorposa -= 1 rotorposa -= 1 _lowerCamelCase = [] # encryption/decryption process -------------------------- for symbol in text: if symbol in abc: # 1st plugboard -------------------------- if symbol in plugboard: _lowerCamelCase = plugboard[symbol] # rotor ra -------------------------- _lowerCamelCase = abc.index(lowercase_ ) + rotorposa _lowerCamelCase = rotora[index % len(lowercase_ )] # rotor rb -------------------------- _lowerCamelCase = abc.index(lowercase_ ) + rotorposa _lowerCamelCase = rotora[index % len(lowercase_ )] # rotor rc -------------------------- _lowerCamelCase = abc.index(lowercase_ ) + rotorposa _lowerCamelCase = rotora[index % len(lowercase_ )] # reflector -------------------------- # this is the reason you don't need another machine to decipher _lowerCamelCase = reflector[symbol] # 2nd rotors _lowerCamelCase = abc[rotora.index(lowercase_ ) - rotorposa] _lowerCamelCase = abc[rotora.index(lowercase_ ) - rotorposa] _lowerCamelCase = abc[rotora.index(lowercase_ ) - rotorposa] # 2nd plugboard if symbol in plugboard: _lowerCamelCase = plugboard[symbol] # moves/resets rotor positions rotorposa += 1 if rotorposa >= len(lowercase_ ): _lowerCamelCase = 0 rotorposa += 1 if rotorposa >= len(lowercase_ ): _lowerCamelCase = 0 rotorposa += 1 if rotorposa >= len(lowercase_ ): _lowerCamelCase = 0 # else: # pass # Error could be also raised # raise ValueError( # 'Invalid symbol('+repr(symbol)+')') result.append(lowercase_ ) return "".join(lowercase_ ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[str] = '''This is my Python script that emulates the Enigma machine from WWII.''' __SCREAMING_SNAKE_CASE : Tuple = (1, 1, 1) __SCREAMING_SNAKE_CASE : Optional[int] = '''pictures''' __SCREAMING_SNAKE_CASE : Optional[int] = (rotora, rotora, rotora) __SCREAMING_SNAKE_CASE : Tuple = enigma(message, rotor_pos, rotor_sel, pb) print('''Encrypted message:''', en) print('''Decrypted message:''', enigma(en, rotor_pos, rotor_sel, pb))
661
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> float: def get_matched_characters(lowercase_ : str , lowercase_ : str ) -> str: _lowerCamelCase = [] _lowerCamelCase = min(len(_stra ) , len(_stra ) ) // 2 for i, l in enumerate(_stra ): _lowerCamelCase = int(max(0 , i - limit ) ) _lowerCamelCase = int(min(i + limit + 1 , len(_stra ) ) ) if l in _stra[left:right]: matched.append(lowercase_ ) _lowerCamelCase = F"""{_stra[0:_stra.index(lowercase_ )]} {_stra[_stra.index(lowercase_ ) + 1:]}""" return "".join(lowercase_ ) # matching characters _lowerCamelCase = get_matched_characters(lowercase_ , lowercase_ ) _lowerCamelCase = get_matched_characters(lowercase_ , lowercase_ ) _lowerCamelCase = len(lowercase_ ) # transposition _lowerCamelCase = ( len([(ca, ca) for ca, ca in zip(lowercase_ , lowercase_ ) if ca != ca] ) // 2 ) if not match_count: _lowerCamelCase = 0.0 else: _lowerCamelCase = ( 1 / 3 * ( match_count / len(lowercase_ ) + match_count / len(lowercase_ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters _lowerCamelCase = 0 for ca, ca in zip(stra[:4] , stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('''hello''', '''world'''))
661
1
"""simple docstring""" import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin class lowerCamelCase_( unittest.TestCase, A__ ): '''simple docstring''' def snake_case__ ( self ): _lowerCamelCase = load_tool('''text-classification''' ) self.tool.setup() _lowerCamelCase = load_tool('''text-classification''' , remote=lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.tool('''That\'s quite cool''' , ['''positive''', '''negative'''] ) self.assertEqual(lowerCamelCase__ , '''positive''' ) def snake_case__ ( self ): _lowerCamelCase = self.remote_tool('''That\'s quite cool''' , ['''positive''', '''negative'''] ) self.assertEqual(lowerCamelCase__ , '''positive''' ) def snake_case__ ( self ): _lowerCamelCase = self.tool(text='''That\'s quite cool''' , labels=['''positive''', '''negative'''] ) self.assertEqual(lowerCamelCase__ , '''positive''' ) def snake_case__ ( self ): _lowerCamelCase = self.remote_tool(text='''That\'s quite cool''' , labels=['''positive''', '''negative'''] ) self.assertEqual(lowerCamelCase__ , '''positive''' )
661
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowerCamelCase_( A__, A__, A__ ): '''simple docstring''' lowercase__ : List[Any] = [r'h\.\d+\.attn\.bias', r'h\.\d+\.attn\.masked_bias'] @register_to_config def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = 5_0_2_5_7 , lowerCamelCase__ = 1_0_2_4 , lowerCamelCase__ = 7_6_8 , lowerCamelCase__ = 1_2 , lowerCamelCase__ = 1_2 , lowerCamelCase__ = None , lowerCamelCase__ = "gelu_new" , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 1e-5 , lowerCamelCase__ = 0.0_2 , lowerCamelCase__ = True , lowerCamelCase__ = True , lowerCamelCase__ = False , lowerCamelCase__ = False , ): super().__init__() _lowerCamelCase = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F"""`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and""" F""" `n_embd`: {n_embd} are not equal.""" ) _lowerCamelCase = prefix_inner_dim _lowerCamelCase = prefix_hidden_dim _lowerCamelCase = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCamelCase = ( nn.Linear(self.prefix_hidden_dim , lowerCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCamelCase = GPTaConfig( vocab_size=lowerCamelCase__ , n_positions=lowerCamelCase__ , n_embd=lowerCamelCase__ , n_layer=lowerCamelCase__ , n_head=lowerCamelCase__ , n_inner=lowerCamelCase__ , activation_function=lowerCamelCase__ , resid_pdrop=lowerCamelCase__ , embd_pdrop=lowerCamelCase__ , attn_pdrop=lowerCamelCase__ , layer_norm_epsilon=lowerCamelCase__ , initializer_range=lowerCamelCase__ , scale_attn_weights=lowerCamelCase__ , use_cache=lowerCamelCase__ , scale_attn_by_inverse_layer_idx=lowerCamelCase__ , reorder_and_upcast_attn=lowerCamelCase__ , ) _lowerCamelCase = GPTaLMHeadModel(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None , ): _lowerCamelCase = self.transformer.transformer.wte(lowerCamelCase__ ) _lowerCamelCase = self.encode_prefix(lowerCamelCase__ ) _lowerCamelCase = self.decode_prefix(lowerCamelCase__ ) _lowerCamelCase = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: _lowerCamelCase = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) _lowerCamelCase = torch.cat((dummy_token, input_ids) , dim=1 ) _lowerCamelCase = self.transformer(inputs_embeds=lowerCamelCase__ , labels=lowerCamelCase__ , attention_mask=lowerCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): return torch.zeros(lowerCamelCase__ , self.prefix_length , dtype=torch.intaa , device=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): return self.encode_prefix(lowerCamelCase__ ) @torch.no_grad() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = torch.split(lowerCamelCase__ , 1 , dim=0 ) _lowerCamelCase = [] _lowerCamelCase = [] for feature in features: _lowerCamelCase = self.decode_prefix(feature.to(lowerCamelCase__ ) ) # back to the clip feature # Only support beam search for now _lowerCamelCase , _lowerCamelCase = self.generate_beam( input_embeds=lowerCamelCase__ , device=lowerCamelCase__ , eos_token_id=lowerCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) _lowerCamelCase = torch.stack(lowerCamelCase__ ) _lowerCamelCase = torch.stack(lowerCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def snake_case__ ( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__ = 5 , lowerCamelCase__ = 6_7 , lowerCamelCase__ = 1.0 , lowerCamelCase__ = None , ): _lowerCamelCase = eos_token_id _lowerCamelCase = None _lowerCamelCase = None _lowerCamelCase = torch.ones(lowerCamelCase__ , device=lowerCamelCase__ , dtype=torch.int ) _lowerCamelCase = torch.zeros(lowerCamelCase__ , device=lowerCamelCase__ , dtype=torch.bool ) if input_embeds is not None: _lowerCamelCase = input_embeds else: _lowerCamelCase = self.transformer.transformer.wte(lowerCamelCase__ ) for i in range(lowerCamelCase__ ): _lowerCamelCase = self.transformer(inputs_embeds=lowerCamelCase__ ) _lowerCamelCase = outputs.logits _lowerCamelCase = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) _lowerCamelCase = logits.softmax(-1 ).log() if scores is None: _lowerCamelCase , _lowerCamelCase = logits.topk(lowerCamelCase__ , -1 ) _lowerCamelCase = generated.expand(lowerCamelCase__ , *generated.shape[1:] ) _lowerCamelCase , _lowerCamelCase = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: _lowerCamelCase = next_tokens else: _lowerCamelCase = tokens.expand(lowerCamelCase__ , *tokens.shape[1:] ) _lowerCamelCase = torch.cat((tokens, next_tokens) , dim=1 ) else: _lowerCamelCase = -float(np.inf ) _lowerCamelCase = 0 _lowerCamelCase = scores[:, None] + logits seq_lengths[~is_stopped] += 1 _lowerCamelCase = scores_sum / seq_lengths[:, None] _lowerCamelCase , _lowerCamelCase = scores_sum_average.view(-1 ).topk(lowerCamelCase__ , -1 ) _lowerCamelCase = next_tokens // scores_sum.shape[1] _lowerCamelCase = seq_lengths[next_tokens_source] _lowerCamelCase = next_tokens % scores_sum.shape[1] _lowerCamelCase = next_tokens.unsqueeze(1 ) _lowerCamelCase = tokens[next_tokens_source] _lowerCamelCase = torch.cat((tokens, next_tokens) , dim=1 ) _lowerCamelCase = generated[next_tokens_source] _lowerCamelCase = scores_sum_average * seq_lengths _lowerCamelCase = is_stopped[next_tokens_source] _lowerCamelCase = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) _lowerCamelCase = torch.cat((generated, next_token_embed) , dim=1 ) _lowerCamelCase = is_stopped + next_tokens.eq(lowerCamelCase__ ).squeeze() if is_stopped.all(): break _lowerCamelCase = scores / seq_lengths _lowerCamelCase = scores.argsort(descending=lowerCamelCase__ ) # tokens tensors are already padded to max_seq_length _lowerCamelCase = [tokens[i] for i in order] _lowerCamelCase = torch.stack(lowerCamelCase__ , dim=0 ) _lowerCamelCase = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
661
1
"""simple docstring""" from dataclasses import dataclass from typing import Dict, Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .attention_processor import AttentionProcessor, AttnProcessor from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, DiagonalGaussianDistribution, Encoder @dataclass class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : "DiagonalGaussianDistribution" class lowerCamelCase_( A__, A__ ): '''simple docstring''' lowercase__ : Tuple = True @register_to_config def __init__( self , lowerCamelCase__ = 3 , lowerCamelCase__ = 3 , lowerCamelCase__ = ("DownEncoderBlock2D",) , lowerCamelCase__ = ("UpDecoderBlock2D",) , lowerCamelCase__ = (6_4,) , lowerCamelCase__ = 1 , lowerCamelCase__ = "silu" , lowerCamelCase__ = 4 , lowerCamelCase__ = 3_2 , lowerCamelCase__ = 3_2 , lowerCamelCase__ = 0.1_8_2_1_5 , ): super().__init__() # pass init params to Encoder _lowerCamelCase = Encoder( in_channels=lowerCamelCase__ , out_channels=lowerCamelCase__ , down_block_types=lowerCamelCase__ , block_out_channels=lowerCamelCase__ , layers_per_block=lowerCamelCase__ , act_fn=lowerCamelCase__ , norm_num_groups=lowerCamelCase__ , double_z=lowerCamelCase__ , ) # pass init params to Decoder _lowerCamelCase = Decoder( in_channels=lowerCamelCase__ , out_channels=lowerCamelCase__ , up_block_types=lowerCamelCase__ , block_out_channels=lowerCamelCase__ , layers_per_block=lowerCamelCase__ , norm_num_groups=lowerCamelCase__ , act_fn=lowerCamelCase__ , ) _lowerCamelCase = nn.Convad(2 * latent_channels , 2 * latent_channels , 1 ) _lowerCamelCase = nn.Convad(lowerCamelCase__ , lowerCamelCase__ , 1 ) _lowerCamelCase = False _lowerCamelCase = False # only relevant if vae tiling is enabled _lowerCamelCase = self.config.sample_size _lowerCamelCase = ( self.config.sample_size[0] if isinstance(self.config.sample_size , (list, tuple) ) else self.config.sample_size ) _lowerCamelCase = int(sample_size / (2 ** (len(self.config.block_out_channels ) - 1)) ) _lowerCamelCase = 0.2_5 def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=False ): if isinstance(lowerCamelCase__ , (Encoder, Decoder) ): _lowerCamelCase = value def snake_case__ ( self , lowerCamelCase__ = True ): _lowerCamelCase = use_tiling def snake_case__ ( self ): self.enable_tiling(lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = True def snake_case__ ( self ): _lowerCamelCase = False @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def snake_case__ ( self ): _lowerCamelCase = {} def fn_recursive_add_processors(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if hasattr(lowerCamelCase__ , '''set_processor''' ): _lowerCamelCase = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(F"""{name}.{sub_name}""" , lowerCamelCase__ , lowerCamelCase__ ) return processors for name, module in self.named_children(): fn_recursive_add_processors(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) return processors def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = len(self.attn_processors.keys() ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) and len(lowerCamelCase__ ) != count: raise ValueError( F"""A dict of processors was passed, but the number of processors {len(lowerCamelCase__ )} does not match the""" F""" number of attention layers: {count}. Please make sure to pass {count} processor classes.""" ) def fn_recursive_attn_processor(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if hasattr(lowerCamelCase__ , '''set_processor''' ): if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): module.set_processor(lowerCamelCase__ ) else: module.set_processor(processor.pop(F"""{name}.processor""" ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(F"""{name}.{sub_name}""" , lowerCamelCase__ , lowerCamelCase__ ) for name, module in self.named_children(): fn_recursive_attn_processor(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def snake_case__ ( self ): self.set_attn_processor(AttnProcessor() ) @apply_forward_hook def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = True ): if self.use_tiling and (x.shape[-1] > self.tile_sample_min_size or x.shape[-2] > self.tile_sample_min_size): return self.tiled_encode(lowerCamelCase__ , return_dict=lowerCamelCase__ ) if self.use_slicing and x.shape[0] > 1: _lowerCamelCase = [self.encoder(lowerCamelCase__ ) for x_slice in x.split(1 )] _lowerCamelCase = torch.cat(lowerCamelCase__ ) else: _lowerCamelCase = self.encoder(lowerCamelCase__ ) _lowerCamelCase = self.quant_conv(lowerCamelCase__ ) _lowerCamelCase = DiagonalGaussianDistribution(lowerCamelCase__ ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = True ): if self.use_tiling and (z.shape[-1] > self.tile_latent_min_size or z.shape[-2] > self.tile_latent_min_size): return self.tiled_decode(lowerCamelCase__ , return_dict=lowerCamelCase__ ) _lowerCamelCase = self.post_quant_conv(lowerCamelCase__ ) _lowerCamelCase = self.decoder(lowerCamelCase__ ) if not return_dict: return (dec,) return DecoderOutput(sample=lowerCamelCase__ ) @apply_forward_hook def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = True ): if self.use_slicing and z.shape[0] > 1: _lowerCamelCase = [self._decode(lowerCamelCase__ ).sample for z_slice in z.split(1 )] _lowerCamelCase = torch.cat(lowerCamelCase__ ) else: _lowerCamelCase = self._decode(lowerCamelCase__ ).sample if not return_dict: return (decoded,) return DecoderOutput(sample=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = min(a.shape[2] , b.shape[2] , lowerCamelCase__ ) for y in range(lowerCamelCase__ ): _lowerCamelCase = a[:, :, -blend_extent + y, :] * (1 - y / blend_extent) + b[:, :, y, :] * (y / blend_extent) return b def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = min(a.shape[3] , b.shape[3] , lowerCamelCase__ ) for x in range(lowerCamelCase__ ): _lowerCamelCase = a[:, :, :, -blend_extent + x] * (1 - x / blend_extent) + b[:, :, :, x] * (x / blend_extent) return b def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = True ): _lowerCamelCase = int(self.tile_sample_min_size * (1 - self.tile_overlap_factor) ) _lowerCamelCase = int(self.tile_latent_min_size * self.tile_overlap_factor ) _lowerCamelCase = self.tile_latent_min_size - blend_extent # Split the image into 512x512 tiles and encode them separately. _lowerCamelCase = [] for i in range(0 , x.shape[2] , lowerCamelCase__ ): _lowerCamelCase = [] for j in range(0 , x.shape[3] , lowerCamelCase__ ): _lowerCamelCase = x[:, :, i : i + self.tile_sample_min_size, j : j + self.tile_sample_min_size] _lowerCamelCase = self.encoder(lowerCamelCase__ ) _lowerCamelCase = self.quant_conv(lowerCamelCase__ ) row.append(lowerCamelCase__ ) rows.append(lowerCamelCase__ ) _lowerCamelCase = [] for i, row in enumerate(lowerCamelCase__ ): _lowerCamelCase = [] for j, tile in enumerate(lowerCamelCase__ ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: _lowerCamelCase = self.blend_v(rows[i - 1][j] , lowerCamelCase__ , lowerCamelCase__ ) if j > 0: _lowerCamelCase = self.blend_h(row[j - 1] , lowerCamelCase__ , lowerCamelCase__ ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(lowerCamelCase__ , dim=3 ) ) _lowerCamelCase = torch.cat(lowerCamelCase__ , dim=2 ) _lowerCamelCase = DiagonalGaussianDistribution(lowerCamelCase__ ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = True ): _lowerCamelCase = int(self.tile_latent_min_size * (1 - self.tile_overlap_factor) ) _lowerCamelCase = int(self.tile_sample_min_size * self.tile_overlap_factor ) _lowerCamelCase = self.tile_sample_min_size - blend_extent # Split z into overlapping 64x64 tiles and decode them separately. # The tiles have an overlap to avoid seams between tiles. _lowerCamelCase = [] for i in range(0 , z.shape[2] , lowerCamelCase__ ): _lowerCamelCase = [] for j in range(0 , z.shape[3] , lowerCamelCase__ ): _lowerCamelCase = z[:, :, i : i + self.tile_latent_min_size, j : j + self.tile_latent_min_size] _lowerCamelCase = self.post_quant_conv(lowerCamelCase__ ) _lowerCamelCase = self.decoder(lowerCamelCase__ ) row.append(lowerCamelCase__ ) rows.append(lowerCamelCase__ ) _lowerCamelCase = [] for i, row in enumerate(lowerCamelCase__ ): _lowerCamelCase = [] for j, tile in enumerate(lowerCamelCase__ ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: _lowerCamelCase = self.blend_v(rows[i - 1][j] , lowerCamelCase__ , lowerCamelCase__ ) if j > 0: _lowerCamelCase = self.blend_h(row[j - 1] , lowerCamelCase__ , lowerCamelCase__ ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(lowerCamelCase__ , dim=3 ) ) _lowerCamelCase = torch.cat(lowerCamelCase__ , dim=2 ) if not return_dict: return (dec,) return DecoderOutput(sample=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = False , lowerCamelCase__ = True , lowerCamelCase__ = None , ): _lowerCamelCase = sample _lowerCamelCase = self.encode(lowerCamelCase__ ).latent_dist if sample_posterior: _lowerCamelCase = posterior.sample(generator=lowerCamelCase__ ) else: _lowerCamelCase = posterior.mode() _lowerCamelCase = self.decode(lowerCamelCase__ ).sample if not return_dict: return (dec,) return DecoderOutput(sample=lowerCamelCase__ )
661
"""simple docstring""" import re import string from collections import Counter import sacrebleu import sacremoses from packaging import version import datasets __SCREAMING_SNAKE_CASE : Optional[int] = ''' @inproceedings{xu-etal-2016-optimizing, title = {Optimizing Statistical Machine Translation for Text Simplification}, authors={Xu, Wei and Napoles, Courtney and Pavlick, Ellie and Chen, Quanze and Callison-Burch, Chris}, journal = {Transactions of the Association for Computational Linguistics}, volume = {4}, year={2016}, url = {https://www.aclweb.org/anthology/Q16-1029}, pages = {401--415 }, @inproceedings{post-2018-call, title = "A Call for Clarity in Reporting {BLEU} Scores", author = "Post, Matt", booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers", month = oct, year = "2018", address = "Belgium, Brussels", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W18-6319", pages = "186--191", } ''' __SCREAMING_SNAKE_CASE : List[str] = '''\ WIKI_SPLIT is the combination of three metrics SARI, EXACT and SACREBLEU It can be used to evaluate the quality of machine-generated texts. ''' __SCREAMING_SNAKE_CASE : Optional[Any] = ''' Calculates sari score (between 0 and 100) given a list of source and predicted sentences, and a list of lists of reference sentences. It also computes the BLEU score as well as the exact match score. Args: sources: list of source sentences where each sentence should be a string. predictions: list of predicted sentences where each sentence should be a string. references: list of lists of reference sentences where each sentence should be a string. Returns: sari: sari score sacrebleu: sacrebleu score exact: exact score Examples: >>> sources=["About 95 species are currently accepted ."] >>> predictions=["About 95 you now get in ."] >>> references=[["About 95 species are currently known ."]] >>> wiki_split = datasets.load_metric("wiki_split") >>> results = wiki_split.compute(sources=sources, predictions=predictions, references=references) >>> print(results) {\'sari\': 21.805555555555557, \'sacrebleu\': 14.535768424205482, \'exact\': 0.0} ''' def lowerCAmelCase_( lowercase_ : Union[str, Any] ) -> str: def remove_articles(lowercase_ : int ): _lowerCamelCase = re.compile(r'''\b(a|an|the)\b''' , re.UNICODE ) return re.sub(lowercase_ , ''' ''' , lowercase_ ) def white_space_fix(lowercase_ : List[Any] ): return " ".join(text.split() ) def remove_punc(lowercase_ : Dict ): _lowerCamelCase = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowercase_ : Union[str, Any] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowercase_ ) ) ) ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[Any] ) -> Union[str, Any]: return int(normalize_answer(lowercase_ ) == normalize_answer(lowercase_ ) ) def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Tuple ) -> Tuple: _lowerCamelCase = [any(compute_exact(lowercase_ , lowercase_ ) for ref in refs ) for pred, refs in zip(lowercase_ , lowercase_ )] return (sum(lowercase_ ) / len(lowercase_ )) * 1_00 def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : str , lowercase_ : Optional[Any] , lowercase_ : str ) -> Optional[int]: _lowerCamelCase = [rgram for rgrams in rgramslist for rgram in rgrams] _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter() for sgram, scount in sgramcounter.items(): _lowerCamelCase = scount * numref _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter() for cgram, ccount in cgramcounter.items(): _lowerCamelCase = ccount * numref # KEEP _lowerCamelCase = sgramcounter_rep & cgramcounter_rep _lowerCamelCase = keepgramcounter_rep & rgramcounter _lowerCamelCase = sgramcounter_rep & rgramcounter _lowerCamelCase = 0 _lowerCamelCase = 0 for keepgram in keepgramcountergood_rep: keeptmpscorea += keepgramcountergood_rep[keepgram] / keepgramcounter_rep[keepgram] # Fix an alleged bug [2] in the keep score computation. # keeptmpscore2 += keepgramcountergood_rep[keepgram] / keepgramcounterall_rep[keepgram] keeptmpscorea += keepgramcountergood_rep[keepgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = keeptmpscorea / len(lowercase_ ) if len(lowercase_ ) > 0: # Fix an alleged bug [2] in the keep score computation. # keepscore_recall = keeptmpscore2 / len(keepgramcounterall_rep) _lowerCamelCase = keeptmpscorea / sum(keepgramcounterall_rep.values() ) _lowerCamelCase = 0 if keepscore_precision > 0 or keepscore_recall > 0: _lowerCamelCase = 2 * keepscore_precision * keepscore_recall / (keepscore_precision + keepscore_recall) # DELETION _lowerCamelCase = sgramcounter_rep - cgramcounter_rep _lowerCamelCase = delgramcounter_rep - rgramcounter _lowerCamelCase = sgramcounter_rep - rgramcounter _lowerCamelCase = 0 _lowerCamelCase = 0 for delgram in delgramcountergood_rep: deltmpscorea += delgramcountergood_rep[delgram] / delgramcounter_rep[delgram] deltmpscorea += delgramcountergood_rep[delgram] / delgramcounterall_rep[delgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = deltmpscorea / len(lowercase_ ) # ADDITION _lowerCamelCase = set(lowercase_ ) - set(lowercase_ ) _lowerCamelCase = set(lowercase_ ) & set(lowercase_ ) _lowerCamelCase = set(lowercase_ ) - set(lowercase_ ) _lowerCamelCase = 0 for addgram in addgramcountergood: addtmpscore += 1 # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = addtmpscore / len(lowercase_ ) if len(lowercase_ ) > 0: _lowerCamelCase = addtmpscore / len(lowercase_ ) _lowerCamelCase = 0 if addscore_precision > 0 or addscore_recall > 0: _lowerCamelCase = 2 * addscore_precision * addscore_recall / (addscore_precision + addscore_recall) return (keepscore, delscore_precision, addscore) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : Optional[Any] , lowercase_ : str ) -> List[str]: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = ssent.split(''' ''' ) _lowerCamelCase = csent.split(''' ''' ) _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] for rsent in rsents: _lowerCamelCase = rsent.split(''' ''' ) _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] ragramslist.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] ragrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] + ''' ''' + ragrams[i + 2] ragrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] + ''' ''' + ragrams[i + 2] + ''' ''' + ragrams[i + 3] ragrams.append(lowercase_ ) ragramslist.append(lowercase_ ) ragramslist.append(lowercase_ ) ragramslist.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] sagrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] + ''' ''' + sagrams[i + 2] sagrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] + ''' ''' + sagrams[i + 2] + ''' ''' + sagrams[i + 3] sagrams.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] cagrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] + ''' ''' + cagrams[i + 2] cagrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] + ''' ''' + cagrams[i + 2] + ''' ''' + cagrams[i + 3] cagrams.append(lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) _lowerCamelCase = sum([keepascore, keepascore, keepascore, keepascore] ) / 4 _lowerCamelCase = sum([delascore, delascore, delascore, delascore] ) / 4 _lowerCamelCase = sum([addascore, addascore, addascore, addascore] ) / 4 _lowerCamelCase = (avgkeepscore + avgdelscore + avgaddscore) / 3 return finalscore def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : bool = True , lowercase_ : str = "13a" , lowercase_ : bool = True ) -> int: # Normalization is requried for the ASSET dataset (one of the primary # datasets in sentence simplification) to allow using space # to split the sentence. Even though Wiki-Auto and TURK datasets, # do not require normalization, we do it for consistency. # Code adapted from the EASSE library [1] written by the authors of the ASSET dataset. # [1] https://github.com/feralvam/easse/blob/580bba7e1378fc8289c663f864e0487188fe8067/easse/utils/preprocessing.py#L7 if lowercase: _lowerCamelCase = sentence.lower() if tokenizer in ["13a", "intl"]: if version.parse(sacrebleu.__version__ ).major >= 2: _lowerCamelCase = sacrebleu.metrics.bleu._get_tokenizer(lowercase_ )()(lowercase_ ) else: _lowerCamelCase = sacrebleu.TOKENIZERS[tokenizer]()(lowercase_ ) elif tokenizer == "moses": _lowerCamelCase = sacremoses.MosesTokenizer().tokenize(lowercase_ , return_str=lowercase_ , escape=lowercase_ ) elif tokenizer == "penn": _lowerCamelCase = sacremoses.MosesTokenizer().penn_tokenize(lowercase_ , return_str=lowercase_ ) else: _lowerCamelCase = sentence if not return_str: _lowerCamelCase = normalized_sent.split() return normalized_sent def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Any , lowercase_ : List[Any] ) -> Optional[int]: if not (len(lowercase_ ) == len(lowercase_ ) == len(lowercase_ )): raise ValueError('''Sources length must match predictions and references lengths.''' ) _lowerCamelCase = 0 for src, pred, refs in zip(lowercase_ , lowercase_ , lowercase_ ): sari_score += SARIsent(normalize(lowercase_ ) , normalize(lowercase_ ) , [normalize(lowercase_ ) for sent in refs] ) _lowerCamelCase = sari_score / len(lowercase_ ) return 1_00 * sari_score def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Any , lowercase_ : List[Any]="exp" , lowercase_ : List[Any]=None , lowercase_ : Optional[Any]=False , lowercase_ : List[Any]=False , lowercase_ : List[Any]=False , ) -> Dict: _lowerCamelCase = len(references[0] ) if any(len(lowercase_ ) != references_per_prediction for refs in references ): raise ValueError('''Sacrebleu requires the same number of references for each prediction''' ) _lowerCamelCase = [[refs[i] for refs in references] for i in range(lowercase_ )] _lowerCamelCase = sacrebleu.corpus_bleu( lowercase_ , lowercase_ , smooth_method=lowercase_ , smooth_value=lowercase_ , force=lowercase_ , lowercase=lowercase_ , use_effective_order=lowercase_ , ) return output.score @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Sequence(datasets.Value('''string''' , id='''sequence''' ) , id='''references''' ), } ) , codebase_urls=[ '''https://github.com/huggingface/transformers/blob/master/src/transformers/data/metrics/squad_metrics.py''', '''https://github.com/cocoxu/simplification/blob/master/SARI.py''', '''https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/sari_hook.py''', '''https://github.com/mjpost/sacreBLEU''', ] , reference_urls=[ '''https://www.aclweb.org/anthology/Q16-1029.pdf''', '''https://github.com/mjpost/sacreBLEU''', '''https://en.wikipedia.org/wiki/BLEU''', '''https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213''', ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = {} result.update({'''sari''': compute_sari(sources=lowerCamelCase__ , predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) result.update({'''sacrebleu''': compute_sacrebleu(predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) result.update({'''exact''': compute_em(predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) return result
661
1
"""simple docstring""" import baseaa def lowerCAmelCase_( lowercase_ : str ) -> bytes: return baseaa.baaencode(string.encode('''utf-8''' ) ) def lowerCAmelCase_( lowercase_ : bytes ) -> str: return baseaa.baadecode(lowercase_ ).decode('''utf-8''' ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : str = '''Hello World!''' __SCREAMING_SNAKE_CASE : Optional[int] = baseaa_encode(test) print(encoded) __SCREAMING_SNAKE_CASE : Any = baseaa_decode(encoded) print(decoded)
661
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''configuration_wav2vec2''': ['''WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Wav2Vec2Config'''], '''feature_extraction_wav2vec2''': ['''Wav2Vec2FeatureExtractor'''], '''processing_wav2vec2''': ['''Wav2Vec2Processor'''], '''tokenization_wav2vec2''': ['''Wav2Vec2CTCTokenizer''', '''Wav2Vec2Tokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = [ '''WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Wav2Vec2ForAudioFrameClassification''', '''Wav2Vec2ForCTC''', '''Wav2Vec2ForMaskedLM''', '''Wav2Vec2ForPreTraining''', '''Wav2Vec2ForSequenceClassification''', '''Wav2Vec2ForXVector''', '''Wav2Vec2Model''', '''Wav2Vec2PreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = [ '''TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWav2Vec2ForCTC''', '''TFWav2Vec2Model''', '''TFWav2Vec2PreTrainedModel''', '''TFWav2Vec2ForSequenceClassification''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = [ '''FlaxWav2Vec2ForCTC''', '''FlaxWav2Vec2ForPreTraining''', '''FlaxWav2Vec2Model''', '''FlaxWav2Vec2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavaveca import WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, WavaVecaConfig from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .processing_wavaveca import WavaVecaProcessor from .tokenization_wavaveca import WavaVecaCTCTokenizer, WavaVecaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavaveca import ( WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaForAudioFrameClassification, WavaVecaForCTC, WavaVecaForMaskedLM, WavaVecaForPreTraining, WavaVecaForSequenceClassification, WavaVecaForXVector, WavaVecaModel, WavaVecaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, TFWavaVecaForCTC, TFWavaVecaForSequenceClassification, TFWavaVecaModel, TFWavaVecaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( FlaxWavaVecaForCTC, FlaxWavaVecaForPreTraining, FlaxWavaVecaModel, FlaxWavaVecaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
661
1
"""simple docstring""" import argparse from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_controlnet_from_original_ckpt if __name__ == "__main__": __SCREAMING_SNAKE_CASE : int = argparse.ArgumentParser() parser.add_argument( '''--checkpoint_path''', default=None, type=str, required=True, help='''Path to the checkpoint to convert.''' ) parser.add_argument( '''--original_config_file''', type=str, required=True, help='''The YAML config file corresponding to the original architecture.''', ) parser.add_argument( '''--num_in_channels''', default=None, type=int, help='''The number of input channels. If `None` number of input channels will be automatically inferred.''', ) parser.add_argument( '''--image_size''', default=5_1_2, type=int, help=( '''The image size that the model was trained on. Use 512 for Stable Diffusion v1.X and Stable Siffusion v2''' ''' Base. Use 768 for Stable Diffusion v2.''' ), ) parser.add_argument( '''--extract_ema''', action='''store_true''', help=( '''Only relevant for checkpoints that have both EMA and non-EMA weights. Whether to extract the EMA weights''' ''' or not. Defaults to `False`. Add `--extract_ema` to extract the EMA weights. EMA weights usually yield''' ''' higher quality images for inference. Non-EMA weights are usually better to continue fine-tuning.''' ), ) parser.add_argument( '''--upcast_attention''', action='''store_true''', help=( '''Whether the attention computation should always be upcasted. This is necessary when running stable''' ''' diffusion 2.1.''' ), ) parser.add_argument( '''--from_safetensors''', action='''store_true''', help='''If `--checkpoint_path` is in `safetensors` format, load checkpoint with safetensors instead of PyTorch.''', ) parser.add_argument( '''--to_safetensors''', action='''store_true''', help='''Whether to store pipeline in safetensors format or not.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') parser.add_argument('''--device''', type=str, help='''Device to use (e.g. cpu, cuda:0, cuda:1, etc.)''') def lowerCAmelCase_( lowercase_ : Any ) -> List[Any]: if string == "True": return True elif string == "False": return False else: raise ValueError(F"""could not parse string as bool {string}""" ) parser.add_argument( '''--use_linear_projection''', help='''Override for use linear projection''', required=False, type=parse_bool ) parser.add_argument('''--cross_attention_dim''', help='''Override for cross attention_dim''', required=False, type=int) __SCREAMING_SNAKE_CASE : Dict = parser.parse_args() __SCREAMING_SNAKE_CASE : List[Any] = download_controlnet_from_original_ckpt( checkpoint_path=args.checkpoint_path, original_config_file=args.original_config_file, image_size=args.image_size, extract_ema=args.extract_ema, num_in_channels=args.num_in_channels, upcast_attention=args.upcast_attention, from_safetensors=args.from_safetensors, device=args.device, use_linear_projection=args.use_linear_projection, cross_attention_dim=args.cross_attention_dim, ) controlnet.save_pretrained(args.dump_path, safe_serialization=args.to_safetensors)
661
"""simple docstring""" __SCREAMING_SNAKE_CASE : Any = { '''A''': '''.-''', '''B''': '''-...''', '''C''': '''-.-.''', '''D''': '''-..''', '''E''': '''.''', '''F''': '''..-.''', '''G''': '''--.''', '''H''': '''....''', '''I''': '''..''', '''J''': '''.---''', '''K''': '''-.-''', '''L''': '''.-..''', '''M''': '''--''', '''N''': '''-.''', '''O''': '''---''', '''P''': '''.--.''', '''Q''': '''--.-''', '''R''': '''.-.''', '''S''': '''...''', '''T''': '''-''', '''U''': '''..-''', '''V''': '''...-''', '''W''': '''.--''', '''X''': '''-..-''', '''Y''': '''-.--''', '''Z''': '''--..''', '''1''': '''.----''', '''2''': '''..---''', '''3''': '''...--''', '''4''': '''....-''', '''5''': '''.....''', '''6''': '''-....''', '''7''': '''--...''', '''8''': '''---..''', '''9''': '''----.''', '''0''': '''-----''', '''&''': '''.-...''', '''@''': '''.--.-.''', ''':''': '''---...''', ''',''': '''--..--''', '''.''': '''.-.-.-''', '''\'''': '''.----.''', '''"''': '''.-..-.''', '''?''': '''..--..''', '''/''': '''-..-.''', '''=''': '''-...-''', '''+''': '''.-.-.''', '''-''': '''-....-''', '''(''': '''-.--.''', ''')''': '''-.--.-''', '''!''': '''-.-.--''', ''' ''': '''/''' } # Exclamation mark is not in ITU-R recommendation # fmt: on __SCREAMING_SNAKE_CASE : List[str] = {value: key for key, value in MORSE_CODE_DICT.items()} def lowerCAmelCase_( lowercase_ : str ) -> str: return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def lowerCAmelCase_( lowercase_ : str ) -> str: return "".join(REVERSE_DICT[char] for char in message.split() ) def lowerCAmelCase_( ) -> None: _lowerCamelCase = '''Morse code here!''' print(lowercase_ ) _lowerCamelCase = encrypt(lowercase_ ) print(lowercase_ ) _lowerCamelCase = decrypt(lowercase_ ) print(lowercase_ ) if __name__ == "__main__": main()
661
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : List[Any] = { '''configuration_mobilebert''': [ '''MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''MobileBertConfig''', '''MobileBertOnnxConfig''', ], '''tokenization_mobilebert''': ['''MobileBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = ['''MobileBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = [ '''MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''MobileBertForMaskedLM''', '''MobileBertForMultipleChoice''', '''MobileBertForNextSentencePrediction''', '''MobileBertForPreTraining''', '''MobileBertForQuestionAnswering''', '''MobileBertForSequenceClassification''', '''MobileBertForTokenClassification''', '''MobileBertLayer''', '''MobileBertModel''', '''MobileBertPreTrainedModel''', '''load_tf_weights_in_mobilebert''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Dict = [ '''TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFMobileBertForMaskedLM''', '''TFMobileBertForMultipleChoice''', '''TFMobileBertForNextSentencePrediction''', '''TFMobileBertForPreTraining''', '''TFMobileBertForQuestionAnswering''', '''TFMobileBertForSequenceClassification''', '''TFMobileBertForTokenClassification''', '''TFMobileBertMainLayer''', '''TFMobileBertModel''', '''TFMobileBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_mobilebert import ( MOBILEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileBertConfig, MobileBertOnnxConfig, ) from .tokenization_mobilebert import MobileBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mobilebert_fast import MobileBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilebert import ( MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileBertForMaskedLM, MobileBertForMultipleChoice, MobileBertForNextSentencePrediction, MobileBertForPreTraining, MobileBertForQuestionAnswering, MobileBertForSequenceClassification, MobileBertForTokenClassification, MobileBertLayer, MobileBertModel, MobileBertPreTrainedModel, load_tf_weights_in_mobilebert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilebert import ( TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertMainLayer, TFMobileBertModel, TFMobileBertPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Union[str, Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
661
"""simple docstring""" from argparse import ArgumentParser, Namespace from typing import Any, List, Optional from ..pipelines import Pipeline, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand try: from fastapi import Body, FastAPI, HTTPException from fastapi.routing import APIRoute from pydantic import BaseModel from starlette.responses import JSONResponse from uvicorn import run __SCREAMING_SNAKE_CASE : List[Any] = True except (ImportError, AttributeError): __SCREAMING_SNAKE_CASE : List[Any] = object def lowerCAmelCase_( *lowercase_ : Dict , **lowercase_ : str ) -> str: pass __SCREAMING_SNAKE_CASE : Tuple = False __SCREAMING_SNAKE_CASE : Any = logging.get_logger('''transformers-cli/serving''') def lowerCAmelCase_( lowercase_ : Namespace ) -> List[Any]: _lowerCamelCase = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) return ServeCommand(lowercase_ , args.host , args.port , args.workers ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : dict class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] lowercase__ : Optional[List[int]] class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : str class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Any class lowerCamelCase_( A__ ): '''simple docstring''' @staticmethod def snake_case__ ( lowerCamelCase__ ): _lowerCamelCase = parser.add_parser( '''serve''' , help='''CLI tool to run inference requests through REST and GraphQL endpoints.''' ) serve_parser.add_argument( '''--task''' , type=lowerCamelCase__ , choices=get_supported_tasks() , help='''The task to run the pipeline on''' , ) serve_parser.add_argument('''--host''' , type=lowerCamelCase__ , default='''localhost''' , help='''Interface the server will listen on.''' ) serve_parser.add_argument('''--port''' , type=lowerCamelCase__ , default=8_8_8_8 , help='''Port the serving will listen to.''' ) serve_parser.add_argument('''--workers''' , type=lowerCamelCase__ , default=1 , help='''Number of http workers''' ) serve_parser.add_argument('''--model''' , type=lowerCamelCase__ , help='''Model\'s name or path to stored model.''' ) serve_parser.add_argument('''--config''' , type=lowerCamelCase__ , help='''Model\'s config name or path to stored model.''' ) serve_parser.add_argument('''--tokenizer''' , type=lowerCamelCase__ , help='''Tokenizer name to use.''' ) serve_parser.add_argument( '''--device''' , type=lowerCamelCase__ , default=-1 , help='''Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)''' , ) serve_parser.set_defaults(func=lowerCamelCase__ ) def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = pipeline _lowerCamelCase = host _lowerCamelCase = port _lowerCamelCase = workers if not _serve_dependencies_installed: raise RuntimeError( '''Using serve command requires FastAPI and uvicorn. ''' '''Please install transformers with [serving]: pip install "transformers[serving]".''' '''Or install FastAPI and uvicorn separately.''' ) else: logger.info(F"""Serving model over {host}:{port}""" ) _lowerCamelCase = FastAPI( routes=[ APIRoute( '''/''' , self.model_info , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''GET'''] , ), APIRoute( '''/tokenize''' , self.tokenize , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), APIRoute( '''/detokenize''' , self.detokenize , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), APIRoute( '''/forward''' , self.forward , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), ] , timeout=6_0_0 , ) def snake_case__ ( self ): run(self._app , host=self.host , port=self.port , workers=self.workers ) def snake_case__ ( self ): return ServeModelInfoResult(infos=vars(self._pipeline.model.config ) ) def snake_case__ ( self , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) ): try: _lowerCamelCase = self._pipeline.tokenizer.tokenize(lowerCamelCase__ ) if return_ids: _lowerCamelCase = self._pipeline.tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) return ServeTokenizeResult(tokens=lowerCamelCase__ , tokens_ids=lowerCamelCase__ ) else: return ServeTokenizeResult(tokens=lowerCamelCase__ ) except Exception as e: raise HTTPException(status_code=5_0_0 , detail={'''model''': '''''', '''error''': str(lowerCamelCase__ )} ) def snake_case__ ( self , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , ): try: _lowerCamelCase = self._pipeline.tokenizer.decode(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) return ServeDeTokenizeResult(model='''''' , text=lowerCamelCase__ ) except Exception as e: raise HTTPException(status_code=5_0_0 , detail={'''model''': '''''', '''error''': str(lowerCamelCase__ )} ) async def snake_case__ ( self , lowerCamelCase__=Body(lowerCamelCase__ , embed=lowerCamelCase__ ) ): # Check we don't have empty string if len(lowerCamelCase__ ) == 0: return ServeForwardResult(output=[] , attention=[] ) try: # Forward through the model _lowerCamelCase = self._pipeline(lowerCamelCase__ ) return ServeForwardResult(output=lowerCamelCase__ ) except Exception as e: raise HTTPException(5_0_0 , {'''error''': str(lowerCamelCase__ )} )
661
1
"""simple docstring""" from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def lowerCAmelCase_( ) -> tuple[list[int], int]: _lowerCamelCase = [randint(-10_00 , 10_00 ) for i in range(10 )] _lowerCamelCase = randint(-50_00 , 50_00 ) return (arr, r) __SCREAMING_SNAKE_CASE : List[str] = make_dataset() def lowerCAmelCase_( lowercase_ : list[int] , lowercase_ : int ) -> tuple[int, ...]: for triplet in permutations(lowercase_ , 3 ): if sum(lowercase_ ) == target: return tuple(sorted(lowercase_ ) ) return (0, 0, 0) def lowerCAmelCase_( lowercase_ : list[int] , lowercase_ : int ) -> tuple[int, int, int]: arr.sort() _lowerCamelCase = len(lowercase_ ) for i in range(n - 1 ): _lowerCamelCase , _lowerCamelCase = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def lowerCAmelCase_( ) -> tuple[float, float]: _lowerCamelCase = ''' from __main__ import dataset, triplet_sum1, triplet_sum2 ''' _lowerCamelCase = ''' triplet_sum1(*dataset) ''' _lowerCamelCase = ''' triplet_sum2(*dataset) ''' _lowerCamelCase = repeat(setup=lowercase_ , stmt=lowercase_ , repeat=5 , number=1_00_00 ) _lowerCamelCase = repeat(setup=lowercase_ , stmt=lowercase_ , repeat=5 , number=1_00_00 ) return (min(lowercase_ ), min(lowercase_ )) if __name__ == "__main__": from doctest import testmod testmod() __SCREAMING_SNAKE_CASE : Dict = solution_times() print(F"""The time for naive implementation is {times[0]}.""") print(F"""The time for optimized implementation is {times[1]}.""")
661
"""simple docstring""" import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def lowerCAmelCase_( lowercase_ : str , lowercase_ : Dict ) -> List[Any]: assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : str , lowercase_ : str ) -> List[Any]: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ , keep_in_memory=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : int ) -> Optional[int]: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = features.copy() if features else default_expected_features _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = JsonDatasetReader(lowercase_ , features=lowercase_ , cache_dir=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''}, ] , ) def lowerCAmelCase_( lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : List[Any] ) -> Tuple: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''} _lowerCamelCase = features.copy() if features else default_expected_features _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = JsonDatasetReader(lowercase_ , features=lowercase_ , cache_dir=lowercase_ ).read() assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Optional[int] ) -> List[str]: # jsonl_312_path features are {"col_3": "float64", "col_1": "string", "col_2": "int64"} _lowerCamelCase = {'''col_2''': '''int64''', '''col_3''': '''float64''', '''col_1''': '''string'''} _lowerCamelCase = features.copy() _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = JsonDatasetReader(lowercase_ , features=lowercase_ , cache_dir=lowercase_ ).read() assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : Optional[Any] ) -> int: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ , split=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('''path_type''' , [str, list] ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[int] , lowercase_ : Tuple ) -> Optional[int]: if issubclass(lowercase_ , lowercase_ ): _lowerCamelCase = jsonl_path elif issubclass(lowercase_ , lowercase_ ): _lowerCamelCase = [jsonl_path] _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : int , lowercase_ : List[str] , lowercase_ : str=("train",) ) -> List[str]: assert isinstance(lowercase_ , lowercase_ ) for split in splits: _lowerCamelCase = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[int] , lowercase_ : int ) -> Dict: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _lowerCamelCase = JsonDatasetReader({'''train''': jsonl_path} , cache_dir=lowercase_ , keep_in_memory=lowercase_ ).read() _check_json_datasetdict(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : Union[str, Any] , lowercase_ : Optional[Any] ) -> List[Any]: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = features.copy() if features else default_expected_features _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = JsonDatasetReader({'''train''': jsonl_path} , features=lowercase_ , cache_dir=lowercase_ ).read() _check_json_datasetdict(lowercase_ , lowercase_ ) @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : str , lowercase_ : List[str] ) -> Optional[Any]: if split: _lowerCamelCase = {split: jsonl_path} else: _lowerCamelCase = '''train''' _lowerCamelCase = {'''train''': jsonl_path, '''test''': jsonl_path} _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ ).read() _check_json_datasetdict(lowercase_ , lowercase_ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def lowerCAmelCase_( lowercase_ : Optional[int] ) -> Optional[Any]: return json.load(lowercase_ ) def lowerCAmelCase_( lowercase_ : Tuple ) -> Tuple: return [json.loads(lowercase_ ) for line in buffer] class lowerCamelCase_: '''simple docstring''' @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ ).write() buffer.seek(0 ) _lowerCamelCase = load_json_function(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) assert isinstance(exported_content[0] , lowerCamelCase__ ) assert len(lowerCamelCase__ ) == 1_0 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ , orient=lowerCamelCase__ ).write() buffer.seek(0 ) _lowerCamelCase = load_json(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowerCamelCase__ , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 1_0 else: assert len(lowerCamelCase__ ) == 1_0 @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ , num_proc=2 ).write() buffer.seek(0 ) _lowerCamelCase = load_json_function(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) assert isinstance(exported_content[0] , lowerCamelCase__ ) assert len(lowerCamelCase__ ) == 1_0 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ , orient=lowerCamelCase__ , num_proc=2 ).write() buffer.seek(0 ) _lowerCamelCase = load_json(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowerCamelCase__ , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 1_0 else: assert len(lowerCamelCase__ ) == 1_0 def snake_case__ ( self , lowerCamelCase__ ): with pytest.raises(lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , num_proc=0 ) @pytest.mark.parametrize('''compression, extension''' , [('''gzip''', '''gz'''), ('''bz2''', '''bz2'''), ('''xz''', '''xz''')] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = tmp_path_factory.mktemp('''data''' ) / F"""test.json.{extension}""" _lowerCamelCase = str(shared_datadir / F"""test_file.json.{extension}""" ) JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , compression=lowerCamelCase__ ).write() with fsspec.open(lowerCamelCase__ , '''rb''' , compression='''infer''' ) as f: _lowerCamelCase = f.read() with fsspec.open(lowerCamelCase__ , '''rb''' , compression='''infer''' ) as f: _lowerCamelCase = f.read() assert exported_content == original_content
661
1
"""simple docstring""" __SCREAMING_SNAKE_CASE : int = 9.8_0665 def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float = g ) -> float: if fluid_density <= 0: raise ValueError('''Impossible fluid density''' ) if volume < 0: raise ValueError('''Impossible Object volume''' ) if gravity <= 0: raise ValueError('''Impossible Gravity''' ) return fluid_density * gravity * volume if __name__ == "__main__": import doctest # run doctest doctest.testmod()
661
"""simple docstring""" import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=3 , lowerCamelCase__=3_2 , lowerCamelCase__=3 , lowerCamelCase__=1_0 , lowerCamelCase__=[8, 1_6, 3_2, 6_4] , lowerCamelCase__=[1, 1, 2, 1] , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__="relu" , lowerCamelCase__=3 , lowerCamelCase__=None , lowerCamelCase__=["stage2", "stage3", "stage4"] , lowerCamelCase__=[2, 3, 4] , lowerCamelCase__=1 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = num_channels _lowerCamelCase = embeddings_size _lowerCamelCase = hidden_sizes _lowerCamelCase = depths _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_act _lowerCamelCase = num_labels _lowerCamelCase = scope _lowerCamelCase = len(lowerCamelCase__ ) _lowerCamelCase = out_features _lowerCamelCase = out_indices _lowerCamelCase = num_groups def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.num_labels ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = BitModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = BitForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = BitBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None _lowerCamelCase = None _lowerCamelCase = BitBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Dict = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () lowercase__ : Any = ( {'feature-extraction': BitModel, 'image-classification': BitForImageClassification} if is_torch_available() else {} ) lowercase__ : Union[str, Any] = False lowercase__ : List[Any] = False lowercase__ : Any = False lowercase__ : List[str] = False lowercase__ : Any = False def snake_case__ ( self ): _lowerCamelCase = BitModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ ) def snake_case__ ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case__ ( self ): return @unittest.skip(reason='''Bit does not output attentions''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''Bit does not use inputs_embeds''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''Bit does not support input and output embeddings''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(config=lowerCamelCase__ ) for name, module in model.named_modules(): if isinstance(lowerCamelCase__ , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) def snake_case__ ( self ): def check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _lowerCamelCase = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase__ ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = ['''preactivation''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: _lowerCamelCase = layer_type _lowerCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) @unittest.skip(reason='''Bit does not use feedforward chunking''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = BitModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def snake_case__ ( self ): _lowerCamelCase = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowerCamelCase__ ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**lowerCamelCase__ ) # verify the logits _lowerCamelCase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([[-0.6_5_2_6, -0.5_2_6_3, -1.4_3_9_8]] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) ) @require_torch class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = (BitBackbone,) if is_torch_available() else () lowercase__ : Tuple = BitConfig lowercase__ : Any = False def snake_case__ ( self ): _lowerCamelCase = BitModelTester(self )
661
1
"""simple docstring""" import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=7 , lowerCamelCase__=3 , lowerCamelCase__=3_0 , lowerCamelCase__=4_0_0 , lowerCamelCase__=True , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__=[0.5, 0.5, 0.5] , lowerCamelCase__=[0.5, 0.5, 0.5] , lowerCamelCase__=True , lowerCamelCase__=1 / 2_5_5 , lowerCamelCase__=True , ): # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p _lowerCamelCase = size if size is not None else {'''shortest_edge''': 1_8, '''longest_edge''': 1_3_3_3} _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = num_channels _lowerCamelCase = min_resolution _lowerCamelCase = max_resolution _lowerCamelCase = do_resize _lowerCamelCase = size _lowerCamelCase = do_normalize _lowerCamelCase = image_mean _lowerCamelCase = image_std _lowerCamelCase = do_rescale _lowerCamelCase = rescale_factor _lowerCamelCase = do_pad def snake_case__ ( self ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=False ): if not batched: _lowerCamelCase = image_inputs[0] if isinstance(lowerCamelCase__ , Image.Image ): _lowerCamelCase , _lowerCamelCase = image.size else: _lowerCamelCase , _lowerCamelCase = image.shape[1], image.shape[2] if w < h: _lowerCamelCase = int(self.size['''shortest_edge'''] * h / w ) _lowerCamelCase = self.size['''shortest_edge'''] elif w > h: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = int(self.size['''shortest_edge'''] * w / h ) else: _lowerCamelCase = self.size['''shortest_edge'''] _lowerCamelCase = self.size['''shortest_edge'''] else: _lowerCamelCase = [] for image in image_inputs: _lowerCamelCase , _lowerCamelCase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _lowerCamelCase = max(lowerCamelCase__ , key=lambda lowerCamelCase__ : item[0] )[0] _lowerCamelCase = max(lowerCamelCase__ , key=lambda lowerCamelCase__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[int] = ConditionalDetrImageProcessor if is_vision_available() else None def snake_case__ ( self ): _lowerCamelCase = ConditionalDetrImageProcessingTester(self ) @property def snake_case__ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def snake_case__ ( self ): _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCamelCase__ , '''image_mean''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''image_std''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_normalize''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''do_resize''' ) ) self.assertTrue(hasattr(lowerCamelCase__ , '''size''' ) ) def snake_case__ ( self ): _lowerCamelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 1_8, '''longest_edge''': 1_3_3_3} ) self.assertEqual(image_processor.do_pad , lowerCamelCase__ ) _lowerCamelCase = self.image_processing_class.from_dict( self.image_processor_dict , size=4_2 , max_size=8_4 , pad_and_return_pixel_mask=lowerCamelCase__ ) self.assertEqual(image_processor.size , {'''shortest_edge''': 4_2, '''longest_edge''': 8_4} ) self.assertEqual(image_processor.do_pad , lowerCamelCase__ ) def snake_case__ ( self ): pass def snake_case__ ( self ): # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , Image.Image ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(lowerCamelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(lowerCamelCase__ , batched=lowerCamelCase__ ) _lowerCamelCase = image_processing(lowerCamelCase__ , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def snake_case__ ( self ): # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ , numpify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , np.ndarray ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(lowerCamelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(lowerCamelCase__ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(lowerCamelCase__ , batched=lowerCamelCase__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def snake_case__ ( self ): # Initialize image_processing _lowerCamelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowerCamelCase = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCamelCase__ , torchify=lowerCamelCase__ ) for image in image_inputs: self.assertIsInstance(lowerCamelCase__ , torch.Tensor ) # Test not batched input _lowerCamelCase = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(lowerCamelCase__ ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _lowerCamelCase = image_processing(lowerCamelCase__ , return_tensors='''pt''' ).pixel_values _lowerCamelCase , _lowerCamelCase = self.image_processor_tester.get_expected_values(lowerCamelCase__ , batched=lowerCamelCase__ ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def snake_case__ ( self ): # prepare image and target _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''image_id''': 3_9_7_6_9, '''annotations''': target} # encode them _lowerCamelCase = ConditionalDetrImageProcessor.from_pretrained('''microsoft/conditional-detr-resnet-50''' ) _lowerCamelCase = image_processing(images=lowerCamelCase__ , annotations=lowerCamelCase__ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_0_0, 1_0_6_6] ) self.assertEqual(encoding['''pixel_values'''].shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , lowerCamelCase__ , atol=1e-4 ) ) # verify area _lowerCamelCase = torch.tensor([5_8_8_7.9_6_0_0, 1_1_2_5_0.2_0_6_1, 4_8_9_3_5_3.8_4_3_8, 8_3_7_1_2_2.7_5_0_0, 1_4_7_9_6_7.5_1_5_6, 1_6_5_7_3_2.3_4_3_8] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , lowerCamelCase__ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([0.5_5_0_3, 0.2_7_6_5, 0.0_6_0_4, 0.2_2_1_5] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , lowerCamelCase__ , atol=1e-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_9_7_6_9] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , lowerCamelCase__ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , lowerCamelCase__ ) ) # verify class_labels _lowerCamelCase = torch.tensor([7_5, 7_5, 6_3, 6_5, 1_7, 1_7] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , lowerCamelCase__ ) ) # verify orig_size _lowerCamelCase = torch.tensor([4_8_0, 6_4_0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , lowerCamelCase__ ) ) # verify size _lowerCamelCase = torch.tensor([8_0_0, 1_0_6_6] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , lowerCamelCase__ ) ) @slow def snake_case__ ( self ): # prepare image, target and masks_path _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: _lowerCamelCase = json.loads(f.read() ) _lowerCamelCase = {'''file_name''': '''000000039769.png''', '''image_id''': 3_9_7_6_9, '''segments_info''': target} _lowerCamelCase = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them _lowerCamelCase = ConditionalDetrImageProcessor(format='''coco_panoptic''' ) _lowerCamelCase = image_processing(images=lowerCamelCase__ , annotations=lowerCamelCase__ , masks_path=lowerCamelCase__ , return_tensors='''pt''' ) # verify pixel values _lowerCamelCase = torch.Size([1, 3, 8_0_0, 1_0_6_6] ) self.assertEqual(encoding['''pixel_values'''].shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , lowerCamelCase__ , atol=1e-4 ) ) # verify area _lowerCamelCase = torch.tensor([1_4_7_9_7_9.6_8_7_5, 1_6_5_5_2_7.0_4_6_9, 4_8_4_6_3_8.5_9_3_8, 1_1_2_9_2.9_3_7_5, 5_8_7_9.6_5_6_2, 7_6_3_4.1_1_4_7] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , lowerCamelCase__ ) ) # verify boxes _lowerCamelCase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([0.2_6_2_5, 0.5_4_3_7, 0.4_6_8_8, 0.8_6_2_5] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , lowerCamelCase__ , atol=1e-3 ) ) # verify image_id _lowerCamelCase = torch.tensor([3_9_7_6_9] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , lowerCamelCase__ ) ) # verify is_crowd _lowerCamelCase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , lowerCamelCase__ ) ) # verify class_labels _lowerCamelCase = torch.tensor([1_7, 1_7, 6_3, 7_5, 7_5, 9_3] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , lowerCamelCase__ ) ) # verify masks _lowerCamelCase = 8_2_2_8_7_3 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , lowerCamelCase__ ) # verify orig_size _lowerCamelCase = torch.tensor([4_8_0, 6_4_0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , lowerCamelCase__ ) ) # verify size _lowerCamelCase = torch.tensor([8_0_0, 1_0_6_6] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , lowerCamelCase__ ) )
661
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=2 , lowerCamelCase__=2_4 , lowerCamelCase__=1_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=3_2 , lowerCamelCase__=5 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1_0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=None , lowerCamelCase__=2 , lowerCamelCase__=2 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = patch_size _lowerCamelCase = max_length _lowerCamelCase = num_mel_bins _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = scope _lowerCamelCase = frequency_stride _lowerCamelCase = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _lowerCamelCase = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _lowerCamelCase = (self.max_length - self.patch_size) // self.time_stride + 1 _lowerCamelCase = frequency_out_dimension * time_out_dimension _lowerCamelCase = num_patches + 2 def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, input_values, labels def snake_case__ ( self ): return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = ASTModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = config_and_inputs _lowerCamelCase = {'''input_values''': input_values} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Any = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) lowercase__ : List[str] = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) lowercase__ : int = False lowercase__ : str = False lowercase__ : Union[str, Any] = False lowercase__ : List[str] = False def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def snake_case__ ( self ): _lowerCamelCase = ASTModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , nn.Linear ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''input_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = ASTModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> str: _lowerCamelCase = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) _lowerCamelCase , _lowerCamelCase = torchaudio.load(lowercase_ ) return audio, sampling_rate @require_torch @require_torchaudio class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def snake_case__ ( self ): _lowerCamelCase = self.default_feature_extractor _lowerCamelCase = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(lowerCamelCase__ ) _lowerCamelCase = self.default_feature_extractor _lowerCamelCase , _lowerCamelCase = prepare_audio() _lowerCamelCase = audio.squeeze().numpy() _lowerCamelCase = feature_extractor(lowerCamelCase__ , sampling_rate=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**lowerCamelCase__ ) # verify the logits _lowerCamelCase = torch.Size((1, 5_2_7) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([-0.8_7_6_0, -7.0_0_4_2, -8.6_6_0_2] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) )
661
1
"""simple docstring""" import argparse import json import pickle from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import MaskFormerConfig, MaskFormerForInstanceSegmentation, MaskFormerImageProcessor, SwinConfig from transformers.utils import logging logging.set_verbosity_info() __SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) def lowerCAmelCase_( lowercase_ : str ) -> Any: _lowerCamelCase = SwinConfig.from_pretrained( '''microsoft/swin-tiny-patch4-window7-224''' , out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] ) _lowerCamelCase = MaskFormerConfig(backbone_config=lowercase_ ) _lowerCamelCase = '''huggingface/label-files''' if "ade20k-full" in model_name: # this should be ok _lowerCamelCase = 8_47 _lowerCamelCase = '''maskformer-ade20k-full-id2label.json''' elif "ade" in model_name: # this should be ok _lowerCamelCase = 1_50 _lowerCamelCase = '''ade20k-id2label.json''' elif "coco-stuff" in model_name: # this should be ok _lowerCamelCase = 1_71 _lowerCamelCase = '''maskformer-coco-stuff-id2label.json''' elif "coco" in model_name: # TODO _lowerCamelCase = 1_33 _lowerCamelCase = '''coco-panoptic-id2label.json''' elif "cityscapes" in model_name: # this should be ok _lowerCamelCase = 19 _lowerCamelCase = '''cityscapes-id2label.json''' elif "vistas" in model_name: # this should be ok _lowerCamelCase = 65 _lowerCamelCase = '''mapillary-vistas-id2label.json''' _lowerCamelCase = json.load(open(hf_hub_download(lowercase_ , lowercase_ , repo_type='''dataset''' ) , '''r''' ) ) _lowerCamelCase = {int(lowercase_ ): v for k, v in idalabel.items()} return config def lowerCAmelCase_( lowercase_ : Dict ) -> Tuple: _lowerCamelCase = [] # stem # fmt: off rename_keys.append(('''backbone.patch_embed.proj.weight''', '''model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.weight''') ) rename_keys.append(('''backbone.patch_embed.proj.bias''', '''model.pixel_level_module.encoder.model.embeddings.patch_embeddings.projection.bias''') ) rename_keys.append(('''backbone.patch_embed.norm.weight''', '''model.pixel_level_module.encoder.model.embeddings.norm.weight''') ) rename_keys.append(('''backbone.patch_embed.norm.bias''', '''model.pixel_level_module.encoder.model.embeddings.norm.bias''') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F"""backbone.layers.{i}.blocks.{j}.norm1.weight""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.weight""") ) rename_keys.append((F"""backbone.layers.{i}.blocks.{j}.norm1.bias""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_before.bias""") ) rename_keys.append((F"""backbone.layers.{i}.blocks.{j}.attn.relative_position_bias_table""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table""") ) rename_keys.append((F"""backbone.layers.{i}.blocks.{j}.attn.relative_position_index""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index""") ) rename_keys.append((F"""backbone.layers.{i}.blocks.{j}.attn.proj.weight""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight""") ) rename_keys.append((F"""backbone.layers.{i}.blocks.{j}.attn.proj.bias""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias""") ) rename_keys.append((F"""backbone.layers.{i}.blocks.{j}.norm2.weight""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.weight""") ) rename_keys.append((F"""backbone.layers.{i}.blocks.{j}.norm2.bias""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.layernorm_after.bias""") ) rename_keys.append((F"""backbone.layers.{i}.blocks.{j}.mlp.fc1.weight""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight""") ) rename_keys.append((F"""backbone.layers.{i}.blocks.{j}.mlp.fc1.bias""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias""") ) rename_keys.append((F"""backbone.layers.{i}.blocks.{j}.mlp.fc2.weight""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.weight""") ) rename_keys.append((F"""backbone.layers.{i}.blocks.{j}.mlp.fc2.bias""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.blocks.{j}.output.dense.bias""") ) if i < 3: rename_keys.append((F"""backbone.layers.{i}.downsample.reduction.weight""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.reduction.weight""") ) rename_keys.append((F"""backbone.layers.{i}.downsample.norm.weight""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.weight""") ) rename_keys.append((F"""backbone.layers.{i}.downsample.norm.bias""", F"""model.pixel_level_module.encoder.model.encoder.layers.{i}.downsample.norm.bias""") ) rename_keys.append((F"""backbone.norm{i}.weight""", F"""model.pixel_level_module.encoder.hidden_states_norms.{i}.weight""") ) rename_keys.append((F"""backbone.norm{i}.bias""", F"""model.pixel_level_module.encoder.hidden_states_norms.{i}.bias""") ) # FPN rename_keys.append(('''sem_seg_head.layer_4.weight''', '''model.pixel_level_module.decoder.fpn.stem.0.weight''') ) rename_keys.append(('''sem_seg_head.layer_4.norm.weight''', '''model.pixel_level_module.decoder.fpn.stem.1.weight''') ) rename_keys.append(('''sem_seg_head.layer_4.norm.bias''', '''model.pixel_level_module.decoder.fpn.stem.1.bias''') ) for source_index, target_index in zip(range(3 , 0 , -1 ) , range(0 , 3 ) ): rename_keys.append((F"""sem_seg_head.adapter_{source_index}.weight""", F"""model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.0.weight""") ) rename_keys.append((F"""sem_seg_head.adapter_{source_index}.norm.weight""", F"""model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.weight""") ) rename_keys.append((F"""sem_seg_head.adapter_{source_index}.norm.bias""", F"""model.pixel_level_module.decoder.fpn.layers.{target_index}.proj.1.bias""") ) rename_keys.append((F"""sem_seg_head.layer_{source_index}.weight""", F"""model.pixel_level_module.decoder.fpn.layers.{target_index}.block.0.weight""") ) rename_keys.append((F"""sem_seg_head.layer_{source_index}.norm.weight""", F"""model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.weight""") ) rename_keys.append((F"""sem_seg_head.layer_{source_index}.norm.bias""", F"""model.pixel_level_module.decoder.fpn.layers.{target_index}.block.1.bias""") ) rename_keys.append(('''sem_seg_head.mask_features.weight''', '''model.pixel_level_module.decoder.mask_projection.weight''') ) rename_keys.append(('''sem_seg_head.mask_features.bias''', '''model.pixel_level_module.decoder.mask_projection.bias''') ) # Transformer decoder for idx in range(config.decoder_config.decoder_layers ): # self-attention out projection rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.weight""", F"""model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.weight""") ) rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.out_proj.bias""", F"""model.transformer_module.decoder.layers.{idx}.self_attn.out_proj.bias""") ) # cross-attention out projection rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.weight""", F"""model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.weight""") ) rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.out_proj.bias""", F"""model.transformer_module.decoder.layers.{idx}.encoder_attn.out_proj.bias""") ) # MLP 1 rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.weight""", F"""model.transformer_module.decoder.layers.{idx}.fc1.weight""") ) rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear1.bias""", F"""model.transformer_module.decoder.layers.{idx}.fc1.bias""") ) # MLP 2 rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.weight""", F"""model.transformer_module.decoder.layers.{idx}.fc2.weight""") ) rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.linear2.bias""", F"""model.transformer_module.decoder.layers.{idx}.fc2.bias""") ) # layernorm 1 (self-attention layernorm) rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.weight""", F"""model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.weight""") ) rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm1.bias""", F"""model.transformer_module.decoder.layers.{idx}.self_attn_layer_norm.bias""") ) # layernorm 2 (cross-attention layernorm) rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.weight""", F"""model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.weight""") ) rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm2.bias""", F"""model.transformer_module.decoder.layers.{idx}.encoder_attn_layer_norm.bias""") ) # layernorm 3 (final layernorm) rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.weight""", F"""model.transformer_module.decoder.layers.{idx}.final_layer_norm.weight""") ) rename_keys.append((F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.norm3.bias""", F"""model.transformer_module.decoder.layers.{idx}.final_layer_norm.bias""") ) rename_keys.append(('''sem_seg_head.predictor.transformer.decoder.norm.weight''', '''model.transformer_module.decoder.layernorm.weight''') ) rename_keys.append(('''sem_seg_head.predictor.transformer.decoder.norm.bias''', '''model.transformer_module.decoder.layernorm.bias''') ) # heads on top rename_keys.append(('''sem_seg_head.predictor.query_embed.weight''', '''model.transformer_module.queries_embedder.weight''') ) rename_keys.append(('''sem_seg_head.predictor.input_proj.weight''', '''model.transformer_module.input_projection.weight''') ) rename_keys.append(('''sem_seg_head.predictor.input_proj.bias''', '''model.transformer_module.input_projection.bias''') ) rename_keys.append(('''sem_seg_head.predictor.class_embed.weight''', '''class_predictor.weight''') ) rename_keys.append(('''sem_seg_head.predictor.class_embed.bias''', '''class_predictor.bias''') ) for i in range(3 ): rename_keys.append((F"""sem_seg_head.predictor.mask_embed.layers.{i}.weight""", F"""mask_embedder.{i}.0.weight""") ) rename_keys.append((F"""sem_seg_head.predictor.mask_embed.layers.{i}.bias""", F"""mask_embedder.{i}.0.bias""") ) # fmt: on return rename_keys def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[Any] ) -> List[str]: _lowerCamelCase = dct.pop(lowercase_ ) _lowerCamelCase = val def lowerCAmelCase_( lowercase_ : List[Any] , lowercase_ : Any ) -> Dict: _lowerCamelCase = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): _lowerCamelCase = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) _lowerCamelCase = state_dict.pop(F"""backbone.layers.{i}.blocks.{j}.attn.qkv.weight""" ) _lowerCamelCase = state_dict.pop(F"""backbone.layers.{i}.blocks.{j}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict _lowerCamelCase = in_proj_weight[:dim, :] _lowerCamelCase = in_proj_bias[: dim] _lowerCamelCase = in_proj_weight[ dim : dim * 2, : ] _lowerCamelCase = in_proj_bias[ dim : dim * 2 ] _lowerCamelCase = in_proj_weight[ -dim :, : ] _lowerCamelCase = in_proj_bias[-dim :] # fmt: on def lowerCAmelCase_( lowercase_ : Any , lowercase_ : List[Any] ) -> List[Any]: # fmt: off _lowerCamelCase = config.decoder_config.hidden_size for idx in range(config.decoder_config.decoder_layers ): # read in weights + bias of self-attention input projection layer (in the original implementation, this is a single matrix + bias) _lowerCamelCase = state_dict.pop(F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_weight""" ) _lowerCamelCase = state_dict.pop(F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict _lowerCamelCase = in_proj_weight[: hidden_size, :] _lowerCamelCase = in_proj_bias[:config.hidden_size] _lowerCamelCase = in_proj_weight[hidden_size : hidden_size * 2, :] _lowerCamelCase = in_proj_bias[hidden_size : hidden_size * 2] _lowerCamelCase = in_proj_weight[-hidden_size :, :] _lowerCamelCase = in_proj_bias[-hidden_size :] # read in weights + bias of cross-attention input projection layer (in the original implementation, this is a single matrix + bias) _lowerCamelCase = state_dict.pop(F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_weight""" ) _lowerCamelCase = state_dict.pop(F"""sem_seg_head.predictor.transformer.decoder.layers.{idx}.multihead_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict _lowerCamelCase = in_proj_weight[: hidden_size, :] _lowerCamelCase = in_proj_bias[:config.hidden_size] _lowerCamelCase = in_proj_weight[hidden_size : hidden_size * 2, :] _lowerCamelCase = in_proj_bias[hidden_size : hidden_size * 2] _lowerCamelCase = in_proj_weight[-hidden_size :, :] _lowerCamelCase = in_proj_bias[-hidden_size :] # fmt: on def lowerCAmelCase_( ) -> torch.Tensor: _lowerCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' _lowerCamelCase = Image.open(requests.get(lowercase_ , stream=lowercase_ ).raw ) return im @torch.no_grad() def lowerCAmelCase_( lowercase_ : str , lowercase_ : str , lowercase_ : str , lowercase_ : bool = False ) -> List[Any]: _lowerCamelCase = get_maskformer_config(lowercase_ ) # load original state_dict with open(lowercase_ , '''rb''' ) as f: _lowerCamelCase = pickle.load(lowercase_ ) _lowerCamelCase = data['''model'''] # for name, param in state_dict.items(): # print(name, param.shape) # rename keys _lowerCamelCase = create_rename_keys(lowercase_ ) for src, dest in rename_keys: rename_key(lowercase_ , lowercase_ , lowercase_ ) read_in_swin_q_k_v(lowercase_ , config.backbone_config ) read_in_decoder_q_k_v(lowercase_ , lowercase_ ) # update to torch tensors for key, value in state_dict.items(): _lowerCamelCase = torch.from_numpy(lowercase_ ) # load 🤗 model _lowerCamelCase = MaskFormerForInstanceSegmentation(lowercase_ ) model.eval() for name, param in model.named_parameters(): print(lowercase_ , param.shape ) _lowerCamelCase , _lowerCamelCase = model.load_state_dict(lowercase_ , strict=lowercase_ ) assert missing_keys == [ "model.pixel_level_module.encoder.model.layernorm.weight", "model.pixel_level_module.encoder.model.layernorm.bias", ] assert len(lowercase_ ) == 0, F"""Unexpected keys: {unexpected_keys}""" # verify results _lowerCamelCase = prepare_img() if "vistas" in model_name: _lowerCamelCase = 65 elif "cityscapes" in model_name: _lowerCamelCase = 6_55_35 else: _lowerCamelCase = 2_55 _lowerCamelCase = True if '''ade''' in model_name else False _lowerCamelCase = MaskFormerImageProcessor(ignore_index=lowercase_ , reduce_labels=lowercase_ ) _lowerCamelCase = image_processor(lowercase_ , return_tensors='''pt''' ) _lowerCamelCase = model(**lowercase_ ) print('''Logits:''' , outputs.class_queries_logits[0, :3, :3] ) if model_name == "maskformer-swin-tiny-ade": _lowerCamelCase = torch.tensor( [[3.6_3_5_3, -4.4_7_7_0, -2.6_0_6_5], [0.5_0_8_1, -4.2_3_9_4, -3.5_3_4_3], [2.1_9_0_9, -5.0_3_5_3, -1.9_3_2_3]] ) assert torch.allclose(outputs.class_queries_logits[0, :3, :3] , lowercase_ , atol=1e-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F"""Saving model and image processor to {pytorch_dump_folder_path}""" ) Path(lowercase_ ).mkdir(exist_ok=lowercase_ ) model.save_pretrained(lowercase_ ) image_processor.save_pretrained(lowercase_ ) if push_to_hub: print('''Pushing model and image processor to the hub...''' ) model.push_to_hub(F"""nielsr/{model_name}""" ) image_processor.push_to_hub(F"""nielsr/{model_name}""" ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''maskformer-swin-tiny-ade''', type=str, help=('''Name of the MaskFormer model you\'d like to convert''',), ) parser.add_argument( '''--checkpoint_path''', default='''/Users/nielsrogge/Documents/MaskFormer_checkpoints/MaskFormer-Swin-tiny-ADE20k/model.pkl''', type=str, help='''Path to the original state dict (.pth file).''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) __SCREAMING_SNAKE_CASE : Dict = parser.parse_args() convert_maskformer_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
661
"""simple docstring""" import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class lowerCamelCase_: '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return None class lowerCamelCase_: '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return None class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = [ # (model_name, model_kwargs) ('bert-base-cased', {}), ('gpt2', {'use_cache': False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase__ , '''tf''' , 1_2 , **lowerCamelCase__ ) @require_torch @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , **lowerCamelCase__ ) @require_torch @slow def snake_case__ ( self ): from transformers import BertModel _lowerCamelCase = ['''[UNK]''', '''[SEP]''', '''[CLS]''', '''[PAD]''', '''[MASK]''', '''some''', '''other''', '''words'''] with NamedTemporaryFile(mode='''w+t''' ) as vocab_file: vocab_file.write('''\n'''.join(lowerCamelCase__ ) ) vocab_file.flush() _lowerCamelCase = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: _lowerCamelCase = BertModel(BertConfig(vocab_size=len(lowerCamelCase__ ) ) ) model.save_pretrained(lowerCamelCase__ ) self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , lowerCamelCase__ ) @require_tf @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCamelCase = self._test_export(lowerCamelCase__ , '''tf''' , 1_2 , **lowerCamelCase__ ) _lowerCamelCase = quantize(Path(lowerCamelCase__ ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) @require_torch @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCamelCase = self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , **lowerCamelCase__ ) _lowerCamelCase = quantize(lowerCamelCase__ ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): try: # Compute path with TemporaryDirectory() as tempdir: _lowerCamelCase = Path(lowerCamelCase__ ).joinpath('''model.onnx''' ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) return path except Exception as e: self.fail(lowerCamelCase__ ) @require_torch @require_tokenizers @slow def snake_case__ ( self ): from transformers import BertModel _lowerCamelCase = BertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) _lowerCamelCase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(lowerCamelCase__ , lowerCamelCase__ , '''pt''' ) @require_tf @require_tokenizers @slow def snake_case__ ( self ): from transformers import TFBertModel _lowerCamelCase = TFBertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) _lowerCamelCase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(lowerCamelCase__ , lowerCamelCase__ , '''tf''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = FeatureExtractionPipeline(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''output_0''', '''output_1'''] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = infer_shapes(lowerCamelCase__ , lowerCamelCase__ ) # Assert all variables are present self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] , lowerCamelCase__ ) self.assertSequenceEqual(variable_names[3:] , lowerCamelCase__ ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] , {0: '''batch''', 1: '''sequence'''} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes['''output_0'''] , {0: '''batch''', 1: '''sequence'''} ) self.assertDictEqual(shapes['''output_1'''] , {0: '''batch'''} ) def snake_case__ ( self ): _lowerCamelCase = ['''input_ids''', '''attention_mask''', '''token_type_ids'''] _lowerCamelCase = {'''input_ids''': [1, 2, 3, 4], '''attention_mask''': [0, 0, 0, 0], '''token_type_ids''': [1, 1, 1, 1]} _lowerCamelCase , _lowerCamelCase = ensure_valid_input(FuncContiguousArgs() , lowerCamelCase__ , lowerCamelCase__ ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(lowerCamelCase__ ) , 3 ) # Should have exactly the same input names self.assertEqual(set(lowerCamelCase__ ) , set(lowerCamelCase__ ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(lowerCamelCase__ , (tokens['''input_ids'''], tokens['''token_type_ids'''], tokens['''attention_mask''']) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) _lowerCamelCase , _lowerCamelCase = ensure_valid_input(FuncNonContiguousArgs() , lowerCamelCase__ , lowerCamelCase__ ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(lowerCamelCase__ ) , 1 ) self.assertEqual(len(lowerCamelCase__ ) , 1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] , tokens['''input_ids'''] ) self.assertEqual(ordered_input_names[0] , '''input_ids''' ) def snake_case__ ( self ): _lowerCamelCase = generate_identified_filename(Path('''/home/something/my_fake_model.onnx''' ) , '''-test''' ) self.assertEqual('''/home/something/my_fake_model-test.onnx''' , generated.as_posix() )
661
1
"""simple docstring""" from math import sqrt import numpy as np from sympy import symbols # Coefficient # Speed of light (m/s) __SCREAMING_SNAKE_CASE : str = 2_9_9_7_9_2_4_5_8 # Symbols __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Union[str, Any] = symbols('''ct x y z''') def lowerCAmelCase_( lowercase_ : float ) -> float: if velocity > c: raise ValueError('''Speed must not exceed light speed 299,792,458 [m/s]!''' ) elif velocity < 1: # Usually the speed should be much higher than 1 (c order of magnitude) raise ValueError('''Speed must be greater than or equal to 1!''' ) return velocity / c def lowerCAmelCase_( lowercase_ : float ) -> float: return 1 / sqrt(1 - beta(lowercase_ ) ** 2 ) def lowerCAmelCase_( lowercase_ : float ) -> np.ndarray: return np.array( [ [gamma(lowercase_ ), -gamma(lowercase_ ) * beta(lowercase_ ), 0, 0], [-gamma(lowercase_ ) * beta(lowercase_ ), gamma(lowercase_ ), 0, 0], [0, 0, 1, 0], [0, 0, 0, 1], ] ) def lowerCAmelCase_( lowercase_ : float , lowercase_ : np.ndarray | None = None ) -> np.ndarray: # Ensure event is not empty if event is None: _lowerCamelCase = np.array([ct, x, y, z] ) # Symbolic four vector else: event[0] *= c # x0 is ct (speed of light * time) return transformation_matrix(lowercase_ ) @ event if __name__ == "__main__": import doctest doctest.testmod() # Example of symbolic vector: __SCREAMING_SNAKE_CASE : List[str] = transform(2_9_9_7_9_2_4_5) print('''Example of four vector: ''') print(F"""ct' = {four_vector[0]}""") print(F"""x' = {four_vector[1]}""") print(F"""y' = {four_vector[2]}""") print(F"""z' = {four_vector[3]}""") # Substitute symbols with numerical values __SCREAMING_SNAKE_CASE : Tuple = {ct: c, x: 1, y: 1, z: 1} __SCREAMING_SNAKE_CASE : int = [four_vector[i].subs(sub_dict) for i in range(4)] print(F"""\n{numerical_vector}""")
661
"""simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = {str(digit): digit**5 for digit in range(1_0)} def lowerCAmelCase_( lowercase_ : int ) -> int: return sum(DIGITS_FIFTH_POWER[digit] for digit in str(lowercase_ ) ) def lowerCAmelCase_( ) -> int: return sum( number for number in range(10_00 , 1_00_00_00 ) if number == digits_fifth_powers_sum(lowercase_ ) ) if __name__ == "__main__": print(solution())
661
1
"""simple docstring""" import importlib import torch import yaml from omegaconf import OmegaConf from taming.models.vqgan import VQModel def lowerCAmelCase_( lowercase_ : Union[str, Any] , lowercase_ : Tuple=False ) -> List[Any]: _lowerCamelCase = OmegaConf.load(lowercase_ ) if display: print(yaml.dump(OmegaConf.to_container(lowercase_ ) ) ) return config def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : Union[str, Any]=None , lowercase_ : Optional[Any]=None ) -> int: if conf_path is None: _lowerCamelCase = '''./model_checkpoints/vqgan_only.yaml''' _lowerCamelCase = load_config(lowercase_ , display=lowercase_ ) _lowerCamelCase = VQModel(**config.model.params ) if ckpt_path is None: _lowerCamelCase = '''./model_checkpoints/vqgan_only.pt''' _lowerCamelCase = torch.load(lowercase_ , map_location=lowercase_ ) if ".ckpt" in ckpt_path: _lowerCamelCase = sd['''state_dict'''] model.load_state_dict(lowercase_ , strict=lowercase_ ) model.to(lowercase_ ) del sd return model def lowerCAmelCase_( lowercase_ : List[Any] , lowercase_ : List[Any] ) -> Tuple: _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = model.encode(lowercase_ ) print(F"""VQGAN --- {model.__class__.__name__}: latent shape: {z.shape[2:]}""" ) _lowerCamelCase = model.decode(lowercase_ ) return xrec def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : Optional[int]=False ) -> Dict: _lowerCamelCase , _lowerCamelCase = string.rsplit('''.''' , 1 ) if reload: _lowerCamelCase = importlib.import_module(lowercase_ ) importlib.reload(lowercase_ ) return getattr(importlib.import_module(lowercase_ , package=lowercase_ ) , cls ) def lowerCAmelCase_( lowercase_ : List[Any] ) -> Any: if "target" not in config: raise KeyError('''Expected key `target` to instantiate.''' ) return get_obj_from_str(config['''target'''] )(**config.get('''params''' , {} ) ) def lowerCAmelCase_( lowercase_ : Any , lowercase_ : List[str] , lowercase_ : List[str]=True , lowercase_ : Any=True ) -> Optional[Any]: _lowerCamelCase = instantiate_from_config(lowercase_ ) if sd is not None: model.load_state_dict(lowercase_ ) if gpu: model.cuda() if eval_mode: model.eval() return {"model": model} def lowerCAmelCase_( lowercase_ : int , lowercase_ : Optional[int] , lowercase_ : Union[str, Any] , lowercase_ : Tuple ) -> str: # load the specified checkpoint if ckpt: _lowerCamelCase = torch.load(lowercase_ , map_location='''cpu''' ) _lowerCamelCase = pl_sd['''global_step'''] print(F"""loaded model from global step {global_step}.""" ) else: _lowerCamelCase = {'''state_dict''': None} _lowerCamelCase = None _lowerCamelCase = load_model_from_config(config.model , pl_sd['''state_dict'''] , gpu=lowercase_ , eval_mode=lowercase_ )['''model'''] return model, global_step
661
"""simple docstring""" from __future__ import annotations import os from collections.abc import Mapping __SCREAMING_SNAKE_CASE : str = tuple[int, int] class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = vertices _lowerCamelCase = { (min(lowerCamelCase__ ), max(lowerCamelCase__ )): weight for edge, weight in edges.items() } def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): self.vertices.add(edge[0] ) self.vertices.add(edge[1] ) _lowerCamelCase = weight def snake_case__ ( self ): _lowerCamelCase = Graph({min(self.vertices )} , {} ) _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 while len(subgraph.vertices ) < len(self.vertices ): _lowerCamelCase = max(self.edges.values() ) + 1 for edge, weight in self.edges.items(): if (edge[0] in subgraph.vertices) ^ (edge[1] in subgraph.vertices): if weight < min_weight: _lowerCamelCase = edge _lowerCamelCase = weight subgraph.add_edge(lowerCamelCase__ , lowerCamelCase__ ) return subgraph def lowerCAmelCase_( lowercase_ : str = "p107_network.txt" ) -> int: _lowerCamelCase = os.path.abspath(os.path.dirname(lowercase_ ) ) _lowerCamelCase = os.path.join(lowercase_ , lowercase_ ) _lowerCamelCase = {} _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 with open(lowercase_ ) as f: _lowerCamelCase = f.read().strip().split('''\n''' ) _lowerCamelCase = [line.split(''',''' ) for line in data] for edgea in range(1 , len(lowercase_ ) ): for edgea in range(lowercase_ ): if adjaceny_matrix[edgea][edgea] != "-": _lowerCamelCase = int(adjaceny_matrix[edgea][edgea] ) _lowerCamelCase = Graph(set(range(len(lowercase_ ) ) ) , lowercase_ ) _lowerCamelCase = graph.prims_algorithm() _lowerCamelCase = sum(graph.edges.values() ) _lowerCamelCase = sum(subgraph.edges.values() ) return initial_total - optimal_total if __name__ == "__main__": print(F"""{solution() = }""")
661
1
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING __SCREAMING_SNAKE_CASE : Tuple = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : List[str] = { '''SenseTime/deformable-detr''': '''https://huggingface.co/sensetime/deformable-detr/resolve/main/config.json''', # See all Deformable DETR models at https://huggingface.co/models?filter=deformable-detr } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Any = 'deformable_detr' lowercase__ : Tuple = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self , lowerCamelCase__=True , lowerCamelCase__=None , lowerCamelCase__=3 , lowerCamelCase__=3_0_0 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=6 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=8 , lowerCamelCase__=6 , lowerCamelCase__=1_0_2_4 , lowerCamelCase__=8 , lowerCamelCase__=0.0 , lowerCamelCase__=True , lowerCamelCase__="relu" , lowerCamelCase__=2_5_6 , lowerCamelCase__=0.1 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1.0 , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__="sine" , lowerCamelCase__="resnet50" , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__=4 , lowerCamelCase__=False , lowerCamelCase__=3_0_0 , lowerCamelCase__=False , lowerCamelCase__=1 , lowerCamelCase__=5 , lowerCamelCase__=2 , lowerCamelCase__=1 , lowerCamelCase__=1 , lowerCamelCase__=5 , lowerCamelCase__=2 , lowerCamelCase__=0.1 , lowerCamelCase__=0.2_5 , lowerCamelCase__=False , **lowerCamelCase__ , ): if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''' ) if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) _lowerCamelCase = CONFIG_MAPPING['''resnet'''](out_features=['''stage4'''] ) elif isinstance(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = backbone_config.get('''model_type''' ) _lowerCamelCase = CONFIG_MAPPING[backbone_model_type] _lowerCamelCase = config_class.from_dict(lowerCamelCase__ ) _lowerCamelCase = use_timm_backbone _lowerCamelCase = backbone_config _lowerCamelCase = num_channels _lowerCamelCase = num_queries _lowerCamelCase = max_position_embeddings _lowerCamelCase = d_model _lowerCamelCase = encoder_ffn_dim _lowerCamelCase = encoder_layers _lowerCamelCase = encoder_attention_heads _lowerCamelCase = decoder_ffn_dim _lowerCamelCase = decoder_layers _lowerCamelCase = decoder_attention_heads _lowerCamelCase = dropout _lowerCamelCase = attention_dropout _lowerCamelCase = activation_dropout _lowerCamelCase = activation_function _lowerCamelCase = init_std _lowerCamelCase = init_xavier_std _lowerCamelCase = encoder_layerdrop _lowerCamelCase = auxiliary_loss _lowerCamelCase = position_embedding_type _lowerCamelCase = backbone _lowerCamelCase = use_pretrained_backbone _lowerCamelCase = dilation # deformable attributes _lowerCamelCase = num_feature_levels _lowerCamelCase = encoder_n_points _lowerCamelCase = decoder_n_points _lowerCamelCase = two_stage _lowerCamelCase = two_stage_num_proposals _lowerCamelCase = with_box_refine if two_stage is True and with_box_refine is False: raise ValueError('''If two_stage is True, with_box_refine must be True.''' ) # Hungarian matcher _lowerCamelCase = class_cost _lowerCamelCase = bbox_cost _lowerCamelCase = giou_cost # Loss coefficients _lowerCamelCase = mask_loss_coefficient _lowerCamelCase = dice_loss_coefficient _lowerCamelCase = bbox_loss_coefficient _lowerCamelCase = giou_loss_coefficient _lowerCamelCase = eos_coefficient _lowerCamelCase = focal_alpha _lowerCamelCase = disable_custom_kernels super().__init__(is_encoder_decoder=lowerCamelCase__ , **lowerCamelCase__ ) @property def snake_case__ ( self ): return self.encoder_attention_heads @property def snake_case__ ( self ): return self.d_model def snake_case__ ( self ): _lowerCamelCase = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: _lowerCamelCase = self.backbone_config.to_dict() _lowerCamelCase = self.__class__.model_type return output
661
"""simple docstring""" from __future__ import annotations from math import ceil, floor, sqrt def lowerCAmelCase_( lowercase_ : int = 2_00_00_00 ) -> int: _lowerCamelCase = [0] _lowerCamelCase = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target _lowerCamelCase = 0 # the area corresponding to the grid that gives the product closest to target _lowerCamelCase = 0 # an estimate of b, using the quadratic formula _lowerCamelCase = 42 # the largest integer less than b_estimate _lowerCamelCase = 42 # the largest integer less than b_estimate _lowerCamelCase = 42 # the triangle number corresponding to b_floor _lowerCamelCase = 42 # the triangle number corresponding to b_ceil _lowerCamelCase = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): _lowerCamelCase = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 _lowerCamelCase = floor(lowercase_ ) _lowerCamelCase = ceil(lowercase_ ) _lowerCamelCase = triangle_numbers[b_floor] _lowerCamelCase = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): _lowerCamelCase = triangle_b_first_guess * triangle_a _lowerCamelCase = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): _lowerCamelCase = triangle_b_second_guess * triangle_a _lowerCamelCase = idx_a * b_ceil return area if __name__ == "__main__": print(F"""{solution() = }""")
661
1
"""simple docstring""" from sklearn.metrics import fa_score import datasets __SCREAMING_SNAKE_CASE : int = ''' The F1 score is the harmonic mean of the precision and recall. It can be computed with the equation: F1 = 2 * (precision * recall) / (precision + recall) ''' __SCREAMING_SNAKE_CASE : Optional[int] = ''' Args: predictions (`list` of `int`): Predicted labels. references (`list` of `int`): Ground truth labels. labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None. pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1. average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`. - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary. - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives. - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account. - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall. - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification). sample_weight (`list` of `float`): Sample weights Defaults to None. Returns: f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better. Examples: Example 1-A simple binary example >>> f1_metric = datasets.load_metric("f1") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0]) >>> print(results) {\'f1\': 0.5} Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`. >>> f1_metric = datasets.load_metric("f1") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0) >>> print(round(results[\'f1\'], 2)) 0.67 Example 3-The same simple binary example as in Example 1, but with `sample_weight` included. >>> f1_metric = datasets.load_metric("f1") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3]) >>> print(round(results[\'f1\'], 2)) 0.35 Example 4-A multiclass example, with different values for the `average` input. >>> predictions = [0, 2, 1, 0, 0, 1] >>> references = [0, 1, 2, 0, 1, 2] >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro") >>> print(round(results[\'f1\'], 2)) 0.27 >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro") >>> print(round(results[\'f1\'], 2)) 0.33 >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted") >>> print(round(results[\'f1\'], 2)) 0.27 >>> results = f1_metric.compute(predictions=predictions, references=references, average=None) >>> print(results) {\'f1\': array([0.8, 0. , 0. ])} ''' __SCREAMING_SNAKE_CASE : Any = ''' @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Sequence(datasets.Value('''int32''' ) ), '''references''': datasets.Sequence(datasets.Value('''int32''' ) ), } if self.config_name == '''multilabel''' else { '''predictions''': datasets.Value('''int32''' ), '''references''': datasets.Value('''int32''' ), } ) , reference_urls=['''https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html'''] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , lowerCamelCase__=1 , lowerCamelCase__="binary" , lowerCamelCase__=None ): _lowerCamelCase = fa_score( lowerCamelCase__ , lowerCamelCase__ , labels=lowerCamelCase__ , pos_label=lowerCamelCase__ , average=lowerCamelCase__ , sample_weight=lowerCamelCase__ ) return {"f1": float(lowerCamelCase__ ) if score.size == 1 else score}
661
"""simple docstring""" import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , *lowerCamelCase__ , **lowerCamelCase__ ): warnings.warn( '''The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use DeiTImageProcessor instead.''' , lowerCamelCase__ , ) super().__init__(*lowerCamelCase__ , **lowerCamelCase__ )
661
1
"""simple docstring""" import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Dict = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''mask_emb''': '''masked_spec_embed''', } __SCREAMING_SNAKE_CASE : List[str] = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', ] def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : Optional[int] , lowercase_ : Any , lowercase_ : Tuple , lowercase_ : Optional[Any] ) -> int: for attribute in key.split('''.''' ): _lowerCamelCase = getattr(lowercase_ , lowercase_ ) if weight_type is not None: _lowerCamelCase = getattr(lowercase_ , lowercase_ ).shape else: _lowerCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( F"""Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be""" F""" {value.shape} for {full_name}""" ) if weight_type == "weight": _lowerCamelCase = value elif weight_type == "weight_g": _lowerCamelCase = value elif weight_type == "weight_v": _lowerCamelCase = value elif weight_type == "bias": _lowerCamelCase = value else: _lowerCamelCase = value logger.info(F"""{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.""" ) def lowerCAmelCase_( lowercase_ : Union[str, Any] , lowercase_ : Optional[Any] ) -> Tuple: _lowerCamelCase = [] _lowerCamelCase = fairseq_model.state_dict() _lowerCamelCase = hf_model.feature_extractor _lowerCamelCase = hf_model.adapter for name, value in fairseq_dict.items(): _lowerCamelCase = False if "conv_layers" in name: load_conv_layer( lowercase_ , lowercase_ , lowercase_ , lowercase_ , hf_model.config.feat_extract_norm == '''group''' , ) _lowerCamelCase = True elif any(x in name for x in ['''adaptor''', '''w2v_encoder.proj.''', '''w2v_proj_ln.'''] ): load_adapter(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) _lowerCamelCase = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: _lowerCamelCase = True if "*" in mapped_key: _lowerCamelCase = name.split(lowercase_ )[0].split('''.''' )[-2] _lowerCamelCase = mapped_key.replace('''*''' , lowercase_ ) if "weight_g" in name: _lowerCamelCase = '''weight_g''' elif "weight_v" in name: _lowerCamelCase = '''weight_v''' elif "bias" in name: _lowerCamelCase = '''bias''' elif "weight" in name: _lowerCamelCase = '''weight''' else: _lowerCamelCase = None set_recursively(lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ) continue if not is_used: unused_weights.append(lowercase_ ) logger.warning(F"""Unused weights: {unused_weights}""" ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Optional[int] , lowercase_ : str , lowercase_ : str ) -> str: _lowerCamelCase = full_name.split('''conv_layers.''' )[-1] _lowerCamelCase = name.split('''.''' ) _lowerCamelCase = int(items[0] ) _lowerCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) _lowerCamelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) _lowerCamelCase = value logger.info(F"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( F"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) _lowerCamelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( F"""{full_name} has size {value.shape}, but""" F""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) _lowerCamelCase = value logger.info(F"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(lowercase_ ) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : int , lowercase_ : Optional[int] , lowercase_ : Optional[int] ) -> List[str]: _lowerCamelCase = full_name.split('''adaptor.''' )[-1] _lowerCamelCase = name.split('''.''' ) if items[1].isdigit(): _lowerCamelCase = int(items[1] ) else: _lowerCamelCase = None if "adaptor" not in full_name: if "proj_ln" in full_name: # has to be layer norm if "bias" in name: assert ( value.shape == adapter.proj_layer_norm.bias.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj_layer_norm.bias.data.shape} was found.""" _lowerCamelCase = value logger.info(F"""Adapter proj layer norm bias was initialized from {full_name}.""" ) if "weight" in name: assert ( value.shape == adapter.proj_layer_norm.weight.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj_layer_norm.weight.data.shape} was found.""" _lowerCamelCase = value else: # has to be projection layer if "bias" in name: assert ( value.shape == adapter.proj.bias.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj.bias.data.shape} was found.""" _lowerCamelCase = value logger.info(F"""Adapter proj layer bias was initialized from {full_name}.""" ) if "weight" in name: assert ( value.shape == adapter.proj.weight.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.proj.weight.data.shape} was found.""" _lowerCamelCase = value logger.info(F"""Adapter proj layer weight was initialized from {full_name}.""" ) elif isinstance(lowercase_ , lowercase_ ): if "bias" in name: assert ( value.shape == adapter.layers[layer_id].conv.bias.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.bias.data.shape} was found.""" _lowerCamelCase = value logger.info(F"""Adapter layer {layer_id} bias was initialized from {full_name}.""" ) elif "weight" in name: assert ( value.shape == adapter.layers[layer_id].conv.weight.data.shape ), F"""{full_name} has size {value.shape}, but {adapter.layers[layer_id].conv.weight.data.shape} was found.""" _lowerCamelCase = value logger.info(F"""Adapter layer {layer_id} bias was initialized from {full_name}.""" ) else: unused_weights.append(lowercase_ ) def lowerCAmelCase_( lowercase_ : List[Any] ) -> List[str]: _lowerCamelCase , _lowerCamelCase = emb.weight.shape _lowerCamelCase = nn.Linear(lowercase_ , lowercase_ , bias=lowercase_ ) _lowerCamelCase = emb.weight.data return lin_layer @torch.no_grad() def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : str , lowercase_ : Any , lowercase_ : Optional[int] , lowercase_ : Dict , lowercase_ : int , lowercase_ : List[Any] , lowercase_ : Optional[Any] , lowercase_ : str , lowercase_ : Any , lowercase_ : List[Any] , ) -> List[str]: _lowerCamelCase = WavaVecaConfig.from_pretrained( lowercase_ , add_adapter=lowercase_ , adapter_stride=lowercase_ , adapter_kernel_size=lowercase_ , use_auth_token=lowercase_ , output_hidden_size=lowercase_ , ) _lowerCamelCase = MBartConfig.from_pretrained(lowercase_ ) # load model _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={ '''config_yaml''': config_yaml_path, '''data''': '''/'''.join(dict_path.split('''/''' )[:-1] ), '''w2v_path''': checkpoint_path, '''load_pretrained_decoder_from''': None, } , ) _lowerCamelCase = model[0].eval() # load feature extractor _lowerCamelCase = WavaVecaFeatureExtractor.from_pretrained(lowercase_ , use_auth_token=lowercase_ ) # set weights for wav2vec2 encoder _lowerCamelCase = WavaVecaModel(lowercase_ ) recursively_load_weights_wavaveca(model.encoder , lowercase_ ) # load decoder weights _lowerCamelCase = MBartForCausalLM(lowercase_ ) _lowerCamelCase , _lowerCamelCase = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=lowercase_ ) logger.warning(F"""The following keys are missing when loading the decoder weights: {missing_keys}""" ) logger.warning(F"""The following keys are unexpected when loading the decoder weights: {unexpected_keys}""" ) _lowerCamelCase = SpeechEncoderDecoderModel(encoder=lowercase_ , decoder=lowercase_ ) _lowerCamelCase = False _lowerCamelCase = MBartaaTokenizer(lowercase_ ) tokenizer.save_pretrained(lowercase_ ) _lowerCamelCase = hf_wavavec.config.to_dict() _lowerCamelCase = tokenizer.pad_token_id _lowerCamelCase = tokenizer.bos_token_id _lowerCamelCase = tokenizer.eos_token_id _lowerCamelCase = '''mbart50''' _lowerCamelCase = '''wav2vec2''' _lowerCamelCase = tokenizer.eos_token_id _lowerCamelCase = 25_00_04 _lowerCamelCase = tokenizer.eos_token_id _lowerCamelCase = SpeechEncoderDecoderConfig.from_dict(lowercase_ ) hf_wavavec.save_pretrained(lowercase_ ) feature_extractor.save_pretrained(lowercase_ ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Tuple = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_yaml_path''', default=None, type=str, help='''Path to yaml file of fine-tuned model''') parser.add_argument( '''--encoder_config_path''', default='''facebook/wav2vec2-xls-r-1b''', type=str, help='''Path to hf encoder wav2vec2 checkpoint config''', ) parser.add_argument( '''--decoder_config_path''', default='''facebook/mbart-large-50-one-to-many-mmt''', type=str, help='''Path to hf decoder checkpoint config''', ) parser.add_argument('''--add_adapter''', default=True, type=bool, help='''whethere to add model adapter layers''') parser.add_argument('''--adapter_stride''', default=2, type=int, help='''stride of adapter layers''') parser.add_argument('''--adapter_kernel_size''', default=3, type=int, help='''kernel size of adapter layers''') parser.add_argument('''--encoder_output_dim''', default=1_0_2_4, type=int, help='''encoder output dim''') parser.add_argument('''--start_token_id''', default=2_5_0_0_0_4, type=int, help='''`decoder_start_token_id` of model config''') __SCREAMING_SNAKE_CASE : List[Any] = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, args.config_yaml_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, add_adapter=args.add_adapter, adapter_kernel_size=args.adapter_kernel_size, adapter_stride=args.adapter_stride, decoder_start_token_id=args.start_token_id, encoder_output_dim=args.encoder_output_dim, )
661
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Optional[Any] = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''vocab_file''': { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json''' ), '''distilbert-base-german-cased''': ( '''https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json''' ), '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json''' ), }, } __SCREAMING_SNAKE_CASE : str = { '''distilbert-base-uncased''': 5_1_2, '''distilbert-base-uncased-distilled-squad''': 5_1_2, '''distilbert-base-cased''': 5_1_2, '''distilbert-base-cased-distilled-squad''': 5_1_2, '''distilbert-base-german-cased''': 5_1_2, '''distilbert-base-multilingual-cased''': 5_1_2, } __SCREAMING_SNAKE_CASE : str = { '''distilbert-base-uncased''': {'''do_lower_case''': True}, '''distilbert-base-uncased-distilled-squad''': {'''do_lower_case''': True}, '''distilbert-base-cased''': {'''do_lower_case''': False}, '''distilbert-base-cased-distilled-squad''': {'''do_lower_case''': False}, '''distilbert-base-german-cased''': {'''do_lower_case''': False}, '''distilbert-base-multilingual-cased''': {'''do_lower_case''': False}, } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[Any] = VOCAB_FILES_NAMES lowercase__ : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP lowercase__ : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : str = PRETRAINED_INIT_CONFIGURATION lowercase__ : int = ['input_ids', 'attention_mask'] lowercase__ : Tuple = DistilBertTokenizer def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__="[UNK]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ): super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) _lowerCamelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , lowerCamelCase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' , lowerCamelCase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , lowerCamelCase__ ) != tokenize_chinese_chars ): _lowerCamelCase = getattr(lowerCamelCase__ , normalizer_state.pop('''type''' ) ) _lowerCamelCase = do_lower_case _lowerCamelCase = strip_accents _lowerCamelCase = tokenize_chinese_chars _lowerCamelCase = normalizer_class(**lowerCamelCase__ ) _lowerCamelCase = do_lower_case def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=None ): _lowerCamelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
661
1
"""simple docstring""" import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all feature extractors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...feature_extraction_utils import FeatureExtractionMixin from ...utils import CONFIG_NAME, FEATURE_EXTRACTOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Any = OrderedDict( [ ('''audio-spectrogram-transformer''', '''ASTFeatureExtractor'''), ('''beit''', '''BeitFeatureExtractor'''), ('''chinese_clip''', '''ChineseCLIPFeatureExtractor'''), ('''clap''', '''ClapFeatureExtractor'''), ('''clip''', '''CLIPFeatureExtractor'''), ('''clipseg''', '''ViTFeatureExtractor'''), ('''conditional_detr''', '''ConditionalDetrFeatureExtractor'''), ('''convnext''', '''ConvNextFeatureExtractor'''), ('''cvt''', '''ConvNextFeatureExtractor'''), ('''data2vec-audio''', '''Wav2Vec2FeatureExtractor'''), ('''data2vec-vision''', '''BeitFeatureExtractor'''), ('''deformable_detr''', '''DeformableDetrFeatureExtractor'''), ('''deit''', '''DeiTFeatureExtractor'''), ('''detr''', '''DetrFeatureExtractor'''), ('''dinat''', '''ViTFeatureExtractor'''), ('''donut-swin''', '''DonutFeatureExtractor'''), ('''dpt''', '''DPTFeatureExtractor'''), ('''encodec''', '''EncodecFeatureExtractor'''), ('''flava''', '''FlavaFeatureExtractor'''), ('''glpn''', '''GLPNFeatureExtractor'''), ('''groupvit''', '''CLIPFeatureExtractor'''), ('''hubert''', '''Wav2Vec2FeatureExtractor'''), ('''imagegpt''', '''ImageGPTFeatureExtractor'''), ('''layoutlmv2''', '''LayoutLMv2FeatureExtractor'''), ('''layoutlmv3''', '''LayoutLMv3FeatureExtractor'''), ('''levit''', '''LevitFeatureExtractor'''), ('''maskformer''', '''MaskFormerFeatureExtractor'''), ('''mctct''', '''MCTCTFeatureExtractor'''), ('''mobilenet_v1''', '''MobileNetV1FeatureExtractor'''), ('''mobilenet_v2''', '''MobileNetV2FeatureExtractor'''), ('''mobilevit''', '''MobileViTFeatureExtractor'''), ('''nat''', '''ViTFeatureExtractor'''), ('''owlvit''', '''OwlViTFeatureExtractor'''), ('''perceiver''', '''PerceiverFeatureExtractor'''), ('''poolformer''', '''PoolFormerFeatureExtractor'''), ('''regnet''', '''ConvNextFeatureExtractor'''), ('''resnet''', '''ConvNextFeatureExtractor'''), ('''segformer''', '''SegformerFeatureExtractor'''), ('''sew''', '''Wav2Vec2FeatureExtractor'''), ('''sew-d''', '''Wav2Vec2FeatureExtractor'''), ('''speech_to_text''', '''Speech2TextFeatureExtractor'''), ('''speecht5''', '''SpeechT5FeatureExtractor'''), ('''swiftformer''', '''ViTFeatureExtractor'''), ('''swin''', '''ViTFeatureExtractor'''), ('''swinv2''', '''ViTFeatureExtractor'''), ('''table-transformer''', '''DetrFeatureExtractor'''), ('''timesformer''', '''VideoMAEFeatureExtractor'''), ('''tvlt''', '''TvltFeatureExtractor'''), ('''unispeech''', '''Wav2Vec2FeatureExtractor'''), ('''unispeech-sat''', '''Wav2Vec2FeatureExtractor'''), ('''van''', '''ConvNextFeatureExtractor'''), ('''videomae''', '''VideoMAEFeatureExtractor'''), ('''vilt''', '''ViltFeatureExtractor'''), ('''vit''', '''ViTFeatureExtractor'''), ('''vit_mae''', '''ViTFeatureExtractor'''), ('''vit_msn''', '''ViTFeatureExtractor'''), ('''wav2vec2''', '''Wav2Vec2FeatureExtractor'''), ('''wav2vec2-conformer''', '''Wav2Vec2FeatureExtractor'''), ('''wavlm''', '''Wav2Vec2FeatureExtractor'''), ('''whisper''', '''WhisperFeatureExtractor'''), ('''xclip''', '''CLIPFeatureExtractor'''), ('''yolos''', '''YolosFeatureExtractor'''), ] ) __SCREAMING_SNAKE_CASE : List[str] = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FEATURE_EXTRACTOR_MAPPING_NAMES) def lowerCAmelCase_( lowercase_ : str ) -> Tuple: for module_name, extractors in FEATURE_EXTRACTOR_MAPPING_NAMES.items(): if class_name in extractors: _lowerCamelCase = model_type_to_module_name(lowercase_ ) _lowerCamelCase = importlib.import_module(F""".{module_name}""" , '''transformers.models''' ) try: return getattr(lowercase_ , lowercase_ ) except AttributeError: continue for _, extractor in FEATURE_EXTRACTOR_MAPPING._extra_content.items(): if getattr(lowercase_ , '''__name__''' , lowercase_ ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. _lowerCamelCase = importlib.import_module('''transformers''' ) if hasattr(lowercase_ , lowercase_ ): return getattr(lowercase_ , lowercase_ ) return None def lowerCAmelCase_( lowercase_ : Union[str, os.PathLike] , lowercase_ : Optional[Union[str, os.PathLike]] = None , lowercase_ : bool = False , lowercase_ : bool = False , lowercase_ : Optional[Dict[str, str]] = None , lowercase_ : Optional[Union[bool, str]] = None , lowercase_ : Optional[str] = None , lowercase_ : bool = False , **lowercase_ : Tuple , ) -> List[str]: _lowerCamelCase = get_file_from_repo( lowercase_ , lowercase_ , cache_dir=lowercase_ , force_download=lowercase_ , resume_download=lowercase_ , proxies=lowercase_ , use_auth_token=lowercase_ , revision=lowercase_ , local_files_only=lowercase_ , ) if resolved_config_file is None: logger.info( '''Could not locate the feature extractor configuration file, will try to use the model config instead.''' ) return {} with open(lowercase_ , encoding='''utf-8''' ) as reader: return json.load(lowercase_ ) class lowerCamelCase_: '''simple docstring''' def __init__( self ): raise EnvironmentError( '''AutoFeatureExtractor is designed to be instantiated ''' '''using the `AutoFeatureExtractor.from_pretrained(pretrained_model_name_or_path)` method.''' ) @classmethod @replace_list_option_in_docstrings(lowerCamelCase__ ) def snake_case__ ( cls , lowerCamelCase__ , **lowerCamelCase__ ): _lowerCamelCase = kwargs.pop('''config''' , lowerCamelCase__ ) _lowerCamelCase = kwargs.pop('''trust_remote_code''' , lowerCamelCase__ ) _lowerCamelCase = True _lowerCamelCase , _lowerCamelCase = FeatureExtractionMixin.get_feature_extractor_dict(lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = config_dict.get('''feature_extractor_type''' , lowerCamelCase__ ) _lowerCamelCase = None if "AutoFeatureExtractor" in config_dict.get('''auto_map''' , {} ): _lowerCamelCase = config_dict['''auto_map''']['''AutoFeatureExtractor'''] # If we don't find the feature extractor class in the feature extractor config, let's try the model config. if feature_extractor_class is None and feature_extractor_auto_map is None: if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = AutoConfig.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) # It could be in `config.feature_extractor_type`` _lowerCamelCase = getattr(lowerCamelCase__ , '''feature_extractor_type''' , lowerCamelCase__ ) if hasattr(lowerCamelCase__ , '''auto_map''' ) and "AutoFeatureExtractor" in config.auto_map: _lowerCamelCase = config.auto_map['''AutoFeatureExtractor'''] if feature_extractor_class is not None: _lowerCamelCase = feature_extractor_class_from_name(lowerCamelCase__ ) _lowerCamelCase = feature_extractor_auto_map is not None _lowerCamelCase = feature_extractor_class is not None or type(lowerCamelCase__ ) in FEATURE_EXTRACTOR_MAPPING _lowerCamelCase = resolve_trust_remote_code( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) if has_remote_code and trust_remote_code: _lowerCamelCase = get_class_from_dynamic_module( lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = kwargs.pop('''code_revision''' , lowerCamelCase__ ) if os.path.isdir(lowerCamelCase__ ): feature_extractor_class.register_for_auto_class() return feature_extractor_class.from_dict(lowerCamelCase__ , **lowerCamelCase__ ) elif feature_extractor_class is not None: return feature_extractor_class.from_dict(lowerCamelCase__ , **lowerCamelCase__ ) # Last try: we use the FEATURE_EXTRACTOR_MAPPING. elif type(lowerCamelCase__ ) in FEATURE_EXTRACTOR_MAPPING: _lowerCamelCase = FEATURE_EXTRACTOR_MAPPING[type(lowerCamelCase__ )] return feature_extractor_class.from_dict(lowerCamelCase__ , **lowerCamelCase__ ) raise ValueError( F"""Unrecognized feature extractor in {pretrained_model_name_or_path}. Should have a """ F"""`feature_extractor_type` key in its {FEATURE_EXTRACTOR_NAME} of {CONFIG_NAME}, or one of the following """ F"""`model_type` keys in its {CONFIG_NAME}: {", ".join(c for c in FEATURE_EXTRACTOR_MAPPING_NAMES.keys() )}""" ) @staticmethod def snake_case__ ( lowerCamelCase__ , lowerCamelCase__ ): FEATURE_EXTRACTOR_MAPPING.register(lowerCamelCase__ , lowerCamelCase__ )
661
"""simple docstring""" from math import sqrt import numpy as np from sympy import symbols # Coefficient # Speed of light (m/s) __SCREAMING_SNAKE_CASE : str = 2_9_9_7_9_2_4_5_8 # Symbols __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Union[str, Any] = symbols('''ct x y z''') def lowerCAmelCase_( lowercase_ : float ) -> float: if velocity > c: raise ValueError('''Speed must not exceed light speed 299,792,458 [m/s]!''' ) elif velocity < 1: # Usually the speed should be much higher than 1 (c order of magnitude) raise ValueError('''Speed must be greater than or equal to 1!''' ) return velocity / c def lowerCAmelCase_( lowercase_ : float ) -> float: return 1 / sqrt(1 - beta(lowercase_ ) ** 2 ) def lowerCAmelCase_( lowercase_ : float ) -> np.ndarray: return np.array( [ [gamma(lowercase_ ), -gamma(lowercase_ ) * beta(lowercase_ ), 0, 0], [-gamma(lowercase_ ) * beta(lowercase_ ), gamma(lowercase_ ), 0, 0], [0, 0, 1, 0], [0, 0, 0, 1], ] ) def lowerCAmelCase_( lowercase_ : float , lowercase_ : np.ndarray | None = None ) -> np.ndarray: # Ensure event is not empty if event is None: _lowerCamelCase = np.array([ct, x, y, z] ) # Symbolic four vector else: event[0] *= c # x0 is ct (speed of light * time) return transformation_matrix(lowercase_ ) @ event if __name__ == "__main__": import doctest doctest.testmod() # Example of symbolic vector: __SCREAMING_SNAKE_CASE : List[str] = transform(2_9_9_7_9_2_4_5) print('''Example of four vector: ''') print(F"""ct' = {four_vector[0]}""") print(F"""x' = {four_vector[1]}""") print(F"""y' = {four_vector[2]}""") print(F"""z' = {four_vector[3]}""") # Substitute symbols with numerical values __SCREAMING_SNAKE_CASE : Tuple = {ct: c, x: 1, y: 1, z: 1} __SCREAMING_SNAKE_CASE : int = [four_vector[i].subs(sub_dict) for i in range(4)] print(F"""\n{numerical_vector}""")
661
1
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Dict = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Union[str, Any] = XLMRobertaTokenizer lowercase__ : Optional[int] = XLMRobertaTokenizerFast lowercase__ : List[str] = True lowercase__ : Union[str, Any] = True def snake_case__ ( self ): super().setUp() # We have a SentencePiece fixture for testing _lowerCamelCase = XLMRobertaTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case__ ( self ): _lowerCamelCase = '''<pad>''' _lowerCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(lowerCamelCase__ ) , 1_0_0_2 ) def snake_case__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_2 ) def snake_case__ ( self ): _lowerCamelCase = XLMRobertaTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) _lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(lowerCamelCase__ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) _lowerCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) _lowerCamelCase = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) _lowerCamelCase = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def snake_case__ ( self ): if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _lowerCamelCase = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-xlm-roberta''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _lowerCamelCase = self.rust_tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = self.tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) _lowerCamelCase = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(lowerCamelCase__ , lowerCamelCase__ ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(lowerCamelCase__ ) # Save tokenizer rust, legacy_format=True _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ , legacy_format=lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it save with the same files self.assertSequenceEqual(lowerCamelCase__ , lowerCamelCase__ ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) shutil.rmtree(lowerCamelCase__ ) # Save tokenizer rust, legacy_format=False _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ , legacy_format=lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) shutil.rmtree(lowerCamelCase__ ) @cached_property def snake_case__ ( self ): return XLMRobertaTokenizer.from_pretrained('''xlm-roberta-base''' ) def snake_case__ ( self ): with tempfile.NamedTemporaryFile() as f: shutil.copyfile(lowerCamelCase__ , f.name ) _lowerCamelCase = XLMRobertaTokenizer(f.name , keep_accents=lowerCamelCase__ ) _lowerCamelCase = pickle.dumps(lowerCamelCase__ ) pickle.loads(lowerCamelCase__ ) def snake_case__ ( self ): if not self.test_rust_tokenizer: return _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = '''I was born in 92000, and this is falsé.''' _lowerCamelCase = tokenizer.tokenize(lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = tokenizer.encode(lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): _lowerCamelCase = '''Hello World!''' _lowerCamelCase = [0, 3_5_3_7_8, 6_6_6_1, 3_8, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(lowerCamelCase__ , self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def snake_case__ ( self ): _lowerCamelCase = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) _lowerCamelCase = [ 0, 3_2_9_3, 8_3, 1_0, 4_5_5_2, 4_9_8_9, 7_9_8_6, 6_7_8, 1_0, 5_9_1_5, 1_1_1, 1_7_9_4_5_9, 1_2_4_8_5_0, 4, 6_0_4_4, 2_3_7, 1_2, 6, 5, 6, 4, 6_7_8_0, 7_0_5, 1_5, 1_3_8_8, 4_4, 3_7_8, 1_0_1_1_4, 7_1_1, 1_5_2, 2_0, 6, 5, 2_2_3_7_6, 6_4_2, 1_2_2_1, 1_5_1_9_0, 3_4_1_5_3, 4_5_0, 5_6_0_8, 9_5_9, 1_1_1_9, 5_7_7_0_2, 1_3_6, 1_8_6, 4_7, 1_0_9_8, 2_9_3_6_7, 4_7, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6_0_4_4, 2_3_7, 6_2_8_4, 5_0_9_0_1, 5_2_8, 3_1, 9_0, 3_4, 9_2_7, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(lowerCamelCase__ , self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def snake_case__ ( self ): # fmt: off _lowerCamelCase = {'''input_ids''': [[0, 1_1_0_6_2, 8_2_7_7_2, 7, 1_5, 8_2_7_7_2, 5_3_8, 5_1_5_2_9, 2_3_7, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 2_1_5_1_7_5, 1_3_1_4, 1_3_6, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 5_6_3_5_9, 4_2, 1_2_2_0_0_9, 9, 1_6_4_6_6, 1_6, 8_7_3_4_4, 4_5_3_7, 9, 4_7_1_7, 7_8_3_8_1, 6, 1_5_9_9_5_8, 7, 1_5, 2_4_4_8_0, 6_1_8, 4, 5_2_7, 2_2_6_9_3, 5_4_2_8, 4, 2_7_7_7, 2_4_4_8_0, 9_8_7_4, 4, 4_3_5_2_3, 5_9_4, 4, 8_0_3, 1_8_3_9_2, 3_3_1_8_9, 1_8, 4, 4_3_5_2_3, 2_4_4_4_7, 1_2_3_9_9, 1_0_0, 2_4_9_5_5, 8_3_6_5_8, 9_6_2_6, 1_4_4_0_5_7, 1_5, 8_3_9, 2_2_3_3_5, 1_6, 1_3_6, 2_4_9_5_5, 8_3_6_5_8, 8_3_4_7_9, 1_5, 3_9_1_0_2, 7_2_4, 1_6, 6_7_8, 6_4_5, 2_7_8_9, 1_3_2_8, 4_5_8_9, 4_2, 1_2_2_0_0_9, 1_1_5_7_7_4, 2_3, 8_0_5, 1_3_2_8, 4_6_8_7_6, 7, 1_3_6, 5_3_8_9_4, 1_9_4_0, 4_2_2_2_7, 4_1_1_5_9, 1_7_7_2_1, 8_2_3, 4_2_5, 4, 2_7_5_1_2, 9_8_7_2_2, 2_0_6, 1_3_6, 5_5_3_1, 4_9_7_0, 9_1_9, 1_7_3_3_6, 5, 2], [0, 2_0_0_8_0, 6_1_8, 8_3, 8_2_7_7_5, 4_7, 4_7_9, 9, 1_5_1_7, 7_3, 5_3_8_9_4, 3_3_3, 8_0_5_8_1, 1_1_0_1_1_7, 1_8_8_1_1, 5_2_5_6, 1_2_9_5, 5_1, 1_5_2_5_2_6, 2_9_7, 7_9_8_6, 3_9_0, 1_2_4_4_1_6, 5_3_8, 3_5_4_3_1, 2_1_4, 9_8, 1_5_0_4_4, 2_5_7_3_7, 1_3_6, 7_1_0_8, 4_3_7_0_1, 2_3, 7_5_6, 1_3_5_3_5_5, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_8_1, 6_3_7_7_3, 1_1_9_4_5_5, 6, 1_4_7_7_9_7, 8_8_2_0_3, 7, 6_4_5, 7_0, 2_1, 3_2_8_5, 1_0_2_6_9, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ , model_name='''xlm-roberta-base''' , revision='''d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3''' , )
661
"""simple docstring""" from __future__ import annotations __SCREAMING_SNAKE_CASE : Dict = 1.6_0_2_1e-1_9 # units = C def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float , ) -> tuple[str, float]: if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif conductivity < 0: raise ValueError('''Conductivity cannot be negative''' ) elif electron_conc < 0: raise ValueError('''Electron concentration cannot be negative''' ) elif mobility < 0: raise ValueError('''mobility cannot be negative''' ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
661
1
"""simple docstring""" import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) __SCREAMING_SNAKE_CASE : Optional[Any] = { '''iou_prediction_head.layers.0''': '''iou_prediction_head.proj_in''', '''iou_prediction_head.layers.1''': '''iou_prediction_head.layers.0''', '''iou_prediction_head.layers.2''': '''iou_prediction_head.proj_out''', '''mask_decoder.output_upscaling.0''': '''mask_decoder.upscale_conv1''', '''mask_decoder.output_upscaling.1''': '''mask_decoder.upscale_layer_norm''', '''mask_decoder.output_upscaling.3''': '''mask_decoder.upscale_conv2''', '''mask_downscaling.0''': '''mask_embed.conv1''', '''mask_downscaling.1''': '''mask_embed.layer_norm1''', '''mask_downscaling.3''': '''mask_embed.conv2''', '''mask_downscaling.4''': '''mask_embed.layer_norm2''', '''mask_downscaling.6''': '''mask_embed.conv3''', '''point_embeddings''': '''point_embed''', '''pe_layer.positional_encoding_gaussian_matrix''': '''shared_embedding.positional_embedding''', '''image_encoder''': '''vision_encoder''', '''neck.0''': '''neck.conv1''', '''neck.1''': '''neck.layer_norm1''', '''neck.2''': '''neck.conv2''', '''neck.3''': '''neck.layer_norm2''', '''patch_embed.proj''': '''patch_embed.projection''', '''.norm''': '''.layer_norm''', '''blocks''': '''layers''', } def lowerCAmelCase_( lowercase_ : str ) -> Dict: _lowerCamelCase = {} state_dict.pop('''pixel_mean''' , lowercase_ ) state_dict.pop('''pixel_std''' , lowercase_ ) _lowerCamelCase = r'''.*.output_hypernetworks_mlps.(\d+).layers.(\d+).*''' for key, value in state_dict.items(): for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: _lowerCamelCase = key.replace(lowercase_ , lowercase_ ) if re.match(lowercase_ , lowercase_ ): _lowerCamelCase = int(re.match(lowercase_ , lowercase_ ).group(2 ) ) if layer_nb == 0: _lowerCamelCase = key.replace('''layers.0''' , '''proj_in''' ) elif layer_nb == 1: _lowerCamelCase = key.replace('''layers.1''' , '''layers.0''' ) elif layer_nb == 2: _lowerCamelCase = key.replace('''layers.2''' , '''proj_out''' ) _lowerCamelCase = value _lowerCamelCase = model_state_dict[ '''prompt_encoder.shared_embedding.positional_embedding''' ] return model_state_dict def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : List[Any]="ybelkada/segment-anything" ) -> str: _lowerCamelCase = hf_hub_download(lowercase_ , F"""checkpoints/{model_name}.pth""" ) if "sam_vit_b" in model_name: _lowerCamelCase = SamConfig() elif "sam_vit_l" in model_name: _lowerCamelCase = SamVisionConfig( hidden_size=10_24 , num_hidden_layers=24 , num_attention_heads=16 , global_attn_indexes=[5, 11, 17, 23] , ) _lowerCamelCase = SamConfig( vision_config=lowercase_ , ) elif "sam_vit_h" in model_name: _lowerCamelCase = SamVisionConfig( hidden_size=12_80 , num_hidden_layers=32 , num_attention_heads=16 , global_attn_indexes=[7, 15, 23, 31] , ) _lowerCamelCase = SamConfig( vision_config=lowercase_ , ) _lowerCamelCase = torch.load(lowercase_ , map_location='''cpu''' ) _lowerCamelCase = replace_keys(lowercase_ ) _lowerCamelCase = SamImageProcessor() _lowerCamelCase = SamProcessor(image_processor=lowercase_ ) _lowerCamelCase = SamModel(lowercase_ ) hf_model.load_state_dict(lowercase_ ) _lowerCamelCase = hf_model.to('''cuda''' ) _lowerCamelCase = '''https://huggingface.co/ybelkada/segment-anything/resolve/main/assets/car.png''' _lowerCamelCase = Image.open(requests.get(lowercase_ , stream=lowercase_ ).raw ).convert('''RGB''' ) _lowerCamelCase = [[[4_00, 6_50]]] _lowerCamelCase = [[1]] _lowerCamelCase = processor(images=np.array(lowercase_ ) , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): _lowerCamelCase = hf_model(**lowercase_ ) _lowerCamelCase = output.iou_scores.squeeze() if model_name == "sam_vit_h_4b8939": assert scores[-1].item() == 0.5_7_9_8_9_0_2_5_1_1_5_9_6_6_8 _lowerCamelCase = processor( images=np.array(lowercase_ ) , input_points=lowercase_ , input_labels=lowercase_ , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): _lowerCamelCase = hf_model(**lowercase_ ) _lowerCamelCase = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_7_1_2_6_0_3_0_9_2_1_9_3_6_0_4 _lowerCamelCase = ((75, 2_75, 17_25, 8_50),) _lowerCamelCase = processor(images=np.array(lowercase_ ) , input_boxes=lowercase_ , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): _lowerCamelCase = hf_model(**lowercase_ ) _lowerCamelCase = output.iou_scores.squeeze() assert scores[-1].item() == 0.8_6_8_6_0_1_5_6_0_5_9_2_6_5_1_4 # Test with 2 points and 1 image. _lowerCamelCase = [[[4_00, 6_50], [8_00, 6_50]]] _lowerCamelCase = [[1, 1]] _lowerCamelCase = processor( images=np.array(lowercase_ ) , input_points=lowercase_ , input_labels=lowercase_ , return_tensors='''pt''' ).to('''cuda''' ) with torch.no_grad(): _lowerCamelCase = hf_model(**lowercase_ ) _lowerCamelCase = output.iou_scores.squeeze() assert scores[-1].item() == 0.9_9_3_6_0_4_7_7_9_2_4_3_4_6_9_2 if __name__ == "__main__": __SCREAMING_SNAKE_CASE : int = argparse.ArgumentParser() __SCREAMING_SNAKE_CASE : List[Any] = ['''sam_vit_b_01ec64''', '''sam_vit_h_4b8939''', '''sam_vit_l_0b3195'''] parser.add_argument( '''--model_name''', default='''sam_vit_h_4b8939''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether to push the model and processor to the hub after converting''', ) parser.add_argument( '''--model_hub_id''', default='''ybelkada/segment-anything''', choices=choices, type=str, help='''Path to hf config.json of model to convert''', ) __SCREAMING_SNAKE_CASE : Any = parser.parse_args() convert_sam_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub, args.model_hub_id)
661
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): _lowerCamelCase = { '''task_specific_params''': { '''summarization''': {'''length_penalty''': 1.0, '''max_length''': 1_2_8, '''min_length''': 1_2, '''num_beams''': 4}, '''summarization_cnn''': {'''length_penalty''': 2.0, '''max_length''': 1_4_2, '''min_length''': 5_6, '''num_beams''': 4}, '''summarization_xsum''': {'''length_penalty''': 1.0, '''max_length''': 6_2, '''min_length''': 1_1, '''num_beams''': 6}, } } _lowerCamelCase = { '''task_specific_params.summarization.length_penalty''': 1.0, '''task_specific_params.summarization.max_length''': 1_2_8, '''task_specific_params.summarization.min_length''': 1_2, '''task_specific_params.summarization.num_beams''': 4, '''task_specific_params.summarization_cnn.length_penalty''': 2.0, '''task_specific_params.summarization_cnn.max_length''': 1_4_2, '''task_specific_params.summarization_cnn.min_length''': 5_6, '''task_specific_params.summarization_cnn.num_beams''': 4, '''task_specific_params.summarization_xsum.length_penalty''': 1.0, '''task_specific_params.summarization_xsum.max_length''': 6_2, '''task_specific_params.summarization_xsum.min_length''': 1_1, '''task_specific_params.summarization_xsum.num_beams''': 6, } self.assertEqual(flatten_dict(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , x.transpose() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , np.asarray(transpose(lowerCamelCase__ ) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , np.asarray(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.reshape(lowerCamelCase__ , (4, 3) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , np.reshape(lowerCamelCase__ , (1_2, 5) ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , reshape(lowerCamelCase__ , (1_2, 5) ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , reshape(lowerCamelCase__ , (1_2, 5) ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.asarray(reshape(lowerCamelCase__ , (4, 3) ) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , np.asarray(reshape(lowerCamelCase__ , (1_2, 5) ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.squeeze(lowerCamelCase__ ) ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.squeeze(lowerCamelCase__ , axis=2 ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.asarray(squeeze(lowerCamelCase__ ) ) ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.asarray(squeeze(lowerCamelCase__ , axis=2 ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.expand_dims(lowerCamelCase__ , axis=1 ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.asarray(expand_dims(lowerCamelCase__ , axis=1 ) ) ) )
661
1
"""simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = '''0.21.0''' from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
661
"""simple docstring""" import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } __SCREAMING_SNAKE_CASE : Tuple = { '''vocab_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json'''}, '''merges_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt'''}, '''tokenizer_config_file''': { '''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json''' }, } __SCREAMING_SNAKE_CASE : str = {'''facebook/blenderbot-3B''': 1_2_8} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def lowerCAmelCase_( ) -> str: _lowerCamelCase = ( list(range(ord('''!''' ) , ord('''~''' ) + 1 ) ) + list(range(ord('''¡''' ) , ord('''¬''' ) + 1 ) ) + list(range(ord('''®''' ) , ord('''ÿ''' ) + 1 ) ) ) _lowerCamelCase = bs[:] _lowerCamelCase = 0 for b in range(2**8 ): if b not in bs: bs.append(lowercase_ ) cs.append(2**8 + n ) n += 1 _lowerCamelCase = [chr(lowercase_ ) for n in cs] return dict(zip(lowercase_ , lowercase_ ) ) def lowerCAmelCase_( lowercase_ : str ) -> Dict: _lowerCamelCase = set() _lowerCamelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _lowerCamelCase = char return pairs class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Union[str, Any] = VOCAB_FILES_NAMES lowercase__ : int = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : Tuple = ['input_ids', 'attention_mask'] def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__="replace" , lowerCamelCase__="<s>" , lowerCamelCase__="</s>" , lowerCamelCase__="</s>" , lowerCamelCase__="<s>" , lowerCamelCase__="<unk>" , lowerCamelCase__="<pad>" , lowerCamelCase__="<mask>" , lowerCamelCase__=False , **lowerCamelCase__ , ): _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else bos_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else eos_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else sep_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else cls_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else unk_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else mask_token super().__init__( errors=lowerCamelCase__ , bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , add_prefix_space=lowerCamelCase__ , **lowerCamelCase__ , ) with open(lowerCamelCase__ , encoding='''utf-8''' ) as vocab_handle: _lowerCamelCase = json.load(lowerCamelCase__ ) _lowerCamelCase = {v: k for k, v in self.encoder.items()} _lowerCamelCase = errors # how to handle errors in decoding _lowerCamelCase = bytes_to_unicode() _lowerCamelCase = {v: k for k, v in self.byte_encoder.items()} with open(lowerCamelCase__ , encoding='''utf-8''' ) as merges_handle: _lowerCamelCase = merges_handle.read().split('''\n''' )[1:-1] _lowerCamelCase = [tuple(merge.split() ) for merge in bpe_merges] _lowerCamelCase = dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) _lowerCamelCase = {} _lowerCamelCase = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions _lowerCamelCase = re.compile(R'''\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+''' ) @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def snake_case__ ( self ): return len(self.encoder ) def snake_case__ ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def snake_case__ ( self , lowerCamelCase__ ): if token in self.cache: return self.cache[token] _lowerCamelCase = tuple(lowerCamelCase__ ) _lowerCamelCase = get_pairs(lowerCamelCase__ ) if not pairs: return token while True: _lowerCamelCase = min(lowerCamelCase__ , key=lambda lowerCamelCase__ : self.bpe_ranks.get(lowerCamelCase__ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break _lowerCamelCase , _lowerCamelCase = bigram _lowerCamelCase = [] _lowerCamelCase = 0 while i < len(lowerCamelCase__ ): try: _lowerCamelCase = word.index(lowerCamelCase__ , lowerCamelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _lowerCamelCase = j if word[i] == first and i < len(lowerCamelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _lowerCamelCase = tuple(lowerCamelCase__ ) _lowerCamelCase = new_word if len(lowerCamelCase__ ) == 1: break else: _lowerCamelCase = get_pairs(lowerCamelCase__ ) _lowerCamelCase = ''' '''.join(lowerCamelCase__ ) _lowerCamelCase = word return word def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] for token in re.findall(self.pat , lowerCamelCase__ ): _lowerCamelCase = ''''''.join( self.byte_encoder[b] for b in token.encode('''utf-8''' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCamelCase__ ).split(''' ''' ) ) return bpe_tokens def snake_case__ ( self , lowerCamelCase__ ): return self.encoder.get(lowerCamelCase__ , self.encoder.get(self.unk_token ) ) def snake_case__ ( self , lowerCamelCase__ ): return self.decoder.get(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = ''''''.join(lowerCamelCase__ ) _lowerCamelCase = bytearray([self.byte_decoder[c] for c in text] ).decode('''utf-8''' , errors=self.errors ) return text def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): if not os.path.isdir(lowerCamelCase__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(lowerCamelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCamelCase__ , ensure_ascii=lowerCamelCase__ ) + '''\n''' ) _lowerCamelCase = 0 with open(lowerCamelCase__ , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCamelCase__ : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ''' Please check that the tokenizer is not corrupted!''' ) _lowerCamelCase = token_index writer.write(''' '''.join(lowerCamelCase__ ) + '''\n''' ) index += 1 return vocab_file, merge_file def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase__ )) + [1] return [1] + ([0] * len(lowerCamelCase__ )) + [1, 1] + ([0] * len(lowerCamelCase__ )) + [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=False , **lowerCamelCase__ ): _lowerCamelCase = kwargs.pop('''add_prefix_space''' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCamelCase__ ) > 0 and not text[0].isspace()): _lowerCamelCase = ''' ''' + text return (text, kwargs) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): return token_ids_a + [self.eos_token_id] def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(''' ''' + text ) else: # Generated responses should contain them already. inputs.append(lowerCamelCase__ ) _lowerCamelCase = ''' '''.join(lowerCamelCase__ ) _lowerCamelCase = self.encode(lowerCamelCase__ ) if len(lowerCamelCase__ ) > self.model_max_length: _lowerCamelCase = input_ids[-self.model_max_length :] logger.warning(F"""Trimmed input from conversation as it was longer than {self.model_max_length} tokens.""" ) return input_ids
661
1
"""simple docstring""" from argparse import ArgumentParser, Namespace from typing import Any, List, Optional from ..pipelines import Pipeline, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand try: from fastapi import Body, FastAPI, HTTPException from fastapi.routing import APIRoute from pydantic import BaseModel from starlette.responses import JSONResponse from uvicorn import run __SCREAMING_SNAKE_CASE : List[Any] = True except (ImportError, AttributeError): __SCREAMING_SNAKE_CASE : List[Any] = object def lowerCAmelCase_( *lowercase_ : Dict , **lowercase_ : str ) -> str: pass __SCREAMING_SNAKE_CASE : Tuple = False __SCREAMING_SNAKE_CASE : Any = logging.get_logger('''transformers-cli/serving''') def lowerCAmelCase_( lowercase_ : Namespace ) -> List[Any]: _lowerCamelCase = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) return ServeCommand(lowercase_ , args.host , args.port , args.workers ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : dict class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] lowercase__ : Optional[List[int]] class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : str class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Any class lowerCamelCase_( A__ ): '''simple docstring''' @staticmethod def snake_case__ ( lowerCamelCase__ ): _lowerCamelCase = parser.add_parser( '''serve''' , help='''CLI tool to run inference requests through REST and GraphQL endpoints.''' ) serve_parser.add_argument( '''--task''' , type=lowerCamelCase__ , choices=get_supported_tasks() , help='''The task to run the pipeline on''' , ) serve_parser.add_argument('''--host''' , type=lowerCamelCase__ , default='''localhost''' , help='''Interface the server will listen on.''' ) serve_parser.add_argument('''--port''' , type=lowerCamelCase__ , default=8_8_8_8 , help='''Port the serving will listen to.''' ) serve_parser.add_argument('''--workers''' , type=lowerCamelCase__ , default=1 , help='''Number of http workers''' ) serve_parser.add_argument('''--model''' , type=lowerCamelCase__ , help='''Model\'s name or path to stored model.''' ) serve_parser.add_argument('''--config''' , type=lowerCamelCase__ , help='''Model\'s config name or path to stored model.''' ) serve_parser.add_argument('''--tokenizer''' , type=lowerCamelCase__ , help='''Tokenizer name to use.''' ) serve_parser.add_argument( '''--device''' , type=lowerCamelCase__ , default=-1 , help='''Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)''' , ) serve_parser.set_defaults(func=lowerCamelCase__ ) def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = pipeline _lowerCamelCase = host _lowerCamelCase = port _lowerCamelCase = workers if not _serve_dependencies_installed: raise RuntimeError( '''Using serve command requires FastAPI and uvicorn. ''' '''Please install transformers with [serving]: pip install "transformers[serving]".''' '''Or install FastAPI and uvicorn separately.''' ) else: logger.info(F"""Serving model over {host}:{port}""" ) _lowerCamelCase = FastAPI( routes=[ APIRoute( '''/''' , self.model_info , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''GET'''] , ), APIRoute( '''/tokenize''' , self.tokenize , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), APIRoute( '''/detokenize''' , self.detokenize , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), APIRoute( '''/forward''' , self.forward , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), ] , timeout=6_0_0 , ) def snake_case__ ( self ): run(self._app , host=self.host , port=self.port , workers=self.workers ) def snake_case__ ( self ): return ServeModelInfoResult(infos=vars(self._pipeline.model.config ) ) def snake_case__ ( self , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) ): try: _lowerCamelCase = self._pipeline.tokenizer.tokenize(lowerCamelCase__ ) if return_ids: _lowerCamelCase = self._pipeline.tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) return ServeTokenizeResult(tokens=lowerCamelCase__ , tokens_ids=lowerCamelCase__ ) else: return ServeTokenizeResult(tokens=lowerCamelCase__ ) except Exception as e: raise HTTPException(status_code=5_0_0 , detail={'''model''': '''''', '''error''': str(lowerCamelCase__ )} ) def snake_case__ ( self , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , ): try: _lowerCamelCase = self._pipeline.tokenizer.decode(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) return ServeDeTokenizeResult(model='''''' , text=lowerCamelCase__ ) except Exception as e: raise HTTPException(status_code=5_0_0 , detail={'''model''': '''''', '''error''': str(lowerCamelCase__ )} ) async def snake_case__ ( self , lowerCamelCase__=Body(lowerCamelCase__ , embed=lowerCamelCase__ ) ): # Check we don't have empty string if len(lowerCamelCase__ ) == 0: return ServeForwardResult(output=[] , attention=[] ) try: # Forward through the model _lowerCamelCase = self._pipeline(lowerCamelCase__ ) return ServeForwardResult(output=lowerCamelCase__ ) except Exception as e: raise HTTPException(5_0_0 , {'''error''': str(lowerCamelCase__ )} )
661
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Dict = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Union[str, Any] = XLMRobertaTokenizer lowercase__ : Optional[int] = XLMRobertaTokenizerFast lowercase__ : List[str] = True lowercase__ : Union[str, Any] = True def snake_case__ ( self ): super().setUp() # We have a SentencePiece fixture for testing _lowerCamelCase = XLMRobertaTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case__ ( self ): _lowerCamelCase = '''<pad>''' _lowerCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(lowerCamelCase__ ) , 1_0_0_2 ) def snake_case__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_2 ) def snake_case__ ( self ): _lowerCamelCase = XLMRobertaTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) _lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(lowerCamelCase__ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) _lowerCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) _lowerCamelCase = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) _lowerCamelCase = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def snake_case__ ( self ): if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _lowerCamelCase = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-xlm-roberta''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _lowerCamelCase = self.rust_tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = self.tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) _lowerCamelCase = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(lowerCamelCase__ , lowerCamelCase__ ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(lowerCamelCase__ ) # Save tokenizer rust, legacy_format=True _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ , legacy_format=lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it save with the same files self.assertSequenceEqual(lowerCamelCase__ , lowerCamelCase__ ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) shutil.rmtree(lowerCamelCase__ ) # Save tokenizer rust, legacy_format=False _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ , legacy_format=lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) shutil.rmtree(lowerCamelCase__ ) @cached_property def snake_case__ ( self ): return XLMRobertaTokenizer.from_pretrained('''xlm-roberta-base''' ) def snake_case__ ( self ): with tempfile.NamedTemporaryFile() as f: shutil.copyfile(lowerCamelCase__ , f.name ) _lowerCamelCase = XLMRobertaTokenizer(f.name , keep_accents=lowerCamelCase__ ) _lowerCamelCase = pickle.dumps(lowerCamelCase__ ) pickle.loads(lowerCamelCase__ ) def snake_case__ ( self ): if not self.test_rust_tokenizer: return _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = '''I was born in 92000, and this is falsé.''' _lowerCamelCase = tokenizer.tokenize(lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = tokenizer.encode(lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): _lowerCamelCase = '''Hello World!''' _lowerCamelCase = [0, 3_5_3_7_8, 6_6_6_1, 3_8, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(lowerCamelCase__ , self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def snake_case__ ( self ): _lowerCamelCase = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) _lowerCamelCase = [ 0, 3_2_9_3, 8_3, 1_0, 4_5_5_2, 4_9_8_9, 7_9_8_6, 6_7_8, 1_0, 5_9_1_5, 1_1_1, 1_7_9_4_5_9, 1_2_4_8_5_0, 4, 6_0_4_4, 2_3_7, 1_2, 6, 5, 6, 4, 6_7_8_0, 7_0_5, 1_5, 1_3_8_8, 4_4, 3_7_8, 1_0_1_1_4, 7_1_1, 1_5_2, 2_0, 6, 5, 2_2_3_7_6, 6_4_2, 1_2_2_1, 1_5_1_9_0, 3_4_1_5_3, 4_5_0, 5_6_0_8, 9_5_9, 1_1_1_9, 5_7_7_0_2, 1_3_6, 1_8_6, 4_7, 1_0_9_8, 2_9_3_6_7, 4_7, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6_0_4_4, 2_3_7, 6_2_8_4, 5_0_9_0_1, 5_2_8, 3_1, 9_0, 3_4, 9_2_7, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(lowerCamelCase__ , self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def snake_case__ ( self ): # fmt: off _lowerCamelCase = {'''input_ids''': [[0, 1_1_0_6_2, 8_2_7_7_2, 7, 1_5, 8_2_7_7_2, 5_3_8, 5_1_5_2_9, 2_3_7, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 2_1_5_1_7_5, 1_3_1_4, 1_3_6, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 5_6_3_5_9, 4_2, 1_2_2_0_0_9, 9, 1_6_4_6_6, 1_6, 8_7_3_4_4, 4_5_3_7, 9, 4_7_1_7, 7_8_3_8_1, 6, 1_5_9_9_5_8, 7, 1_5, 2_4_4_8_0, 6_1_8, 4, 5_2_7, 2_2_6_9_3, 5_4_2_8, 4, 2_7_7_7, 2_4_4_8_0, 9_8_7_4, 4, 4_3_5_2_3, 5_9_4, 4, 8_0_3, 1_8_3_9_2, 3_3_1_8_9, 1_8, 4, 4_3_5_2_3, 2_4_4_4_7, 1_2_3_9_9, 1_0_0, 2_4_9_5_5, 8_3_6_5_8, 9_6_2_6, 1_4_4_0_5_7, 1_5, 8_3_9, 2_2_3_3_5, 1_6, 1_3_6, 2_4_9_5_5, 8_3_6_5_8, 8_3_4_7_9, 1_5, 3_9_1_0_2, 7_2_4, 1_6, 6_7_8, 6_4_5, 2_7_8_9, 1_3_2_8, 4_5_8_9, 4_2, 1_2_2_0_0_9, 1_1_5_7_7_4, 2_3, 8_0_5, 1_3_2_8, 4_6_8_7_6, 7, 1_3_6, 5_3_8_9_4, 1_9_4_0, 4_2_2_2_7, 4_1_1_5_9, 1_7_7_2_1, 8_2_3, 4_2_5, 4, 2_7_5_1_2, 9_8_7_2_2, 2_0_6, 1_3_6, 5_5_3_1, 4_9_7_0, 9_1_9, 1_7_3_3_6, 5, 2], [0, 2_0_0_8_0, 6_1_8, 8_3, 8_2_7_7_5, 4_7, 4_7_9, 9, 1_5_1_7, 7_3, 5_3_8_9_4, 3_3_3, 8_0_5_8_1, 1_1_0_1_1_7, 1_8_8_1_1, 5_2_5_6, 1_2_9_5, 5_1, 1_5_2_5_2_6, 2_9_7, 7_9_8_6, 3_9_0, 1_2_4_4_1_6, 5_3_8, 3_5_4_3_1, 2_1_4, 9_8, 1_5_0_4_4, 2_5_7_3_7, 1_3_6, 7_1_0_8, 4_3_7_0_1, 2_3, 7_5_6, 1_3_5_3_5_5, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_8_1, 6_3_7_7_3, 1_1_9_4_5_5, 6, 1_4_7_7_9_7, 8_8_2_0_3, 7, 6_4_5, 7_0, 2_1, 3_2_8_5, 1_0_2_6_9, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ , model_name='''xlm-roberta-base''' , revision='''d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3''' , )
661
1
"""simple docstring""" import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=6_4 , lowerCamelCase__=5 , lowerCamelCase__=4 , lowerCamelCase__=6_4 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=4 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = seq_length _lowerCamelCase = is_training _lowerCamelCase = use_input_mask _lowerCamelCase = use_token_type_ids _lowerCamelCase = use_labels _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = num_labels _lowerCamelCase = num_choices _lowerCamelCase = scope def snake_case__ ( self ): return MPNetConfig.from_pretrained('''microsoft/mpnet-base''' ) def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_input_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None _lowerCamelCase = None _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) _lowerCamelCase = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case__ ( self ): return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = MPNetModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = MPNetForQuestionAnswering(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model( lowerCamelCase__ , attention_mask=lowerCamelCase__ , start_positions=lowerCamelCase__ , end_positions=lowerCamelCase__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = MPNetForSequenceClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_choices _lowerCamelCase = MPNetForMultipleChoice(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase = model( lowerCamelCase__ , attention_mask=lowerCamelCase__ , labels=lowerCamelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = MPNetForTokenClassification(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : List[Any] = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) lowercase__ : Any = ( { 'feature-extraction': MPNetModel, 'fill-mask': MPNetForMaskedLM, 'question-answering': MPNetForQuestionAnswering, 'text-classification': MPNetForSequenceClassification, 'token-classification': MPNetForTokenClassification, 'zero-shot': MPNetForSequenceClassification, } if is_torch_available() else {} ) lowercase__ : Tuple = False lowercase__ : Dict = True def snake_case__ ( self ): _lowerCamelCase = MPNetModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*lowerCamelCase__ ) @require_torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @slow def snake_case__ ( self ): _lowerCamelCase = MPNetModel.from_pretrained('''microsoft/mpnet-base''' ) _lowerCamelCase = torch.tensor([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) _lowerCamelCase = model(lowerCamelCase__ )[0] _lowerCamelCase = torch.Size((1, 1_1, 7_6_8) ) self.assertEqual(output.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor( [[[-0.0_5_5_0, 0.1_9_4_3, -0.0_7_4_0], [-0.0_5_6_2, 0.2_2_1_1, -0.0_5_7_9], [-0.0_4_3_7, 0.3_3_3_7, -0.0_6_4_1]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , lowerCamelCase__ , atol=1e-4 ) )
661
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int = 10 ) -> str: if not isinstance(lowercase_ , lowercase_ ) or n < 0: raise ValueError('''Invalid input''' ) _lowerCamelCase = 10**n _lowerCamelCase = 2_84_33 * (pow(2 , 7_83_04_57 , lowercase_ )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(F"""{solution(1_0) = }""")
661
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : list , lowercase_ : list ) -> float: _validate_point(lowercase_ ) _validate_point(lowercase_ ) if len(lowercase_ ) != len(lowercase_ ): raise ValueError('''Both points must be in the same n-dimensional space''' ) return float(sum(abs(a - b ) for a, b in zip(lowercase_ , lowercase_ ) ) ) def lowerCAmelCase_( lowercase_ : list[float] ) -> None: if point: if isinstance(lowercase_ , lowercase_ ): for item in point: if not isinstance(lowercase_ , (int, float) ): _lowerCamelCase = ( '''Expected a list of numbers as input, found ''' F"""{type(lowercase_ ).__name__}""" ) raise TypeError(lowercase_ ) else: _lowerCamelCase = F"""Expected a list of numbers as input, found {type(lowercase_ ).__name__}""" raise TypeError(lowercase_ ) else: raise ValueError('''Missing an input''' ) def lowerCAmelCase_( lowercase_ : list , lowercase_ : list ) -> float: _validate_point(lowercase_ ) _validate_point(lowercase_ ) if len(lowercase_ ) != len(lowercase_ ): raise ValueError('''Both points must be in the same n-dimensional space''' ) return float(sum(abs(x - y ) for x, y in zip(lowercase_ , lowercase_ ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
661
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __SCREAMING_SNAKE_CASE : Optional[int] = { '''configuration_informer''': [ '''INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ '''INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InformerForPrediction''', '''InformerModel''', '''InformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
661
1
"""simple docstring""" import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class lowerCamelCase_( nn.Module ): '''simple docstring''' lowercase__ : int lowercase__ : int lowercase__ : float = 0.0 lowercase__ : int = 1 lowercase__ : int = 1 lowercase__ : bool = True lowercase__ : bool = False lowercase__ : bool = False lowercase__ : bool = False lowercase__ : jnp.dtype = jnp.floataa def snake_case__ ( self ): _lowerCamelCase = [] _lowerCamelCase = [] for i in range(self.num_layers ): _lowerCamelCase = self.in_channels if i == 0 else self.out_channels _lowerCamelCase = FlaxResnetBlockaD( in_channels=lowerCamelCase__ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(lowerCamelCase__ ) _lowerCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(lowerCamelCase__ ) _lowerCamelCase = resnets _lowerCamelCase = attentions if self.add_downsample: _lowerCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=True ): _lowerCamelCase = () for resnet, attn in zip(self.resnets , self.attentions ): _lowerCamelCase = resnet(lowerCamelCase__ , lowerCamelCase__ , deterministic=lowerCamelCase__ ) _lowerCamelCase = attn(lowerCamelCase__ , lowerCamelCase__ , deterministic=lowerCamelCase__ ) output_states += (hidden_states,) if self.add_downsample: _lowerCamelCase = self.downsamplers_a(lowerCamelCase__ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCamelCase_( nn.Module ): '''simple docstring''' lowercase__ : int lowercase__ : int lowercase__ : float = 0.0 lowercase__ : int = 1 lowercase__ : bool = True lowercase__ : jnp.dtype = jnp.floataa def snake_case__ ( self ): _lowerCamelCase = [] for i in range(self.num_layers ): _lowerCamelCase = self.in_channels if i == 0 else self.out_channels _lowerCamelCase = FlaxResnetBlockaD( in_channels=lowerCamelCase__ , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(lowerCamelCase__ ) _lowerCamelCase = resnets if self.add_downsample: _lowerCamelCase = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=True ): _lowerCamelCase = () for resnet in self.resnets: _lowerCamelCase = resnet(lowerCamelCase__ , lowerCamelCase__ , deterministic=lowerCamelCase__ ) output_states += (hidden_states,) if self.add_downsample: _lowerCamelCase = self.downsamplers_a(lowerCamelCase__ ) output_states += (hidden_states,) return hidden_states, output_states class lowerCamelCase_( nn.Module ): '''simple docstring''' lowercase__ : int lowercase__ : int lowercase__ : int lowercase__ : float = 0.0 lowercase__ : int = 1 lowercase__ : int = 1 lowercase__ : bool = True lowercase__ : bool = False lowercase__ : bool = False lowercase__ : bool = False lowercase__ : jnp.dtype = jnp.floataa def snake_case__ ( self ): _lowerCamelCase = [] _lowerCamelCase = [] for i in range(self.num_layers ): _lowerCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels _lowerCamelCase = self.prev_output_channel if i == 0 else self.out_channels _lowerCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(lowerCamelCase__ ) _lowerCamelCase = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(lowerCamelCase__ ) _lowerCamelCase = resnets _lowerCamelCase = attentions if self.add_upsample: _lowerCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=True ): for resnet, attn in zip(self.resnets , self.attentions ): # pop res hidden states _lowerCamelCase = res_hidden_states_tuple[-1] _lowerCamelCase = res_hidden_states_tuple[:-1] _lowerCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) _lowerCamelCase = resnet(lowerCamelCase__ , lowerCamelCase__ , deterministic=lowerCamelCase__ ) _lowerCamelCase = attn(lowerCamelCase__ , lowerCamelCase__ , deterministic=lowerCamelCase__ ) if self.add_upsample: _lowerCamelCase = self.upsamplers_a(lowerCamelCase__ ) return hidden_states class lowerCamelCase_( nn.Module ): '''simple docstring''' lowercase__ : int lowercase__ : int lowercase__ : int lowercase__ : float = 0.0 lowercase__ : int = 1 lowercase__ : bool = True lowercase__ : jnp.dtype = jnp.floataa def snake_case__ ( self ): _lowerCamelCase = [] for i in range(self.num_layers ): _lowerCamelCase = self.in_channels if (i == self.num_layers - 1) else self.out_channels _lowerCamelCase = self.prev_output_channel if i == 0 else self.out_channels _lowerCamelCase = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(lowerCamelCase__ ) _lowerCamelCase = resnets if self.add_upsample: _lowerCamelCase = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=True ): for resnet in self.resnets: # pop res hidden states _lowerCamelCase = res_hidden_states_tuple[-1] _lowerCamelCase = res_hidden_states_tuple[:-1] _lowerCamelCase = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) _lowerCamelCase = resnet(lowerCamelCase__ , lowerCamelCase__ , deterministic=lowerCamelCase__ ) if self.add_upsample: _lowerCamelCase = self.upsamplers_a(lowerCamelCase__ ) return hidden_states class lowerCamelCase_( nn.Module ): '''simple docstring''' lowercase__ : int lowercase__ : float = 0.0 lowercase__ : int = 1 lowercase__ : int = 1 lowercase__ : bool = False lowercase__ : bool = False lowercase__ : jnp.dtype = jnp.floataa def snake_case__ ( self ): # there is always at least one resnet _lowerCamelCase = [ FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) ] _lowerCamelCase = [] for _ in range(self.num_layers ): _lowerCamelCase = FlaxTransformeraDModel( in_channels=self.in_channels , n_heads=self.num_attention_heads , d_head=self.in_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(lowerCamelCase__ ) _lowerCamelCase = FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(lowerCamelCase__ ) _lowerCamelCase = resnets _lowerCamelCase = attentions def __call__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=True ): _lowerCamelCase = self.resnets[0](lowerCamelCase__ , lowerCamelCase__ ) for attn, resnet in zip(self.attentions , self.resnets[1:] ): _lowerCamelCase = attn(lowerCamelCase__ , lowerCamelCase__ , deterministic=lowerCamelCase__ ) _lowerCamelCase = resnet(lowerCamelCase__ , lowerCamelCase__ , deterministic=lowerCamelCase__ ) return hidden_states
661
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> float: def get_matched_characters(lowercase_ : str , lowercase_ : str ) -> str: _lowerCamelCase = [] _lowerCamelCase = min(len(_stra ) , len(_stra ) ) // 2 for i, l in enumerate(_stra ): _lowerCamelCase = int(max(0 , i - limit ) ) _lowerCamelCase = int(min(i + limit + 1 , len(_stra ) ) ) if l in _stra[left:right]: matched.append(lowercase_ ) _lowerCamelCase = F"""{_stra[0:_stra.index(lowercase_ )]} {_stra[_stra.index(lowercase_ ) + 1:]}""" return "".join(lowercase_ ) # matching characters _lowerCamelCase = get_matched_characters(lowercase_ , lowercase_ ) _lowerCamelCase = get_matched_characters(lowercase_ , lowercase_ ) _lowerCamelCase = len(lowercase_ ) # transposition _lowerCamelCase = ( len([(ca, ca) for ca, ca in zip(lowercase_ , lowercase_ ) if ca != ca] ) // 2 ) if not match_count: _lowerCamelCase = 0.0 else: _lowerCamelCase = ( 1 / 3 * ( match_count / len(lowercase_ ) + match_count / len(lowercase_ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters _lowerCamelCase = 0 for ca, ca in zip(stra[:4] , stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('''hello''', '''world'''))
661
1
"""simple docstring""" from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Tuple = ['image_processor', 'tokenizer'] lowercase__ : List[str] = 'BlipImageProcessor' lowercase__ : Union[str, Any] = ('BertTokenizer', 'BertTokenizerFast') def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = False super().__init__(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = self.image_processor def __call__( self , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = True , lowerCamelCase__ = False , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = 0 , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = False , lowerCamelCase__ = False , lowerCamelCase__ = False , lowerCamelCase__ = False , lowerCamelCase__ = False , lowerCamelCase__ = True , lowerCamelCase__ = None , **lowerCamelCase__ , ): if images is None and text is None: raise ValueError('''You have to specify either images or text.''' ) # Get only text if images is None: _lowerCamelCase = self.tokenizer _lowerCamelCase = self.tokenizer( text=lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , padding=lowerCamelCase__ , truncation=lowerCamelCase__ , max_length=lowerCamelCase__ , stride=lowerCamelCase__ , pad_to_multiple_of=lowerCamelCase__ , return_attention_mask=lowerCamelCase__ , return_overflowing_tokens=lowerCamelCase__ , return_special_tokens_mask=lowerCamelCase__ , return_offsets_mapping=lowerCamelCase__ , return_token_type_ids=lowerCamelCase__ , return_length=lowerCamelCase__ , verbose=lowerCamelCase__ , return_tensors=lowerCamelCase__ , **lowerCamelCase__ , ) return text_encoding # add pixel_values _lowerCamelCase = self.image_processor(lowerCamelCase__ , return_tensors=lowerCamelCase__ ) if text is not None: _lowerCamelCase = self.tokenizer( text=lowerCamelCase__ , add_special_tokens=lowerCamelCase__ , padding=lowerCamelCase__ , truncation=lowerCamelCase__ , max_length=lowerCamelCase__ , stride=lowerCamelCase__ , pad_to_multiple_of=lowerCamelCase__ , return_attention_mask=lowerCamelCase__ , return_overflowing_tokens=lowerCamelCase__ , return_special_tokens_mask=lowerCamelCase__ , return_offsets_mapping=lowerCamelCase__ , return_token_type_ids=lowerCamelCase__ , return_length=lowerCamelCase__ , verbose=lowerCamelCase__ , return_tensors=lowerCamelCase__ , **lowerCamelCase__ , ) else: _lowerCamelCase = None if text_encoding is not None: encoding_image_processor.update(lowerCamelCase__ ) return encoding_image_processor def snake_case__ ( self , *lowerCamelCase__ , **lowerCamelCase__ ): return self.tokenizer.batch_decode(*lowerCamelCase__ , **lowerCamelCase__ ) def snake_case__ ( self , *lowerCamelCase__ , **lowerCamelCase__ ): return self.tokenizer.decode(*lowerCamelCase__ , **lowerCamelCase__ ) @property def snake_case__ ( self ): _lowerCamelCase = self.tokenizer.model_input_names _lowerCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
661
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowerCamelCase_( A__, A__, A__ ): '''simple docstring''' lowercase__ : List[Any] = [r'h\.\d+\.attn\.bias', r'h\.\d+\.attn\.masked_bias'] @register_to_config def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = 5_0_2_5_7 , lowerCamelCase__ = 1_0_2_4 , lowerCamelCase__ = 7_6_8 , lowerCamelCase__ = 1_2 , lowerCamelCase__ = 1_2 , lowerCamelCase__ = None , lowerCamelCase__ = "gelu_new" , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 1e-5 , lowerCamelCase__ = 0.0_2 , lowerCamelCase__ = True , lowerCamelCase__ = True , lowerCamelCase__ = False , lowerCamelCase__ = False , ): super().__init__() _lowerCamelCase = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F"""`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and""" F""" `n_embd`: {n_embd} are not equal.""" ) _lowerCamelCase = prefix_inner_dim _lowerCamelCase = prefix_hidden_dim _lowerCamelCase = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCamelCase = ( nn.Linear(self.prefix_hidden_dim , lowerCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCamelCase = GPTaConfig( vocab_size=lowerCamelCase__ , n_positions=lowerCamelCase__ , n_embd=lowerCamelCase__ , n_layer=lowerCamelCase__ , n_head=lowerCamelCase__ , n_inner=lowerCamelCase__ , activation_function=lowerCamelCase__ , resid_pdrop=lowerCamelCase__ , embd_pdrop=lowerCamelCase__ , attn_pdrop=lowerCamelCase__ , layer_norm_epsilon=lowerCamelCase__ , initializer_range=lowerCamelCase__ , scale_attn_weights=lowerCamelCase__ , use_cache=lowerCamelCase__ , scale_attn_by_inverse_layer_idx=lowerCamelCase__ , reorder_and_upcast_attn=lowerCamelCase__ , ) _lowerCamelCase = GPTaLMHeadModel(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None , ): _lowerCamelCase = self.transformer.transformer.wte(lowerCamelCase__ ) _lowerCamelCase = self.encode_prefix(lowerCamelCase__ ) _lowerCamelCase = self.decode_prefix(lowerCamelCase__ ) _lowerCamelCase = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: _lowerCamelCase = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) _lowerCamelCase = torch.cat((dummy_token, input_ids) , dim=1 ) _lowerCamelCase = self.transformer(inputs_embeds=lowerCamelCase__ , labels=lowerCamelCase__ , attention_mask=lowerCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): return torch.zeros(lowerCamelCase__ , self.prefix_length , dtype=torch.intaa , device=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): return self.encode_prefix(lowerCamelCase__ ) @torch.no_grad() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = torch.split(lowerCamelCase__ , 1 , dim=0 ) _lowerCamelCase = [] _lowerCamelCase = [] for feature in features: _lowerCamelCase = self.decode_prefix(feature.to(lowerCamelCase__ ) ) # back to the clip feature # Only support beam search for now _lowerCamelCase , _lowerCamelCase = self.generate_beam( input_embeds=lowerCamelCase__ , device=lowerCamelCase__ , eos_token_id=lowerCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) _lowerCamelCase = torch.stack(lowerCamelCase__ ) _lowerCamelCase = torch.stack(lowerCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def snake_case__ ( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__ = 5 , lowerCamelCase__ = 6_7 , lowerCamelCase__ = 1.0 , lowerCamelCase__ = None , ): _lowerCamelCase = eos_token_id _lowerCamelCase = None _lowerCamelCase = None _lowerCamelCase = torch.ones(lowerCamelCase__ , device=lowerCamelCase__ , dtype=torch.int ) _lowerCamelCase = torch.zeros(lowerCamelCase__ , device=lowerCamelCase__ , dtype=torch.bool ) if input_embeds is not None: _lowerCamelCase = input_embeds else: _lowerCamelCase = self.transformer.transformer.wte(lowerCamelCase__ ) for i in range(lowerCamelCase__ ): _lowerCamelCase = self.transformer(inputs_embeds=lowerCamelCase__ ) _lowerCamelCase = outputs.logits _lowerCamelCase = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) _lowerCamelCase = logits.softmax(-1 ).log() if scores is None: _lowerCamelCase , _lowerCamelCase = logits.topk(lowerCamelCase__ , -1 ) _lowerCamelCase = generated.expand(lowerCamelCase__ , *generated.shape[1:] ) _lowerCamelCase , _lowerCamelCase = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: _lowerCamelCase = next_tokens else: _lowerCamelCase = tokens.expand(lowerCamelCase__ , *tokens.shape[1:] ) _lowerCamelCase = torch.cat((tokens, next_tokens) , dim=1 ) else: _lowerCamelCase = -float(np.inf ) _lowerCamelCase = 0 _lowerCamelCase = scores[:, None] + logits seq_lengths[~is_stopped] += 1 _lowerCamelCase = scores_sum / seq_lengths[:, None] _lowerCamelCase , _lowerCamelCase = scores_sum_average.view(-1 ).topk(lowerCamelCase__ , -1 ) _lowerCamelCase = next_tokens // scores_sum.shape[1] _lowerCamelCase = seq_lengths[next_tokens_source] _lowerCamelCase = next_tokens % scores_sum.shape[1] _lowerCamelCase = next_tokens.unsqueeze(1 ) _lowerCamelCase = tokens[next_tokens_source] _lowerCamelCase = torch.cat((tokens, next_tokens) , dim=1 ) _lowerCamelCase = generated[next_tokens_source] _lowerCamelCase = scores_sum_average * seq_lengths _lowerCamelCase = is_stopped[next_tokens_source] _lowerCamelCase = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) _lowerCamelCase = torch.cat((generated, next_token_embed) , dim=1 ) _lowerCamelCase = is_stopped + next_tokens.eq(lowerCamelCase__ ).squeeze() if is_stopped.all(): break _lowerCamelCase = scores / seq_lengths _lowerCamelCase = scores.argsort(descending=lowerCamelCase__ ) # tokens tensors are already padded to max_seq_length _lowerCamelCase = [tokens[i] for i in order] _lowerCamelCase = torch.stack(lowerCamelCase__ , dim=0 ) _lowerCamelCase = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
661
1
"""simple docstring""" import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. __SCREAMING_SNAKE_CASE : Optional[int] = {'''LayoutLMv2Config''', '''LayoutLMv3Config'''} @is_pipeline_test class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' lowercase__ : Dict = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING lowercase__ : Optional[Any] = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: lowercase__ : Tuple = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: lowercase__ : List[str] = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = ZeroShotClassificationPipeline( model=lowerCamelCase__ , tokenizer=lowerCamelCase__ , candidate_labels=['''polics''', '''health'''] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = classifier('''Who are you voting for in 2020?''' , candidate_labels='''politics''' ) self.assertEqual(lowerCamelCase__ , {'''sequence''': ANY(lowerCamelCase__ ), '''labels''': [ANY(lowerCamelCase__ )], '''scores''': [ANY(lowerCamelCase__ )]} ) # No kwarg _lowerCamelCase = classifier('''Who are you voting for in 2020?''' , ['''politics'''] ) self.assertEqual(lowerCamelCase__ , {'''sequence''': ANY(lowerCamelCase__ ), '''labels''': [ANY(lowerCamelCase__ )], '''scores''': [ANY(lowerCamelCase__ )]} ) _lowerCamelCase = classifier('''Who are you voting for in 2020?''' , candidate_labels=['''politics'''] ) self.assertEqual(lowerCamelCase__ , {'''sequence''': ANY(lowerCamelCase__ ), '''labels''': [ANY(lowerCamelCase__ )], '''scores''': [ANY(lowerCamelCase__ )]} ) _lowerCamelCase = classifier('''Who are you voting for in 2020?''' , candidate_labels='''politics, public health''' ) self.assertEqual( lowerCamelCase__ , {'''sequence''': ANY(lowerCamelCase__ ), '''labels''': [ANY(lowerCamelCase__ ), ANY(lowerCamelCase__ )], '''scores''': [ANY(lowerCamelCase__ ), ANY(lowerCamelCase__ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs['''scores'''] ) ) , 1.0 ) _lowerCamelCase = classifier('''Who are you voting for in 2020?''' , candidate_labels=['''politics''', '''public health'''] ) self.assertEqual( lowerCamelCase__ , {'''sequence''': ANY(lowerCamelCase__ ), '''labels''': [ANY(lowerCamelCase__ ), ANY(lowerCamelCase__ )], '''scores''': [ANY(lowerCamelCase__ ), ANY(lowerCamelCase__ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs['''scores'''] ) ) , 1.0 ) _lowerCamelCase = classifier( '''Who are you voting for in 2020?''' , candidate_labels='''politics''' , hypothesis_template='''This text is about {}''' ) self.assertEqual(lowerCamelCase__ , {'''sequence''': ANY(lowerCamelCase__ ), '''labels''': [ANY(lowerCamelCase__ )], '''scores''': [ANY(lowerCamelCase__ )]} ) # https://github.com/huggingface/transformers/issues/13846 _lowerCamelCase = classifier(['''I am happy'''] , ['''positive''', '''negative'''] ) self.assertEqual( lowerCamelCase__ , [ {'''sequence''': ANY(lowerCamelCase__ ), '''labels''': [ANY(lowerCamelCase__ ), ANY(lowerCamelCase__ )], '''scores''': [ANY(lowerCamelCase__ ), ANY(lowerCamelCase__ )]} for i in range(1 ) ] , ) _lowerCamelCase = classifier(['''I am happy''', '''I am sad'''] , ['''positive''', '''negative'''] ) self.assertEqual( lowerCamelCase__ , [ {'''sequence''': ANY(lowerCamelCase__ ), '''labels''': [ANY(lowerCamelCase__ ), ANY(lowerCamelCase__ )], '''scores''': [ANY(lowerCamelCase__ ), ANY(lowerCamelCase__ )]} for i in range(2 ) ] , ) with self.assertRaises(lowerCamelCase__ ): classifier('''''' , candidate_labels='''politics''' ) with self.assertRaises(lowerCamelCase__ ): classifier(lowerCamelCase__ , candidate_labels='''politics''' ) with self.assertRaises(lowerCamelCase__ ): classifier('''Who are you voting for in 2020?''' , candidate_labels='''''' ) with self.assertRaises(lowerCamelCase__ ): classifier('''Who are you voting for in 2020?''' , candidate_labels=lowerCamelCase__ ) with self.assertRaises(lowerCamelCase__ ): classifier( '''Who are you voting for in 2020?''' , candidate_labels='''politics''' , hypothesis_template='''Not formatting template''' , ) with self.assertRaises(lowerCamelCase__ ): classifier( '''Who are you voting for in 2020?''' , candidate_labels='''politics''' , hypothesis_template=lowerCamelCase__ , ) self.run_entailment_id(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = zero_shot_classifier.model.config _lowerCamelCase = config.labelaid _lowerCamelCase = zero_shot_classifier.entailment_id _lowerCamelCase = {'''LABEL_0''': 0, '''LABEL_1''': 1, '''LABEL_2''': 2} self.assertEqual(zero_shot_classifier.entailment_id , -1 ) _lowerCamelCase = {'''entailment''': 0, '''neutral''': 1, '''contradiction''': 2} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) _lowerCamelCase = {'''ENTAIL''': 0, '''NON-ENTAIL''': 1} self.assertEqual(zero_shot_classifier.entailment_id , 0 ) _lowerCamelCase = {'''ENTAIL''': 2, '''NEUTRAL''': 1, '''CONTR''': 0} self.assertEqual(zero_shot_classifier.entailment_id , 2 ) _lowerCamelCase = original_labelaid self.assertEqual(lowerCamelCase__ , zero_shot_classifier.entailment_id ) @require_torch def snake_case__ ( self ): _lowerCamelCase = pipeline( '''zero-shot-classification''' , model='''sshleifer/tiny-distilbert-base-cased-distilled-squad''' , framework='''pt''' , ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( '''Who are you voting for in 2020?''' * 1_0_0 , candidate_labels=['''politics''', '''public health''', '''science'''] ) @require_torch def snake_case__ ( self ): _lowerCamelCase = pipeline( '''zero-shot-classification''' , model='''sshleifer/tiny-distilbert-base-cased-distilled-squad''' , framework='''pt''' , ) _lowerCamelCase = zero_shot_classifier( '''Who are you voting for in 2020?''' , candidate_labels=['''politics''', '''public health''', '''science'''] ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , { '''sequence''': '''Who are you voting for in 2020?''', '''labels''': ['''science''', '''public health''', '''politics'''], '''scores''': [0.3_3_3, 0.3_3_3, 0.3_3_3], } , ) @require_tf def snake_case__ ( self ): _lowerCamelCase = pipeline( '''zero-shot-classification''' , model='''sshleifer/tiny-distilbert-base-cased-distilled-squad''' , framework='''tf''' , ) _lowerCamelCase = zero_shot_classifier( '''Who are you voting for in 2020?''' , candidate_labels=['''politics''', '''public health''', '''science'''] ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , { '''sequence''': '''Who are you voting for in 2020?''', '''labels''': ['''science''', '''public health''', '''politics'''], '''scores''': [0.3_3_3, 0.3_3_3, 0.3_3_3], } , ) @slow @require_torch def snake_case__ ( self ): _lowerCamelCase = pipeline('''zero-shot-classification''' , model='''roberta-large-mnli''' , framework='''pt''' ) _lowerCamelCase = zero_shot_classifier( '''Who are you voting for in 2020?''' , candidate_labels=['''politics''', '''public health''', '''science'''] ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , { '''sequence''': '''Who are you voting for in 2020?''', '''labels''': ['''politics''', '''public health''', '''science'''], '''scores''': [0.9_7_6, 0.0_1_5, 0.0_0_9], } , ) _lowerCamelCase = zero_shot_classifier( '''The dominant sequence transduction models are based on complex recurrent or convolutional neural networks''' ''' in an encoder-decoder configuration. The best performing models also connect the encoder and decoder''' ''' through an attention mechanism. We propose a new simple network architecture, the Transformer, based''' ''' solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two''' ''' machine translation tasks show these models to be superior in quality while being more parallelizable''' ''' and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014''' ''' English-to-German translation task, improving over the existing best results, including ensembles by''' ''' over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new''' ''' single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small''' ''' fraction of the training costs of the best models from the literature. We show that the Transformer''' ''' generalizes well to other tasks by applying it successfully to English constituency parsing both with''' ''' large and limited training data.''' , candidate_labels=['''machine learning''', '''statistics''', '''translation''', '''vision'''] , multi_label=lowerCamelCase__ , ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , { '''sequence''': ( '''The dominant sequence transduction models are based on complex recurrent or convolutional neural''' ''' networks in an encoder-decoder configuration. The best performing models also connect the''' ''' encoder and decoder through an attention mechanism. We propose a new simple network''' ''' architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence''' ''' and convolutions entirely. Experiments on two machine translation tasks show these models to be''' ''' superior in quality while being more parallelizable and requiring significantly less time to''' ''' train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,''' ''' improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014''' ''' English-to-French translation task, our model establishes a new single-model state-of-the-art''' ''' BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training''' ''' costs of the best models from the literature. We show that the Transformer generalizes well to''' ''' other tasks by applying it successfully to English constituency parsing both with large and''' ''' limited training data.''' ), '''labels''': ['''translation''', '''machine learning''', '''vision''', '''statistics'''], '''scores''': [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], } , ) @slow @require_tf def snake_case__ ( self ): _lowerCamelCase = pipeline('''zero-shot-classification''' , model='''roberta-large-mnli''' , framework='''tf''' ) _lowerCamelCase = zero_shot_classifier( '''Who are you voting for in 2020?''' , candidate_labels=['''politics''', '''public health''', '''science'''] ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , { '''sequence''': '''Who are you voting for in 2020?''', '''labels''': ['''politics''', '''public health''', '''science'''], '''scores''': [0.9_7_6, 0.0_1_5, 0.0_0_9], } , ) _lowerCamelCase = zero_shot_classifier( '''The dominant sequence transduction models are based on complex recurrent or convolutional neural networks''' ''' in an encoder-decoder configuration. The best performing models also connect the encoder and decoder''' ''' through an attention mechanism. We propose a new simple network architecture, the Transformer, based''' ''' solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two''' ''' machine translation tasks show these models to be superior in quality while being more parallelizable''' ''' and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014''' ''' English-to-German translation task, improving over the existing best results, including ensembles by''' ''' over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new''' ''' single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small''' ''' fraction of the training costs of the best models from the literature. We show that the Transformer''' ''' generalizes well to other tasks by applying it successfully to English constituency parsing both with''' ''' large and limited training data.''' , candidate_labels=['''machine learning''', '''statistics''', '''translation''', '''vision'''] , multi_label=lowerCamelCase__ , ) self.assertEqual( nested_simplify(lowerCamelCase__ ) , { '''sequence''': ( '''The dominant sequence transduction models are based on complex recurrent or convolutional neural''' ''' networks in an encoder-decoder configuration. The best performing models also connect the''' ''' encoder and decoder through an attention mechanism. We propose a new simple network''' ''' architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence''' ''' and convolutions entirely. Experiments on two machine translation tasks show these models to be''' ''' superior in quality while being more parallelizable and requiring significantly less time to''' ''' train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,''' ''' improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014''' ''' English-to-French translation task, our model establishes a new single-model state-of-the-art''' ''' BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training''' ''' costs of the best models from the literature. We show that the Transformer generalizes well to''' ''' other tasks by applying it successfully to English constituency parsing both with large and''' ''' limited training data.''' ), '''labels''': ['''translation''', '''machine learning''', '''vision''', '''statistics'''], '''scores''': [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], } , )
661
"""simple docstring""" import re import string from collections import Counter import sacrebleu import sacremoses from packaging import version import datasets __SCREAMING_SNAKE_CASE : Optional[int] = ''' @inproceedings{xu-etal-2016-optimizing, title = {Optimizing Statistical Machine Translation for Text Simplification}, authors={Xu, Wei and Napoles, Courtney and Pavlick, Ellie and Chen, Quanze and Callison-Burch, Chris}, journal = {Transactions of the Association for Computational Linguistics}, volume = {4}, year={2016}, url = {https://www.aclweb.org/anthology/Q16-1029}, pages = {401--415 }, @inproceedings{post-2018-call, title = "A Call for Clarity in Reporting {BLEU} Scores", author = "Post, Matt", booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers", month = oct, year = "2018", address = "Belgium, Brussels", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W18-6319", pages = "186--191", } ''' __SCREAMING_SNAKE_CASE : List[str] = '''\ WIKI_SPLIT is the combination of three metrics SARI, EXACT and SACREBLEU It can be used to evaluate the quality of machine-generated texts. ''' __SCREAMING_SNAKE_CASE : Optional[Any] = ''' Calculates sari score (between 0 and 100) given a list of source and predicted sentences, and a list of lists of reference sentences. It also computes the BLEU score as well as the exact match score. Args: sources: list of source sentences where each sentence should be a string. predictions: list of predicted sentences where each sentence should be a string. references: list of lists of reference sentences where each sentence should be a string. Returns: sari: sari score sacrebleu: sacrebleu score exact: exact score Examples: >>> sources=["About 95 species are currently accepted ."] >>> predictions=["About 95 you now get in ."] >>> references=[["About 95 species are currently known ."]] >>> wiki_split = datasets.load_metric("wiki_split") >>> results = wiki_split.compute(sources=sources, predictions=predictions, references=references) >>> print(results) {\'sari\': 21.805555555555557, \'sacrebleu\': 14.535768424205482, \'exact\': 0.0} ''' def lowerCAmelCase_( lowercase_ : Union[str, Any] ) -> str: def remove_articles(lowercase_ : int ): _lowerCamelCase = re.compile(r'''\b(a|an|the)\b''' , re.UNICODE ) return re.sub(lowercase_ , ''' ''' , lowercase_ ) def white_space_fix(lowercase_ : List[Any] ): return " ".join(text.split() ) def remove_punc(lowercase_ : Dict ): _lowerCamelCase = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowercase_ : Union[str, Any] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowercase_ ) ) ) ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[Any] ) -> Union[str, Any]: return int(normalize_answer(lowercase_ ) == normalize_answer(lowercase_ ) ) def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Tuple ) -> Tuple: _lowerCamelCase = [any(compute_exact(lowercase_ , lowercase_ ) for ref in refs ) for pred, refs in zip(lowercase_ , lowercase_ )] return (sum(lowercase_ ) / len(lowercase_ )) * 1_00 def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : str , lowercase_ : Optional[Any] , lowercase_ : str ) -> Optional[int]: _lowerCamelCase = [rgram for rgrams in rgramslist for rgram in rgrams] _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter() for sgram, scount in sgramcounter.items(): _lowerCamelCase = scount * numref _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter() for cgram, ccount in cgramcounter.items(): _lowerCamelCase = ccount * numref # KEEP _lowerCamelCase = sgramcounter_rep & cgramcounter_rep _lowerCamelCase = keepgramcounter_rep & rgramcounter _lowerCamelCase = sgramcounter_rep & rgramcounter _lowerCamelCase = 0 _lowerCamelCase = 0 for keepgram in keepgramcountergood_rep: keeptmpscorea += keepgramcountergood_rep[keepgram] / keepgramcounter_rep[keepgram] # Fix an alleged bug [2] in the keep score computation. # keeptmpscore2 += keepgramcountergood_rep[keepgram] / keepgramcounterall_rep[keepgram] keeptmpscorea += keepgramcountergood_rep[keepgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = keeptmpscorea / len(lowercase_ ) if len(lowercase_ ) > 0: # Fix an alleged bug [2] in the keep score computation. # keepscore_recall = keeptmpscore2 / len(keepgramcounterall_rep) _lowerCamelCase = keeptmpscorea / sum(keepgramcounterall_rep.values() ) _lowerCamelCase = 0 if keepscore_precision > 0 or keepscore_recall > 0: _lowerCamelCase = 2 * keepscore_precision * keepscore_recall / (keepscore_precision + keepscore_recall) # DELETION _lowerCamelCase = sgramcounter_rep - cgramcounter_rep _lowerCamelCase = delgramcounter_rep - rgramcounter _lowerCamelCase = sgramcounter_rep - rgramcounter _lowerCamelCase = 0 _lowerCamelCase = 0 for delgram in delgramcountergood_rep: deltmpscorea += delgramcountergood_rep[delgram] / delgramcounter_rep[delgram] deltmpscorea += delgramcountergood_rep[delgram] / delgramcounterall_rep[delgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = deltmpscorea / len(lowercase_ ) # ADDITION _lowerCamelCase = set(lowercase_ ) - set(lowercase_ ) _lowerCamelCase = set(lowercase_ ) & set(lowercase_ ) _lowerCamelCase = set(lowercase_ ) - set(lowercase_ ) _lowerCamelCase = 0 for addgram in addgramcountergood: addtmpscore += 1 # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = addtmpscore / len(lowercase_ ) if len(lowercase_ ) > 0: _lowerCamelCase = addtmpscore / len(lowercase_ ) _lowerCamelCase = 0 if addscore_precision > 0 or addscore_recall > 0: _lowerCamelCase = 2 * addscore_precision * addscore_recall / (addscore_precision + addscore_recall) return (keepscore, delscore_precision, addscore) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : Optional[Any] , lowercase_ : str ) -> List[str]: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = ssent.split(''' ''' ) _lowerCamelCase = csent.split(''' ''' ) _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] for rsent in rsents: _lowerCamelCase = rsent.split(''' ''' ) _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] ragramslist.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] ragrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] + ''' ''' + ragrams[i + 2] ragrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] + ''' ''' + ragrams[i + 2] + ''' ''' + ragrams[i + 3] ragrams.append(lowercase_ ) ragramslist.append(lowercase_ ) ragramslist.append(lowercase_ ) ragramslist.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] sagrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] + ''' ''' + sagrams[i + 2] sagrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] + ''' ''' + sagrams[i + 2] + ''' ''' + sagrams[i + 3] sagrams.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] cagrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] + ''' ''' + cagrams[i + 2] cagrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] + ''' ''' + cagrams[i + 2] + ''' ''' + cagrams[i + 3] cagrams.append(lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) _lowerCamelCase = sum([keepascore, keepascore, keepascore, keepascore] ) / 4 _lowerCamelCase = sum([delascore, delascore, delascore, delascore] ) / 4 _lowerCamelCase = sum([addascore, addascore, addascore, addascore] ) / 4 _lowerCamelCase = (avgkeepscore + avgdelscore + avgaddscore) / 3 return finalscore def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : bool = True , lowercase_ : str = "13a" , lowercase_ : bool = True ) -> int: # Normalization is requried for the ASSET dataset (one of the primary # datasets in sentence simplification) to allow using space # to split the sentence. Even though Wiki-Auto and TURK datasets, # do not require normalization, we do it for consistency. # Code adapted from the EASSE library [1] written by the authors of the ASSET dataset. # [1] https://github.com/feralvam/easse/blob/580bba7e1378fc8289c663f864e0487188fe8067/easse/utils/preprocessing.py#L7 if lowercase: _lowerCamelCase = sentence.lower() if tokenizer in ["13a", "intl"]: if version.parse(sacrebleu.__version__ ).major >= 2: _lowerCamelCase = sacrebleu.metrics.bleu._get_tokenizer(lowercase_ )()(lowercase_ ) else: _lowerCamelCase = sacrebleu.TOKENIZERS[tokenizer]()(lowercase_ ) elif tokenizer == "moses": _lowerCamelCase = sacremoses.MosesTokenizer().tokenize(lowercase_ , return_str=lowercase_ , escape=lowercase_ ) elif tokenizer == "penn": _lowerCamelCase = sacremoses.MosesTokenizer().penn_tokenize(lowercase_ , return_str=lowercase_ ) else: _lowerCamelCase = sentence if not return_str: _lowerCamelCase = normalized_sent.split() return normalized_sent def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Any , lowercase_ : List[Any] ) -> Optional[int]: if not (len(lowercase_ ) == len(lowercase_ ) == len(lowercase_ )): raise ValueError('''Sources length must match predictions and references lengths.''' ) _lowerCamelCase = 0 for src, pred, refs in zip(lowercase_ , lowercase_ , lowercase_ ): sari_score += SARIsent(normalize(lowercase_ ) , normalize(lowercase_ ) , [normalize(lowercase_ ) for sent in refs] ) _lowerCamelCase = sari_score / len(lowercase_ ) return 1_00 * sari_score def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Any , lowercase_ : List[Any]="exp" , lowercase_ : List[Any]=None , lowercase_ : Optional[Any]=False , lowercase_ : List[Any]=False , lowercase_ : List[Any]=False , ) -> Dict: _lowerCamelCase = len(references[0] ) if any(len(lowercase_ ) != references_per_prediction for refs in references ): raise ValueError('''Sacrebleu requires the same number of references for each prediction''' ) _lowerCamelCase = [[refs[i] for refs in references] for i in range(lowercase_ )] _lowerCamelCase = sacrebleu.corpus_bleu( lowercase_ , lowercase_ , smooth_method=lowercase_ , smooth_value=lowercase_ , force=lowercase_ , lowercase=lowercase_ , use_effective_order=lowercase_ , ) return output.score @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Sequence(datasets.Value('''string''' , id='''sequence''' ) , id='''references''' ), } ) , codebase_urls=[ '''https://github.com/huggingface/transformers/blob/master/src/transformers/data/metrics/squad_metrics.py''', '''https://github.com/cocoxu/simplification/blob/master/SARI.py''', '''https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/sari_hook.py''', '''https://github.com/mjpost/sacreBLEU''', ] , reference_urls=[ '''https://www.aclweb.org/anthology/Q16-1029.pdf''', '''https://github.com/mjpost/sacreBLEU''', '''https://en.wikipedia.org/wiki/BLEU''', '''https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213''', ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = {} result.update({'''sari''': compute_sari(sources=lowerCamelCase__ , predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) result.update({'''sacrebleu''': compute_sacrebleu(predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) result.update({'''exact''': compute_em(predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) return result
661
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_videomae import VideoMAEImageProcessor __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , *lowerCamelCase__ , **lowerCamelCase__ ): warnings.warn( '''The class VideoMAEFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use VideoMAEImageProcessor instead.''' , lowerCamelCase__ , ) super().__init__(*lowerCamelCase__ , **lowerCamelCase__ )
661
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''configuration_wav2vec2''': ['''WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Wav2Vec2Config'''], '''feature_extraction_wav2vec2''': ['''Wav2Vec2FeatureExtractor'''], '''processing_wav2vec2''': ['''Wav2Vec2Processor'''], '''tokenization_wav2vec2''': ['''Wav2Vec2CTCTokenizer''', '''Wav2Vec2Tokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = [ '''WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Wav2Vec2ForAudioFrameClassification''', '''Wav2Vec2ForCTC''', '''Wav2Vec2ForMaskedLM''', '''Wav2Vec2ForPreTraining''', '''Wav2Vec2ForSequenceClassification''', '''Wav2Vec2ForXVector''', '''Wav2Vec2Model''', '''Wav2Vec2PreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = [ '''TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWav2Vec2ForCTC''', '''TFWav2Vec2Model''', '''TFWav2Vec2PreTrainedModel''', '''TFWav2Vec2ForSequenceClassification''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = [ '''FlaxWav2Vec2ForCTC''', '''FlaxWav2Vec2ForPreTraining''', '''FlaxWav2Vec2Model''', '''FlaxWav2Vec2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavaveca import WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, WavaVecaConfig from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .processing_wavaveca import WavaVecaProcessor from .tokenization_wavaveca import WavaVecaCTCTokenizer, WavaVecaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavaveca import ( WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaForAudioFrameClassification, WavaVecaForCTC, WavaVecaForMaskedLM, WavaVecaForPreTraining, WavaVecaForSequenceClassification, WavaVecaForXVector, WavaVecaModel, WavaVecaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, TFWavaVecaForCTC, TFWavaVecaForSequenceClassification, TFWavaVecaModel, TFWavaVecaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( FlaxWavaVecaForCTC, FlaxWavaVecaForPreTraining, FlaxWavaVecaModel, FlaxWavaVecaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
661
1
"""simple docstring""" import unittest from transformers import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING, is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class lowerCamelCase_: '''simple docstring''' @staticmethod def snake_case__ ( *lowerCamelCase__ , **lowerCamelCase__ ): pass @is_pipeline_test @require_torch @require_vision class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = pipeline('''visual-question-answering''' , model='''hf-internal-testing/tiny-vilt-random-vqa''' ) _lowerCamelCase = [ { '''image''': Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''question''': '''How many cats are there?''', }, { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''question''': '''How many cats are there?''', }, ] return vqa_pipeline, examples def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = vqa_pipeline(lowerCamelCase__ , top_k=1 ) self.assertEqual( lowerCamelCase__ , [ [{'''score''': ANY(lowerCamelCase__ ), '''answer''': ANY(lowerCamelCase__ )}], [{'''score''': ANY(lowerCamelCase__ ), '''answer''': ANY(lowerCamelCase__ )}], ] , ) @require_torch def snake_case__ ( self ): _lowerCamelCase = pipeline('''visual-question-answering''' , model='''hf-internal-testing/tiny-vilt-random-vqa''' ) _lowerCamelCase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' _lowerCamelCase = '''How many cats are there?''' _lowerCamelCase = vqa_pipeline(image=lowerCamelCase__ , question='''How many cats are there?''' , top_k=2 ) self.assertEqual( lowerCamelCase__ , [{'''score''': ANY(lowerCamelCase__ ), '''answer''': ANY(lowerCamelCase__ )}, {'''score''': ANY(lowerCamelCase__ ), '''answer''': ANY(lowerCamelCase__ )}] ) _lowerCamelCase = vqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( lowerCamelCase__ , [{'''score''': ANY(lowerCamelCase__ ), '''answer''': ANY(lowerCamelCase__ )}, {'''score''': ANY(lowerCamelCase__ ), '''answer''': ANY(lowerCamelCase__ )}] ) @slow @require_torch def snake_case__ ( self ): _lowerCamelCase = pipeline('''visual-question-answering''' , model='''dandelin/vilt-b32-finetuned-vqa''' ) _lowerCamelCase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' _lowerCamelCase = '''How many cats are there?''' _lowerCamelCase = vqa_pipeline(image=lowerCamelCase__ , question=lowerCamelCase__ , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase__ , decimals=4 ) , [{'''score''': 0.8_7_9_9, '''answer''': '''2'''}, {'''score''': 0.2_9_6, '''answer''': '''1'''}] ) _lowerCamelCase = vqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase__ , decimals=4 ) , [{'''score''': 0.8_7_9_9, '''answer''': '''2'''}, {'''score''': 0.2_9_6, '''answer''': '''1'''}] ) _lowerCamelCase = vqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(lowerCamelCase__ , decimals=4 ) , [[{'''score''': 0.8_7_9_9, '''answer''': '''2'''}, {'''score''': 0.2_9_6, '''answer''': '''1'''}]] * 2 , ) @require_tf @unittest.skip('''Visual question answering not implemented in TF''' ) def snake_case__ ( self ): pass
661
"""simple docstring""" __SCREAMING_SNAKE_CASE : Any = { '''A''': '''.-''', '''B''': '''-...''', '''C''': '''-.-.''', '''D''': '''-..''', '''E''': '''.''', '''F''': '''..-.''', '''G''': '''--.''', '''H''': '''....''', '''I''': '''..''', '''J''': '''.---''', '''K''': '''-.-''', '''L''': '''.-..''', '''M''': '''--''', '''N''': '''-.''', '''O''': '''---''', '''P''': '''.--.''', '''Q''': '''--.-''', '''R''': '''.-.''', '''S''': '''...''', '''T''': '''-''', '''U''': '''..-''', '''V''': '''...-''', '''W''': '''.--''', '''X''': '''-..-''', '''Y''': '''-.--''', '''Z''': '''--..''', '''1''': '''.----''', '''2''': '''..---''', '''3''': '''...--''', '''4''': '''....-''', '''5''': '''.....''', '''6''': '''-....''', '''7''': '''--...''', '''8''': '''---..''', '''9''': '''----.''', '''0''': '''-----''', '''&''': '''.-...''', '''@''': '''.--.-.''', ''':''': '''---...''', ''',''': '''--..--''', '''.''': '''.-.-.-''', '''\'''': '''.----.''', '''"''': '''.-..-.''', '''?''': '''..--..''', '''/''': '''-..-.''', '''=''': '''-...-''', '''+''': '''.-.-.''', '''-''': '''-....-''', '''(''': '''-.--.''', ''')''': '''-.--.-''', '''!''': '''-.-.--''', ''' ''': '''/''' } # Exclamation mark is not in ITU-R recommendation # fmt: on __SCREAMING_SNAKE_CASE : List[str] = {value: key for key, value in MORSE_CODE_DICT.items()} def lowerCAmelCase_( lowercase_ : str ) -> str: return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def lowerCAmelCase_( lowercase_ : str ) -> str: return "".join(REVERSE_DICT[char] for char in message.split() ) def lowerCAmelCase_( ) -> None: _lowerCamelCase = '''Morse code here!''' print(lowercase_ ) _lowerCamelCase = encrypt(lowercase_ ) print(lowercase_ ) _lowerCamelCase = decrypt(lowercase_ ) print(lowercase_ ) if __name__ == "__main__": main()
661
1
"""simple docstring""" from typing import Dict, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import flip_channel_order, resize, to_channel_dimension_format, to_pil_image from ...image_utils import ( ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_pytesseract_available, is_vision_available, logging, requires_backends if is_vision_available(): import PIL # soft dependency if is_pytesseract_available(): import pytesseract __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : List[Any] , lowercase_ : Union[str, Any] ) -> Optional[Any]: return [ int(10_00 * (box[0] / width) ), int(10_00 * (box[1] / height) ), int(10_00 * (box[2] / width) ), int(10_00 * (box[3] / height) ), ] def lowerCAmelCase_( lowercase_ : np.ndarray , lowercase_ : Optional[str] , lowercase_ : Optional[str] = None ) -> Union[str, Any]: _lowerCamelCase = tesseract_config if tesseract_config is not None else '''''' # apply OCR _lowerCamelCase = to_pil_image(lowercase_ ) _lowerCamelCase , _lowerCamelCase = pil_image.size _lowerCamelCase = pytesseract.image_to_data(lowercase_ , lang=lowercase_ , output_type='''dict''' , config=lowercase_ ) _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = data['''text'''], data['''left'''], data['''top'''], data['''width'''], data['''height'''] # filter empty words and corresponding coordinates _lowerCamelCase = [idx for idx, word in enumerate(lowercase_ ) if not word.strip()] _lowerCamelCase = [word for idx, word in enumerate(lowercase_ ) if idx not in irrelevant_indices] _lowerCamelCase = [coord for idx, coord in enumerate(lowercase_ ) if idx not in irrelevant_indices] _lowerCamelCase = [coord for idx, coord in enumerate(lowercase_ ) if idx not in irrelevant_indices] _lowerCamelCase = [coord for idx, coord in enumerate(lowercase_ ) if idx not in irrelevant_indices] _lowerCamelCase = [coord for idx, coord in enumerate(lowercase_ ) if idx not in irrelevant_indices] # turn coordinates into (left, top, left+width, top+height) format _lowerCamelCase = [] for x, y, w, h in zip(lowercase_ , lowercase_ , lowercase_ , lowercase_ ): _lowerCamelCase = [x, y, x + w, y + h] actual_boxes.append(lowercase_ ) # finally, normalize the bounding boxes _lowerCamelCase = [] for box in actual_boxes: normalized_boxes.append(normalize_box(lowercase_ , lowercase_ , lowercase_ ) ) assert len(lowercase_ ) == len(lowercase_ ), "Not as many words as there are bounding boxes" return words, normalized_boxes class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Dict = ['pixel_values'] def __init__( self , lowerCamelCase__ = True , lowerCamelCase__ = None , lowerCamelCase__ = PILImageResampling.BILINEAR , lowerCamelCase__ = True , lowerCamelCase__ = None , lowerCamelCase__ = "" , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) _lowerCamelCase = size if size is not None else {'''height''': 2_2_4, '''width''': 2_2_4} _lowerCamelCase = get_size_dict(lowerCamelCase__ ) _lowerCamelCase = do_resize _lowerCamelCase = size _lowerCamelCase = resample _lowerCamelCase = apply_ocr _lowerCamelCase = ocr_lang _lowerCamelCase = tesseract_config def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = PILImageResampling.BILINEAR , lowerCamelCase__ = None , **lowerCamelCase__ , ): _lowerCamelCase = get_size_dict(lowerCamelCase__ ) if "height" not in size or "width" not in size: raise ValueError(F"""The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}""" ) _lowerCamelCase = (size['''height'''], size['''width''']) return resize(lowerCamelCase__ , size=lowerCamelCase__ , resample=lowerCamelCase__ , data_format=lowerCamelCase__ , **lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = None , lowerCamelCase__ = ChannelDimension.FIRST , **lowerCamelCase__ , ): _lowerCamelCase = do_resize if do_resize is not None else self.do_resize _lowerCamelCase = size if size is not None else self.size _lowerCamelCase = get_size_dict(lowerCamelCase__ ) _lowerCamelCase = resample if resample is not None else self.resample _lowerCamelCase = apply_ocr if apply_ocr is not None else self.apply_ocr _lowerCamelCase = ocr_lang if ocr_lang is not None else self.ocr_lang _lowerCamelCase = tesseract_config if tesseract_config is not None else self.tesseract_config _lowerCamelCase = make_list_of_images(lowerCamelCase__ ) if not valid_images(lowerCamelCase__ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) # All transformations expect numpy arrays. _lowerCamelCase = [to_numpy_array(lowerCamelCase__ ) for image in images] if apply_ocr: requires_backends(self , '''pytesseract''' ) _lowerCamelCase = [] _lowerCamelCase = [] for image in images: _lowerCamelCase , _lowerCamelCase = apply_tesseract(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) words_batch.append(lowerCamelCase__ ) boxes_batch.append(lowerCamelCase__ ) if do_resize: _lowerCamelCase = [self.resize(image=lowerCamelCase__ , size=lowerCamelCase__ , resample=lowerCamelCase__ ) for image in images] # flip color channels from RGB to BGR (as Detectron2 requires this) _lowerCamelCase = [flip_channel_order(lowerCamelCase__ ) for image in images] _lowerCamelCase = [to_channel_dimension_format(lowerCamelCase__ , lowerCamelCase__ ) for image in images] _lowerCamelCase = BatchFeature(data={'''pixel_values''': images} , tensor_type=lowerCamelCase__ ) if apply_ocr: _lowerCamelCase = words_batch _lowerCamelCase = boxes_batch return data
661
"""simple docstring""" from argparse import ArgumentParser, Namespace from typing import Any, List, Optional from ..pipelines import Pipeline, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand try: from fastapi import Body, FastAPI, HTTPException from fastapi.routing import APIRoute from pydantic import BaseModel from starlette.responses import JSONResponse from uvicorn import run __SCREAMING_SNAKE_CASE : List[Any] = True except (ImportError, AttributeError): __SCREAMING_SNAKE_CASE : List[Any] = object def lowerCAmelCase_( *lowercase_ : Dict , **lowercase_ : str ) -> str: pass __SCREAMING_SNAKE_CASE : Tuple = False __SCREAMING_SNAKE_CASE : Any = logging.get_logger('''transformers-cli/serving''') def lowerCAmelCase_( lowercase_ : Namespace ) -> List[Any]: _lowerCamelCase = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) return ServeCommand(lowercase_ , args.host , args.port , args.workers ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : dict class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] lowercase__ : Optional[List[int]] class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : str class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Any class lowerCamelCase_( A__ ): '''simple docstring''' @staticmethod def snake_case__ ( lowerCamelCase__ ): _lowerCamelCase = parser.add_parser( '''serve''' , help='''CLI tool to run inference requests through REST and GraphQL endpoints.''' ) serve_parser.add_argument( '''--task''' , type=lowerCamelCase__ , choices=get_supported_tasks() , help='''The task to run the pipeline on''' , ) serve_parser.add_argument('''--host''' , type=lowerCamelCase__ , default='''localhost''' , help='''Interface the server will listen on.''' ) serve_parser.add_argument('''--port''' , type=lowerCamelCase__ , default=8_8_8_8 , help='''Port the serving will listen to.''' ) serve_parser.add_argument('''--workers''' , type=lowerCamelCase__ , default=1 , help='''Number of http workers''' ) serve_parser.add_argument('''--model''' , type=lowerCamelCase__ , help='''Model\'s name or path to stored model.''' ) serve_parser.add_argument('''--config''' , type=lowerCamelCase__ , help='''Model\'s config name or path to stored model.''' ) serve_parser.add_argument('''--tokenizer''' , type=lowerCamelCase__ , help='''Tokenizer name to use.''' ) serve_parser.add_argument( '''--device''' , type=lowerCamelCase__ , default=-1 , help='''Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)''' , ) serve_parser.set_defaults(func=lowerCamelCase__ ) def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = pipeline _lowerCamelCase = host _lowerCamelCase = port _lowerCamelCase = workers if not _serve_dependencies_installed: raise RuntimeError( '''Using serve command requires FastAPI and uvicorn. ''' '''Please install transformers with [serving]: pip install "transformers[serving]".''' '''Or install FastAPI and uvicorn separately.''' ) else: logger.info(F"""Serving model over {host}:{port}""" ) _lowerCamelCase = FastAPI( routes=[ APIRoute( '''/''' , self.model_info , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''GET'''] , ), APIRoute( '''/tokenize''' , self.tokenize , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), APIRoute( '''/detokenize''' , self.detokenize , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), APIRoute( '''/forward''' , self.forward , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), ] , timeout=6_0_0 , ) def snake_case__ ( self ): run(self._app , host=self.host , port=self.port , workers=self.workers ) def snake_case__ ( self ): return ServeModelInfoResult(infos=vars(self._pipeline.model.config ) ) def snake_case__ ( self , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) ): try: _lowerCamelCase = self._pipeline.tokenizer.tokenize(lowerCamelCase__ ) if return_ids: _lowerCamelCase = self._pipeline.tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) return ServeTokenizeResult(tokens=lowerCamelCase__ , tokens_ids=lowerCamelCase__ ) else: return ServeTokenizeResult(tokens=lowerCamelCase__ ) except Exception as e: raise HTTPException(status_code=5_0_0 , detail={'''model''': '''''', '''error''': str(lowerCamelCase__ )} ) def snake_case__ ( self , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , ): try: _lowerCamelCase = self._pipeline.tokenizer.decode(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) return ServeDeTokenizeResult(model='''''' , text=lowerCamelCase__ ) except Exception as e: raise HTTPException(status_code=5_0_0 , detail={'''model''': '''''', '''error''': str(lowerCamelCase__ )} ) async def snake_case__ ( self , lowerCamelCase__=Body(lowerCamelCase__ , embed=lowerCamelCase__ ) ): # Check we don't have empty string if len(lowerCamelCase__ ) == 0: return ServeForwardResult(output=[] , attention=[] ) try: # Forward through the model _lowerCamelCase = self._pipeline(lowerCamelCase__ ) return ServeForwardResult(output=lowerCamelCase__ ) except Exception as e: raise HTTPException(5_0_0 , {'''error''': str(lowerCamelCase__ )} )
661
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : Dict ) -> str: _lowerCamelCase = [0] * len(lowercase_ ) _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = 0 for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(lowercase_ ) ): if indegree[i] == 0: queue.append(lowercase_ ) while queue: _lowerCamelCase = queue.pop(0 ) cnt += 1 topo.append(lowercase_ ) for x in graph[vertex]: indegree[x] -= 1 if indegree[x] == 0: queue.append(lowercase_ ) if cnt != len(lowercase_ ): print('''Cycle exists''' ) else: print(lowercase_ ) # Adjacency List of Graph __SCREAMING_SNAKE_CASE : Optional[Any] = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []} topological_sort(graph)
661
"""simple docstring""" import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def lowerCAmelCase_( lowercase_ : str , lowercase_ : Dict ) -> List[Any]: assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : str , lowercase_ : str ) -> List[Any]: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ , keep_in_memory=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : int ) -> Optional[int]: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = features.copy() if features else default_expected_features _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = JsonDatasetReader(lowercase_ , features=lowercase_ , cache_dir=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''}, ] , ) def lowerCAmelCase_( lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : List[Any] ) -> Tuple: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''} _lowerCamelCase = features.copy() if features else default_expected_features _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = JsonDatasetReader(lowercase_ , features=lowercase_ , cache_dir=lowercase_ ).read() assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Optional[int] ) -> List[str]: # jsonl_312_path features are {"col_3": "float64", "col_1": "string", "col_2": "int64"} _lowerCamelCase = {'''col_2''': '''int64''', '''col_3''': '''float64''', '''col_1''': '''string'''} _lowerCamelCase = features.copy() _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = JsonDatasetReader(lowercase_ , features=lowercase_ , cache_dir=lowercase_ ).read() assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : Optional[Any] ) -> int: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ , split=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('''path_type''' , [str, list] ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[int] , lowercase_ : Tuple ) -> Optional[int]: if issubclass(lowercase_ , lowercase_ ): _lowerCamelCase = jsonl_path elif issubclass(lowercase_ , lowercase_ ): _lowerCamelCase = [jsonl_path] _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : int , lowercase_ : List[str] , lowercase_ : str=("train",) ) -> List[str]: assert isinstance(lowercase_ , lowercase_ ) for split in splits: _lowerCamelCase = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[int] , lowercase_ : int ) -> Dict: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _lowerCamelCase = JsonDatasetReader({'''train''': jsonl_path} , cache_dir=lowercase_ , keep_in_memory=lowercase_ ).read() _check_json_datasetdict(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : Union[str, Any] , lowercase_ : Optional[Any] ) -> List[Any]: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = features.copy() if features else default_expected_features _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = JsonDatasetReader({'''train''': jsonl_path} , features=lowercase_ , cache_dir=lowercase_ ).read() _check_json_datasetdict(lowercase_ , lowercase_ ) @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : str , lowercase_ : List[str] ) -> Optional[Any]: if split: _lowerCamelCase = {split: jsonl_path} else: _lowerCamelCase = '''train''' _lowerCamelCase = {'''train''': jsonl_path, '''test''': jsonl_path} _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ ).read() _check_json_datasetdict(lowercase_ , lowercase_ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def lowerCAmelCase_( lowercase_ : Optional[int] ) -> Optional[Any]: return json.load(lowercase_ ) def lowerCAmelCase_( lowercase_ : Tuple ) -> Tuple: return [json.loads(lowercase_ ) for line in buffer] class lowerCamelCase_: '''simple docstring''' @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ ).write() buffer.seek(0 ) _lowerCamelCase = load_json_function(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) assert isinstance(exported_content[0] , lowerCamelCase__ ) assert len(lowerCamelCase__ ) == 1_0 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ , orient=lowerCamelCase__ ).write() buffer.seek(0 ) _lowerCamelCase = load_json(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowerCamelCase__ , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 1_0 else: assert len(lowerCamelCase__ ) == 1_0 @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ , num_proc=2 ).write() buffer.seek(0 ) _lowerCamelCase = load_json_function(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) assert isinstance(exported_content[0] , lowerCamelCase__ ) assert len(lowerCamelCase__ ) == 1_0 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ , orient=lowerCamelCase__ , num_proc=2 ).write() buffer.seek(0 ) _lowerCamelCase = load_json(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowerCamelCase__ , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 1_0 else: assert len(lowerCamelCase__ ) == 1_0 def snake_case__ ( self , lowerCamelCase__ ): with pytest.raises(lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , num_proc=0 ) @pytest.mark.parametrize('''compression, extension''' , [('''gzip''', '''gz'''), ('''bz2''', '''bz2'''), ('''xz''', '''xz''')] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = tmp_path_factory.mktemp('''data''' ) / F"""test.json.{extension}""" _lowerCamelCase = str(shared_datadir / F"""test_file.json.{extension}""" ) JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , compression=lowerCamelCase__ ).write() with fsspec.open(lowerCamelCase__ , '''rb''' , compression='''infer''' ) as f: _lowerCamelCase = f.read() with fsspec.open(lowerCamelCase__ , '''rb''' , compression='''infer''' ) as f: _lowerCamelCase = f.read() assert exported_content == original_content
661
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Dict = { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json''' ), } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Any = 'dpr' def __init__( self , lowerCamelCase__=3_0_5_2_2 , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-12 , lowerCamelCase__=0 , lowerCamelCase__="absolute" , lowerCamelCase__ = 0 , **lowerCamelCase__ , ): super().__init__(pad_token_id=lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = hidden_act _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = initializer_range _lowerCamelCase = layer_norm_eps _lowerCamelCase = projection_dim _lowerCamelCase = position_embedding_type
661
"""simple docstring""" import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=3 , lowerCamelCase__=3_2 , lowerCamelCase__=3 , lowerCamelCase__=1_0 , lowerCamelCase__=[8, 1_6, 3_2, 6_4] , lowerCamelCase__=[1, 1, 2, 1] , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__="relu" , lowerCamelCase__=3 , lowerCamelCase__=None , lowerCamelCase__=["stage2", "stage3", "stage4"] , lowerCamelCase__=[2, 3, 4] , lowerCamelCase__=1 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = num_channels _lowerCamelCase = embeddings_size _lowerCamelCase = hidden_sizes _lowerCamelCase = depths _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_act _lowerCamelCase = num_labels _lowerCamelCase = scope _lowerCamelCase = len(lowerCamelCase__ ) _lowerCamelCase = out_features _lowerCamelCase = out_indices _lowerCamelCase = num_groups def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.num_labels ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = BitModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = BitForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = BitBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None _lowerCamelCase = None _lowerCamelCase = BitBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Dict = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () lowercase__ : Any = ( {'feature-extraction': BitModel, 'image-classification': BitForImageClassification} if is_torch_available() else {} ) lowercase__ : Union[str, Any] = False lowercase__ : List[Any] = False lowercase__ : Any = False lowercase__ : List[str] = False lowercase__ : Any = False def snake_case__ ( self ): _lowerCamelCase = BitModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ ) def snake_case__ ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case__ ( self ): return @unittest.skip(reason='''Bit does not output attentions''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''Bit does not use inputs_embeds''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''Bit does not support input and output embeddings''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(config=lowerCamelCase__ ) for name, module in model.named_modules(): if isinstance(lowerCamelCase__ , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) def snake_case__ ( self ): def check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _lowerCamelCase = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase__ ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = ['''preactivation''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: _lowerCamelCase = layer_type _lowerCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) @unittest.skip(reason='''Bit does not use feedforward chunking''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = BitModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def snake_case__ ( self ): _lowerCamelCase = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowerCamelCase__ ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**lowerCamelCase__ ) # verify the logits _lowerCamelCase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([[-0.6_5_2_6, -0.5_2_6_3, -1.4_3_9_8]] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) ) @require_torch class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = (BitBackbone,) if is_torch_available() else () lowercase__ : Tuple = BitConfig lowercase__ : Any = False def snake_case__ ( self ): _lowerCamelCase = BitModelTester(self )
661
1
"""simple docstring""" import argparse import datetime import json import time import warnings from logging import getLogger from pathlib import Path from typing import Dict, List import torch from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from utils import calculate_bleu, calculate_rouge, chunks, parse_numeric_n_bool_cl_kwargs, use_task_specific_params __SCREAMING_SNAKE_CASE : Optional[int] = getLogger(__name__) __SCREAMING_SNAKE_CASE : str = '''cuda''' if torch.cuda.is_available() else '''cpu''' def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : str , lowercase_ : str , lowercase_ : int = 8 , lowercase_ : str = DEFAULT_DEVICE , lowercase_ : Union[str, Any]=False , lowercase_ : Tuple="summarization" , lowercase_ : Union[str, Any]=None , **lowercase_ : Optional[Any] , ) -> Dict: _lowerCamelCase = Path(lowercase_ ).open('''w''' , encoding='''utf-8''' ) _lowerCamelCase = str(lowercase_ ) _lowerCamelCase = AutoModelForSeqaSeqLM.from_pretrained(lowercase_ ).to(lowercase_ ) if fpaa: _lowerCamelCase = model.half() _lowerCamelCase = AutoTokenizer.from_pretrained(lowercase_ ) logger.info(F"""Inferred tokenizer type: {tokenizer.__class__}""" ) # if this is wrong, check config.model_type. _lowerCamelCase = time.time() # update config with task specific params use_task_specific_params(lowercase_ , lowercase_ ) if prefix is None: _lowerCamelCase = prefix or getattr(model.config , '''prefix''' , '''''' ) or '''''' for examples_chunk in tqdm(list(chunks(lowercase_ , lowercase_ ) ) ): _lowerCamelCase = [prefix + text for text in examples_chunk] _lowerCamelCase = tokenizer(lowercase_ , return_tensors='''pt''' , truncation=lowercase_ , padding='''longest''' ).to(lowercase_ ) _lowerCamelCase = model.generate( input_ids=batch.input_ids , attention_mask=batch.attention_mask , **lowercase_ , ) _lowerCamelCase = tokenizer.batch_decode(lowercase_ , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ ) for hypothesis in dec: fout.write(hypothesis + '''\n''' ) fout.flush() fout.close() _lowerCamelCase = int(time.time() - start_time ) # seconds _lowerCamelCase = len(lowercase_ ) return {"n_obs": n_obs, "runtime": runtime, "seconds_per_sample": round(runtime / n_obs , 4 )} def lowerCAmelCase_( ) -> Tuple: return datetime.datetime.now().strftime('''%Y-%m-%d %H:%M:%S''' ) def lowerCAmelCase_( lowercase_ : List[Any]=True ) -> List[str]: _lowerCamelCase = argparse.ArgumentParser() parser.add_argument('''model_name''' , type=lowercase_ , help='''like facebook/bart-large-cnn,t5-base, etc.''' ) parser.add_argument('''input_path''' , type=lowercase_ , help='''like cnn_dm/test.source''' ) parser.add_argument('''save_path''' , type=lowercase_ , help='''where to save summaries''' ) parser.add_argument('''--reference_path''' , type=lowercase_ , required=lowercase_ , help='''like cnn_dm/test.target''' ) parser.add_argument('''--score_path''' , type=lowercase_ , required=lowercase_ , default='''metrics.json''' , help='''where to save metrics''' ) parser.add_argument('''--device''' , type=lowercase_ , required=lowercase_ , default=lowercase_ , help='''cuda, cuda:1, cpu etc.''' ) parser.add_argument( '''--prefix''' , type=lowercase_ , required=lowercase_ , default=lowercase_ , help='''will be added to the begininng of src examples''' ) parser.add_argument('''--task''' , type=lowercase_ , default='''summarization''' , help='''used for task_specific_params + metrics''' ) parser.add_argument('''--bs''' , type=lowercase_ , default=8 , required=lowercase_ , help='''batch size''' ) parser.add_argument( '''--n_obs''' , type=lowercase_ , default=-1 , required=lowercase_ , help='''How many observations. Defaults to all.''' ) parser.add_argument('''--fp16''' , action='''store_true''' ) parser.add_argument('''--dump-args''' , action='''store_true''' , help='''print the custom hparams with the results''' ) parser.add_argument( '''--info''' , nargs='''?''' , type=lowercase_ , const=datetime_now() , help=( '''use in conjunction w/ --dump-args to print with the results whatever other info you\'d like, e.g.''' ''' lang=en-ru. If no value is passed, the current datetime string will be used.''' ) , ) # Unspecified args like --num_beams=2 --decoder_start_token_id=4 are passed to model.generate _lowerCamelCase , _lowerCamelCase = parser.parse_known_args() _lowerCamelCase = parse_numeric_n_bool_cl_kwargs(lowercase_ ) if parsed_args and verbose: print(F"""parsed the following generate kwargs: {parsed_args}""" ) _lowerCamelCase = [''' ''' + x.rstrip() if '''t5''' in args.model_name else x.rstrip() for x in open(args.input_path ).readlines()] if args.n_obs > 0: _lowerCamelCase = examples[: args.n_obs] Path(args.save_path ).parent.mkdir(exist_ok=lowercase_ ) if args.reference_path is None and Path(args.score_path ).exists(): warnings.warn(F"""score_path {args.score_path} will be overwritten unless you type ctrl-c.""" ) if args.device == "cpu" and args.fpaa: # this mix leads to RuntimeError: "threshold_cpu" not implemented for 'Half' raise ValueError('''Can\'t mix --fp16 and --device cpu''' ) _lowerCamelCase = generate_summaries_or_translations( lowercase_ , args.save_path , args.model_name , batch_size=args.bs , device=args.device , fpaa=args.fpaa , task=args.task , prefix=args.prefix , **lowercase_ , ) if args.reference_path is None: return {} # Compute scores _lowerCamelCase = calculate_bleu if '''translation''' in args.task else calculate_rouge _lowerCamelCase = [x.rstrip() for x in open(args.save_path ).readlines()] _lowerCamelCase = [x.rstrip() for x in open(args.reference_path ).readlines()][: len(lowercase_ )] _lowerCamelCase = score_fn(lowercase_ , lowercase_ ) scores.update(lowercase_ ) if args.dump_args: scores.update(lowercase_ ) if args.info: _lowerCamelCase = args.info if verbose: print(lowercase_ ) if args.score_path is not None: json.dump(lowercase_ , open(args.score_path , '''w''' ) ) return scores if __name__ == "__main__": # Usage for MT: # python run_eval.py MODEL_NAME $DATA_DIR/test.source $save_dir/test_translations.txt --reference_path $DATA_DIR/test.target --score_path $save_dir/test_bleu.json --task translation $@ run_generate(verbose=True)
661
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=2 , lowerCamelCase__=2_4 , lowerCamelCase__=1_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=3_2 , lowerCamelCase__=5 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1_0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=None , lowerCamelCase__=2 , lowerCamelCase__=2 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = patch_size _lowerCamelCase = max_length _lowerCamelCase = num_mel_bins _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = scope _lowerCamelCase = frequency_stride _lowerCamelCase = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _lowerCamelCase = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _lowerCamelCase = (self.max_length - self.patch_size) // self.time_stride + 1 _lowerCamelCase = frequency_out_dimension * time_out_dimension _lowerCamelCase = num_patches + 2 def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, input_values, labels def snake_case__ ( self ): return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = ASTModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = config_and_inputs _lowerCamelCase = {'''input_values''': input_values} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Any = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) lowercase__ : List[str] = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) lowercase__ : int = False lowercase__ : str = False lowercase__ : Union[str, Any] = False lowercase__ : List[str] = False def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def snake_case__ ( self ): _lowerCamelCase = ASTModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , nn.Linear ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''input_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = ASTModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> str: _lowerCamelCase = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) _lowerCamelCase , _lowerCamelCase = torchaudio.load(lowercase_ ) return audio, sampling_rate @require_torch @require_torchaudio class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def snake_case__ ( self ): _lowerCamelCase = self.default_feature_extractor _lowerCamelCase = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(lowerCamelCase__ ) _lowerCamelCase = self.default_feature_extractor _lowerCamelCase , _lowerCamelCase = prepare_audio() _lowerCamelCase = audio.squeeze().numpy() _lowerCamelCase = feature_extractor(lowerCamelCase__ , sampling_rate=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**lowerCamelCase__ ) # verify the logits _lowerCamelCase = torch.Size((1, 5_2_7) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([-0.8_7_6_0, -7.0_0_4_2, -8.6_6_0_2] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) )
661
1
"""simple docstring""" from __future__ import annotations from collections.abc import MutableSequence class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): if len(lowerCamelCase__ ) != degree + 1: raise ValueError( '''The number of coefficients should be equal to the degree + 1.''' ) _lowerCamelCase = list(lowerCamelCase__ ) _lowerCamelCase = degree def __add__( self , lowerCamelCase__ ): if self.degree > polynomial_a.degree: _lowerCamelCase = self.coefficients[:] for i in range(polynomial_a.degree + 1 ): coefficients[i] += polynomial_a.coefficients[i] return Polynomial(self.degree , lowerCamelCase__ ) else: _lowerCamelCase = polynomial_a.coefficients[:] for i in range(self.degree + 1 ): coefficients[i] += self.coefficients[i] return Polynomial(polynomial_a.degree , lowerCamelCase__ ) def __sub__( self , lowerCamelCase__ ): return self + polynomial_a * Polynomial(0 , [-1] ) def __neg__( self ): return Polynomial(self.degree , [-c for c in self.coefficients] ) def __mul__( self , lowerCamelCase__ ): _lowerCamelCase = [0] * (self.degree + polynomial_a.degree + 1) for i in range(self.degree + 1 ): for j in range(polynomial_a.degree + 1 ): coefficients[i + j] += ( self.coefficients[i] * polynomial_a.coefficients[j] ) return Polynomial(self.degree + polynomial_a.degree , lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = 0 for i in range(self.degree + 1 ): result += self.coefficients[i] * (substitution**i) return result def __str__( self ): _lowerCamelCase = '''''' for i in range(self.degree , -1 , -1 ): if self.coefficients[i] == 0: continue elif self.coefficients[i] > 0: if polynomial: polynomial += " + " else: polynomial += " - " if i == 0: polynomial += str(abs(self.coefficients[i] ) ) elif i == 1: polynomial += str(abs(self.coefficients[i] ) ) + "x" else: polynomial += str(abs(self.coefficients[i] ) ) + "x^" + str(lowerCamelCase__ ) return polynomial def __repr__( self ): return self.__str__() def snake_case__ ( self ): _lowerCamelCase = [0] * self.degree for i in range(self.degree ): _lowerCamelCase = self.coefficients[i + 1] * (i + 1) return Polynomial(self.degree - 1 , lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ = 0 ): _lowerCamelCase = [0] * (self.degree + 2) _lowerCamelCase = constant for i in range(self.degree + 1 ): _lowerCamelCase = self.coefficients[i] / (i + 1) return Polynomial(self.degree + 1 , lowerCamelCase__ ) def __eq__( self , lowerCamelCase__ ): if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): return False if self.degree != polynomial_a.degree: return False for i in range(self.degree + 1 ): if self.coefficients[i] != polynomial_a.coefficients[i]: return False return True def __ne__( self , lowerCamelCase__ ): return not self.__eq__(lowerCamelCase__ )
661
"""simple docstring""" import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class lowerCamelCase_: '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return None class lowerCamelCase_: '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return None class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = [ # (model_name, model_kwargs) ('bert-base-cased', {}), ('gpt2', {'use_cache': False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase__ , '''tf''' , 1_2 , **lowerCamelCase__ ) @require_torch @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , **lowerCamelCase__ ) @require_torch @slow def snake_case__ ( self ): from transformers import BertModel _lowerCamelCase = ['''[UNK]''', '''[SEP]''', '''[CLS]''', '''[PAD]''', '''[MASK]''', '''some''', '''other''', '''words'''] with NamedTemporaryFile(mode='''w+t''' ) as vocab_file: vocab_file.write('''\n'''.join(lowerCamelCase__ ) ) vocab_file.flush() _lowerCamelCase = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: _lowerCamelCase = BertModel(BertConfig(vocab_size=len(lowerCamelCase__ ) ) ) model.save_pretrained(lowerCamelCase__ ) self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , lowerCamelCase__ ) @require_tf @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCamelCase = self._test_export(lowerCamelCase__ , '''tf''' , 1_2 , **lowerCamelCase__ ) _lowerCamelCase = quantize(Path(lowerCamelCase__ ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) @require_torch @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCamelCase = self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , **lowerCamelCase__ ) _lowerCamelCase = quantize(lowerCamelCase__ ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): try: # Compute path with TemporaryDirectory() as tempdir: _lowerCamelCase = Path(lowerCamelCase__ ).joinpath('''model.onnx''' ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) return path except Exception as e: self.fail(lowerCamelCase__ ) @require_torch @require_tokenizers @slow def snake_case__ ( self ): from transformers import BertModel _lowerCamelCase = BertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) _lowerCamelCase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(lowerCamelCase__ , lowerCamelCase__ , '''pt''' ) @require_tf @require_tokenizers @slow def snake_case__ ( self ): from transformers import TFBertModel _lowerCamelCase = TFBertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) _lowerCamelCase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(lowerCamelCase__ , lowerCamelCase__ , '''tf''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = FeatureExtractionPipeline(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''output_0''', '''output_1'''] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = infer_shapes(lowerCamelCase__ , lowerCamelCase__ ) # Assert all variables are present self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] , lowerCamelCase__ ) self.assertSequenceEqual(variable_names[3:] , lowerCamelCase__ ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] , {0: '''batch''', 1: '''sequence'''} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes['''output_0'''] , {0: '''batch''', 1: '''sequence'''} ) self.assertDictEqual(shapes['''output_1'''] , {0: '''batch'''} ) def snake_case__ ( self ): _lowerCamelCase = ['''input_ids''', '''attention_mask''', '''token_type_ids'''] _lowerCamelCase = {'''input_ids''': [1, 2, 3, 4], '''attention_mask''': [0, 0, 0, 0], '''token_type_ids''': [1, 1, 1, 1]} _lowerCamelCase , _lowerCamelCase = ensure_valid_input(FuncContiguousArgs() , lowerCamelCase__ , lowerCamelCase__ ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(lowerCamelCase__ ) , 3 ) # Should have exactly the same input names self.assertEqual(set(lowerCamelCase__ ) , set(lowerCamelCase__ ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(lowerCamelCase__ , (tokens['''input_ids'''], tokens['''token_type_ids'''], tokens['''attention_mask''']) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) _lowerCamelCase , _lowerCamelCase = ensure_valid_input(FuncNonContiguousArgs() , lowerCamelCase__ , lowerCamelCase__ ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(lowerCamelCase__ ) , 1 ) self.assertEqual(len(lowerCamelCase__ ) , 1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] , tokens['''input_ids'''] ) self.assertEqual(ordered_input_names[0] , '''input_ids''' ) def snake_case__ ( self ): _lowerCamelCase = generate_identified_filename(Path('''/home/something/my_fake_model.onnx''' ) , '''-test''' ) self.assertEqual('''/home/something/my_fake_model-test.onnx''' , generated.as_posix() )
661
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : List[str] = { '''google/vivit-b-16x2-kinetics400''': ( '''https://huggingface.co/google/vivit-b-16x2-kinetics400/resolve/main/config.json''' ), # See all Vivit models at https://huggingface.co/models?filter=vivit } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] = 'vivit' def __init__( self , lowerCamelCase__=2_2_4 , lowerCamelCase__=3_2 , lowerCamelCase__=[2, 1_6, 1_6] , lowerCamelCase__=3 , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__="gelu_fast" , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-06 , lowerCamelCase__=True , **lowerCamelCase__ , ): _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = initializer_range _lowerCamelCase = layer_norm_eps _lowerCamelCase = image_size _lowerCamelCase = num_frames _lowerCamelCase = tubelet_size _lowerCamelCase = num_channels _lowerCamelCase = qkv_bias super().__init__(**lowerCamelCase__ )
661
"""simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = {str(digit): digit**5 for digit in range(1_0)} def lowerCAmelCase_( lowercase_ : int ) -> int: return sum(DIGITS_FIFTH_POWER[digit] for digit in str(lowercase_ ) ) def lowerCAmelCase_( ) -> int: return sum( number for number in range(10_00 , 1_00_00_00 ) if number == digits_fifth_powers_sum(lowercase_ ) ) if __name__ == "__main__": print(solution())
661
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import RoFormerConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerModel, ) from transformers.models.roformer.modeling_tf_roformer import ( TFRoFormerSelfAttention, TFRoFormerSinusoidalPositionalEmbedding, ) class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=9_9 , lowerCamelCase__=3_2 , lowerCamelCase__=2 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_6 , lowerCamelCase__=2 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3 , lowerCamelCase__=4 , lowerCamelCase__=None , ): _lowerCamelCase = parent _lowerCamelCase = 1_3 _lowerCamelCase = 7 _lowerCamelCase = True _lowerCamelCase = True _lowerCamelCase = True _lowerCamelCase = True _lowerCamelCase = 9_9 _lowerCamelCase = 3_2 _lowerCamelCase = 2 _lowerCamelCase = 4 _lowerCamelCase = 3_7 _lowerCamelCase = '''gelu''' _lowerCamelCase = 0.1 _lowerCamelCase = 0.1 _lowerCamelCase = 5_1_2 _lowerCamelCase = 1_6 _lowerCamelCase = 2 _lowerCamelCase = 0.0_2 _lowerCamelCase = 3 _lowerCamelCase = 4 _lowerCamelCase = None def snake_case__ ( self ): _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase = None if self.use_input_mask: _lowerCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) _lowerCamelCase = None if self.use_token_type_ids: _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase = None _lowerCamelCase = None _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase = ids_tensor([self.batch_size] , self.num_choices ) _lowerCamelCase = RoFormerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=lowerCamelCase__ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFRoFormerModel(config=lowerCamelCase__ ) _lowerCamelCase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} _lowerCamelCase = [input_ids, input_mask] _lowerCamelCase = model(lowerCamelCase__ ) _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = True _lowerCamelCase = TFRoFormerForCausalLM(config=lowerCamelCase__ ) _lowerCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _lowerCamelCase = model(lowerCamelCase__ )['''logits'''] self.parent.assertListEqual( list(prediction_scores.numpy().shape ) , [self.batch_size, self.seq_length, self.vocab_size] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFRoFormerForMaskedLM(config=lowerCamelCase__ ) _lowerCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = TFRoFormerForSequenceClassification(config=lowerCamelCase__ ) _lowerCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_choices _lowerCamelCase = TFRoFormerForMultipleChoice(config=lowerCamelCase__ ) _lowerCamelCase = tf.tile(tf.expand_dims(lowerCamelCase__ , 1 ) , (1, self.num_choices, 1) ) _lowerCamelCase = tf.tile(tf.expand_dims(lowerCamelCase__ , 1 ) , (1, self.num_choices, 1) ) _lowerCamelCase = tf.tile(tf.expand_dims(lowerCamelCase__ , 1 ) , (1, self.num_choices, 1) ) _lowerCamelCase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = TFRoFormerForTokenClassification(config=lowerCamelCase__ ) _lowerCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = TFRoFormerForQuestionAnswering(config=lowerCamelCase__ ) _lowerCamelCase = { '''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids, } _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = config_and_inputs _lowerCamelCase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Any = ( ( TFRoFormerModel, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerForMultipleChoice, ) if is_tf_available() else () ) lowercase__ : Optional[int] = ( { 'feature-extraction': TFRoFormerModel, 'fill-mask': TFRoFormerForMaskedLM, 'question-answering': TFRoFormerForQuestionAnswering, 'text-classification': TFRoFormerForSequenceClassification, 'text-generation': TFRoFormerForCausalLM, 'token-classification': TFRoFormerForTokenClassification, 'zero-shot': TFRoFormerForSequenceClassification, } if is_tf_available() else {} ) lowercase__ : Optional[Any] = False lowercase__ : Dict = False def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if pipeline_test_casse_name == "TextGenerationPipelineTests": return True return False def snake_case__ ( self ): _lowerCamelCase = TFRoFormerModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowerCamelCase__ ) @slow def snake_case__ ( self ): _lowerCamelCase = TFRoFormerModel.from_pretrained('''junnyu/roformer_chinese_base''' ) self.assertIsNotNone(lowerCamelCase__ ) @require_tf class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @slow def snake_case__ ( self ): _lowerCamelCase = TFRoFormerForMaskedLM.from_pretrained('''junnyu/roformer_chinese_base''' ) _lowerCamelCase = tf.constant([[0, 1, 2, 3, 4, 5]] ) _lowerCamelCase = model(lowerCamelCase__ )[0] # TODO Replace vocab size _lowerCamelCase = 5_0_0_0_0 _lowerCamelCase = [1, 6, vocab_size] self.assertEqual(output.shape , lowerCamelCase__ ) print(output[:, :3, :3] ) # TODO Replace values below with what was printed above. _lowerCamelCase = tf.constant( [ [ [-0.1_2_0_5_3_3_4_1, -1.0_2_6_4_9_0_1, 0.2_9_2_2_1_9_4_6], [-1.5_1_3_3_7_8_3, 0.1_9_7_4_3_3, 0.1_5_1_9_0_6_0_7], [-5.0_1_3_5_4_0_3, -3.9_0_0_2_5_6, -0.8_4_0_3_8_7_6_4], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , lowerCamelCase__ , atol=1e-4 ) @require_tf class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' lowercase__ : int = 1E-4 def snake_case__ ( self ): _lowerCamelCase = tf.constant([[4, 1_0]] ) _lowerCamelCase = TFRoFormerSinusoidalPositionalEmbedding(num_positions=6 , embedding_dim=6 ) _lowerCamelCase = emba(input_ids.shape ) _lowerCamelCase = tf.constant( [[0.0_0_0_0, 0.0_0_0_0, 0.0_0_0_0, 1.0_0_0_0, 1.0_0_0_0, 1.0_0_0_0], [0.8_4_1_5, 0.0_4_6_4, 0.0_0_2_2, 0.5_4_0_3, 0.9_9_8_9, 1.0_0_0_0]] ) tf.debugging.assert_near(lowerCamelCase__ , lowerCamelCase__ , atol=self.tolerance ) def snake_case__ ( self ): _lowerCamelCase = tf.constant( [ [0.0_0_0_0, 0.0_0_0_0, 0.0_0_0_0, 0.0_0_0_0, 0.0_0_0_0], [0.8_4_1_5, 0.8_2_1_9, 0.8_0_2_0, 0.7_8_1_9, 0.7_6_1_7], [0.9_0_9_3, 0.9_3_6_4, 0.9_5_8_1, 0.9_7_4_9, 0.9_8_7_0], ] ) _lowerCamelCase = TFRoFormerSinusoidalPositionalEmbedding(num_positions=5_1_2 , embedding_dim=5_1_2 ) emba([2, 1_6, 5_1_2] ) _lowerCamelCase = emba.weight[:3, :5] tf.debugging.assert_near(lowerCamelCase__ , lowerCamelCase__ , atol=self.tolerance ) @require_tf class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = 1E-4 def snake_case__ ( self ): # 2,12,16,64 _lowerCamelCase = tf.reshape(tf.range(2 * 1_2 * 1_6 * 6_4 , dtype=tf.floataa ) , shape=(2, 1_2, 1_6, 6_4) ) / 1_0_0 _lowerCamelCase = -tf.reshape(tf.range(2 * 1_2 * 1_6 * 6_4 , dtype=tf.floataa ) , shape=(2, 1_2, 1_6, 6_4) ) / 1_0_0 _lowerCamelCase = TFRoFormerSinusoidalPositionalEmbedding(num_positions=3_2 , embedding_dim=6_4 ) _lowerCamelCase = embed_positions([2, 1_6, 7_6_8] )[None, None, :, :] _lowerCamelCase , _lowerCamelCase = TFRoFormerSelfAttention.apply_rotary_position_embeddings( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = tf.constant( [ [0.0_0_0_0, 0.0_1_0_0, 0.0_2_0_0, 0.0_3_0_0, 0.0_4_0_0, 0.0_5_0_0, 0.0_6_0_0, 0.0_7_0_0], [-0.2_0_1_2, 0.8_8_9_7, 0.0_2_6_3, 0.9_4_0_1, 0.2_0_7_4, 0.9_4_6_3, 0.3_4_8_1, 0.9_3_4_3], [-1.7_0_5_7, 0.6_2_7_1, -1.2_1_4_5, 1.3_8_9_7, -0.6_3_0_3, 1.7_6_4_7, -0.1_1_7_3, 1.8_9_8_5], [-2.1_7_3_1, -1.6_3_9_7, -2.7_3_5_8, 0.2_8_5_4, -2.1_8_4_0, 1.7_1_8_3, -1.3_0_1_8, 2.4_8_7_1], [0.2_7_1_7, -3.6_1_7_3, -2.9_2_0_6, -2.1_9_8_8, -3.6_6_3_8, 0.3_8_5_8, -2.9_1_5_5, 2.2_9_8_0], [3.9_8_5_9, -2.1_5_8_0, -0.7_9_8_4, -4.4_9_0_4, -4.1_1_8_1, -2.0_2_5_2, -4.4_7_8_2, 1.1_2_5_3], ] ) _lowerCamelCase = tf.constant( [ [0.0_0_0_0, -0.0_1_0_0, -0.0_2_0_0, -0.0_3_0_0, -0.0_4_0_0, -0.0_5_0_0, -0.0_6_0_0, -0.0_7_0_0], [0.2_0_1_2, -0.8_8_9_7, -0.0_2_6_3, -0.9_4_0_1, -0.2_0_7_4, -0.9_4_6_3, -0.3_4_8_1, -0.9_3_4_3], [1.7_0_5_7, -0.6_2_7_1, 1.2_1_4_5, -1.3_8_9_7, 0.6_3_0_3, -1.7_6_4_7, 0.1_1_7_3, -1.8_9_8_5], [2.1_7_3_1, 1.6_3_9_7, 2.7_3_5_8, -0.2_8_5_4, 2.1_8_4_0, -1.7_1_8_3, 1.3_0_1_8, -2.4_8_7_1], [-0.2_7_1_7, 3.6_1_7_3, 2.9_2_0_6, 2.1_9_8_8, 3.6_6_3_8, -0.3_8_5_8, 2.9_1_5_5, -2.2_9_8_0], [-3.9_8_5_9, 2.1_5_8_0, 0.7_9_8_4, 4.4_9_0_4, 4.1_1_8_1, 2.0_2_5_2, 4.4_7_8_2, -1.1_2_5_3], ] ) tf.debugging.assert_near(query_layer[0, 0, :6, :8] , lowerCamelCase__ , atol=self.tolerance ) tf.debugging.assert_near(key_layer[0, 0, :6, :8] , lowerCamelCase__ , atol=self.tolerance )
661
"""simple docstring""" from __future__ import annotations import os from collections.abc import Mapping __SCREAMING_SNAKE_CASE : str = tuple[int, int] class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = vertices _lowerCamelCase = { (min(lowerCamelCase__ ), max(lowerCamelCase__ )): weight for edge, weight in edges.items() } def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): self.vertices.add(edge[0] ) self.vertices.add(edge[1] ) _lowerCamelCase = weight def snake_case__ ( self ): _lowerCamelCase = Graph({min(self.vertices )} , {} ) _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 while len(subgraph.vertices ) < len(self.vertices ): _lowerCamelCase = max(self.edges.values() ) + 1 for edge, weight in self.edges.items(): if (edge[0] in subgraph.vertices) ^ (edge[1] in subgraph.vertices): if weight < min_weight: _lowerCamelCase = edge _lowerCamelCase = weight subgraph.add_edge(lowerCamelCase__ , lowerCamelCase__ ) return subgraph def lowerCAmelCase_( lowercase_ : str = "p107_network.txt" ) -> int: _lowerCamelCase = os.path.abspath(os.path.dirname(lowercase_ ) ) _lowerCamelCase = os.path.join(lowercase_ , lowercase_ ) _lowerCamelCase = {} _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 with open(lowercase_ ) as f: _lowerCamelCase = f.read().strip().split('''\n''' ) _lowerCamelCase = [line.split(''',''' ) for line in data] for edgea in range(1 , len(lowercase_ ) ): for edgea in range(lowercase_ ): if adjaceny_matrix[edgea][edgea] != "-": _lowerCamelCase = int(adjaceny_matrix[edgea][edgea] ) _lowerCamelCase = Graph(set(range(len(lowercase_ ) ) ) , lowercase_ ) _lowerCamelCase = graph.prims_algorithm() _lowerCamelCase = sum(graph.edges.values() ) _lowerCamelCase = sum(subgraph.edges.values() ) return initial_total - optimal_total if __name__ == "__main__": print(F"""{solution() = }""")
661
1
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, PreTrainedTokenizerBase, TensorType __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Optional[Any] = { '''microsoft/deberta-v2-xlarge''': '''https://huggingface.co/microsoft/deberta-v2-xlarge/resolve/main/config.json''', '''microsoft/deberta-v2-xxlarge''': '''https://huggingface.co/microsoft/deberta-v2-xxlarge/resolve/main/config.json''', '''microsoft/deberta-v2-xlarge-mnli''': ( '''https://huggingface.co/microsoft/deberta-v2-xlarge-mnli/resolve/main/config.json''' ), '''microsoft/deberta-v2-xxlarge-mnli''': ( '''https://huggingface.co/microsoft/deberta-v2-xxlarge-mnli/resolve/main/config.json''' ), } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[Any] = 'deberta-v2' def __init__( self , lowerCamelCase__=1_2_8_1_0_0 , lowerCamelCase__=1_5_3_6 , lowerCamelCase__=2_4 , lowerCamelCase__=2_4 , lowerCamelCase__=6_1_4_4 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=5_1_2 , lowerCamelCase__=0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=1e-7 , lowerCamelCase__=False , lowerCamelCase__=-1 , lowerCamelCase__=0 , lowerCamelCase__=True , lowerCamelCase__=None , lowerCamelCase__=0 , lowerCamelCase__="gelu" , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = max_position_embeddings _lowerCamelCase = type_vocab_size _lowerCamelCase = initializer_range _lowerCamelCase = relative_attention _lowerCamelCase = max_relative_positions _lowerCamelCase = pad_token_id _lowerCamelCase = position_biased_input # Backwards compatibility if type(lowerCamelCase__ ) == str: _lowerCamelCase = [x.strip() for x in pos_att_type.lower().split('''|''' )] _lowerCamelCase = pos_att_type _lowerCamelCase = vocab_size _lowerCamelCase = layer_norm_eps _lowerCamelCase = kwargs.get('''pooler_hidden_size''' , lowerCamelCase__ ) _lowerCamelCase = pooler_dropout _lowerCamelCase = pooler_hidden_act class lowerCamelCase_( A__ ): '''simple docstring''' @property def snake_case__ ( self ): if self.task == "multiple-choice": _lowerCamelCase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: _lowerCamelCase = {0: '''batch''', 1: '''sequence'''} if self._config.type_vocab_size > 0: return OrderedDict( [('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis)] ) else: return OrderedDict([('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis)] ) @property def snake_case__ ( self ): return 1_2 def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = -1 , lowerCamelCase__ = -1 , lowerCamelCase__ = -1 , lowerCamelCase__ = False , lowerCamelCase__ = None , lowerCamelCase__ = 3 , lowerCamelCase__ = 4_0 , lowerCamelCase__ = 4_0 , lowerCamelCase__ = None , ): _lowerCamelCase = super().generate_dummy_inputs(preprocessor=lowerCamelCase__ , framework=lowerCamelCase__ ) if self._config.type_vocab_size == 0 and "token_type_ids" in dummy_inputs: del dummy_inputs["token_type_ids"] return dummy_inputs
661
"""simple docstring""" from __future__ import annotations from math import ceil, floor, sqrt def lowerCAmelCase_( lowercase_ : int = 2_00_00_00 ) -> int: _lowerCamelCase = [0] _lowerCamelCase = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target _lowerCamelCase = 0 # the area corresponding to the grid that gives the product closest to target _lowerCamelCase = 0 # an estimate of b, using the quadratic formula _lowerCamelCase = 42 # the largest integer less than b_estimate _lowerCamelCase = 42 # the largest integer less than b_estimate _lowerCamelCase = 42 # the triangle number corresponding to b_floor _lowerCamelCase = 42 # the triangle number corresponding to b_ceil _lowerCamelCase = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): _lowerCamelCase = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 _lowerCamelCase = floor(lowercase_ ) _lowerCamelCase = ceil(lowercase_ ) _lowerCamelCase = triangle_numbers[b_floor] _lowerCamelCase = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): _lowerCamelCase = triangle_b_first_guess * triangle_a _lowerCamelCase = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): _lowerCamelCase = triangle_b_second_guess * triangle_a _lowerCamelCase = idx_a * b_ceil return area if __name__ == "__main__": print(F"""{solution() = }""")
661
1
"""simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = {str(digit): digit**5 for digit in range(1_0)} def lowerCAmelCase_( lowercase_ : int ) -> int: return sum(DIGITS_FIFTH_POWER[digit] for digit in str(lowercase_ ) ) def lowerCAmelCase_( ) -> int: return sum( number for number in range(10_00 , 1_00_00_00 ) if number == digits_fifth_powers_sum(lowercase_ ) ) if __name__ == "__main__": print(solution())
661
"""simple docstring""" import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , *lowerCamelCase__ , **lowerCamelCase__ ): warnings.warn( '''The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use DeiTImageProcessor instead.''' , lowerCamelCase__ , ) super().__init__(*lowerCamelCase__ , **lowerCamelCase__ )
661
1
"""simple docstring""" from __future__ import annotations __SCREAMING_SNAKE_CASE : Any = [] def lowerCAmelCase_( lowercase_ : list[list[int]] , lowercase_ : int , lowercase_ : int ) -> bool: for i in range(len(lowercase_ ) ): if board[row][i] == 1: return False for i in range(len(lowercase_ ) ): if board[i][column] == 1: return False for i, j in zip(range(lowercase_ , -1 , -1 ) , range(lowercase_ , -1 , -1 ) ): if board[i][j] == 1: return False for i, j in zip(range(lowercase_ , -1 , -1 ) , range(lowercase_ , len(lowercase_ ) ) ): if board[i][j] == 1: return False return True def lowerCAmelCase_( lowercase_ : list[list[int]] , lowercase_ : int ) -> bool: if row >= len(lowercase_ ): solution.append(lowercase_ ) printboard(lowercase_ ) print() return True for i in range(len(lowercase_ ) ): if is_safe(lowercase_ , lowercase_ , lowercase_ ): _lowerCamelCase = 1 solve(lowercase_ , row + 1 ) _lowerCamelCase = 0 return False def lowerCAmelCase_( lowercase_ : list[list[int]] ) -> None: for i in range(len(lowercase_ ) ): for j in range(len(lowercase_ ) ): if board[i][j] == 1: print('''Q''' , end=''' ''' ) else: print('''.''' , end=''' ''' ) print() # n=int(input("The no. of queens")) __SCREAMING_SNAKE_CASE : str = 8 __SCREAMING_SNAKE_CASE : str = [[0 for i in range(n)] for j in range(n)] solve(board, 0) print('''The total no. of solutions are :''', len(solution))
661
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Optional[Any] = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''vocab_file''': { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json''' ), '''distilbert-base-german-cased''': ( '''https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json''' ), '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json''' ), }, } __SCREAMING_SNAKE_CASE : str = { '''distilbert-base-uncased''': 5_1_2, '''distilbert-base-uncased-distilled-squad''': 5_1_2, '''distilbert-base-cased''': 5_1_2, '''distilbert-base-cased-distilled-squad''': 5_1_2, '''distilbert-base-german-cased''': 5_1_2, '''distilbert-base-multilingual-cased''': 5_1_2, } __SCREAMING_SNAKE_CASE : str = { '''distilbert-base-uncased''': {'''do_lower_case''': True}, '''distilbert-base-uncased-distilled-squad''': {'''do_lower_case''': True}, '''distilbert-base-cased''': {'''do_lower_case''': False}, '''distilbert-base-cased-distilled-squad''': {'''do_lower_case''': False}, '''distilbert-base-german-cased''': {'''do_lower_case''': False}, '''distilbert-base-multilingual-cased''': {'''do_lower_case''': False}, } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[Any] = VOCAB_FILES_NAMES lowercase__ : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP lowercase__ : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : str = PRETRAINED_INIT_CONFIGURATION lowercase__ : int = ['input_ids', 'attention_mask'] lowercase__ : Tuple = DistilBertTokenizer def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__="[UNK]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ): super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) _lowerCamelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , lowerCamelCase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' , lowerCamelCase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , lowerCamelCase__ ) != tokenize_chinese_chars ): _lowerCamelCase = getattr(lowerCamelCase__ , normalizer_state.pop('''type''' ) ) _lowerCamelCase = do_lower_case _lowerCamelCase = strip_accents _lowerCamelCase = tokenize_chinese_chars _lowerCamelCase = normalizer_class(**lowerCamelCase__ ) _lowerCamelCase = do_lower_case def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=None ): _lowerCamelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
661
1
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int , lowercase_ : int ) -> int: _lowerCamelCase = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): _lowerCamelCase = n - k # Calculate C(n,k) for i in range(lowercase_ ): result *= n - i result //= i + 1 return result def lowerCAmelCase_( lowercase_ : int ) -> int: return binomial_coefficient(2 * node_count , lowercase_ ) // (node_count + 1) def lowerCAmelCase_( lowercase_ : int ) -> int: if n < 0: raise ValueError('''factorial() not defined for negative values''' ) _lowerCamelCase = 1 for i in range(1 , n + 1 ): result *= i return result def lowerCAmelCase_( lowercase_ : int ) -> int: return catalan_number(lowercase_ ) * factorial(lowercase_ ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : Tuple = int(input('''Enter the number of nodes: ''').strip() or 0) if node_count <= 0: raise ValueError('''We need some nodes to work with.''') print( F"""Given {node_count} nodes, there are {binary_tree_count(node_count)} """ F"""binary trees and {catalan_number(node_count)} binary search trees.""" )
661
"""simple docstring""" from math import sqrt import numpy as np from sympy import symbols # Coefficient # Speed of light (m/s) __SCREAMING_SNAKE_CASE : str = 2_9_9_7_9_2_4_5_8 # Symbols __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Union[str, Any] = symbols('''ct x y z''') def lowerCAmelCase_( lowercase_ : float ) -> float: if velocity > c: raise ValueError('''Speed must not exceed light speed 299,792,458 [m/s]!''' ) elif velocity < 1: # Usually the speed should be much higher than 1 (c order of magnitude) raise ValueError('''Speed must be greater than or equal to 1!''' ) return velocity / c def lowerCAmelCase_( lowercase_ : float ) -> float: return 1 / sqrt(1 - beta(lowercase_ ) ** 2 ) def lowerCAmelCase_( lowercase_ : float ) -> np.ndarray: return np.array( [ [gamma(lowercase_ ), -gamma(lowercase_ ) * beta(lowercase_ ), 0, 0], [-gamma(lowercase_ ) * beta(lowercase_ ), gamma(lowercase_ ), 0, 0], [0, 0, 1, 0], [0, 0, 0, 1], ] ) def lowerCAmelCase_( lowercase_ : float , lowercase_ : np.ndarray | None = None ) -> np.ndarray: # Ensure event is not empty if event is None: _lowerCamelCase = np.array([ct, x, y, z] ) # Symbolic four vector else: event[0] *= c # x0 is ct (speed of light * time) return transformation_matrix(lowercase_ ) @ event if __name__ == "__main__": import doctest doctest.testmod() # Example of symbolic vector: __SCREAMING_SNAKE_CASE : List[str] = transform(2_9_9_7_9_2_4_5) print('''Example of four vector: ''') print(F"""ct' = {four_vector[0]}""") print(F"""x' = {four_vector[1]}""") print(F"""y' = {four_vector[2]}""") print(F"""z' = {four_vector[3]}""") # Substitute symbols with numerical values __SCREAMING_SNAKE_CASE : Tuple = {ct: c, x: 1, y: 1, z: 1} __SCREAMING_SNAKE_CASE : int = [four_vector[i].subs(sub_dict) for i in range(4)] print(F"""\n{numerical_vector}""")
661
1
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : Dict = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Dict = { '''Salesforce/blip-vqa-base''': '''https://huggingface.co/Salesforce/blip-vqa-base/resolve/main/config.json''', '''Salesforce/blip-vqa-capfit-large''': ( '''https://huggingface.co/Salesforce/blip-vqa-base-capfit/resolve/main/config.json''' ), '''Salesforce/blip-image-captioning-base''': ( '''https://huggingface.co/Salesforce/blip-image-captioning-base/resolve/main/config.json''' ), '''Salesforce/blip-image-captioning-large''': ( '''https://huggingface.co/Salesforce/blip-image-captioning-large/resolve/main/config.json''' ), '''Salesforce/blip-itm-base-coco''': '''https://huggingface.co/Salesforce/blip-itm-base-coco/resolve/main/config.json''', '''Salesforce/blip-itm-large-coco''': '''https://huggingface.co/Salesforce/blip-itm-large-coco/resolve/main/config.json''', '''Salesforce/blip-itm-base-flikr''': '''https://huggingface.co/Salesforce/blip-itm-base-flikr/resolve/main/config.json''', '''Salesforce/blip-itm-large-flikr''': ( '''https://huggingface.co/Salesforce/blip-itm-large-flikr/resolve/main/config.json''' ), } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Optional[int] = 'blip_text_model' def __init__( self , lowerCamelCase__=3_0_5_2_4 , lowerCamelCase__=7_6_8 , lowerCamelCase__=7_6_8 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2 , lowerCamelCase__=8 , lowerCamelCase__=5_1_2 , lowerCamelCase__="gelu" , lowerCamelCase__=1e-12 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=3_0_5_2_2 , lowerCamelCase__=2 , lowerCamelCase__=0 , lowerCamelCase__=1_0_2 , lowerCamelCase__=True , lowerCamelCase__=True , **lowerCamelCase__ , ): super().__init__( pad_token_id=lowerCamelCase__ , bos_token_id=lowerCamelCase__ , eos_token_id=lowerCamelCase__ , sep_token_id=lowerCamelCase__ , **lowerCamelCase__ , ) _lowerCamelCase = vocab_size _lowerCamelCase = hidden_size _lowerCamelCase = encoder_hidden_size _lowerCamelCase = intermediate_size _lowerCamelCase = projection_dim _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = max_position_embeddings _lowerCamelCase = layer_norm_eps _lowerCamelCase = hidden_act _lowerCamelCase = initializer_range _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = is_decoder _lowerCamelCase = use_cache @classmethod def snake_case__ ( cls , lowerCamelCase__ , **lowerCamelCase__ ): cls._set_token_in_kwargs(lowerCamelCase__ ) _lowerCamelCase , _lowerCamelCase = cls.get_config_dict(lowerCamelCase__ , **lowerCamelCase__ ) # get the text config dict if we are loading from BlipConfig if config_dict.get('''model_type''' ) == "blip": _lowerCamelCase = config_dict['''text_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(lowerCamelCase__ , **lowerCamelCase__ ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Dict = 'blip_vision_model' def __init__( self , lowerCamelCase__=7_6_8 , lowerCamelCase__=3_0_7_2 , lowerCamelCase__=5_1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=3_8_4 , lowerCamelCase__=1_6 , lowerCamelCase__="gelu" , lowerCamelCase__=1e-5 , lowerCamelCase__=0.0 , lowerCamelCase__=1e-10 , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) _lowerCamelCase = hidden_size _lowerCamelCase = intermediate_size _lowerCamelCase = projection_dim _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = patch_size _lowerCamelCase = image_size _lowerCamelCase = initializer_range _lowerCamelCase = attention_dropout _lowerCamelCase = layer_norm_eps _lowerCamelCase = hidden_act @classmethod def snake_case__ ( cls , lowerCamelCase__ , **lowerCamelCase__ ): cls._set_token_in_kwargs(lowerCamelCase__ ) _lowerCamelCase , _lowerCamelCase = cls.get_config_dict(lowerCamelCase__ , **lowerCamelCase__ ) # get the vision config dict if we are loading from BlipConfig if config_dict.get('''model_type''' ) == "blip": _lowerCamelCase = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ F"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(lowerCamelCase__ , **lowerCamelCase__ ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Union[str, Any] = 'blip' lowercase__ : int = True def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=5_1_2 , lowerCamelCase__=2.6_5_9_2 , lowerCamelCase__=2_5_6 , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) if text_config is None: _lowerCamelCase = {} logger.info('''`text_config` is `None`. Initializing the `BlipTextConfig` with default values.''' ) if vision_config is None: _lowerCamelCase = {} logger.info('''`vision_config` is `None`. Initializing the `BlipVisionConfig` with default values.''' ) _lowerCamelCase = BlipTextConfig(**lowerCamelCase__ ) _lowerCamelCase = BlipVisionConfig(**lowerCamelCase__ ) _lowerCamelCase = self.vision_config.hidden_size _lowerCamelCase = projection_dim _lowerCamelCase = logit_scale_init_value _lowerCamelCase = 1.0 _lowerCamelCase = 0.0_2 _lowerCamelCase = image_text_hidden_size @classmethod def snake_case__ ( cls , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ): return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = copy.deepcopy(self.__dict__ ) _lowerCamelCase = self.text_config.to_dict() _lowerCamelCase = self.vision_config.to_dict() _lowerCamelCase = self.__class__.model_type return output
661
"""simple docstring""" from __future__ import annotations __SCREAMING_SNAKE_CASE : Dict = 1.6_0_2_1e-1_9 # units = C def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float , ) -> tuple[str, float]: if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif conductivity < 0: raise ValueError('''Conductivity cannot be negative''' ) elif electron_conc < 0: raise ValueError('''Electron concentration cannot be negative''' ) elif mobility < 0: raise ValueError('''mobility cannot be negative''' ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
661
1
"""simple docstring""" from __future__ import annotations from typing import Any class lowerCamelCase_( A__ ): '''simple docstring''' pass class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ ): _lowerCamelCase = data _lowerCamelCase = None def __iter__( self ): _lowerCamelCase = self _lowerCamelCase = [] while node: if node in visited: raise ContainsLoopError visited.append(lowerCamelCase__ ) yield node.data _lowerCamelCase = node.next_node @property def snake_case__ ( self ): try: list(self ) return False except ContainsLoopError: return True if __name__ == "__main__": __SCREAMING_SNAKE_CASE : str = Node(1) __SCREAMING_SNAKE_CASE : Optional[Any] = Node(2) __SCREAMING_SNAKE_CASE : Tuple = Node(3) __SCREAMING_SNAKE_CASE : Any = Node(4) print(root_node.has_loop) # False __SCREAMING_SNAKE_CASE : List[Any] = root_node.next_node print(root_node.has_loop) # True __SCREAMING_SNAKE_CASE : Optional[int] = Node(5) __SCREAMING_SNAKE_CASE : Union[str, Any] = Node(6) __SCREAMING_SNAKE_CASE : int = Node(5) __SCREAMING_SNAKE_CASE : int = Node(6) print(root_node.has_loop) # False __SCREAMING_SNAKE_CASE : Tuple = Node(1) print(root_node.has_loop) # False
661
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): _lowerCamelCase = { '''task_specific_params''': { '''summarization''': {'''length_penalty''': 1.0, '''max_length''': 1_2_8, '''min_length''': 1_2, '''num_beams''': 4}, '''summarization_cnn''': {'''length_penalty''': 2.0, '''max_length''': 1_4_2, '''min_length''': 5_6, '''num_beams''': 4}, '''summarization_xsum''': {'''length_penalty''': 1.0, '''max_length''': 6_2, '''min_length''': 1_1, '''num_beams''': 6}, } } _lowerCamelCase = { '''task_specific_params.summarization.length_penalty''': 1.0, '''task_specific_params.summarization.max_length''': 1_2_8, '''task_specific_params.summarization.min_length''': 1_2, '''task_specific_params.summarization.num_beams''': 4, '''task_specific_params.summarization_cnn.length_penalty''': 2.0, '''task_specific_params.summarization_cnn.max_length''': 1_4_2, '''task_specific_params.summarization_cnn.min_length''': 5_6, '''task_specific_params.summarization_cnn.num_beams''': 4, '''task_specific_params.summarization_xsum.length_penalty''': 1.0, '''task_specific_params.summarization_xsum.max_length''': 6_2, '''task_specific_params.summarization_xsum.min_length''': 1_1, '''task_specific_params.summarization_xsum.num_beams''': 6, } self.assertEqual(flatten_dict(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , x.transpose() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , np.asarray(transpose(lowerCamelCase__ ) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , np.asarray(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.reshape(lowerCamelCase__ , (4, 3) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , np.reshape(lowerCamelCase__ , (1_2, 5) ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , reshape(lowerCamelCase__ , (1_2, 5) ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , reshape(lowerCamelCase__ , (1_2, 5) ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.asarray(reshape(lowerCamelCase__ , (4, 3) ) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , np.asarray(reshape(lowerCamelCase__ , (1_2, 5) ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.squeeze(lowerCamelCase__ ) ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.squeeze(lowerCamelCase__ , axis=2 ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.asarray(squeeze(lowerCamelCase__ ) ) ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.asarray(squeeze(lowerCamelCase__ , axis=2 ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.expand_dims(lowerCamelCase__ , axis=1 ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.asarray(expand_dims(lowerCamelCase__ , axis=1 ) ) ) )
661
1
"""simple docstring""" from unittest.mock import Mock, patch from file_transfer.send_file import send_file @patch('''socket.socket''' ) @patch('''builtins.open''' ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : List[str] ) -> List[str]: # ===== initialization ===== _lowerCamelCase = Mock() _lowerCamelCase = conn, Mock() _lowerCamelCase = iter([1, None] ) _lowerCamelCase = lambda lowercase_ : next(lowercase_ ) # ===== invoke ===== send_file(filename='''mytext.txt''' , testing=lowercase_ ) # ===== ensurance ===== sock.assert_called_once() sock.return_value.bind.assert_called_once() sock.return_value.listen.assert_called_once() sock.return_value.accept.assert_called_once() conn.recv.assert_called_once() file.return_value.__enter__.assert_called_once() file.return_value.__enter__.return_value.read.assert_called() conn.send.assert_called_once() conn.close.assert_called_once() sock.return_value.shutdown.assert_called_once() sock.return_value.close.assert_called_once()
661
"""simple docstring""" import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } __SCREAMING_SNAKE_CASE : Tuple = { '''vocab_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json'''}, '''merges_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt'''}, '''tokenizer_config_file''': { '''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json''' }, } __SCREAMING_SNAKE_CASE : str = {'''facebook/blenderbot-3B''': 1_2_8} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def lowerCAmelCase_( ) -> str: _lowerCamelCase = ( list(range(ord('''!''' ) , ord('''~''' ) + 1 ) ) + list(range(ord('''¡''' ) , ord('''¬''' ) + 1 ) ) + list(range(ord('''®''' ) , ord('''ÿ''' ) + 1 ) ) ) _lowerCamelCase = bs[:] _lowerCamelCase = 0 for b in range(2**8 ): if b not in bs: bs.append(lowercase_ ) cs.append(2**8 + n ) n += 1 _lowerCamelCase = [chr(lowercase_ ) for n in cs] return dict(zip(lowercase_ , lowercase_ ) ) def lowerCAmelCase_( lowercase_ : str ) -> Dict: _lowerCamelCase = set() _lowerCamelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _lowerCamelCase = char return pairs class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Union[str, Any] = VOCAB_FILES_NAMES lowercase__ : int = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : Tuple = ['input_ids', 'attention_mask'] def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__="replace" , lowerCamelCase__="<s>" , lowerCamelCase__="</s>" , lowerCamelCase__="</s>" , lowerCamelCase__="<s>" , lowerCamelCase__="<unk>" , lowerCamelCase__="<pad>" , lowerCamelCase__="<mask>" , lowerCamelCase__=False , **lowerCamelCase__ , ): _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else bos_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else eos_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else sep_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else cls_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else unk_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else mask_token super().__init__( errors=lowerCamelCase__ , bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , add_prefix_space=lowerCamelCase__ , **lowerCamelCase__ , ) with open(lowerCamelCase__ , encoding='''utf-8''' ) as vocab_handle: _lowerCamelCase = json.load(lowerCamelCase__ ) _lowerCamelCase = {v: k for k, v in self.encoder.items()} _lowerCamelCase = errors # how to handle errors in decoding _lowerCamelCase = bytes_to_unicode() _lowerCamelCase = {v: k for k, v in self.byte_encoder.items()} with open(lowerCamelCase__ , encoding='''utf-8''' ) as merges_handle: _lowerCamelCase = merges_handle.read().split('''\n''' )[1:-1] _lowerCamelCase = [tuple(merge.split() ) for merge in bpe_merges] _lowerCamelCase = dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) _lowerCamelCase = {} _lowerCamelCase = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions _lowerCamelCase = re.compile(R'''\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+''' ) @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def snake_case__ ( self ): return len(self.encoder ) def snake_case__ ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def snake_case__ ( self , lowerCamelCase__ ): if token in self.cache: return self.cache[token] _lowerCamelCase = tuple(lowerCamelCase__ ) _lowerCamelCase = get_pairs(lowerCamelCase__ ) if not pairs: return token while True: _lowerCamelCase = min(lowerCamelCase__ , key=lambda lowerCamelCase__ : self.bpe_ranks.get(lowerCamelCase__ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break _lowerCamelCase , _lowerCamelCase = bigram _lowerCamelCase = [] _lowerCamelCase = 0 while i < len(lowerCamelCase__ ): try: _lowerCamelCase = word.index(lowerCamelCase__ , lowerCamelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _lowerCamelCase = j if word[i] == first and i < len(lowerCamelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _lowerCamelCase = tuple(lowerCamelCase__ ) _lowerCamelCase = new_word if len(lowerCamelCase__ ) == 1: break else: _lowerCamelCase = get_pairs(lowerCamelCase__ ) _lowerCamelCase = ''' '''.join(lowerCamelCase__ ) _lowerCamelCase = word return word def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] for token in re.findall(self.pat , lowerCamelCase__ ): _lowerCamelCase = ''''''.join( self.byte_encoder[b] for b in token.encode('''utf-8''' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCamelCase__ ).split(''' ''' ) ) return bpe_tokens def snake_case__ ( self , lowerCamelCase__ ): return self.encoder.get(lowerCamelCase__ , self.encoder.get(self.unk_token ) ) def snake_case__ ( self , lowerCamelCase__ ): return self.decoder.get(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = ''''''.join(lowerCamelCase__ ) _lowerCamelCase = bytearray([self.byte_decoder[c] for c in text] ).decode('''utf-8''' , errors=self.errors ) return text def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): if not os.path.isdir(lowerCamelCase__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(lowerCamelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCamelCase__ , ensure_ascii=lowerCamelCase__ ) + '''\n''' ) _lowerCamelCase = 0 with open(lowerCamelCase__ , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCamelCase__ : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ''' Please check that the tokenizer is not corrupted!''' ) _lowerCamelCase = token_index writer.write(''' '''.join(lowerCamelCase__ ) + '''\n''' ) index += 1 return vocab_file, merge_file def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase__ )) + [1] return [1] + ([0] * len(lowerCamelCase__ )) + [1, 1] + ([0] * len(lowerCamelCase__ )) + [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=False , **lowerCamelCase__ ): _lowerCamelCase = kwargs.pop('''add_prefix_space''' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCamelCase__ ) > 0 and not text[0].isspace()): _lowerCamelCase = ''' ''' + text return (text, kwargs) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): return token_ids_a + [self.eos_token_id] def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(''' ''' + text ) else: # Generated responses should contain them already. inputs.append(lowerCamelCase__ ) _lowerCamelCase = ''' '''.join(lowerCamelCase__ ) _lowerCamelCase = self.encode(lowerCamelCase__ ) if len(lowerCamelCase__ ) > self.model_max_length: _lowerCamelCase = input_ids[-self.model_max_length :] logger.warning(F"""Trimmed input from conversation as it was longer than {self.model_max_length} tokens.""" ) return input_ids
661
1
"""simple docstring""" import re import string from collections import Counter import sacrebleu import sacremoses from packaging import version import datasets __SCREAMING_SNAKE_CASE : Optional[int] = ''' @inproceedings{xu-etal-2016-optimizing, title = {Optimizing Statistical Machine Translation for Text Simplification}, authors={Xu, Wei and Napoles, Courtney and Pavlick, Ellie and Chen, Quanze and Callison-Burch, Chris}, journal = {Transactions of the Association for Computational Linguistics}, volume = {4}, year={2016}, url = {https://www.aclweb.org/anthology/Q16-1029}, pages = {401--415 }, @inproceedings{post-2018-call, title = "A Call for Clarity in Reporting {BLEU} Scores", author = "Post, Matt", booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers", month = oct, year = "2018", address = "Belgium, Brussels", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W18-6319", pages = "186--191", } ''' __SCREAMING_SNAKE_CASE : List[str] = '''\ WIKI_SPLIT is the combination of three metrics SARI, EXACT and SACREBLEU It can be used to evaluate the quality of machine-generated texts. ''' __SCREAMING_SNAKE_CASE : Optional[Any] = ''' Calculates sari score (between 0 and 100) given a list of source and predicted sentences, and a list of lists of reference sentences. It also computes the BLEU score as well as the exact match score. Args: sources: list of source sentences where each sentence should be a string. predictions: list of predicted sentences where each sentence should be a string. references: list of lists of reference sentences where each sentence should be a string. Returns: sari: sari score sacrebleu: sacrebleu score exact: exact score Examples: >>> sources=["About 95 species are currently accepted ."] >>> predictions=["About 95 you now get in ."] >>> references=[["About 95 species are currently known ."]] >>> wiki_split = datasets.load_metric("wiki_split") >>> results = wiki_split.compute(sources=sources, predictions=predictions, references=references) >>> print(results) {\'sari\': 21.805555555555557, \'sacrebleu\': 14.535768424205482, \'exact\': 0.0} ''' def lowerCAmelCase_( lowercase_ : Union[str, Any] ) -> str: def remove_articles(lowercase_ : int ): _lowerCamelCase = re.compile(r'''\b(a|an|the)\b''' , re.UNICODE ) return re.sub(lowercase_ , ''' ''' , lowercase_ ) def white_space_fix(lowercase_ : List[Any] ): return " ".join(text.split() ) def remove_punc(lowercase_ : Dict ): _lowerCamelCase = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowercase_ : Union[str, Any] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowercase_ ) ) ) ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[Any] ) -> Union[str, Any]: return int(normalize_answer(lowercase_ ) == normalize_answer(lowercase_ ) ) def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Tuple ) -> Tuple: _lowerCamelCase = [any(compute_exact(lowercase_ , lowercase_ ) for ref in refs ) for pred, refs in zip(lowercase_ , lowercase_ )] return (sum(lowercase_ ) / len(lowercase_ )) * 1_00 def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : str , lowercase_ : Optional[Any] , lowercase_ : str ) -> Optional[int]: _lowerCamelCase = [rgram for rgrams in rgramslist for rgram in rgrams] _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter() for sgram, scount in sgramcounter.items(): _lowerCamelCase = scount * numref _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter() for cgram, ccount in cgramcounter.items(): _lowerCamelCase = ccount * numref # KEEP _lowerCamelCase = sgramcounter_rep & cgramcounter_rep _lowerCamelCase = keepgramcounter_rep & rgramcounter _lowerCamelCase = sgramcounter_rep & rgramcounter _lowerCamelCase = 0 _lowerCamelCase = 0 for keepgram in keepgramcountergood_rep: keeptmpscorea += keepgramcountergood_rep[keepgram] / keepgramcounter_rep[keepgram] # Fix an alleged bug [2] in the keep score computation. # keeptmpscore2 += keepgramcountergood_rep[keepgram] / keepgramcounterall_rep[keepgram] keeptmpscorea += keepgramcountergood_rep[keepgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = keeptmpscorea / len(lowercase_ ) if len(lowercase_ ) > 0: # Fix an alleged bug [2] in the keep score computation. # keepscore_recall = keeptmpscore2 / len(keepgramcounterall_rep) _lowerCamelCase = keeptmpscorea / sum(keepgramcounterall_rep.values() ) _lowerCamelCase = 0 if keepscore_precision > 0 or keepscore_recall > 0: _lowerCamelCase = 2 * keepscore_precision * keepscore_recall / (keepscore_precision + keepscore_recall) # DELETION _lowerCamelCase = sgramcounter_rep - cgramcounter_rep _lowerCamelCase = delgramcounter_rep - rgramcounter _lowerCamelCase = sgramcounter_rep - rgramcounter _lowerCamelCase = 0 _lowerCamelCase = 0 for delgram in delgramcountergood_rep: deltmpscorea += delgramcountergood_rep[delgram] / delgramcounter_rep[delgram] deltmpscorea += delgramcountergood_rep[delgram] / delgramcounterall_rep[delgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = deltmpscorea / len(lowercase_ ) # ADDITION _lowerCamelCase = set(lowercase_ ) - set(lowercase_ ) _lowerCamelCase = set(lowercase_ ) & set(lowercase_ ) _lowerCamelCase = set(lowercase_ ) - set(lowercase_ ) _lowerCamelCase = 0 for addgram in addgramcountergood: addtmpscore += 1 # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = addtmpscore / len(lowercase_ ) if len(lowercase_ ) > 0: _lowerCamelCase = addtmpscore / len(lowercase_ ) _lowerCamelCase = 0 if addscore_precision > 0 or addscore_recall > 0: _lowerCamelCase = 2 * addscore_precision * addscore_recall / (addscore_precision + addscore_recall) return (keepscore, delscore_precision, addscore) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : Optional[Any] , lowercase_ : str ) -> List[str]: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = ssent.split(''' ''' ) _lowerCamelCase = csent.split(''' ''' ) _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] for rsent in rsents: _lowerCamelCase = rsent.split(''' ''' ) _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] ragramslist.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] ragrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] + ''' ''' + ragrams[i + 2] ragrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] + ''' ''' + ragrams[i + 2] + ''' ''' + ragrams[i + 3] ragrams.append(lowercase_ ) ragramslist.append(lowercase_ ) ragramslist.append(lowercase_ ) ragramslist.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] sagrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] + ''' ''' + sagrams[i + 2] sagrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] + ''' ''' + sagrams[i + 2] + ''' ''' + sagrams[i + 3] sagrams.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] cagrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] + ''' ''' + cagrams[i + 2] cagrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] + ''' ''' + cagrams[i + 2] + ''' ''' + cagrams[i + 3] cagrams.append(lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) _lowerCamelCase = sum([keepascore, keepascore, keepascore, keepascore] ) / 4 _lowerCamelCase = sum([delascore, delascore, delascore, delascore] ) / 4 _lowerCamelCase = sum([addascore, addascore, addascore, addascore] ) / 4 _lowerCamelCase = (avgkeepscore + avgdelscore + avgaddscore) / 3 return finalscore def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : bool = True , lowercase_ : str = "13a" , lowercase_ : bool = True ) -> int: # Normalization is requried for the ASSET dataset (one of the primary # datasets in sentence simplification) to allow using space # to split the sentence. Even though Wiki-Auto and TURK datasets, # do not require normalization, we do it for consistency. # Code adapted from the EASSE library [1] written by the authors of the ASSET dataset. # [1] https://github.com/feralvam/easse/blob/580bba7e1378fc8289c663f864e0487188fe8067/easse/utils/preprocessing.py#L7 if lowercase: _lowerCamelCase = sentence.lower() if tokenizer in ["13a", "intl"]: if version.parse(sacrebleu.__version__ ).major >= 2: _lowerCamelCase = sacrebleu.metrics.bleu._get_tokenizer(lowercase_ )()(lowercase_ ) else: _lowerCamelCase = sacrebleu.TOKENIZERS[tokenizer]()(lowercase_ ) elif tokenizer == "moses": _lowerCamelCase = sacremoses.MosesTokenizer().tokenize(lowercase_ , return_str=lowercase_ , escape=lowercase_ ) elif tokenizer == "penn": _lowerCamelCase = sacremoses.MosesTokenizer().penn_tokenize(lowercase_ , return_str=lowercase_ ) else: _lowerCamelCase = sentence if not return_str: _lowerCamelCase = normalized_sent.split() return normalized_sent def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Any , lowercase_ : List[Any] ) -> Optional[int]: if not (len(lowercase_ ) == len(lowercase_ ) == len(lowercase_ )): raise ValueError('''Sources length must match predictions and references lengths.''' ) _lowerCamelCase = 0 for src, pred, refs in zip(lowercase_ , lowercase_ , lowercase_ ): sari_score += SARIsent(normalize(lowercase_ ) , normalize(lowercase_ ) , [normalize(lowercase_ ) for sent in refs] ) _lowerCamelCase = sari_score / len(lowercase_ ) return 1_00 * sari_score def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Any , lowercase_ : List[Any]="exp" , lowercase_ : List[Any]=None , lowercase_ : Optional[Any]=False , lowercase_ : List[Any]=False , lowercase_ : List[Any]=False , ) -> Dict: _lowerCamelCase = len(references[0] ) if any(len(lowercase_ ) != references_per_prediction for refs in references ): raise ValueError('''Sacrebleu requires the same number of references for each prediction''' ) _lowerCamelCase = [[refs[i] for refs in references] for i in range(lowercase_ )] _lowerCamelCase = sacrebleu.corpus_bleu( lowercase_ , lowercase_ , smooth_method=lowercase_ , smooth_value=lowercase_ , force=lowercase_ , lowercase=lowercase_ , use_effective_order=lowercase_ , ) return output.score @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Sequence(datasets.Value('''string''' , id='''sequence''' ) , id='''references''' ), } ) , codebase_urls=[ '''https://github.com/huggingface/transformers/blob/master/src/transformers/data/metrics/squad_metrics.py''', '''https://github.com/cocoxu/simplification/blob/master/SARI.py''', '''https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/sari_hook.py''', '''https://github.com/mjpost/sacreBLEU''', ] , reference_urls=[ '''https://www.aclweb.org/anthology/Q16-1029.pdf''', '''https://github.com/mjpost/sacreBLEU''', '''https://en.wikipedia.org/wiki/BLEU''', '''https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213''', ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = {} result.update({'''sari''': compute_sari(sources=lowerCamelCase__ , predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) result.update({'''sacrebleu''': compute_sacrebleu(predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) result.update({'''exact''': compute_em(predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) return result
661
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Dict = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Union[str, Any] = XLMRobertaTokenizer lowercase__ : Optional[int] = XLMRobertaTokenizerFast lowercase__ : List[str] = True lowercase__ : Union[str, Any] = True def snake_case__ ( self ): super().setUp() # We have a SentencePiece fixture for testing _lowerCamelCase = XLMRobertaTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case__ ( self ): _lowerCamelCase = '''<pad>''' _lowerCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(lowerCamelCase__ ) , 1_0_0_2 ) def snake_case__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_2 ) def snake_case__ ( self ): _lowerCamelCase = XLMRobertaTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) _lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(lowerCamelCase__ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) _lowerCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) _lowerCamelCase = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) _lowerCamelCase = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def snake_case__ ( self ): if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _lowerCamelCase = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-xlm-roberta''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _lowerCamelCase = self.rust_tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = self.tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) _lowerCamelCase = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(lowerCamelCase__ , lowerCamelCase__ ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(lowerCamelCase__ ) # Save tokenizer rust, legacy_format=True _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ , legacy_format=lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it save with the same files self.assertSequenceEqual(lowerCamelCase__ , lowerCamelCase__ ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) shutil.rmtree(lowerCamelCase__ ) # Save tokenizer rust, legacy_format=False _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ , legacy_format=lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) shutil.rmtree(lowerCamelCase__ ) @cached_property def snake_case__ ( self ): return XLMRobertaTokenizer.from_pretrained('''xlm-roberta-base''' ) def snake_case__ ( self ): with tempfile.NamedTemporaryFile() as f: shutil.copyfile(lowerCamelCase__ , f.name ) _lowerCamelCase = XLMRobertaTokenizer(f.name , keep_accents=lowerCamelCase__ ) _lowerCamelCase = pickle.dumps(lowerCamelCase__ ) pickle.loads(lowerCamelCase__ ) def snake_case__ ( self ): if not self.test_rust_tokenizer: return _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = '''I was born in 92000, and this is falsé.''' _lowerCamelCase = tokenizer.tokenize(lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = tokenizer.encode(lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): _lowerCamelCase = '''Hello World!''' _lowerCamelCase = [0, 3_5_3_7_8, 6_6_6_1, 3_8, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(lowerCamelCase__ , self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def snake_case__ ( self ): _lowerCamelCase = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) _lowerCamelCase = [ 0, 3_2_9_3, 8_3, 1_0, 4_5_5_2, 4_9_8_9, 7_9_8_6, 6_7_8, 1_0, 5_9_1_5, 1_1_1, 1_7_9_4_5_9, 1_2_4_8_5_0, 4, 6_0_4_4, 2_3_7, 1_2, 6, 5, 6, 4, 6_7_8_0, 7_0_5, 1_5, 1_3_8_8, 4_4, 3_7_8, 1_0_1_1_4, 7_1_1, 1_5_2, 2_0, 6, 5, 2_2_3_7_6, 6_4_2, 1_2_2_1, 1_5_1_9_0, 3_4_1_5_3, 4_5_0, 5_6_0_8, 9_5_9, 1_1_1_9, 5_7_7_0_2, 1_3_6, 1_8_6, 4_7, 1_0_9_8, 2_9_3_6_7, 4_7, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6_0_4_4, 2_3_7, 6_2_8_4, 5_0_9_0_1, 5_2_8, 3_1, 9_0, 3_4, 9_2_7, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(lowerCamelCase__ , self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def snake_case__ ( self ): # fmt: off _lowerCamelCase = {'''input_ids''': [[0, 1_1_0_6_2, 8_2_7_7_2, 7, 1_5, 8_2_7_7_2, 5_3_8, 5_1_5_2_9, 2_3_7, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 2_1_5_1_7_5, 1_3_1_4, 1_3_6, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 5_6_3_5_9, 4_2, 1_2_2_0_0_9, 9, 1_6_4_6_6, 1_6, 8_7_3_4_4, 4_5_3_7, 9, 4_7_1_7, 7_8_3_8_1, 6, 1_5_9_9_5_8, 7, 1_5, 2_4_4_8_0, 6_1_8, 4, 5_2_7, 2_2_6_9_3, 5_4_2_8, 4, 2_7_7_7, 2_4_4_8_0, 9_8_7_4, 4, 4_3_5_2_3, 5_9_4, 4, 8_0_3, 1_8_3_9_2, 3_3_1_8_9, 1_8, 4, 4_3_5_2_3, 2_4_4_4_7, 1_2_3_9_9, 1_0_0, 2_4_9_5_5, 8_3_6_5_8, 9_6_2_6, 1_4_4_0_5_7, 1_5, 8_3_9, 2_2_3_3_5, 1_6, 1_3_6, 2_4_9_5_5, 8_3_6_5_8, 8_3_4_7_9, 1_5, 3_9_1_0_2, 7_2_4, 1_6, 6_7_8, 6_4_5, 2_7_8_9, 1_3_2_8, 4_5_8_9, 4_2, 1_2_2_0_0_9, 1_1_5_7_7_4, 2_3, 8_0_5, 1_3_2_8, 4_6_8_7_6, 7, 1_3_6, 5_3_8_9_4, 1_9_4_0, 4_2_2_2_7, 4_1_1_5_9, 1_7_7_2_1, 8_2_3, 4_2_5, 4, 2_7_5_1_2, 9_8_7_2_2, 2_0_6, 1_3_6, 5_5_3_1, 4_9_7_0, 9_1_9, 1_7_3_3_6, 5, 2], [0, 2_0_0_8_0, 6_1_8, 8_3, 8_2_7_7_5, 4_7, 4_7_9, 9, 1_5_1_7, 7_3, 5_3_8_9_4, 3_3_3, 8_0_5_8_1, 1_1_0_1_1_7, 1_8_8_1_1, 5_2_5_6, 1_2_9_5, 5_1, 1_5_2_5_2_6, 2_9_7, 7_9_8_6, 3_9_0, 1_2_4_4_1_6, 5_3_8, 3_5_4_3_1, 2_1_4, 9_8, 1_5_0_4_4, 2_5_7_3_7, 1_3_6, 7_1_0_8, 4_3_7_0_1, 2_3, 7_5_6, 1_3_5_3_5_5, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_8_1, 6_3_7_7_3, 1_1_9_4_5_5, 6, 1_4_7_7_9_7, 8_8_2_0_3, 7, 6_4_5, 7_0, 2_1, 3_2_8_5, 1_0_2_6_9, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ , model_name='''xlm-roberta-base''' , revision='''d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3''' , )
661
1
"""simple docstring""" __SCREAMING_SNAKE_CASE : List[str] = ''' # Installazione di Transformers ! pip install transformers datasets # Per installare dalla fonte invece dell\'ultima versione rilasciata, commenta il comando sopra e # rimuovi la modalità commento al comando seguente. # ! pip install git+https://github.com/huggingface/transformers.git ''' __SCREAMING_SNAKE_CASE : Dict = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] __SCREAMING_SNAKE_CASE : List[str] = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
661
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int = 10 ) -> str: if not isinstance(lowercase_ , lowercase_ ) or n < 0: raise ValueError('''Invalid input''' ) _lowerCamelCase = 10**n _lowerCamelCase = 2_84_33 * (pow(2 , 7_83_04_57 , lowercase_ )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(F"""{solution(1_0) = }""")
661
1
"""simple docstring""" import unittest import numpy as np import torch from diffusers import VersatileDiffusionImageVariationPipeline from diffusers.utils.testing_utils import load_image, require_torch_gpu, slow, torch_device __SCREAMING_SNAKE_CASE : Optional[int] = False class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' pass @slow @require_torch_gpu class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): _lowerCamelCase = VersatileDiffusionImageVariationPipeline.from_pretrained('''shi-labs/versatile-diffusion''' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _lowerCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) _lowerCamelCase = torch.manual_seed(0 ) _lowerCamelCase = pipe( image=lowerCamelCase__ , generator=lowerCamelCase__ , guidance_scale=7.5 , num_inference_steps=5_0 , output_type='''numpy''' , ).images _lowerCamelCase = image[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowerCamelCase = np.array([0.0_4_4_1, 0.0_4_6_9, 0.0_5_0_7, 0.0_5_7_5, 0.0_6_3_2, 0.0_6_5_0, 0.0_8_6_5, 0.0_9_0_9, 0.0_9_4_5] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
661
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __SCREAMING_SNAKE_CASE : Optional[int] = { '''configuration_informer''': [ '''INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ '''INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InformerForPrediction''', '''InformerModel''', '''InformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
661
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, is_vision_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = {'''configuration_vit''': ['''VIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTConfig''', '''ViTOnnxConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[str] = ['''ViTFeatureExtractor'''] __SCREAMING_SNAKE_CASE : Dict = ['''ViTImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ '''VIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTForImageClassification''', '''ViTForMaskedImageModeling''', '''ViTModel''', '''ViTPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = [ '''TFViTForImageClassification''', '''TFViTModel''', '''TFViTPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = [ '''FlaxViTForImageClassification''', '''FlaxViTModel''', '''FlaxViTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit import VIT_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTConfig, ViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_vit import ViTFeatureExtractor from .image_processing_vit import ViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit import ( VIT_PRETRAINED_MODEL_ARCHIVE_LIST, ViTForImageClassification, ViTForMaskedImageModeling, ViTModel, ViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit import TFViTForImageClassification, TFViTModel, TFViTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel, FlaxViTPreTrainedModel else: import sys __SCREAMING_SNAKE_CASE : Any = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
661
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> float: def get_matched_characters(lowercase_ : str , lowercase_ : str ) -> str: _lowerCamelCase = [] _lowerCamelCase = min(len(_stra ) , len(_stra ) ) // 2 for i, l in enumerate(_stra ): _lowerCamelCase = int(max(0 , i - limit ) ) _lowerCamelCase = int(min(i + limit + 1 , len(_stra ) ) ) if l in _stra[left:right]: matched.append(lowercase_ ) _lowerCamelCase = F"""{_stra[0:_stra.index(lowercase_ )]} {_stra[_stra.index(lowercase_ ) + 1:]}""" return "".join(lowercase_ ) # matching characters _lowerCamelCase = get_matched_characters(lowercase_ , lowercase_ ) _lowerCamelCase = get_matched_characters(lowercase_ , lowercase_ ) _lowerCamelCase = len(lowercase_ ) # transposition _lowerCamelCase = ( len([(ca, ca) for ca, ca in zip(lowercase_ , lowercase_ ) if ca != ca] ) // 2 ) if not match_count: _lowerCamelCase = 0.0 else: _lowerCamelCase = ( 1 / 3 * ( match_count / len(lowercase_ ) + match_count / len(lowercase_ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters _lowerCamelCase = 0 for ca, ca in zip(stra[:4] , stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('''hello''', '''world'''))
661
1
import random from typing import Any def __lowercase ( snake_case ): """simple docstring""" for _ in range(len(snake_case ) ): __magic_name__ :Optional[int] = random.randint(0, len(snake_case ) - 1 ) __magic_name__ :Union[str, Any] = random.randint(0, len(snake_case ) - 1 ) __magic_name__ , __magic_name__ :List[Any] = data[b], data[a] return data if __name__ == "__main__": SCREAMING_SNAKE_CASE__ : List[Any] = [0, 1, 2, 3, 4, 5, 6, 7] SCREAMING_SNAKE_CASE__ : int = ["""python""", """says""", """hello""", """!"""] print("""Fisher-Yates Shuffle:""") print("""List""", integers, strings) print("""FY Shuffle""", fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
0
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowerCamelCase_( A__, A__, A__ ): '''simple docstring''' lowercase__ : List[Any] = [r'h\.\d+\.attn\.bias', r'h\.\d+\.attn\.masked_bias'] @register_to_config def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = 5_0_2_5_7 , lowerCamelCase__ = 1_0_2_4 , lowerCamelCase__ = 7_6_8 , lowerCamelCase__ = 1_2 , lowerCamelCase__ = 1_2 , lowerCamelCase__ = None , lowerCamelCase__ = "gelu_new" , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 1e-5 , lowerCamelCase__ = 0.0_2 , lowerCamelCase__ = True , lowerCamelCase__ = True , lowerCamelCase__ = False , lowerCamelCase__ = False , ): super().__init__() _lowerCamelCase = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F"""`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and""" F""" `n_embd`: {n_embd} are not equal.""" ) _lowerCamelCase = prefix_inner_dim _lowerCamelCase = prefix_hidden_dim _lowerCamelCase = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCamelCase = ( nn.Linear(self.prefix_hidden_dim , lowerCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCamelCase = GPTaConfig( vocab_size=lowerCamelCase__ , n_positions=lowerCamelCase__ , n_embd=lowerCamelCase__ , n_layer=lowerCamelCase__ , n_head=lowerCamelCase__ , n_inner=lowerCamelCase__ , activation_function=lowerCamelCase__ , resid_pdrop=lowerCamelCase__ , embd_pdrop=lowerCamelCase__ , attn_pdrop=lowerCamelCase__ , layer_norm_epsilon=lowerCamelCase__ , initializer_range=lowerCamelCase__ , scale_attn_weights=lowerCamelCase__ , use_cache=lowerCamelCase__ , scale_attn_by_inverse_layer_idx=lowerCamelCase__ , reorder_and_upcast_attn=lowerCamelCase__ , ) _lowerCamelCase = GPTaLMHeadModel(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None , ): _lowerCamelCase = self.transformer.transformer.wte(lowerCamelCase__ ) _lowerCamelCase = self.encode_prefix(lowerCamelCase__ ) _lowerCamelCase = self.decode_prefix(lowerCamelCase__ ) _lowerCamelCase = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: _lowerCamelCase = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) _lowerCamelCase = torch.cat((dummy_token, input_ids) , dim=1 ) _lowerCamelCase = self.transformer(inputs_embeds=lowerCamelCase__ , labels=lowerCamelCase__ , attention_mask=lowerCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): return torch.zeros(lowerCamelCase__ , self.prefix_length , dtype=torch.intaa , device=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): return self.encode_prefix(lowerCamelCase__ ) @torch.no_grad() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = torch.split(lowerCamelCase__ , 1 , dim=0 ) _lowerCamelCase = [] _lowerCamelCase = [] for feature in features: _lowerCamelCase = self.decode_prefix(feature.to(lowerCamelCase__ ) ) # back to the clip feature # Only support beam search for now _lowerCamelCase , _lowerCamelCase = self.generate_beam( input_embeds=lowerCamelCase__ , device=lowerCamelCase__ , eos_token_id=lowerCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) _lowerCamelCase = torch.stack(lowerCamelCase__ ) _lowerCamelCase = torch.stack(lowerCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def snake_case__ ( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__ = 5 , lowerCamelCase__ = 6_7 , lowerCamelCase__ = 1.0 , lowerCamelCase__ = None , ): _lowerCamelCase = eos_token_id _lowerCamelCase = None _lowerCamelCase = None _lowerCamelCase = torch.ones(lowerCamelCase__ , device=lowerCamelCase__ , dtype=torch.int ) _lowerCamelCase = torch.zeros(lowerCamelCase__ , device=lowerCamelCase__ , dtype=torch.bool ) if input_embeds is not None: _lowerCamelCase = input_embeds else: _lowerCamelCase = self.transformer.transformer.wte(lowerCamelCase__ ) for i in range(lowerCamelCase__ ): _lowerCamelCase = self.transformer(inputs_embeds=lowerCamelCase__ ) _lowerCamelCase = outputs.logits _lowerCamelCase = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) _lowerCamelCase = logits.softmax(-1 ).log() if scores is None: _lowerCamelCase , _lowerCamelCase = logits.topk(lowerCamelCase__ , -1 ) _lowerCamelCase = generated.expand(lowerCamelCase__ , *generated.shape[1:] ) _lowerCamelCase , _lowerCamelCase = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: _lowerCamelCase = next_tokens else: _lowerCamelCase = tokens.expand(lowerCamelCase__ , *tokens.shape[1:] ) _lowerCamelCase = torch.cat((tokens, next_tokens) , dim=1 ) else: _lowerCamelCase = -float(np.inf ) _lowerCamelCase = 0 _lowerCamelCase = scores[:, None] + logits seq_lengths[~is_stopped] += 1 _lowerCamelCase = scores_sum / seq_lengths[:, None] _lowerCamelCase , _lowerCamelCase = scores_sum_average.view(-1 ).topk(lowerCamelCase__ , -1 ) _lowerCamelCase = next_tokens // scores_sum.shape[1] _lowerCamelCase = seq_lengths[next_tokens_source] _lowerCamelCase = next_tokens % scores_sum.shape[1] _lowerCamelCase = next_tokens.unsqueeze(1 ) _lowerCamelCase = tokens[next_tokens_source] _lowerCamelCase = torch.cat((tokens, next_tokens) , dim=1 ) _lowerCamelCase = generated[next_tokens_source] _lowerCamelCase = scores_sum_average * seq_lengths _lowerCamelCase = is_stopped[next_tokens_source] _lowerCamelCase = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) _lowerCamelCase = torch.cat((generated, next_token_embed) , dim=1 ) _lowerCamelCase = is_stopped + next_tokens.eq(lowerCamelCase__ ).squeeze() if is_stopped.all(): break _lowerCamelCase = scores / seq_lengths _lowerCamelCase = scores.argsort(descending=lowerCamelCase__ ) # tokens tensors are already padded to max_seq_length _lowerCamelCase = [tokens[i] for i in order] _lowerCamelCase = torch.stack(lowerCamelCase__ , dim=0 ) _lowerCamelCase = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
661
0
import doctest import glob import importlib import inspect import os import re from contextlib import contextmanager from functools import wraps from unittest.mock import patch import numpy as np import pytest from absl.testing import parameterized import datasets from datasets import load_metric from .utils import for_all_test_methods, local, slow # mark all tests as integration __snake_case = pytest.mark.integration __snake_case = {'''comet'''} __snake_case = importlib.util.find_spec('''fairseq''') is not None __snake_case = {'''code_eval'''} __snake_case = os.name == '''nt''' __snake_case = {'''bertscore''', '''frugalscore''', '''perplexity'''} __snake_case = importlib.util.find_spec('''transformers''') is not None def _A ( _lowercase ) -> Optional[Any]: """simple docstring""" @wraps(_lowercase ) def wrapper(self , _lowercase ): if not _has_fairseq and metric_name in REQUIRE_FAIRSEQ: self.skipTest('"test requires Fairseq"' ) else: test_case(self , _lowercase ) return wrapper def _A ( _lowercase ) -> List[str]: """simple docstring""" @wraps(_lowercase ) def wrapper(self , _lowercase ): if not _has_transformers and metric_name in REQUIRE_TRANSFORMERS: self.skipTest('"test requires transformers"' ) else: test_case(self , _lowercase ) return wrapper def _A ( _lowercase ) -> List[str]: """simple docstring""" @wraps(_lowercase ) def wrapper(self , _lowercase ): if _on_windows and metric_name in UNSUPPORTED_ON_WINDOWS: self.skipTest('"test not supported on Windows"' ) else: test_case(self , _lowercase ) return wrapper def _A ( ) -> List[str]: """simple docstring""" __UpperCamelCase = [metric_dir.split(os.sep )[-2] for metric_dir in glob.glob('./metrics/*/' )] return [{"testcase_name": x, "metric_name": x} for x in metrics if x != "gleu"] # gleu is unfinished @parameterized.named_parameters(get_local_metric_names() ) @for_all_test_methods( _a , _a , _a ) @local class __lowerCamelCase (parameterized.TestCase ): _lowercase = {} _lowercase = None @pytest.mark.filterwarnings('ignore:metric_module_factory is deprecated:FutureWarning' ) @pytest.mark.filterwarnings('ignore:load_metric is deprecated:FutureWarning' ) def snake_case_ ( self: int,A_: List[Any] ): '''simple docstring''' __UpperCamelCase = '[...]' __UpperCamelCase = importlib.import_module( datasets.load.metric_module_factory(os.path.join('metrics',A_ ) ).module_path ) __UpperCamelCase = datasets.load.import_main_class(metric_module.__name__,dataset=A_ ) # check parameters __UpperCamelCase = inspect.signature(metric._compute ).parameters self.assertTrue(all(p.kind != p.VAR_KEYWORD for p in parameters.values() ) ) # no **kwargs # run doctest with self.patch_intensive_calls(A_,metric_module.__name__ ): with self.use_local_metrics(): try: __UpperCamelCase = doctest.testmod(A_,verbose=A_,raise_on_error=A_ ) except doctest.UnexpectedException as e: raise e.exc_info[1] # raise the exception that doctest caught self.assertEqual(results.failed,0 ) self.assertGreater(results.attempted,1 ) @slow def snake_case_ ( self: Optional[Any],A_: List[str] ): '''simple docstring''' __UpperCamelCase = '[...]' __UpperCamelCase = importlib.import_module( datasets.load.metric_module_factory(os.path.join('metrics',A_ ) ).module_path ) # run doctest with self.use_local_metrics(): __UpperCamelCase = doctest.testmod(A_,verbose=A_,raise_on_error=A_ ) self.assertEqual(results.failed,0 ) self.assertGreater(results.attempted,1 ) @contextmanager def snake_case_ ( self: Dict,A_: Optional[Any],A_: Any ): '''simple docstring''' if metric_name in self.INTENSIVE_CALLS_PATCHER: with self.INTENSIVE_CALLS_PATCHER[metric_name](A_ ): yield else: yield @contextmanager def snake_case_ ( self: Tuple ): '''simple docstring''' def load_local_metric(A_: List[Any],*A_: str,**A_: List[Any] ): return load_metric(os.path.join('metrics',A_ ),*A_,**A_ ) with patch('datasets.load_metric' ) as mock_load_metric: __UpperCamelCase = load_local_metric yield @classmethod def snake_case_ ( cls: Any,A_: str ): '''simple docstring''' def wrapper(A_: Optional[Any] ): __UpperCamelCase = contextmanager(A_ ) __UpperCamelCase = patcher return patcher return wrapper @LocalMetricTest.register_intensive_calls_patcher('bleurt' ) def _A ( _lowercase ) -> Optional[Any]: """simple docstring""" import tensorflow.compat.va as tf from bleurt.score import Predictor tf.flags.DEFINE_string('sv' , '' , '' ) # handle pytest cli flags class __lowerCamelCase (_a ): def snake_case_ ( self: List[str],A_: List[Any] ): '''simple docstring''' assert len(input_dict['input_ids'] ) == 2 return np.array([1.0_3, 1.0_4] ) # mock predict_fn which is supposed to do a forward pass with a bleurt model with patch('bleurt.score._create_predictor' ) as mock_create_predictor: __UpperCamelCase = MockedPredictor() yield @LocalMetricTest.register_intensive_calls_patcher('bertscore' ) def _A ( _lowercase ) -> Union[str, Any]: """simple docstring""" import torch def bert_cos_score_idf(_lowercase , _lowercase , *_lowercase , **_lowercase ): return torch.tensor([[1.0, 1.0, 1.0]] * len(_lowercase ) ) # mock get_model which is supposed to do download a bert model # mock bert_cos_score_idf which is supposed to do a forward pass with a bert model with patch('bert_score.scorer.get_model' ), patch( 'bert_score.scorer.bert_cos_score_idf' ) as mock_bert_cos_score_idf: __UpperCamelCase = bert_cos_score_idf yield @LocalMetricTest.register_intensive_calls_patcher('comet' ) def _A ( _lowercase ) -> Optional[Any]: """simple docstring""" def load_from_checkpoint(_lowercase ): class __lowerCamelCase : def snake_case_ ( self: Optional[int],A_: int,*A_: Optional[int],**A_: List[str] ): '''simple docstring''' assert len(A_ ) == 2 __UpperCamelCase = [0.1_9, 0.9_2] return scores, sum(A_ ) / len(A_ ) return Model() # mock load_from_checkpoint which is supposed to do download a bert model # mock load_from_checkpoint which is supposed to do download a bert model with patch('comet.download_model' ) as mock_download_model: __UpperCamelCase = None with patch('comet.load_from_checkpoint' ) as mock_load_from_checkpoint: __UpperCamelCase = load_from_checkpoint yield def _A ( ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = load_metric(os.path.join('metrics' , 'seqeval' ) ) __UpperCamelCase = 'ERROR' __UpperCamelCase = f'''Scheme should be one of [IOB1, IOB2, IOE1, IOE2, IOBES, BILOU], got {wrong_scheme}''' with pytest.raises(_lowercase , match=re.escape(_lowercase ) ): metric.compute(predictions=[] , references=[] , scheme=_lowercase )
1
"""simple docstring""" import re import string from collections import Counter import sacrebleu import sacremoses from packaging import version import datasets __SCREAMING_SNAKE_CASE : Optional[int] = ''' @inproceedings{xu-etal-2016-optimizing, title = {Optimizing Statistical Machine Translation for Text Simplification}, authors={Xu, Wei and Napoles, Courtney and Pavlick, Ellie and Chen, Quanze and Callison-Burch, Chris}, journal = {Transactions of the Association for Computational Linguistics}, volume = {4}, year={2016}, url = {https://www.aclweb.org/anthology/Q16-1029}, pages = {401--415 }, @inproceedings{post-2018-call, title = "A Call for Clarity in Reporting {BLEU} Scores", author = "Post, Matt", booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers", month = oct, year = "2018", address = "Belgium, Brussels", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W18-6319", pages = "186--191", } ''' __SCREAMING_SNAKE_CASE : List[str] = '''\ WIKI_SPLIT is the combination of three metrics SARI, EXACT and SACREBLEU It can be used to evaluate the quality of machine-generated texts. ''' __SCREAMING_SNAKE_CASE : Optional[Any] = ''' Calculates sari score (between 0 and 100) given a list of source and predicted sentences, and a list of lists of reference sentences. It also computes the BLEU score as well as the exact match score. Args: sources: list of source sentences where each sentence should be a string. predictions: list of predicted sentences where each sentence should be a string. references: list of lists of reference sentences where each sentence should be a string. Returns: sari: sari score sacrebleu: sacrebleu score exact: exact score Examples: >>> sources=["About 95 species are currently accepted ."] >>> predictions=["About 95 you now get in ."] >>> references=[["About 95 species are currently known ."]] >>> wiki_split = datasets.load_metric("wiki_split") >>> results = wiki_split.compute(sources=sources, predictions=predictions, references=references) >>> print(results) {\'sari\': 21.805555555555557, \'sacrebleu\': 14.535768424205482, \'exact\': 0.0} ''' def lowerCAmelCase_( lowercase_ : Union[str, Any] ) -> str: def remove_articles(lowercase_ : int ): _lowerCamelCase = re.compile(r'''\b(a|an|the)\b''' , re.UNICODE ) return re.sub(lowercase_ , ''' ''' , lowercase_ ) def white_space_fix(lowercase_ : List[Any] ): return " ".join(text.split() ) def remove_punc(lowercase_ : Dict ): _lowerCamelCase = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowercase_ : Union[str, Any] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowercase_ ) ) ) ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[Any] ) -> Union[str, Any]: return int(normalize_answer(lowercase_ ) == normalize_answer(lowercase_ ) ) def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Tuple ) -> Tuple: _lowerCamelCase = [any(compute_exact(lowercase_ , lowercase_ ) for ref in refs ) for pred, refs in zip(lowercase_ , lowercase_ )] return (sum(lowercase_ ) / len(lowercase_ )) * 1_00 def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : str , lowercase_ : Optional[Any] , lowercase_ : str ) -> Optional[int]: _lowerCamelCase = [rgram for rgrams in rgramslist for rgram in rgrams] _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter() for sgram, scount in sgramcounter.items(): _lowerCamelCase = scount * numref _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter() for cgram, ccount in cgramcounter.items(): _lowerCamelCase = ccount * numref # KEEP _lowerCamelCase = sgramcounter_rep & cgramcounter_rep _lowerCamelCase = keepgramcounter_rep & rgramcounter _lowerCamelCase = sgramcounter_rep & rgramcounter _lowerCamelCase = 0 _lowerCamelCase = 0 for keepgram in keepgramcountergood_rep: keeptmpscorea += keepgramcountergood_rep[keepgram] / keepgramcounter_rep[keepgram] # Fix an alleged bug [2] in the keep score computation. # keeptmpscore2 += keepgramcountergood_rep[keepgram] / keepgramcounterall_rep[keepgram] keeptmpscorea += keepgramcountergood_rep[keepgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = keeptmpscorea / len(lowercase_ ) if len(lowercase_ ) > 0: # Fix an alleged bug [2] in the keep score computation. # keepscore_recall = keeptmpscore2 / len(keepgramcounterall_rep) _lowerCamelCase = keeptmpscorea / sum(keepgramcounterall_rep.values() ) _lowerCamelCase = 0 if keepscore_precision > 0 or keepscore_recall > 0: _lowerCamelCase = 2 * keepscore_precision * keepscore_recall / (keepscore_precision + keepscore_recall) # DELETION _lowerCamelCase = sgramcounter_rep - cgramcounter_rep _lowerCamelCase = delgramcounter_rep - rgramcounter _lowerCamelCase = sgramcounter_rep - rgramcounter _lowerCamelCase = 0 _lowerCamelCase = 0 for delgram in delgramcountergood_rep: deltmpscorea += delgramcountergood_rep[delgram] / delgramcounter_rep[delgram] deltmpscorea += delgramcountergood_rep[delgram] / delgramcounterall_rep[delgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = deltmpscorea / len(lowercase_ ) # ADDITION _lowerCamelCase = set(lowercase_ ) - set(lowercase_ ) _lowerCamelCase = set(lowercase_ ) & set(lowercase_ ) _lowerCamelCase = set(lowercase_ ) - set(lowercase_ ) _lowerCamelCase = 0 for addgram in addgramcountergood: addtmpscore += 1 # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = addtmpscore / len(lowercase_ ) if len(lowercase_ ) > 0: _lowerCamelCase = addtmpscore / len(lowercase_ ) _lowerCamelCase = 0 if addscore_precision > 0 or addscore_recall > 0: _lowerCamelCase = 2 * addscore_precision * addscore_recall / (addscore_precision + addscore_recall) return (keepscore, delscore_precision, addscore) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : Optional[Any] , lowercase_ : str ) -> List[str]: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = ssent.split(''' ''' ) _lowerCamelCase = csent.split(''' ''' ) _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] for rsent in rsents: _lowerCamelCase = rsent.split(''' ''' ) _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] ragramslist.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] ragrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] + ''' ''' + ragrams[i + 2] ragrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] + ''' ''' + ragrams[i + 2] + ''' ''' + ragrams[i + 3] ragrams.append(lowercase_ ) ragramslist.append(lowercase_ ) ragramslist.append(lowercase_ ) ragramslist.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] sagrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] + ''' ''' + sagrams[i + 2] sagrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] + ''' ''' + sagrams[i + 2] + ''' ''' + sagrams[i + 3] sagrams.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] cagrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] + ''' ''' + cagrams[i + 2] cagrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] + ''' ''' + cagrams[i + 2] + ''' ''' + cagrams[i + 3] cagrams.append(lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) _lowerCamelCase = sum([keepascore, keepascore, keepascore, keepascore] ) / 4 _lowerCamelCase = sum([delascore, delascore, delascore, delascore] ) / 4 _lowerCamelCase = sum([addascore, addascore, addascore, addascore] ) / 4 _lowerCamelCase = (avgkeepscore + avgdelscore + avgaddscore) / 3 return finalscore def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : bool = True , lowercase_ : str = "13a" , lowercase_ : bool = True ) -> int: # Normalization is requried for the ASSET dataset (one of the primary # datasets in sentence simplification) to allow using space # to split the sentence. Even though Wiki-Auto and TURK datasets, # do not require normalization, we do it for consistency. # Code adapted from the EASSE library [1] written by the authors of the ASSET dataset. # [1] https://github.com/feralvam/easse/blob/580bba7e1378fc8289c663f864e0487188fe8067/easse/utils/preprocessing.py#L7 if lowercase: _lowerCamelCase = sentence.lower() if tokenizer in ["13a", "intl"]: if version.parse(sacrebleu.__version__ ).major >= 2: _lowerCamelCase = sacrebleu.metrics.bleu._get_tokenizer(lowercase_ )()(lowercase_ ) else: _lowerCamelCase = sacrebleu.TOKENIZERS[tokenizer]()(lowercase_ ) elif tokenizer == "moses": _lowerCamelCase = sacremoses.MosesTokenizer().tokenize(lowercase_ , return_str=lowercase_ , escape=lowercase_ ) elif tokenizer == "penn": _lowerCamelCase = sacremoses.MosesTokenizer().penn_tokenize(lowercase_ , return_str=lowercase_ ) else: _lowerCamelCase = sentence if not return_str: _lowerCamelCase = normalized_sent.split() return normalized_sent def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Any , lowercase_ : List[Any] ) -> Optional[int]: if not (len(lowercase_ ) == len(lowercase_ ) == len(lowercase_ )): raise ValueError('''Sources length must match predictions and references lengths.''' ) _lowerCamelCase = 0 for src, pred, refs in zip(lowercase_ , lowercase_ , lowercase_ ): sari_score += SARIsent(normalize(lowercase_ ) , normalize(lowercase_ ) , [normalize(lowercase_ ) for sent in refs] ) _lowerCamelCase = sari_score / len(lowercase_ ) return 1_00 * sari_score def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Any , lowercase_ : List[Any]="exp" , lowercase_ : List[Any]=None , lowercase_ : Optional[Any]=False , lowercase_ : List[Any]=False , lowercase_ : List[Any]=False , ) -> Dict: _lowerCamelCase = len(references[0] ) if any(len(lowercase_ ) != references_per_prediction for refs in references ): raise ValueError('''Sacrebleu requires the same number of references for each prediction''' ) _lowerCamelCase = [[refs[i] for refs in references] for i in range(lowercase_ )] _lowerCamelCase = sacrebleu.corpus_bleu( lowercase_ , lowercase_ , smooth_method=lowercase_ , smooth_value=lowercase_ , force=lowercase_ , lowercase=lowercase_ , use_effective_order=lowercase_ , ) return output.score @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Sequence(datasets.Value('''string''' , id='''sequence''' ) , id='''references''' ), } ) , codebase_urls=[ '''https://github.com/huggingface/transformers/blob/master/src/transformers/data/metrics/squad_metrics.py''', '''https://github.com/cocoxu/simplification/blob/master/SARI.py''', '''https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/sari_hook.py''', '''https://github.com/mjpost/sacreBLEU''', ] , reference_urls=[ '''https://www.aclweb.org/anthology/Q16-1029.pdf''', '''https://github.com/mjpost/sacreBLEU''', '''https://en.wikipedia.org/wiki/BLEU''', '''https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213''', ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = {} result.update({'''sari''': compute_sari(sources=lowerCamelCase__ , predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) result.update({'''sacrebleu''': compute_sacrebleu(predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) result.update({'''exact''': compute_em(predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) return result
661
0
import unittest import torch from diffusers import VQModel from diffusers.utils import floats_tensor, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin enable_full_determinism() class lowerCamelCase__ ( _A , _A , unittest.TestCase): """simple docstring""" a__ : Optional[Any] = VQModel a__ : Optional[int] = "sample" @property def snake_case_ ( self : Optional[int] , __lowerCAmelCase : Any=(32, 32) ) -> int: _A = 4 _A = 3 _A = floats_tensor((batch_size, num_channels) + sizes ).to(__lowerCAmelCase ) return {"sample": image} @property def snake_case_ ( self : Optional[Any] ) -> List[Any]: return (3, 32, 32) @property def snake_case_ ( self : Tuple ) -> Tuple: return (3, 32, 32) def snake_case_ ( self : Any ) -> Tuple: _A = { '''block_out_channels''': [32, 64], '''in_channels''': 3, '''out_channels''': 3, '''down_block_types''': ['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''], '''up_block_types''': ['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''], '''latent_channels''': 3, } _A = self.dummy_input return init_dict, inputs_dict def snake_case_ ( self : Optional[Any] ) -> Tuple: pass def snake_case_ ( self : Any ) -> Optional[int]: pass def snake_case_ ( self : Union[str, Any] ) -> Any: _A , _A = VQModel.from_pretrained('''fusing/vqgan-dummy''' , output_loading_info=__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) self.assertEqual(len(loading_info['''missing_keys'''] ) , 0 ) model.to(__lowerCAmelCase ) _A = model(**self.dummy_input ) assert image is not None, "Make sure output is not None" def snake_case_ ( self : List[Any] ) -> Optional[Any]: _A = VQModel.from_pretrained('''fusing/vqgan-dummy''' ) model.to(__lowerCAmelCase ).eval() torch.manual_seed(0 ) if torch.cuda.is_available(): torch.cuda.manual_seed_all(0 ) _A = torch.randn(1 , model.config.in_channels , model.config.sample_size , model.config.sample_size ) _A = image.to(__lowerCAmelCase ) with torch.no_grad(): _A = model(__lowerCAmelCase ).sample _A = output[0, -1, -3:, -3:].flatten().cpu() # fmt: off _A = torch.tensor([-0.0153, -0.4044, -0.1880, -0.5161, -0.2418, -0.4072, -0.1612, -0.0633, -0.0143] ) # fmt: on self.assertTrue(torch.allclose(__lowerCAmelCase , __lowerCAmelCase , atol=1E-3 ) )
2
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''configuration_wav2vec2''': ['''WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Wav2Vec2Config'''], '''feature_extraction_wav2vec2''': ['''Wav2Vec2FeatureExtractor'''], '''processing_wav2vec2''': ['''Wav2Vec2Processor'''], '''tokenization_wav2vec2''': ['''Wav2Vec2CTCTokenizer''', '''Wav2Vec2Tokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = [ '''WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Wav2Vec2ForAudioFrameClassification''', '''Wav2Vec2ForCTC''', '''Wav2Vec2ForMaskedLM''', '''Wav2Vec2ForPreTraining''', '''Wav2Vec2ForSequenceClassification''', '''Wav2Vec2ForXVector''', '''Wav2Vec2Model''', '''Wav2Vec2PreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = [ '''TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWav2Vec2ForCTC''', '''TFWav2Vec2Model''', '''TFWav2Vec2PreTrainedModel''', '''TFWav2Vec2ForSequenceClassification''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = [ '''FlaxWav2Vec2ForCTC''', '''FlaxWav2Vec2ForPreTraining''', '''FlaxWav2Vec2Model''', '''FlaxWav2Vec2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavaveca import WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, WavaVecaConfig from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .processing_wavaveca import WavaVecaProcessor from .tokenization_wavaveca import WavaVecaCTCTokenizer, WavaVecaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavaveca import ( WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaForAudioFrameClassification, WavaVecaForCTC, WavaVecaForMaskedLM, WavaVecaForPreTraining, WavaVecaForSequenceClassification, WavaVecaForXVector, WavaVecaModel, WavaVecaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, TFWavaVecaForCTC, TFWavaVecaForSequenceClassification, TFWavaVecaModel, TFWavaVecaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( FlaxWavaVecaForCTC, FlaxWavaVecaForPreTraining, FlaxWavaVecaModel, FlaxWavaVecaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
661
0
'''simple docstring''' def A_( A : int): UpperCamelCase = int(A) if decimal in (0, 1): # Exit cases for the recursion return str(A) UpperCamelCase , UpperCamelCase = divmod(A , 2) return binary_recursive(A) + str(A) def A_( A : str): UpperCamelCase = str(A).strip() if not number: raise ValueError('No input value was provided') UpperCamelCase = '-' if number.startswith('-') else '' UpperCamelCase = number.lstrip('-') if not number.isnumeric(): raise ValueError('Input value is not an integer') return f'''{negative}0b{binary_recursive(int(A))}''' if __name__ == "__main__": from doctest import testmod testmod()
3
"""simple docstring""" __SCREAMING_SNAKE_CASE : Any = { '''A''': '''.-''', '''B''': '''-...''', '''C''': '''-.-.''', '''D''': '''-..''', '''E''': '''.''', '''F''': '''..-.''', '''G''': '''--.''', '''H''': '''....''', '''I''': '''..''', '''J''': '''.---''', '''K''': '''-.-''', '''L''': '''.-..''', '''M''': '''--''', '''N''': '''-.''', '''O''': '''---''', '''P''': '''.--.''', '''Q''': '''--.-''', '''R''': '''.-.''', '''S''': '''...''', '''T''': '''-''', '''U''': '''..-''', '''V''': '''...-''', '''W''': '''.--''', '''X''': '''-..-''', '''Y''': '''-.--''', '''Z''': '''--..''', '''1''': '''.----''', '''2''': '''..---''', '''3''': '''...--''', '''4''': '''....-''', '''5''': '''.....''', '''6''': '''-....''', '''7''': '''--...''', '''8''': '''---..''', '''9''': '''----.''', '''0''': '''-----''', '''&''': '''.-...''', '''@''': '''.--.-.''', ''':''': '''---...''', ''',''': '''--..--''', '''.''': '''.-.-.-''', '''\'''': '''.----.''', '''"''': '''.-..-.''', '''?''': '''..--..''', '''/''': '''-..-.''', '''=''': '''-...-''', '''+''': '''.-.-.''', '''-''': '''-....-''', '''(''': '''-.--.''', ''')''': '''-.--.-''', '''!''': '''-.-.--''', ''' ''': '''/''' } # Exclamation mark is not in ITU-R recommendation # fmt: on __SCREAMING_SNAKE_CASE : List[str] = {value: key for key, value in MORSE_CODE_DICT.items()} def lowerCAmelCase_( lowercase_ : str ) -> str: return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def lowerCAmelCase_( lowercase_ : str ) -> str: return "".join(REVERSE_DICT[char] for char in message.split() ) def lowerCAmelCase_( ) -> None: _lowerCamelCase = '''Morse code here!''' print(lowercase_ ) _lowerCamelCase = encrypt(lowercase_ ) print(lowercase_ ) _lowerCamelCase = decrypt(lowercase_ ) print(lowercase_ ) if __name__ == "__main__": main()
661
0
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ConvNextConfig, UperNetConfig from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torch_device from transformers.utils import is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import UperNetForSemanticSegmentation from transformers.models.upernet.modeling_upernet import UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class a : def __init__( self , _snake_case , _snake_case=13 , _snake_case=32 , _snake_case=3 , _snake_case=4 , _snake_case=[10, 20, 30, 40] , _snake_case=[2, 2, 3, 2] , _snake_case=True , _snake_case=True , _snake_case=37 , _snake_case="gelu" , _snake_case=10 , _snake_case=0.02 , _snake_case=["stage2", "stage3", "stage4"] , _snake_case=3 , _snake_case=None , ): """simple docstring""" lowerCAmelCase = parent lowerCAmelCase = batch_size lowerCAmelCase = image_size lowerCAmelCase = num_channels lowerCAmelCase = num_stages lowerCAmelCase = hidden_sizes lowerCAmelCase = depths lowerCAmelCase = is_training lowerCAmelCase = use_labels lowerCAmelCase = intermediate_size lowerCAmelCase = hidden_act lowerCAmelCase = type_sequence_label_size lowerCAmelCase = initializer_range lowerCAmelCase = out_features lowerCAmelCase = num_labels lowerCAmelCase = scope lowerCAmelCase = num_stages def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCAmelCase = None if self.use_labels: lowerCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowerCAmelCase = self.get_config() return config, pixel_values, labels def UpperCamelCase__ ( self ): """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , num_stages=self.num_stages , hidden_sizes=self.hidden_sizes , depths=self.depths , is_training=self.is_training , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , out_features=self.out_features , ) def UpperCamelCase__ ( self ): """simple docstring""" return UperNetConfig( backbone_config=self.get_backbone_config() , hidden_size=5_12 , pool_scales=[1, 2, 3, 6] , use_auxiliary_head=_snake_case , auxiliary_loss_weight=0.4 , auxiliary_in_channels=40 , auxiliary_channels=2_56 , auxiliary_num_convs=1 , auxiliary_concat_input=_snake_case , loss_ignore_index=2_55 , num_labels=self.num_labels , ) def UpperCamelCase__ ( self , _snake_case , _snake_case , _snake_case ): """simple docstring""" lowerCAmelCase = UperNetForSemanticSegmentation(config=_snake_case ) model.to(_snake_case ) model.eval() lowerCAmelCase = model(_snake_case ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = self.prepare_config_and_inputs() ( ( lowerCAmelCase ) ,( lowerCAmelCase ) ,( lowerCAmelCase ) , ) = config_and_inputs lowerCAmelCase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class a ( a__ , a__ , unittest.TestCase ): snake_case__ = (UperNetForSemanticSegmentation,) if is_torch_available() else () snake_case__ = {'''image-segmentation''': UperNetForSemanticSegmentation} if is_torch_available() else {} snake_case__ = False snake_case__ = False snake_case__ = False snake_case__ = False snake_case__ = False snake_case__ = False def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = UperNetModelTester(self ) lowerCAmelCase = ConfigTester(self , config_class=_snake_case , has_text_modality=_snake_case , hidden_size=37 ) def UpperCamelCase__ ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase__ ( self ): """simple docstring""" return def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase ,lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = model_class(_snake_case ) lowerCAmelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCAmelCase = [*signature.parameters.keys()] lowerCAmelCase = ['pixel_values'] self.assertListEqual(arg_names[:1] , _snake_case ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*_snake_case ) @unittest.skip(reason='UperNet does not use inputs_embeds' ) def UpperCamelCase__ ( self ): """simple docstring""" pass @unittest.skip(reason='UperNet does not support input and output embeddings' ) def UpperCamelCase__ ( self ): """simple docstring""" pass @unittest.skip(reason='UperNet does not have a base model' ) def UpperCamelCase__ ( self ): """simple docstring""" pass @unittest.skip(reason='UperNet does not have a base model' ) def UpperCamelCase__ ( self ): """simple docstring""" pass @require_torch_multi_gpu @unittest.skip(reason='UperNet has some layers using `add_module` which doesn\'t work well with `nn.DataParallel`' ) def UpperCamelCase__ ( self ): """simple docstring""" pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self ): """simple docstring""" def check_hidden_states_output(_snake_case , _snake_case , _snake_case ): lowerCAmelCase = model_class(_snake_case ) model.to(_snake_case ) model.eval() with torch.no_grad(): lowerCAmelCase = model(**self._prepare_for_class(_snake_case , _snake_case ) ) lowerCAmelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states lowerCAmelCase = self.model_tester.num_stages self.assertEqual(len(_snake_case ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) lowerCAmelCase ,lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCAmelCase = True check_hidden_states_output(_snake_case , _snake_case , _snake_case ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCAmelCase = True check_hidden_states_output(_snake_case , _snake_case , _snake_case ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase ,lowerCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() lowerCAmelCase = _config_zero_init(_snake_case ) lowerCAmelCase = _config_zero_init(configs_no_init.backbone_config ) for model_class in self.all_model_classes: lowerCAmelCase = model_class(config=_snake_case ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) @unittest.skip(reason='UperNet does not have tied weights' ) def UpperCamelCase__ ( self ): """simple docstring""" pass @slow def UpperCamelCase__ ( self ): """simple docstring""" for model_name in UPERNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCAmelCase = UperNetForSemanticSegmentation.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) def _SCREAMING_SNAKE_CASE (): lowerCAmelCase = hf_hub_download( repo_id='hf-internal-testing/fixtures_ade20k' , repo_type='dataset' , filename='ADE_val_00000001.jpg' ) lowerCAmelCase = Image.open(_UpperCAmelCase ).convert('RGB' ) return image @require_torch @require_vision @slow class a ( unittest.TestCase ): def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = AutoImageProcessor.from_pretrained('openmmlab/upernet-swin-tiny' ) lowerCAmelCase = UperNetForSemanticSegmentation.from_pretrained('openmmlab/upernet-swin-tiny' ).to(_snake_case ) lowerCAmelCase = prepare_img() lowerCAmelCase = processor(images=_snake_case , return_tensors='pt' ).to(_snake_case ) with torch.no_grad(): lowerCAmelCase = model(**_snake_case ) lowerCAmelCase = torch.Size((1, model.config.num_labels, 5_12, 5_12) ) self.assertEqual(outputs.logits.shape , _snake_case ) lowerCAmelCase = torch.tensor( [[-7.5_958, -7.5_958, -7.4_302], [-7.5_958, -7.5_958, -7.4_302], [-7.4_797, -7.4_797, -7.3_068]] ).to(_snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , _snake_case , atol=1E-4 ) ) def UpperCamelCase__ ( self ): """simple docstring""" lowerCAmelCase = AutoImageProcessor.from_pretrained('openmmlab/upernet-convnext-tiny' ) lowerCAmelCase = UperNetForSemanticSegmentation.from_pretrained('openmmlab/upernet-convnext-tiny' ).to(_snake_case ) lowerCAmelCase = prepare_img() lowerCAmelCase = processor(images=_snake_case , return_tensors='pt' ).to(_snake_case ) with torch.no_grad(): lowerCAmelCase = model(**_snake_case ) lowerCAmelCase = torch.Size((1, model.config.num_labels, 5_12, 5_12) ) self.assertEqual(outputs.logits.shape , _snake_case ) lowerCAmelCase = torch.tensor( [[-8.8_110, -8.8_110, -8.6_521], [-8.8_110, -8.8_110, -8.6_521], [-8.7_746, -8.7_746, -8.6_130]] ).to(_snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, 0, :3, :3] , _snake_case , atol=1E-4 ) )
4
"""simple docstring""" from argparse import ArgumentParser, Namespace from typing import Any, List, Optional from ..pipelines import Pipeline, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand try: from fastapi import Body, FastAPI, HTTPException from fastapi.routing import APIRoute from pydantic import BaseModel from starlette.responses import JSONResponse from uvicorn import run __SCREAMING_SNAKE_CASE : List[Any] = True except (ImportError, AttributeError): __SCREAMING_SNAKE_CASE : List[Any] = object def lowerCAmelCase_( *lowercase_ : Dict , **lowercase_ : str ) -> str: pass __SCREAMING_SNAKE_CASE : Tuple = False __SCREAMING_SNAKE_CASE : Any = logging.get_logger('''transformers-cli/serving''') def lowerCAmelCase_( lowercase_ : Namespace ) -> List[Any]: _lowerCamelCase = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) return ServeCommand(lowercase_ , args.host , args.port , args.workers ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : dict class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] lowercase__ : Optional[List[int]] class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : str class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Any class lowerCamelCase_( A__ ): '''simple docstring''' @staticmethod def snake_case__ ( lowerCamelCase__ ): _lowerCamelCase = parser.add_parser( '''serve''' , help='''CLI tool to run inference requests through REST and GraphQL endpoints.''' ) serve_parser.add_argument( '''--task''' , type=lowerCamelCase__ , choices=get_supported_tasks() , help='''The task to run the pipeline on''' , ) serve_parser.add_argument('''--host''' , type=lowerCamelCase__ , default='''localhost''' , help='''Interface the server will listen on.''' ) serve_parser.add_argument('''--port''' , type=lowerCamelCase__ , default=8_8_8_8 , help='''Port the serving will listen to.''' ) serve_parser.add_argument('''--workers''' , type=lowerCamelCase__ , default=1 , help='''Number of http workers''' ) serve_parser.add_argument('''--model''' , type=lowerCamelCase__ , help='''Model\'s name or path to stored model.''' ) serve_parser.add_argument('''--config''' , type=lowerCamelCase__ , help='''Model\'s config name or path to stored model.''' ) serve_parser.add_argument('''--tokenizer''' , type=lowerCamelCase__ , help='''Tokenizer name to use.''' ) serve_parser.add_argument( '''--device''' , type=lowerCamelCase__ , default=-1 , help='''Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)''' , ) serve_parser.set_defaults(func=lowerCamelCase__ ) def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = pipeline _lowerCamelCase = host _lowerCamelCase = port _lowerCamelCase = workers if not _serve_dependencies_installed: raise RuntimeError( '''Using serve command requires FastAPI and uvicorn. ''' '''Please install transformers with [serving]: pip install "transformers[serving]".''' '''Or install FastAPI and uvicorn separately.''' ) else: logger.info(F"""Serving model over {host}:{port}""" ) _lowerCamelCase = FastAPI( routes=[ APIRoute( '''/''' , self.model_info , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''GET'''] , ), APIRoute( '''/tokenize''' , self.tokenize , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), APIRoute( '''/detokenize''' , self.detokenize , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), APIRoute( '''/forward''' , self.forward , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), ] , timeout=6_0_0 , ) def snake_case__ ( self ): run(self._app , host=self.host , port=self.port , workers=self.workers ) def snake_case__ ( self ): return ServeModelInfoResult(infos=vars(self._pipeline.model.config ) ) def snake_case__ ( self , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) ): try: _lowerCamelCase = self._pipeline.tokenizer.tokenize(lowerCamelCase__ ) if return_ids: _lowerCamelCase = self._pipeline.tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) return ServeTokenizeResult(tokens=lowerCamelCase__ , tokens_ids=lowerCamelCase__ ) else: return ServeTokenizeResult(tokens=lowerCamelCase__ ) except Exception as e: raise HTTPException(status_code=5_0_0 , detail={'''model''': '''''', '''error''': str(lowerCamelCase__ )} ) def snake_case__ ( self , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , ): try: _lowerCamelCase = self._pipeline.tokenizer.decode(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) return ServeDeTokenizeResult(model='''''' , text=lowerCamelCase__ ) except Exception as e: raise HTTPException(status_code=5_0_0 , detail={'''model''': '''''', '''error''': str(lowerCamelCase__ )} ) async def snake_case__ ( self , lowerCamelCase__=Body(lowerCamelCase__ , embed=lowerCamelCase__ ) ): # Check we don't have empty string if len(lowerCamelCase__ ) == 0: return ServeForwardResult(output=[] , attention=[] ) try: # Forward through the model _lowerCamelCase = self._pipeline(lowerCamelCase__ ) return ServeForwardResult(output=lowerCamelCase__ ) except Exception as e: raise HTTPException(5_0_0 , {'''error''': str(lowerCamelCase__ )} )
661
0
'''simple docstring''' from typing import List from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { """snap-research/efficientformer-l1-300""": ( """https://huggingface.co/snap-research/efficientformer-l1-300/resolve/main/config.json""" ), } class UpperCAmelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' _lowercase : Tuple = '''efficientformer''' def __init__( self , _lowercase = [3, 2, 6, 4] , _lowercase = [48, 96, 224, 448] , _lowercase = [True, True, True, True] , _lowercase = 448 , _lowercase = 32 , _lowercase = 4 , _lowercase = 7 , _lowercase = 5 , _lowercase = 8 , _lowercase = 4 , _lowercase = 0.0 , _lowercase = 16 , _lowercase = 3 , _lowercase = 3 , _lowercase = 3 , _lowercase = 2 , _lowercase = 1 , _lowercase = 0.0 , _lowercase = 1 , _lowercase = True , _lowercase = True , _lowercase = 1e-5 , _lowercase = "gelu" , _lowercase = 0.02 , _lowercase = 1e-12 , _lowercase = 224 , _lowercase = 1e-05 , **_lowercase , ): """simple docstring""" super().__init__(**_lowercase ) _lowerCAmelCase = hidden_act _lowerCAmelCase = hidden_dropout_prob _lowerCAmelCase = hidden_sizes _lowerCAmelCase = num_hidden_layers _lowerCAmelCase = num_attention_heads _lowerCAmelCase = initializer_range _lowerCAmelCase = layer_norm_eps _lowerCAmelCase = patch_size _lowerCAmelCase = num_channels _lowerCAmelCase = depths _lowerCAmelCase = mlp_expansion_ratio _lowerCAmelCase = downsamples _lowerCAmelCase = dim _lowerCAmelCase = key_dim _lowerCAmelCase = attention_ratio _lowerCAmelCase = resolution _lowerCAmelCase = pool_size _lowerCAmelCase = downsample_patch_size _lowerCAmelCase = downsample_stride _lowerCAmelCase = downsample_pad _lowerCAmelCase = drop_path_rate _lowerCAmelCase = num_metaad_blocks _lowerCAmelCase = distillation _lowerCAmelCase = use_layer_scale _lowerCAmelCase = layer_scale_init_value _lowerCAmelCase = image_size _lowerCAmelCase = batch_norm_eps
5
"""simple docstring""" import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def lowerCAmelCase_( lowercase_ : str , lowercase_ : Dict ) -> List[Any]: assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : str , lowercase_ : str ) -> List[Any]: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ , keep_in_memory=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : int ) -> Optional[int]: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = features.copy() if features else default_expected_features _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = JsonDatasetReader(lowercase_ , features=lowercase_ , cache_dir=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''}, ] , ) def lowerCAmelCase_( lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : List[Any] ) -> Tuple: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''} _lowerCamelCase = features.copy() if features else default_expected_features _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = JsonDatasetReader(lowercase_ , features=lowercase_ , cache_dir=lowercase_ ).read() assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Optional[int] ) -> List[str]: # jsonl_312_path features are {"col_3": "float64", "col_1": "string", "col_2": "int64"} _lowerCamelCase = {'''col_2''': '''int64''', '''col_3''': '''float64''', '''col_1''': '''string'''} _lowerCamelCase = features.copy() _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = JsonDatasetReader(lowercase_ , features=lowercase_ , cache_dir=lowercase_ ).read() assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : Optional[Any] ) -> int: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ , split=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('''path_type''' , [str, list] ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[int] , lowercase_ : Tuple ) -> Optional[int]: if issubclass(lowercase_ , lowercase_ ): _lowerCamelCase = jsonl_path elif issubclass(lowercase_ , lowercase_ ): _lowerCamelCase = [jsonl_path] _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : int , lowercase_ : List[str] , lowercase_ : str=("train",) ) -> List[str]: assert isinstance(lowercase_ , lowercase_ ) for split in splits: _lowerCamelCase = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[int] , lowercase_ : int ) -> Dict: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _lowerCamelCase = JsonDatasetReader({'''train''': jsonl_path} , cache_dir=lowercase_ , keep_in_memory=lowercase_ ).read() _check_json_datasetdict(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : Union[str, Any] , lowercase_ : Optional[Any] ) -> List[Any]: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = features.copy() if features else default_expected_features _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = JsonDatasetReader({'''train''': jsonl_path} , features=lowercase_ , cache_dir=lowercase_ ).read() _check_json_datasetdict(lowercase_ , lowercase_ ) @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : str , lowercase_ : List[str] ) -> Optional[Any]: if split: _lowerCamelCase = {split: jsonl_path} else: _lowerCamelCase = '''train''' _lowerCamelCase = {'''train''': jsonl_path, '''test''': jsonl_path} _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ ).read() _check_json_datasetdict(lowercase_ , lowercase_ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def lowerCAmelCase_( lowercase_ : Optional[int] ) -> Optional[Any]: return json.load(lowercase_ ) def lowerCAmelCase_( lowercase_ : Tuple ) -> Tuple: return [json.loads(lowercase_ ) for line in buffer] class lowerCamelCase_: '''simple docstring''' @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ ).write() buffer.seek(0 ) _lowerCamelCase = load_json_function(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) assert isinstance(exported_content[0] , lowerCamelCase__ ) assert len(lowerCamelCase__ ) == 1_0 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ , orient=lowerCamelCase__ ).write() buffer.seek(0 ) _lowerCamelCase = load_json(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowerCamelCase__ , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 1_0 else: assert len(lowerCamelCase__ ) == 1_0 @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ , num_proc=2 ).write() buffer.seek(0 ) _lowerCamelCase = load_json_function(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) assert isinstance(exported_content[0] , lowerCamelCase__ ) assert len(lowerCamelCase__ ) == 1_0 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ , orient=lowerCamelCase__ , num_proc=2 ).write() buffer.seek(0 ) _lowerCamelCase = load_json(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowerCamelCase__ , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 1_0 else: assert len(lowerCamelCase__ ) == 1_0 def snake_case__ ( self , lowerCamelCase__ ): with pytest.raises(lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , num_proc=0 ) @pytest.mark.parametrize('''compression, extension''' , [('''gzip''', '''gz'''), ('''bz2''', '''bz2'''), ('''xz''', '''xz''')] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = tmp_path_factory.mktemp('''data''' ) / F"""test.json.{extension}""" _lowerCamelCase = str(shared_datadir / F"""test_file.json.{extension}""" ) JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , compression=lowerCamelCase__ ).write() with fsspec.open(lowerCamelCase__ , '''rb''' , compression='''infer''' ) as f: _lowerCamelCase = f.read() with fsspec.open(lowerCamelCase__ , '''rb''' , compression='''infer''' ) as f: _lowerCamelCase = f.read() assert exported_content == original_content
661
0
import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class UpperCamelCase_ : def _snake_case ( self :int ) -> List[str]: """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = TaEncoderModel.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = UNetaDConditionModel( sample_size=32 , layers_per_block=1 , block_out_channels=[32, 64] , down_block_types=[ """ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D""", ] , mid_block_type="""UNetMidBlock2DSimpleCrossAttn""" , up_block_types=["""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""] , in_channels=3 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="""text""" , addition_embed_type_num_heads=2 , cross_attention_norm="""group_norm""" , resnet_time_scale_shift="""scale_shift""" , act_fn="""gelu""" , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = DDPMScheduler( num_train_timesteps=1000 , beta_schedule="""squaredcos_cap_v2""" , beta_start=0.0_0_0_1 , beta_end=0.0_2 , thresholding=__A , dynamic_thresholding_ratio=0.9_5 , sample_max_value=1.0 , prediction_type="""epsilon""" , variance_type="""learned_range""" , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def _snake_case ( self :Optional[int] ) -> Optional[Any]: """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = TaEncoderModel.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-random-t5""" ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = UNetaDConditionModel( sample_size=32 , layers_per_block=[1, 2] , block_out_channels=[32, 64] , down_block_types=[ """ResnetDownsampleBlock2D""", """SimpleCrossAttnDownBlock2D""", ] , mid_block_type="""UNetMidBlock2DSimpleCrossAttn""" , up_block_types=["""SimpleCrossAttnUpBlock2D""", """ResnetUpsampleBlock2D"""] , in_channels=6 , out_channels=6 , cross_attention_dim=32 , encoder_hid_dim=32 , attention_head_dim=8 , addition_embed_type="""text""" , addition_embed_type_num_heads=2 , cross_attention_norm="""group_norm""" , resnet_time_scale_shift="""scale_shift""" , act_fn="""gelu""" , class_embed_type="""timestep""" , mid_block_scale_factor=1.4_1_4 , time_embedding_act_fn="""gelu""" , time_embedding_dim=32 , ) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = DDPMScheduler( num_train_timesteps=1000 , beta_schedule="""squaredcos_cap_v2""" , beta_start=0.0_0_0_1 , beta_end=0.0_2 , thresholding=__A , dynamic_thresholding_ratio=0.9_5 , sample_max_value=1.0 , prediction_type="""epsilon""" , variance_type="""learned_range""" , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = DDPMScheduler( num_train_timesteps=1000 , beta_schedule="""squaredcos_cap_v2""" , beta_start=0.0_0_0_1 , beta_end=0.0_2 , ) torch.manual_seed(0 ) SCREAMING_SNAKE_CASE__ = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def _snake_case ( self :List[str] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = self.pipeline_class(**__A ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(__A ) SCREAMING_SNAKE_CASE__ = inputs["""prompt"""] SCREAMING_SNAKE_CASE__ = inputs["""generator"""] SCREAMING_SNAKE_CASE__ = inputs["""num_inference_steps"""] SCREAMING_SNAKE_CASE__ = inputs["""output_type"""] if "image" in inputs: SCREAMING_SNAKE_CASE__ = inputs["""image"""] else: SCREAMING_SNAKE_CASE__ = None if "mask_image" in inputs: SCREAMING_SNAKE_CASE__ = inputs["""mask_image"""] else: SCREAMING_SNAKE_CASE__ = None if "original_image" in inputs: SCREAMING_SNAKE_CASE__ = inputs["""original_image"""] else: SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = pipe.encode_prompt(__A ) # inputs with prompt converted to embeddings SCREAMING_SNAKE_CASE__ = { """prompt_embeds""": prompt_embeds, """negative_prompt_embeds""": negative_prompt_embeds, """generator""": generator, """num_inference_steps""": num_inference_steps, """output_type""": output_type, } if image is not None: SCREAMING_SNAKE_CASE__ = image if mask_image is not None: SCREAMING_SNAKE_CASE__ = mask_image if original_image is not None: SCREAMING_SNAKE_CASE__ = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(__A , __A , __A ) SCREAMING_SNAKE_CASE__ = pipe(**__A )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__A ) SCREAMING_SNAKE_CASE__ = self.pipeline_class.from_pretrained(__A ) pipe_loaded.to(__A ) pipe_loaded.set_progress_bar_config(disable=__A ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(__A , __A ) is None , f'''`{optional_component}` did not stay set to None after loading.''' , ) SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(__A ) SCREAMING_SNAKE_CASE__ = inputs["""generator"""] SCREAMING_SNAKE_CASE__ = inputs["""num_inference_steps"""] SCREAMING_SNAKE_CASE__ = inputs["""output_type"""] # inputs with prompt converted to embeddings SCREAMING_SNAKE_CASE__ = { """prompt_embeds""": prompt_embeds, """negative_prompt_embeds""": negative_prompt_embeds, """generator""": generator, """num_inference_steps""": num_inference_steps, """output_type""": output_type, } if image is not None: SCREAMING_SNAKE_CASE__ = image if mask_image is not None: SCREAMING_SNAKE_CASE__ = mask_image if original_image is not None: SCREAMING_SNAKE_CASE__ = original_image SCREAMING_SNAKE_CASE__ = pipe_loaded(**__A )[0] SCREAMING_SNAKE_CASE__ = np.abs(to_np(__A ) - to_np(__A ) ).max() self.assertLess(__A , 1E-4 ) def _snake_case ( self :Optional[int] ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE__ = self.get_dummy_components() SCREAMING_SNAKE_CASE__ = self.pipeline_class(**__A ) pipe.to(__A ) pipe.set_progress_bar_config(disable=__A ) SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(__A ) SCREAMING_SNAKE_CASE__ = pipe(**__A )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(__A ) SCREAMING_SNAKE_CASE__ = self.pipeline_class.from_pretrained(__A ) pipe_loaded.to(__A ) pipe_loaded.set_progress_bar_config(disable=__A ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests SCREAMING_SNAKE_CASE__ = self.get_dummy_inputs(__A ) SCREAMING_SNAKE_CASE__ = pipe_loaded(**__A )[0] SCREAMING_SNAKE_CASE__ = np.abs(to_np(__A ) - to_np(__A ) ).max() self.assertLess(__A , 1E-4 )
6
"""simple docstring""" import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=3 , lowerCamelCase__=3_2 , lowerCamelCase__=3 , lowerCamelCase__=1_0 , lowerCamelCase__=[8, 1_6, 3_2, 6_4] , lowerCamelCase__=[1, 1, 2, 1] , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__="relu" , lowerCamelCase__=3 , lowerCamelCase__=None , lowerCamelCase__=["stage2", "stage3", "stage4"] , lowerCamelCase__=[2, 3, 4] , lowerCamelCase__=1 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = num_channels _lowerCamelCase = embeddings_size _lowerCamelCase = hidden_sizes _lowerCamelCase = depths _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_act _lowerCamelCase = num_labels _lowerCamelCase = scope _lowerCamelCase = len(lowerCamelCase__ ) _lowerCamelCase = out_features _lowerCamelCase = out_indices _lowerCamelCase = num_groups def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.num_labels ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = BitModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = BitForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = BitBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None _lowerCamelCase = None _lowerCamelCase = BitBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Dict = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () lowercase__ : Any = ( {'feature-extraction': BitModel, 'image-classification': BitForImageClassification} if is_torch_available() else {} ) lowercase__ : Union[str, Any] = False lowercase__ : List[Any] = False lowercase__ : Any = False lowercase__ : List[str] = False lowercase__ : Any = False def snake_case__ ( self ): _lowerCamelCase = BitModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ ) def snake_case__ ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case__ ( self ): return @unittest.skip(reason='''Bit does not output attentions''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''Bit does not use inputs_embeds''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''Bit does not support input and output embeddings''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(config=lowerCamelCase__ ) for name, module in model.named_modules(): if isinstance(lowerCamelCase__ , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) def snake_case__ ( self ): def check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _lowerCamelCase = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase__ ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = ['''preactivation''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: _lowerCamelCase = layer_type _lowerCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) @unittest.skip(reason='''Bit does not use feedforward chunking''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = BitModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def snake_case__ ( self ): _lowerCamelCase = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowerCamelCase__ ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**lowerCamelCase__ ) # verify the logits _lowerCamelCase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([[-0.6_5_2_6, -0.5_2_6_3, -1.4_3_9_8]] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) ) @require_torch class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = (BitBackbone,) if is_torch_available() else () lowercase__ : Tuple = BitConfig lowercase__ : Any = False def snake_case__ ( self ): _lowerCamelCase = BitModelTester(self )
661
0
"""simple docstring""" def _snake_case ( _snake_case : list ) -> list: '''simple docstring''' for i in range(len(_snake_case ) - 1 , 0 , -1 ): _A = False for j in range(_snake_case , 0 , -1 ): if unsorted[j] < unsorted[j - 1]: _A , _A = unsorted[j - 1], unsorted[j] _A = True for j in range(_snake_case ): if unsorted[j] > unsorted[j + 1]: _A , _A = unsorted[j + 1], unsorted[j] _A = True if not swapped: break return unsorted if __name__ == "__main__": import doctest doctest.testmod() a = input('''Enter numbers separated by a comma:\n''').strip() a = [int(item) for item in user_input.split(''',''')] print(F'''{cocktail_shaker_sort(unsorted) = }''')
7
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=2 , lowerCamelCase__=2_4 , lowerCamelCase__=1_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=3_2 , lowerCamelCase__=5 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1_0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=None , lowerCamelCase__=2 , lowerCamelCase__=2 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = patch_size _lowerCamelCase = max_length _lowerCamelCase = num_mel_bins _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = scope _lowerCamelCase = frequency_stride _lowerCamelCase = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _lowerCamelCase = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _lowerCamelCase = (self.max_length - self.patch_size) // self.time_stride + 1 _lowerCamelCase = frequency_out_dimension * time_out_dimension _lowerCamelCase = num_patches + 2 def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, input_values, labels def snake_case__ ( self ): return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = ASTModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = config_and_inputs _lowerCamelCase = {'''input_values''': input_values} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Any = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) lowercase__ : List[str] = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) lowercase__ : int = False lowercase__ : str = False lowercase__ : Union[str, Any] = False lowercase__ : List[str] = False def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def snake_case__ ( self ): _lowerCamelCase = ASTModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , nn.Linear ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''input_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = ASTModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> str: _lowerCamelCase = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) _lowerCamelCase , _lowerCamelCase = torchaudio.load(lowercase_ ) return audio, sampling_rate @require_torch @require_torchaudio class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def snake_case__ ( self ): _lowerCamelCase = self.default_feature_extractor _lowerCamelCase = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(lowerCamelCase__ ) _lowerCamelCase = self.default_feature_extractor _lowerCamelCase , _lowerCamelCase = prepare_audio() _lowerCamelCase = audio.squeeze().numpy() _lowerCamelCase = feature_extractor(lowerCamelCase__ , sampling_rate=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**lowerCamelCase__ ) # verify the logits _lowerCamelCase = torch.Size((1, 5_2_7) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([-0.8_7_6_0, -7.0_0_4_2, -8.6_6_0_2] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) )
661
0
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class SCREAMING_SNAKE_CASE (a__ ): lowerCAmelCase = '''naver-clova-ix/donut-base-finetuned-docvqa''' lowerCAmelCase = ( '''This is a tool that answers a question about an document (pdf). It takes an input named `document` which ''' '''should be the document containing the information, as well as a `question` that is the question about the ''' '''document. It returns a text that contains the answer to the question.''' ) lowerCAmelCase = '''document_qa''' lowerCAmelCase = AutoProcessor lowerCAmelCase = VisionEncoderDecoderModel lowerCAmelCase = ['''image''', '''text'''] lowerCAmelCase = ['''text'''] def __init__( self , *_UpperCAmelCase , **_UpperCAmelCase): '''simple docstring''' if not is_vision_available(): raise ValueError('Pillow must be installed to use the DocumentQuestionAnsweringTool.') super().__init__(*_UpperCAmelCase , **_UpperCAmelCase) def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase , _UpperCAmelCase): '''simple docstring''' __A : Optional[int] = '<s_docvqa><s_question>{user_input}</s_question><s_answer>' __A : List[Any] = task_prompt.replace('{user_input}' , _UpperCAmelCase) __A : Tuple = self.pre_processor.tokenizer( _UpperCAmelCase , add_special_tokens=_UpperCAmelCase , return_tensors='pt').input_ids __A : Any = self.pre_processor(_UpperCAmelCase , return_tensors='pt').pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase): '''simple docstring''' return self.model.generate( inputs['pixel_values'].to(self.device) , decoder_input_ids=inputs['decoder_input_ids'].to(self.device) , max_length=self.model.decoder.config.max_position_embeddings , early_stopping=_UpperCAmelCase , pad_token_id=self.pre_processor.tokenizer.pad_token_id , eos_token_id=self.pre_processor.tokenizer.eos_token_id , use_cache=_UpperCAmelCase , num_beams=1 , bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] , return_dict_in_generate=_UpperCAmelCase , ).sequences def SCREAMING_SNAKE_CASE ( self , _UpperCAmelCase): '''simple docstring''' __A : Optional[Any] = self.pre_processor.batch_decode(_UpperCAmelCase)[0] __A : Optional[Any] = sequence.replace(self.pre_processor.tokenizer.eos_token , '') __A : int = sequence.replace(self.pre_processor.tokenizer.pad_token , '') __A : Any = re.sub(R'<.*?>' , '' , _UpperCAmelCase , count=1).strip() # remove first task start token __A : List[str] = self.pre_processor.tokenajson(_UpperCAmelCase) return sequence["answer"]
8
"""simple docstring""" import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class lowerCamelCase_: '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return None class lowerCamelCase_: '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return None class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = [ # (model_name, model_kwargs) ('bert-base-cased', {}), ('gpt2', {'use_cache': False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase__ , '''tf''' , 1_2 , **lowerCamelCase__ ) @require_torch @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , **lowerCamelCase__ ) @require_torch @slow def snake_case__ ( self ): from transformers import BertModel _lowerCamelCase = ['''[UNK]''', '''[SEP]''', '''[CLS]''', '''[PAD]''', '''[MASK]''', '''some''', '''other''', '''words'''] with NamedTemporaryFile(mode='''w+t''' ) as vocab_file: vocab_file.write('''\n'''.join(lowerCamelCase__ ) ) vocab_file.flush() _lowerCamelCase = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: _lowerCamelCase = BertModel(BertConfig(vocab_size=len(lowerCamelCase__ ) ) ) model.save_pretrained(lowerCamelCase__ ) self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , lowerCamelCase__ ) @require_tf @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCamelCase = self._test_export(lowerCamelCase__ , '''tf''' , 1_2 , **lowerCamelCase__ ) _lowerCamelCase = quantize(Path(lowerCamelCase__ ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) @require_torch @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCamelCase = self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , **lowerCamelCase__ ) _lowerCamelCase = quantize(lowerCamelCase__ ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): try: # Compute path with TemporaryDirectory() as tempdir: _lowerCamelCase = Path(lowerCamelCase__ ).joinpath('''model.onnx''' ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) return path except Exception as e: self.fail(lowerCamelCase__ ) @require_torch @require_tokenizers @slow def snake_case__ ( self ): from transformers import BertModel _lowerCamelCase = BertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) _lowerCamelCase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(lowerCamelCase__ , lowerCamelCase__ , '''pt''' ) @require_tf @require_tokenizers @slow def snake_case__ ( self ): from transformers import TFBertModel _lowerCamelCase = TFBertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) _lowerCamelCase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(lowerCamelCase__ , lowerCamelCase__ , '''tf''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = FeatureExtractionPipeline(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''output_0''', '''output_1'''] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = infer_shapes(lowerCamelCase__ , lowerCamelCase__ ) # Assert all variables are present self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] , lowerCamelCase__ ) self.assertSequenceEqual(variable_names[3:] , lowerCamelCase__ ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] , {0: '''batch''', 1: '''sequence'''} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes['''output_0'''] , {0: '''batch''', 1: '''sequence'''} ) self.assertDictEqual(shapes['''output_1'''] , {0: '''batch'''} ) def snake_case__ ( self ): _lowerCamelCase = ['''input_ids''', '''attention_mask''', '''token_type_ids'''] _lowerCamelCase = {'''input_ids''': [1, 2, 3, 4], '''attention_mask''': [0, 0, 0, 0], '''token_type_ids''': [1, 1, 1, 1]} _lowerCamelCase , _lowerCamelCase = ensure_valid_input(FuncContiguousArgs() , lowerCamelCase__ , lowerCamelCase__ ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(lowerCamelCase__ ) , 3 ) # Should have exactly the same input names self.assertEqual(set(lowerCamelCase__ ) , set(lowerCamelCase__ ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(lowerCamelCase__ , (tokens['''input_ids'''], tokens['''token_type_ids'''], tokens['''attention_mask''']) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) _lowerCamelCase , _lowerCamelCase = ensure_valid_input(FuncNonContiguousArgs() , lowerCamelCase__ , lowerCamelCase__ ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(lowerCamelCase__ ) , 1 ) self.assertEqual(len(lowerCamelCase__ ) , 1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] , tokens['''input_ids'''] ) self.assertEqual(ordered_input_names[0] , '''input_ids''' ) def snake_case__ ( self ): _lowerCamelCase = generate_identified_filename(Path('''/home/something/my_fake_model.onnx''' ) , '''-test''' ) self.assertEqual('''/home/something/my_fake_model-test.onnx''' , generated.as_posix() )
661
0
import os import string import sys SCREAMING_SNAKE_CASE__ = 1 << 8 SCREAMING_SNAKE_CASE__ = { '''tab''': ord('''\t'''), '''newline''': ord('''\r'''), '''esc''': 2_7, '''up''': 6_5 + ARROW_KEY_FLAG, '''down''': 6_6 + ARROW_KEY_FLAG, '''right''': 6_7 + ARROW_KEY_FLAG, '''left''': 6_8 + ARROW_KEY_FLAG, '''mod_int''': 9_1, '''undefined''': sys.maxsize, '''interrupt''': 3, '''insert''': 5_0, '''delete''': 5_1, '''pg_up''': 5_3, '''pg_down''': 5_4, } SCREAMING_SNAKE_CASE__ = KEYMAP['''up'''] SCREAMING_SNAKE_CASE__ = KEYMAP['''left'''] if sys.platform == "win32": SCREAMING_SNAKE_CASE__ = [] SCREAMING_SNAKE_CASE__ = { b'''\xe0H''': KEYMAP['''up'''] - ARROW_KEY_FLAG, b'''\x00H''': KEYMAP['''up'''] - ARROW_KEY_FLAG, b'''\xe0P''': KEYMAP['''down'''] - ARROW_KEY_FLAG, b'''\x00P''': KEYMAP['''down'''] - ARROW_KEY_FLAG, b'''\xe0M''': KEYMAP['''right'''] - ARROW_KEY_FLAG, b'''\x00M''': KEYMAP['''right'''] - ARROW_KEY_FLAG, b'''\xe0K''': KEYMAP['''left'''] - ARROW_KEY_FLAG, b'''\x00K''': KEYMAP['''left'''] - ARROW_KEY_FLAG, } for i in range(1_0): SCREAMING_SNAKE_CASE__ = ord(str(i)) def A ( ) -> Union[str, Any]: if os.name == "nt": import msvcrt A__ = 'mbcs' # Flush the keyboard buffer while msvcrt.kbhit(): msvcrt.getch() if len(__UpperCamelCase ) == 0: # Read the keystroke A__ = msvcrt.getch() # If it is a prefix char, get second part if ch in (b"\x00", b"\xe0"): A__ = ch + msvcrt.getch() # Translate actual Win chars to bullet char types try: A__ = chr(WIN_KEYMAP[cha] ) WIN_CH_BUFFER.append(chr(KEYMAP['mod_int'] ) ) WIN_CH_BUFFER.append(__UpperCamelCase ) if ord(__UpperCamelCase ) in ( KEYMAP["insert"] - 1 << 9, KEYMAP["delete"] - 1 << 9, KEYMAP["pg_up"] - 1 << 9, KEYMAP["pg_down"] - 1 << 9, ): WIN_CH_BUFFER.append(chr(126 ) ) A__ = chr(KEYMAP['esc'] ) except KeyError: A__ = cha[1] else: A__ = ch.decode(__UpperCamelCase ) else: A__ = WIN_CH_BUFFER.pop(0 ) elif os.name == "posix": import termios import tty A__ = sys.stdin.fileno() A__ = termios.tcgetattr(__UpperCamelCase ) try: tty.setraw(__UpperCamelCase ) A__ = sys.stdin.read(1 ) finally: termios.tcsetattr(__UpperCamelCase , termios.TCSADRAIN , __UpperCamelCase ) return ch def A ( ) -> Dict: A__ = get_raw_chars() if ord(__UpperCamelCase ) in [KEYMAP["interrupt"], KEYMAP["newline"]]: return char elif ord(__UpperCamelCase ) == KEYMAP["esc"]: A__ = get_raw_chars() if ord(__UpperCamelCase ) == KEYMAP["mod_int"]: A__ = get_raw_chars() if ord(__UpperCamelCase ) >= KEYMAP["arrow_begin"] - ARROW_KEY_FLAG and ord(__UpperCamelCase ) <= KEYMAP["arrow_end"] - ARROW_KEY_FLAG: return chr(ord(__UpperCamelCase ) + ARROW_KEY_FLAG ) else: return KEYMAP["undefined"] else: return get_raw_chars() else: if char in string.printable: return char else: return KEYMAP["undefined"]
9
"""simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = {str(digit): digit**5 for digit in range(1_0)} def lowerCAmelCase_( lowercase_ : int ) -> int: return sum(DIGITS_FIFTH_POWER[digit] for digit in str(lowercase_ ) ) def lowerCAmelCase_( ) -> int: return sum( number for number in range(10_00 , 1_00_00_00 ) if number == digits_fifth_powers_sum(lowercase_ ) ) if __name__ == "__main__": print(solution())
661
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCAmelCase = { "configuration_rembert": ["REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "RemBertConfig", "RemBertOnnxConfig"] } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = ["RemBertTokenizer"] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = ["RemBertTokenizerFast"] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = [ "REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "RemBertForCausalLM", "RemBertForMaskedLM", "RemBertForMultipleChoice", "RemBertForQuestionAnswering", "RemBertForSequenceClassification", "RemBertForTokenClassification", "RemBertLayer", "RemBertModel", "RemBertPreTrainedModel", "load_tf_weights_in_rembert", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase = [ "TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST", "TFRemBertForCausalLM", "TFRemBertForMaskedLM", "TFRemBertForMultipleChoice", "TFRemBertForQuestionAnswering", "TFRemBertForSequenceClassification", "TFRemBertForTokenClassification", "TFRemBertLayer", "TFRemBertModel", "TFRemBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_rembert import REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RemBertConfig, RemBertOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert import RemBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert_fast import RemBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rembert import ( REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, RemBertForCausalLM, RemBertForMaskedLM, RemBertForMultipleChoice, RemBertForQuestionAnswering, RemBertForSequenceClassification, RemBertForTokenClassification, RemBertLayer, RemBertModel, RemBertPreTrainedModel, load_tf_weights_in_rembert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rembert import ( TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFRemBertForCausalLM, TFRemBertForMaskedLM, TFRemBertForMultipleChoice, TFRemBertForQuestionAnswering, TFRemBertForSequenceClassification, TFRemBertForTokenClassification, TFRemBertLayer, TFRemBertModel, TFRemBertPreTrainedModel, ) else: import sys _lowerCAmelCase = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
10
"""simple docstring""" from __future__ import annotations import os from collections.abc import Mapping __SCREAMING_SNAKE_CASE : str = tuple[int, int] class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = vertices _lowerCamelCase = { (min(lowerCamelCase__ ), max(lowerCamelCase__ )): weight for edge, weight in edges.items() } def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): self.vertices.add(edge[0] ) self.vertices.add(edge[1] ) _lowerCamelCase = weight def snake_case__ ( self ): _lowerCamelCase = Graph({min(self.vertices )} , {} ) _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 while len(subgraph.vertices ) < len(self.vertices ): _lowerCamelCase = max(self.edges.values() ) + 1 for edge, weight in self.edges.items(): if (edge[0] in subgraph.vertices) ^ (edge[1] in subgraph.vertices): if weight < min_weight: _lowerCamelCase = edge _lowerCamelCase = weight subgraph.add_edge(lowerCamelCase__ , lowerCamelCase__ ) return subgraph def lowerCAmelCase_( lowercase_ : str = "p107_network.txt" ) -> int: _lowerCamelCase = os.path.abspath(os.path.dirname(lowercase_ ) ) _lowerCamelCase = os.path.join(lowercase_ , lowercase_ ) _lowerCamelCase = {} _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 with open(lowercase_ ) as f: _lowerCamelCase = f.read().strip().split('''\n''' ) _lowerCamelCase = [line.split(''',''' ) for line in data] for edgea in range(1 , len(lowercase_ ) ): for edgea in range(lowercase_ ): if adjaceny_matrix[edgea][edgea] != "-": _lowerCamelCase = int(adjaceny_matrix[edgea][edgea] ) _lowerCamelCase = Graph(set(range(len(lowercase_ ) ) ) , lowercase_ ) _lowerCamelCase = graph.prims_algorithm() _lowerCamelCase = sum(graph.edges.values() ) _lowerCamelCase = sum(subgraph.edges.values() ) return initial_total - optimal_total if __name__ == "__main__": print(F"""{solution() = }""")
661
0
'''simple docstring''' import argparse import collections import torch from flax import traverse_util from tax import checkpoints from transformers import TaConfig, TaEncoderModel, TaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase (__A , __A , __A , __A="attention"): """simple docstring""" _a = params[F'''{prefix}/layers_{i}/{layer_name}/key/kernel'''] _a = params[F'''{prefix}/layers_{i}/{layer_name}/out/kernel'''] _a = params[F'''{prefix}/layers_{i}/{layer_name}/query/kernel'''] _a = params[F'''{prefix}/layers_{i}/{layer_name}/value/kernel'''] return k, o, q, v def lowerCAmelCase (__A , __A , __A , __A=False): """simple docstring""" if split_mlp_wi: _a = params[F'''{prefix}/layers_{i}/mlp/wi_0/kernel'''] _a = params[F'''{prefix}/layers_{i}/mlp/wi_1/kernel'''] _a = (wi_a, wi_a) else: _a = params[F'''{prefix}/layers_{i}/mlp/wi/kernel'''] _a = params[F'''{prefix}/layers_{i}/mlp/wo/kernel'''] return wi, wo def lowerCAmelCase (__A , __A , __A , __A): """simple docstring""" return params[F'''{prefix}/layers_{i}/{layer_name}/scale'''] def lowerCAmelCase (__A , *, __A , __A): """simple docstring""" _a = traverse_util.flatten_dict(variables['''target''']) _a = {'''/'''.join(__A): v for k, v in old.items()} # v1.1 models have a gated GeLU with wi_0 and wi_1 instead of wi _a = '''encoder/layers_0/mlp/wi_0/kernel''' in old print('''Split MLP:''' , __A) _a = collections.OrderedDict() # Shared embeddings. _a = old['''token_embedder/embedding'''] # Encoder. for i in range(__A): # Block i, layer 0 (Self Attention). _a = tax_layer_norm_lookup(__A , __A , '''encoder''' , '''pre_attention_layer_norm''') _a , _a , _a , _a = tax_attention_lookup(__A , __A , '''encoder''' , '''attention''') _a = layer_norm _a = k.T _a = o.T _a = q.T _a = v.T # Block i, layer 1 (MLP). _a = tax_layer_norm_lookup(__A , __A , '''encoder''' , '''pre_mlp_layer_norm''') _a , _a = tax_mlp_lookup(__A , __A , '''encoder''' , __A) _a = layer_norm if split_mlp_wi: _a = wi[0].T _a = wi[1].T else: _a = wi.T _a = wo.T _a = old[ '''encoder/relpos_bias/rel_embedding''' ].T _a = old['''encoder/encoder_norm/scale'''] if not is_encoder_only: # Decoder. for i in range(__A): # Block i, layer 0 (Self Attention). _a = tax_layer_norm_lookup(__A , __A , '''decoder''' , '''pre_self_attention_layer_norm''') _a , _a , _a , _a = tax_attention_lookup(__A , __A , '''decoder''' , '''self_attention''') _a = layer_norm _a = k.T _a = o.T _a = q.T _a = v.T # Block i, layer 1 (Cross Attention). _a = tax_layer_norm_lookup(__A , __A , '''decoder''' , '''pre_cross_attention_layer_norm''') _a , _a , _a , _a = tax_attention_lookup(__A , __A , '''decoder''' , '''encoder_decoder_attention''') _a = layer_norm _a = k.T _a = o.T _a = q.T _a = v.T # Block i, layer 2 (MLP). _a = tax_layer_norm_lookup(__A , __A , '''decoder''' , '''pre_mlp_layer_norm''') _a , _a = tax_mlp_lookup(__A , __A , '''decoder''' , __A) _a = layer_norm if split_mlp_wi: _a = wi[0].T _a = wi[1].T else: _a = wi.T _a = wo.T _a = old['''decoder/decoder_norm/scale'''] _a = old[ '''decoder/relpos_bias/rel_embedding''' ].T # LM Head (only in v1.1 checkpoints, in v1.0 embeddings are used instead) if "decoder/logits_dense/kernel" in old: _a = old['''decoder/logits_dense/kernel'''].T return new def lowerCAmelCase (__A , __A): """simple docstring""" _a = collections.OrderedDict([(k, torch.from_numpy(v.copy())) for (k, v) in converted_params.items()]) # Add what is missing. if "encoder.embed_tokens.weight" not in state_dict: _a = state_dict['''shared.weight'''] if not is_encoder_only: if "decoder.embed_tokens.weight" not in state_dict: _a = state_dict['''shared.weight'''] if "lm_head.weight" not in state_dict: # For old 1.0 models. print('''Using shared word embeddings as lm_head.''') _a = state_dict['''shared.weight'''] return state_dict def lowerCAmelCase (__A , __A , __A , __A): """simple docstring""" _a = checkpoints.load_tax_checkpoint(__A) _a = convert_tax_to_pytorch(__A , num_layers=config.num_layers , is_encoder_only=__A) _a = make_state_dict(__A , __A) model.load_state_dict(__A , strict=__A) def lowerCAmelCase (__A , __A , __A , __A = False): """simple docstring""" _a = TaConfig.from_json_file(__A) print(F'''Building PyTorch model from configuration: {config}''') # Non-v1.1 checkpoints could also use T5Model, but this works for all. # The v1.0 checkpoints will simply have an LM head that is the word embeddings. if is_encoder_only: _a = TaEncoderModel(__A) else: _a = TaForConditionalGeneration(__A) # Load weights from tf checkpoint load_tax_weights_in_ta(__A , __A , __A , __A) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''') model.save_pretrained(__A) # Verify that we can load the checkpoint. model.from_pretrained(__A) print('''Done''') if __name__ == "__main__": lowercase_ = argparse.ArgumentParser(description="Converts a native T5X checkpoint into a PyTorch checkpoint.") # Required parameters parser.add_argument( "--t5x_checkpoint_path", default=None, type=str, required=True, help="Path to the T5X checkpoint." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The config json file corresponding to the pre-trained T5 model.\nThis specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--is_encoder_only", action="store_true", help="Check if the model is encoder-decoder model", default=False ) lowercase_ = parser.parse_args() convert_tax_checkpoint_to_pytorch( args.tax_checkpoint_path, args.config_file, args.pytorch_dump_path, args.is_encoder_only )
11
"""simple docstring""" from __future__ import annotations from math import ceil, floor, sqrt def lowerCAmelCase_( lowercase_ : int = 2_00_00_00 ) -> int: _lowerCamelCase = [0] _lowerCamelCase = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target _lowerCamelCase = 0 # the area corresponding to the grid that gives the product closest to target _lowerCamelCase = 0 # an estimate of b, using the quadratic formula _lowerCamelCase = 42 # the largest integer less than b_estimate _lowerCamelCase = 42 # the largest integer less than b_estimate _lowerCamelCase = 42 # the triangle number corresponding to b_floor _lowerCamelCase = 42 # the triangle number corresponding to b_ceil _lowerCamelCase = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): _lowerCamelCase = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 _lowerCamelCase = floor(lowercase_ ) _lowerCamelCase = ceil(lowercase_ ) _lowerCamelCase = triangle_numbers[b_floor] _lowerCamelCase = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): _lowerCamelCase = triangle_b_first_guess * triangle_a _lowerCamelCase = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): _lowerCamelCase = triangle_b_second_guess * triangle_a _lowerCamelCase = idx_a * b_ceil return area if __name__ == "__main__": print(F"""{solution() = }""")
661
0
import contextlib import os import sqlitea import pytest from datasets import Dataset, Features, Value from datasets.io.sql import SqlDatasetReader, SqlDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases, require_sqlalchemy def UpperCamelCase ( lowercase_ , lowercase_ ) -> Union[str, Any]: '''simple docstring''' assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @require_sqlalchemy @pytest.mark.parametrize("""keep_in_memory""" , [False, True] ) def UpperCamelCase ( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> List[Any]: '''simple docstring''' lowercase__ : Tuple = tmp_path / """cache""" lowercase__ : Union[str, Any] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): lowercase__ : Optional[int] = SqlDatasetReader( """dataset""" , """sqlite:///""" + sqlite_path , cache_dir=lowercase_ , keep_in_memory=lowercase_ ).read() _check_sql_dataset(lowercase_ , lowercase_ ) @require_sqlalchemy @pytest.mark.parametrize( """features""" , [ None, {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""}, {"""col_1""": """string""", """col_2""": """string""", """col_3""": """string"""}, {"""col_1""": """int32""", """col_2""": """int32""", """col_3""": """int32"""}, {"""col_1""": """float32""", """col_2""": """float32""", """col_3""": """float32"""}, ] , ) def UpperCamelCase ( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> Dict: '''simple docstring''' lowercase__ : Optional[int] = tmp_path / """cache""" lowercase__ : List[Any] = {"""col_1""": """string""", """col_2""": """int64""", """col_3""": """float64"""} lowercase__ : Optional[int] = features.copy() if features else default_expected_features lowercase__ : str = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) lowercase__ : Tuple = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , features=lowercase_ , cache_dir=lowercase_ ).read() _check_sql_dataset(lowercase_ , lowercase_ ) def UpperCamelCase ( lowercase_ ) -> int: '''simple docstring''' with contextlib.closing(sqlitea.connect(lowercase_ ) ) as con: lowercase__ : str = con.cursor() cur.execute("""SELECT * FROM dataset""" ) for row in cur: yield row @require_sqlalchemy def UpperCamelCase ( lowercase_ , lowercase_ , lowercase_ ) -> Optional[int]: '''simple docstring''' lowercase__ : str = tmp_path / """cache""" lowercase__ : Optional[Any] = os.path.join(lowercase_ , """tmp.sql""" ) lowercase__ : Union[str, Any] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=lowercase_ ).read() SqlDatasetWriter(lowercase_ , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=1 ).write() lowercase__ : Optional[Any] = iter_sql_file(lowercase_ ) lowercase__ : int = iter_sql_file(lowercase_ ) for rowa, rowa in zip(lowercase_ , lowercase_ ): assert rowa == rowa @require_sqlalchemy def UpperCamelCase ( lowercase_ , lowercase_ , lowercase_ ) -> str: '''simple docstring''' lowercase__ : List[str] = tmp_path / """cache""" lowercase__ : str = os.path.join(lowercase_ , """tmp.sql""" ) lowercase__ : List[str] = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=lowercase_ ).read() SqlDatasetWriter(lowercase_ , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=2 ).write() lowercase__ : List[Any] = iter_sql_file(lowercase_ ) lowercase__ : Tuple = iter_sql_file(lowercase_ ) for rowa, rowa in zip(lowercase_ , lowercase_ ): assert rowa == rowa @require_sqlalchemy def UpperCamelCase ( lowercase_ , lowercase_ , lowercase_ ) -> Optional[Any]: '''simple docstring''' lowercase__ : Tuple = tmp_path / """cache""" lowercase__ : Union[str, Any] = os.path.join(lowercase_ , """tmp.sql""" ) lowercase__ : Dict = SqlDatasetReader("""dataset""" , """sqlite:///""" + sqlite_path , cache_dir=lowercase_ ).read() with pytest.raises(lowercase_ ): SqlDatasetWriter(lowercase_ , """dataset""" , """sqlite:///""" + output_sqlite_path , num_proc=0 ).write()
12
"""simple docstring""" import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , *lowerCamelCase__ , **lowerCamelCase__ ): warnings.warn( '''The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use DeiTImageProcessor instead.''' , lowerCamelCase__ , ) super().__init__(*lowerCamelCase__ , **lowerCamelCase__ )
661
0
'''simple docstring''' import os import zipfile import requests from get_ci_error_statistics import download_artifact, get_artifacts_links def UpperCAmelCase__ ( UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Union[str, Any]=7 ) -> Dict: __lowerCamelCase : Dict = None if token is not None: __lowerCamelCase : Dict = {'Accept': 'application/vnd.github+json', 'Authorization': F'Bearer {token}'} # The id of a workflow (not of a workflow run) __lowerCamelCase : List[str] = '636036' __lowerCamelCase : Optional[int] = F'https://api.github.com/repos/huggingface/transformers/actions/workflows/{workflow_id}/runs' # On `main` branch + event being `schedule` + not returning PRs + only `num_runs` results url += F'?branch=main&event=schedule&exclude_pull_requests=true&per_page={num_runs}' __lowerCamelCase : List[str] = requests.get(UpperCAmelCase_ , headers=UpperCAmelCase_ ).json() return result["workflow_runs"] def UpperCAmelCase__ ( UpperCAmelCase_ : str ) -> List[str]: __lowerCamelCase : Optional[Any] = get_daily_ci_runs(UpperCAmelCase_ ) __lowerCamelCase : Dict = None for workflow_run in workflow_runs: if workflow_run["status"] == "completed": __lowerCamelCase : List[Any] = workflow_run['id'] break return workflow_run_id def UpperCAmelCase__ ( UpperCAmelCase_ : str , UpperCAmelCase_ : str , UpperCAmelCase_ : Dict ) -> List[str]: __lowerCamelCase : Any = get_last_daily_ci_runs(UpperCAmelCase_ ) if workflow_run_id is not None: __lowerCamelCase : Union[str, Any] = get_artifacts_links(worflow_run_id=UpperCAmelCase_ , token=UpperCAmelCase_ ) for artifact_name in artifact_names: if artifact_name in artifacts_links: __lowerCamelCase : Optional[int] = artifacts_links[artifact_name] download_artifact( artifact_name=UpperCAmelCase_ , artifact_url=UpperCAmelCase_ , output_dir=UpperCAmelCase_ , token=UpperCAmelCase_ ) def UpperCAmelCase__ ( UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Union[str, Any] ) -> str: get_last_daily_ci_artifacts(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) __lowerCamelCase : List[str] = {} for artifact_name in artifact_names: __lowerCamelCase : List[str] = os.path.join(UpperCAmelCase_ , F'{artifact_name}.zip' ) if os.path.isfile(UpperCAmelCase_ ): __lowerCamelCase : Any = {} with zipfile.ZipFile(UpperCAmelCase_ ) as z: for filename in z.namelist(): if not os.path.isdir(UpperCAmelCase_ ): # read the file with z.open(UpperCAmelCase_ ) as f: __lowerCamelCase : str = f.read().decode('UTF-8' ) return results
13
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Optional[Any] = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''vocab_file''': { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json''' ), '''distilbert-base-german-cased''': ( '''https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json''' ), '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json''' ), }, } __SCREAMING_SNAKE_CASE : str = { '''distilbert-base-uncased''': 5_1_2, '''distilbert-base-uncased-distilled-squad''': 5_1_2, '''distilbert-base-cased''': 5_1_2, '''distilbert-base-cased-distilled-squad''': 5_1_2, '''distilbert-base-german-cased''': 5_1_2, '''distilbert-base-multilingual-cased''': 5_1_2, } __SCREAMING_SNAKE_CASE : str = { '''distilbert-base-uncased''': {'''do_lower_case''': True}, '''distilbert-base-uncased-distilled-squad''': {'''do_lower_case''': True}, '''distilbert-base-cased''': {'''do_lower_case''': False}, '''distilbert-base-cased-distilled-squad''': {'''do_lower_case''': False}, '''distilbert-base-german-cased''': {'''do_lower_case''': False}, '''distilbert-base-multilingual-cased''': {'''do_lower_case''': False}, } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[Any] = VOCAB_FILES_NAMES lowercase__ : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP lowercase__ : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : str = PRETRAINED_INIT_CONFIGURATION lowercase__ : int = ['input_ids', 'attention_mask'] lowercase__ : Tuple = DistilBertTokenizer def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__="[UNK]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ): super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) _lowerCamelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , lowerCamelCase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' , lowerCamelCase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , lowerCamelCase__ ) != tokenize_chinese_chars ): _lowerCamelCase = getattr(lowerCamelCase__ , normalizer_state.pop('''type''' ) ) _lowerCamelCase = do_lower_case _lowerCamelCase = strip_accents _lowerCamelCase = tokenize_chinese_chars _lowerCamelCase = normalizer_class(**lowerCamelCase__ ) _lowerCamelCase = do_lower_case def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=None ): _lowerCamelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
661
0
from __future__ import annotations a__ = list[list[int]] # assigning initial values to the grid a__ = [ [3, 0, 6, 5, 0, 8, 4, 0, 0], [5, 2, 0, 0, 0, 0, 0, 0, 0], [0, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] # a grid with no solution a__ = [ [5, 0, 6, 5, 0, 8, 4, 0, 3], [5, 2, 0, 0, 0, 0, 0, 0, 2], [1, 8, 7, 0, 0, 0, 0, 3, 1], [0, 0, 3, 0, 1, 0, 0, 8, 0], [9, 0, 0, 8, 6, 3, 0, 0, 5], [0, 5, 0, 0, 9, 0, 6, 0, 0], [1, 3, 0, 0, 0, 0, 2, 5, 0], [0, 0, 0, 0, 0, 0, 0, 7, 4], [0, 0, 5, 2, 0, 6, 3, 0, 0], ] def __UpperCAmelCase ( __a : Matrix ,__a : int ,__a : int ,__a : int ) -> bool: """simple docstring""" for i in range(9 ): if grid[row][i] == n or grid[i][column] == n: return False for i in range(3 ): for j in range(3 ): if grid[(row - row % 3) + i][(column - column % 3) + j] == n: return False return True def __UpperCAmelCase ( __a : Matrix ) -> tuple[int, int] | None: """simple docstring""" for i in range(9 ): for j in range(9 ): if grid[i][j] == 0: return i, j return None def __UpperCAmelCase ( __a : Matrix ) -> Matrix | None: """simple docstring""" if location := find_empty_location(__a ): _a , _a : Optional[Any] = location else: # If the location is ``None``, then the grid is solved. return grid for digit in range(1 ,10 ): if is_safe(__a ,__a ,__a ,__a ): _a : List[str] = digit if sudoku(__a ) is not None: return grid _a : Optional[int] = 0 return None def __UpperCAmelCase ( __a : Matrix ) -> None: """simple docstring""" for row in grid: for cell in row: print(__a ,end=''' ''' ) print() if __name__ == "__main__": # make a copy of grid so that you can compare with the unmodified grid for example_grid in (initial_grid, no_solution): print('''\nExample grid:\n''' + '''=''' * 20) print_solution(example_grid) print('''\nExample grid solution:''') a__ = sudoku(example_grid) if solution is not None: print_solution(solution) else: print('''Cannot find a solution.''')
14
"""simple docstring""" from math import sqrt import numpy as np from sympy import symbols # Coefficient # Speed of light (m/s) __SCREAMING_SNAKE_CASE : str = 2_9_9_7_9_2_4_5_8 # Symbols __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Union[str, Any] = symbols('''ct x y z''') def lowerCAmelCase_( lowercase_ : float ) -> float: if velocity > c: raise ValueError('''Speed must not exceed light speed 299,792,458 [m/s]!''' ) elif velocity < 1: # Usually the speed should be much higher than 1 (c order of magnitude) raise ValueError('''Speed must be greater than or equal to 1!''' ) return velocity / c def lowerCAmelCase_( lowercase_ : float ) -> float: return 1 / sqrt(1 - beta(lowercase_ ) ** 2 ) def lowerCAmelCase_( lowercase_ : float ) -> np.ndarray: return np.array( [ [gamma(lowercase_ ), -gamma(lowercase_ ) * beta(lowercase_ ), 0, 0], [-gamma(lowercase_ ) * beta(lowercase_ ), gamma(lowercase_ ), 0, 0], [0, 0, 1, 0], [0, 0, 0, 1], ] ) def lowerCAmelCase_( lowercase_ : float , lowercase_ : np.ndarray | None = None ) -> np.ndarray: # Ensure event is not empty if event is None: _lowerCamelCase = np.array([ct, x, y, z] ) # Symbolic four vector else: event[0] *= c # x0 is ct (speed of light * time) return transformation_matrix(lowercase_ ) @ event if __name__ == "__main__": import doctest doctest.testmod() # Example of symbolic vector: __SCREAMING_SNAKE_CASE : List[str] = transform(2_9_9_7_9_2_4_5) print('''Example of four vector: ''') print(F"""ct' = {four_vector[0]}""") print(F"""x' = {four_vector[1]}""") print(F"""y' = {four_vector[2]}""") print(F"""z' = {four_vector[3]}""") # Substitute symbols with numerical values __SCREAMING_SNAKE_CASE : Tuple = {ct: c, x: 1, y: 1, z: 1} __SCREAMING_SNAKE_CASE : int = [four_vector[i].subs(sub_dict) for i in range(4)] print(F"""\n{numerical_vector}""")
661
0
import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class A ( unittest.TestCase ): '''simple docstring''' def lowerCamelCase__ (self : int ) -> Union[str, Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @property def lowerCamelCase__ (self : Dict ) -> List[Any]: """simple docstring""" torch.manual_seed(0 ) lowercase__ = UNetaDModel( sample_size=(32, 64) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("""AttnDownBlock2D""", """DownBlock2D""") , up_block_types=("""UpBlock2D""", """AttnUpBlock2D""") , ) return model @property def lowerCamelCase__ (self : List[Any] ) -> List[str]: """simple docstring""" torch.manual_seed(0 ) lowercase__ = UNetaDConditionModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("""CrossAttnDownBlock2D""", """DownBlock2D""") , up_block_types=("""UpBlock2D""", """CrossAttnUpBlock2D""") , cross_attention_dim=10 , ) return model @property def lowerCamelCase__ (self : Dict ) -> Any: """simple docstring""" torch.manual_seed(0 ) lowercase__ = AutoencoderKL( sample_size=(128, 64) , in_channels=1 , out_channels=1 , latent_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("""DownEncoderBlock2D""", """DownEncoderBlock2D""") , up_block_types=("""UpDecoderBlock2D""", """UpDecoderBlock2D""") , ) lowercase__ = UNetaDModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("""AttnDownBlock2D""", """DownBlock2D""") , up_block_types=("""UpBlock2D""", """AttnUpBlock2D""") , ) return vqvae, unet @slow def lowerCamelCase__ (self : Optional[int] ) -> Any: """simple docstring""" lowercase__ = """cpu""" # ensure determinism for the device-dependent torch.Generator lowercase__ = Mel( x_res=self.dummy_unet.config.sample_size[1] , y_res=self.dummy_unet.config.sample_size[0] , ) lowercase__ = DDPMScheduler() lowercase__ = AudioDiffusionPipeline(vqvae=_UpperCAmelCase , unet=self.dummy_unet , mel=_UpperCAmelCase , scheduler=_UpperCAmelCase ) lowercase__ = pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) lowercase__ = torch.Generator(device=_UpperCAmelCase ).manual_seed(42 ) lowercase__ = pipe(generator=_UpperCAmelCase , steps=4 ) lowercase__ = output.audios[0] lowercase__ = output.images[0] lowercase__ = torch.Generator(device=_UpperCAmelCase ).manual_seed(42 ) lowercase__ = pipe(generator=_UpperCAmelCase , steps=4 , return_dict=_UpperCAmelCase ) lowercase__ = output[0][0] assert audio.shape == (1, (self.dummy_unet.config.sample_size[1] - 1) * mel.hop_length) assert ( image.height == self.dummy_unet.config.sample_size[0] and image.width == self.dummy_unet.config.sample_size[1] ) lowercase__ = np.frombuffer(image.tobytes() , dtype="""uint8""" )[:10] lowercase__ = np.frombuffer(image_from_tuple.tobytes() , dtype="""uint8""" )[:10] lowercase__ = np.array([69, 255, 255, 255, 0, 0, 77, 181, 12, 127] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() == 0 lowercase__ = Mel( x_res=self.dummy_vqvae_and_unet[0].config.sample_size[1] , y_res=self.dummy_vqvae_and_unet[0].config.sample_size[0] , ) lowercase__ = DDIMScheduler() lowercase__ = self.dummy_vqvae_and_unet lowercase__ = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=dummy_vqvae_and_unet[1] , mel=_UpperCAmelCase , scheduler=_UpperCAmelCase ) lowercase__ = pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) np.random.seed(0 ) lowercase__ = np.random.uniform(-1 , 1 , ((dummy_vqvae_and_unet[0].config.sample_size[1] - 1) * mel.hop_length,) ) lowercase__ = torch.Generator(device=_UpperCAmelCase ).manual_seed(42 ) lowercase__ = pipe(raw_audio=_UpperCAmelCase , generator=_UpperCAmelCase , start_step=5 , steps=10 ) lowercase__ = output.images[0] assert ( image.height == self.dummy_vqvae_and_unet[0].config.sample_size[0] and image.width == self.dummy_vqvae_and_unet[0].config.sample_size[1] ) lowercase__ = np.frombuffer(image.tobytes() , dtype="""uint8""" )[:10] lowercase__ = np.array([120, 117, 110, 109, 138, 167, 138, 148, 132, 121] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 lowercase__ = self.dummy_unet_condition lowercase__ = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=_UpperCAmelCase , mel=_UpperCAmelCase , scheduler=_UpperCAmelCase ) lowercase__ = pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) np.random.seed(0 ) lowercase__ = torch.rand((1, 1, 10) ) lowercase__ = pipe(generator=_UpperCAmelCase , encoding=_UpperCAmelCase ) lowercase__ = output.images[0] lowercase__ = np.frombuffer(image.tobytes() , dtype="""uint8""" )[:10] lowercase__ = np.array([107, 103, 120, 127, 142, 122, 113, 122, 97, 111] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 @slow @require_torch_gpu class A ( unittest.TestCase ): '''simple docstring''' def lowerCamelCase__ (self : Dict ) -> List[str]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase__ (self : str ) -> List[str]: """simple docstring""" lowercase__ = torch_device lowercase__ = DiffusionPipeline.from_pretrained("""teticio/audio-diffusion-ddim-256""" ) lowercase__ = pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) lowercase__ = torch.Generator(device=_UpperCAmelCase ).manual_seed(42 ) lowercase__ = pipe(generator=_UpperCAmelCase ) lowercase__ = output.audios[0] lowercase__ = output.images[0] assert audio.shape == (1, (pipe.unet.config.sample_size[1] - 1) * pipe.mel.hop_length) assert image.height == pipe.unet.config.sample_size[0] and image.width == pipe.unet.config.sample_size[1] lowercase__ = np.frombuffer(image.tobytes() , dtype="""uint8""" )[:10] lowercase__ = np.array([151, 167, 154, 144, 122, 134, 121, 105, 70, 26] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0
15
"""simple docstring""" from __future__ import annotations __SCREAMING_SNAKE_CASE : Dict = 1.6_0_2_1e-1_9 # units = C def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float , ) -> tuple[str, float]: if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif conductivity < 0: raise ValueError('''Conductivity cannot be negative''' ) elif electron_conc < 0: raise ValueError('''Electron concentration cannot be negative''' ) elif mobility < 0: raise ValueError('''mobility cannot be negative''' ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
661
0
def __a ( A__ : int = 1000 ): SCREAMING_SNAKE_CASE = 3 SCREAMING_SNAKE_CASE = 0 while a < n: if a % 3 == 0 or a % 5 == 0: result += a elif a % 15 == 0: result -= a a += 1 return result if __name__ == "__main__": print(f'{solution() = }')
16
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): _lowerCamelCase = { '''task_specific_params''': { '''summarization''': {'''length_penalty''': 1.0, '''max_length''': 1_2_8, '''min_length''': 1_2, '''num_beams''': 4}, '''summarization_cnn''': {'''length_penalty''': 2.0, '''max_length''': 1_4_2, '''min_length''': 5_6, '''num_beams''': 4}, '''summarization_xsum''': {'''length_penalty''': 1.0, '''max_length''': 6_2, '''min_length''': 1_1, '''num_beams''': 6}, } } _lowerCamelCase = { '''task_specific_params.summarization.length_penalty''': 1.0, '''task_specific_params.summarization.max_length''': 1_2_8, '''task_specific_params.summarization.min_length''': 1_2, '''task_specific_params.summarization.num_beams''': 4, '''task_specific_params.summarization_cnn.length_penalty''': 2.0, '''task_specific_params.summarization_cnn.max_length''': 1_4_2, '''task_specific_params.summarization_cnn.min_length''': 5_6, '''task_specific_params.summarization_cnn.num_beams''': 4, '''task_specific_params.summarization_xsum.length_penalty''': 1.0, '''task_specific_params.summarization_xsum.max_length''': 6_2, '''task_specific_params.summarization_xsum.min_length''': 1_1, '''task_specific_params.summarization_xsum.num_beams''': 6, } self.assertEqual(flatten_dict(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , x.transpose() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , np.asarray(transpose(lowerCamelCase__ ) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , np.asarray(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.reshape(lowerCamelCase__ , (4, 3) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , np.reshape(lowerCamelCase__ , (1_2, 5) ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , reshape(lowerCamelCase__ , (1_2, 5) ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , reshape(lowerCamelCase__ , (1_2, 5) ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.asarray(reshape(lowerCamelCase__ , (4, 3) ) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , np.asarray(reshape(lowerCamelCase__ , (1_2, 5) ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.squeeze(lowerCamelCase__ ) ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.squeeze(lowerCamelCase__ , axis=2 ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.asarray(squeeze(lowerCamelCase__ ) ) ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.asarray(squeeze(lowerCamelCase__ , axis=2 ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.expand_dims(lowerCamelCase__ , axis=1 ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.asarray(expand_dims(lowerCamelCase__ , axis=1 ) ) ) )
661
0
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('''TEST_SAGEMAKER''' , '''False''' ) ) is not True , reason='''Skipping test because should only be run when releasing minor transformers version''' , ) @pytest.mark.usefixtures('''sm_env''' ) @parameterized_class( [ { '''framework''': '''pytorch''', '''script''': '''run_glue.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 650, '''eval_accuracy''': 0.6, '''eval_loss''': 0.9}, }, { '''framework''': '''tensorflow''', '''script''': '''run_tf.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.g4dn.xlarge''', '''results''': {'''train_runtime''': 600, '''eval_accuracy''': 0.3, '''eval_loss''': 0.9}, }, ] ) class lowerCamelCase_ ( unittest.TestCase ): def lowerCAmelCase_ ( self : Optional[Any] ): if self.framework == "pytorch": subprocess.run( F"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="""utf-8""" , check=__A , ) assert hasattr(self , """env""" ) def lowerCAmelCase_ ( self : List[Any] , __A : List[Any]=1 ): # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=F"""{self.env.base_job_name}-single""" , instance_count=__A , instance_type=self.instance_type , debugger_hook_config=__A , hyperparameters={**self.env.hyperparameters, """model_name_or_path""": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="""py36""" , ) def lowerCAmelCase_ ( self : Dict , __A : List[Any] ): TrainingJobAnalytics(__A ).export_csv(F"""{self.env.test_path}/{job_name}_metrics.csv""" ) def lowerCAmelCase_ ( self : Tuple ): # create estimator __A : Tuple = self.create_estimator() # run training estimator.fit() # result dataframe __A : Optional[Any] = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis __A : List[str] = list(result_metrics_df[result_metrics_df.metric_name == """eval_accuracy"""]["""value"""] ) __A : Union[str, Any] = list(result_metrics_df[result_metrics_df.metric_name == """eval_loss"""]["""value"""] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping __A : List[str] = ( Session().describe_training_job(estimator.latest_training_job.name ).get("""TrainingTimeInSeconds""" , 99_9999 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["""eval_accuracy"""] for t in eval_accuracy ) assert all(t <= self.results["""eval_loss"""] for t in eval_loss ) # dump tests result into json file to share in PR with open(F"""{estimator.latest_training_job.name}.json""" , """w""" ) as outfile: json.dump({"""train_time""": train_runtime, """eval_accuracy""": eval_accuracy, """eval_loss""": eval_loss} , __A )
17
"""simple docstring""" import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } __SCREAMING_SNAKE_CASE : Tuple = { '''vocab_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json'''}, '''merges_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt'''}, '''tokenizer_config_file''': { '''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json''' }, } __SCREAMING_SNAKE_CASE : str = {'''facebook/blenderbot-3B''': 1_2_8} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def lowerCAmelCase_( ) -> str: _lowerCamelCase = ( list(range(ord('''!''' ) , ord('''~''' ) + 1 ) ) + list(range(ord('''¡''' ) , ord('''¬''' ) + 1 ) ) + list(range(ord('''®''' ) , ord('''ÿ''' ) + 1 ) ) ) _lowerCamelCase = bs[:] _lowerCamelCase = 0 for b in range(2**8 ): if b not in bs: bs.append(lowercase_ ) cs.append(2**8 + n ) n += 1 _lowerCamelCase = [chr(lowercase_ ) for n in cs] return dict(zip(lowercase_ , lowercase_ ) ) def lowerCAmelCase_( lowercase_ : str ) -> Dict: _lowerCamelCase = set() _lowerCamelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _lowerCamelCase = char return pairs class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Union[str, Any] = VOCAB_FILES_NAMES lowercase__ : int = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : Tuple = ['input_ids', 'attention_mask'] def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__="replace" , lowerCamelCase__="<s>" , lowerCamelCase__="</s>" , lowerCamelCase__="</s>" , lowerCamelCase__="<s>" , lowerCamelCase__="<unk>" , lowerCamelCase__="<pad>" , lowerCamelCase__="<mask>" , lowerCamelCase__=False , **lowerCamelCase__ , ): _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else bos_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else eos_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else sep_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else cls_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else unk_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else mask_token super().__init__( errors=lowerCamelCase__ , bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , add_prefix_space=lowerCamelCase__ , **lowerCamelCase__ , ) with open(lowerCamelCase__ , encoding='''utf-8''' ) as vocab_handle: _lowerCamelCase = json.load(lowerCamelCase__ ) _lowerCamelCase = {v: k for k, v in self.encoder.items()} _lowerCamelCase = errors # how to handle errors in decoding _lowerCamelCase = bytes_to_unicode() _lowerCamelCase = {v: k for k, v in self.byte_encoder.items()} with open(lowerCamelCase__ , encoding='''utf-8''' ) as merges_handle: _lowerCamelCase = merges_handle.read().split('''\n''' )[1:-1] _lowerCamelCase = [tuple(merge.split() ) for merge in bpe_merges] _lowerCamelCase = dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) _lowerCamelCase = {} _lowerCamelCase = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions _lowerCamelCase = re.compile(R'''\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+''' ) @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def snake_case__ ( self ): return len(self.encoder ) def snake_case__ ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def snake_case__ ( self , lowerCamelCase__ ): if token in self.cache: return self.cache[token] _lowerCamelCase = tuple(lowerCamelCase__ ) _lowerCamelCase = get_pairs(lowerCamelCase__ ) if not pairs: return token while True: _lowerCamelCase = min(lowerCamelCase__ , key=lambda lowerCamelCase__ : self.bpe_ranks.get(lowerCamelCase__ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break _lowerCamelCase , _lowerCamelCase = bigram _lowerCamelCase = [] _lowerCamelCase = 0 while i < len(lowerCamelCase__ ): try: _lowerCamelCase = word.index(lowerCamelCase__ , lowerCamelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _lowerCamelCase = j if word[i] == first and i < len(lowerCamelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _lowerCamelCase = tuple(lowerCamelCase__ ) _lowerCamelCase = new_word if len(lowerCamelCase__ ) == 1: break else: _lowerCamelCase = get_pairs(lowerCamelCase__ ) _lowerCamelCase = ''' '''.join(lowerCamelCase__ ) _lowerCamelCase = word return word def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] for token in re.findall(self.pat , lowerCamelCase__ ): _lowerCamelCase = ''''''.join( self.byte_encoder[b] for b in token.encode('''utf-8''' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCamelCase__ ).split(''' ''' ) ) return bpe_tokens def snake_case__ ( self , lowerCamelCase__ ): return self.encoder.get(lowerCamelCase__ , self.encoder.get(self.unk_token ) ) def snake_case__ ( self , lowerCamelCase__ ): return self.decoder.get(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = ''''''.join(lowerCamelCase__ ) _lowerCamelCase = bytearray([self.byte_decoder[c] for c in text] ).decode('''utf-8''' , errors=self.errors ) return text def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): if not os.path.isdir(lowerCamelCase__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(lowerCamelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCamelCase__ , ensure_ascii=lowerCamelCase__ ) + '''\n''' ) _lowerCamelCase = 0 with open(lowerCamelCase__ , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCamelCase__ : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ''' Please check that the tokenizer is not corrupted!''' ) _lowerCamelCase = token_index writer.write(''' '''.join(lowerCamelCase__ ) + '''\n''' ) index += 1 return vocab_file, merge_file def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase__ )) + [1] return [1] + ([0] * len(lowerCamelCase__ )) + [1, 1] + ([0] * len(lowerCamelCase__ )) + [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=False , **lowerCamelCase__ ): _lowerCamelCase = kwargs.pop('''add_prefix_space''' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCamelCase__ ) > 0 and not text[0].isspace()): _lowerCamelCase = ''' ''' + text return (text, kwargs) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): return token_ids_a + [self.eos_token_id] def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(''' ''' + text ) else: # Generated responses should contain them already. inputs.append(lowerCamelCase__ ) _lowerCamelCase = ''' '''.join(lowerCamelCase__ ) _lowerCamelCase = self.encode(lowerCamelCase__ ) if len(lowerCamelCase__ ) > self.model_max_length: _lowerCamelCase = input_ids[-self.model_max_length :] logger.warning(F"""Trimmed input from conversation as it was longer than {self.model_max_length} tokens.""" ) return input_ids
661
0
'''simple docstring''' from __future__ import annotations import string from itertools import cycle, product from pathlib import Path _SCREAMING_SNAKE_CASE = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) _SCREAMING_SNAKE_CASE = [ord(letter) for letter in string.ascii_lowercase] _SCREAMING_SNAKE_CASE = {ord(char) for char in VALID_CHARS} _SCREAMING_SNAKE_CASE = ["the", "be", "to", "of", "and", "in", "that", "have"] def __a(SCREAMING_SNAKE_CASE_ : list[int] , SCREAMING_SNAKE_CASE_ : tuple[int, ...] ): '''simple docstring''' _lowerCAmelCase = "" _lowerCAmelCase = 42 _lowerCAmelCase = 42 _lowerCAmelCase = 42 for keychar, cipherchar in zip(cycle(SCREAMING_SNAKE_CASE_ ) , SCREAMING_SNAKE_CASE_ ): _lowerCAmelCase = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(SCREAMING_SNAKE_CASE_ ) return decoded def __a(SCREAMING_SNAKE_CASE_ : list[int] ): '''simple docstring''' _lowerCAmelCase = [] for key in product(SCREAMING_SNAKE_CASE_ , repeat=3 ): _lowerCAmelCase = try_key(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if encoded is not None: possibles.append(SCREAMING_SNAKE_CASE_ ) return possibles def __a(SCREAMING_SNAKE_CASE_ : list[str] , SCREAMING_SNAKE_CASE_ : str ): '''simple docstring''' return [possible for possible in possibles if common_word in possible.lower()] def __a(SCREAMING_SNAKE_CASE_ : str = "p059_cipher.txt" ): '''simple docstring''' _lowerCAmelCase = 42 _lowerCAmelCase = 42 _lowerCAmelCase = 42 _lowerCAmelCase = 42 _lowerCAmelCase = Path(SCREAMING_SNAKE_CASE_ ).parent.joinpath(SCREAMING_SNAKE_CASE_ ).read_text(encoding="utf-8" ) _lowerCAmelCase = [int(SCREAMING_SNAKE_CASE_ ) for number in data.strip().split("," )] _lowerCAmelCase = filter_valid_chars(SCREAMING_SNAKE_CASE_ ) for common_word in COMMON_WORDS: _lowerCAmelCase = filter_common_word(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if len(SCREAMING_SNAKE_CASE_ ) == 1: break _lowerCAmelCase = possibles[0] return sum(ord(SCREAMING_SNAKE_CASE_ ) for char in decoded_text ) if __name__ == "__main__": print(f'''{solution() = }''')
18
"""simple docstring""" import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin __SCREAMING_SNAKE_CASE : Dict = get_tests_dir('''fixtures/test_sentencepiece.model''') @require_sentencepiece @require_tokenizers class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Union[str, Any] = XLMRobertaTokenizer lowercase__ : Optional[int] = XLMRobertaTokenizerFast lowercase__ : List[str] = True lowercase__ : Union[str, Any] = True def snake_case__ ( self ): super().setUp() # We have a SentencePiece fixture for testing _lowerCamelCase = XLMRobertaTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def snake_case__ ( self ): _lowerCamelCase = '''<pad>''' _lowerCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<s>''' ) self.assertEqual(vocab_keys[1] , '''<pad>''' ) self.assertEqual(vocab_keys[-1] , '''<mask>''' ) self.assertEqual(len(lowerCamelCase__ ) , 1_0_0_2 ) def snake_case__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 1_0_0_2 ) def snake_case__ ( self ): _lowerCamelCase = XLMRobertaTokenizer(lowerCamelCase__ , keep_accents=lowerCamelCase__ ) _lowerCamelCase = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(lowerCamelCase__ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) , [value + tokenizer.fairseq_offset for value in [2_8_5, 4_6, 1_0, 1_7_0, 3_8_2]] , ) _lowerCamelCase = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) _lowerCamelCase = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ value + tokenizer.fairseq_offset for value in [8, 2_1, 8_4, 5_5, 2_4, 1_9, 7, 2, 6_0_2, 3_4_7, 3_4_7, 3_4_7, 3, 1_2, 6_6, 4_6, 7_2, 8_0, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] , ) _lowerCamelCase = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def snake_case__ ( self ): if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _lowerCamelCase = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-xlm-roberta''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): _lowerCamelCase = self.rust_tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = self.tokenizer_class.from_pretrained(lowerCamelCase__ , **lowerCamelCase__ ) _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) _lowerCamelCase = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(lowerCamelCase__ , lowerCamelCase__ ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(lowerCamelCase__ ) # Save tokenizer rust, legacy_format=True _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ , legacy_format=lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it save with the same files self.assertSequenceEqual(lowerCamelCase__ , lowerCamelCase__ ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) shutil.rmtree(lowerCamelCase__ ) # Save tokenizer rust, legacy_format=False _lowerCamelCase = tempfile.mkdtemp() _lowerCamelCase = tokenizer_r.save_pretrained(lowerCamelCase__ , legacy_format=lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _lowerCamelCase = tokenizer_r.from_pretrained(lowerCamelCase__ ) _lowerCamelCase = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ , lowerCamelCase__ ) ) shutil.rmtree(lowerCamelCase__ ) @cached_property def snake_case__ ( self ): return XLMRobertaTokenizer.from_pretrained('''xlm-roberta-base''' ) def snake_case__ ( self ): with tempfile.NamedTemporaryFile() as f: shutil.copyfile(lowerCamelCase__ , f.name ) _lowerCamelCase = XLMRobertaTokenizer(f.name , keep_accents=lowerCamelCase__ ) _lowerCamelCase = pickle.dumps(lowerCamelCase__ ) pickle.loads(lowerCamelCase__ ) def snake_case__ ( self ): if not self.test_rust_tokenizer: return _lowerCamelCase = self.get_tokenizer() _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = '''I was born in 92000, and this is falsé.''' _lowerCamelCase = tokenizer.tokenize(lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ , add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = self.get_rust_tokenizer() _lowerCamelCase = tokenizer.encode(lowerCamelCase__ ) _lowerCamelCase = rust_tokenizer.encode(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ , lowerCamelCase__ ) @slow def snake_case__ ( self ): _lowerCamelCase = '''Hello World!''' _lowerCamelCase = [0, 3_5_3_7_8, 6_6_6_1, 3_8, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(lowerCamelCase__ , self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def snake_case__ ( self ): _lowerCamelCase = ( '''This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will''' ''' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth''' ) _lowerCamelCase = [ 0, 3_2_9_3, 8_3, 1_0, 4_5_5_2, 4_9_8_9, 7_9_8_6, 6_7_8, 1_0, 5_9_1_5, 1_1_1, 1_7_9_4_5_9, 1_2_4_8_5_0, 4, 6_0_4_4, 2_3_7, 1_2, 6, 5, 6, 4, 6_7_8_0, 7_0_5, 1_5, 1_3_8_8, 4_4, 3_7_8, 1_0_1_1_4, 7_1_1, 1_5_2, 2_0, 6, 5, 2_2_3_7_6, 6_4_2, 1_2_2_1, 1_5_1_9_0, 3_4_1_5_3, 4_5_0, 5_6_0_8, 9_5_9, 1_1_1_9, 5_7_7_0_2, 1_3_6, 1_8_6, 4_7, 1_0_9_8, 2_9_3_6_7, 4_7, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6_0_4_4, 2_3_7, 6_2_8_4, 5_0_9_0_1, 5_2_8, 3_1, 9_0, 3_4, 9_2_7, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(lowerCamelCase__ , self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def snake_case__ ( self ): # fmt: off _lowerCamelCase = {'''input_ids''': [[0, 1_1_0_6_2, 8_2_7_7_2, 7, 1_5, 8_2_7_7_2, 5_3_8, 5_1_5_2_9, 2_3_7, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 2_1_5_1_7_5, 1_3_1_4, 1_3_6, 1_7_1_9_8, 1_2_9_0, 2_0_6, 9, 5_6_3_5_9, 4_2, 1_2_2_0_0_9, 9, 1_6_4_6_6, 1_6, 8_7_3_4_4, 4_5_3_7, 9, 4_7_1_7, 7_8_3_8_1, 6, 1_5_9_9_5_8, 7, 1_5, 2_4_4_8_0, 6_1_8, 4, 5_2_7, 2_2_6_9_3, 5_4_2_8, 4, 2_7_7_7, 2_4_4_8_0, 9_8_7_4, 4, 4_3_5_2_3, 5_9_4, 4, 8_0_3, 1_8_3_9_2, 3_3_1_8_9, 1_8, 4, 4_3_5_2_3, 2_4_4_4_7, 1_2_3_9_9, 1_0_0, 2_4_9_5_5, 8_3_6_5_8, 9_6_2_6, 1_4_4_0_5_7, 1_5, 8_3_9, 2_2_3_3_5, 1_6, 1_3_6, 2_4_9_5_5, 8_3_6_5_8, 8_3_4_7_9, 1_5, 3_9_1_0_2, 7_2_4, 1_6, 6_7_8, 6_4_5, 2_7_8_9, 1_3_2_8, 4_5_8_9, 4_2, 1_2_2_0_0_9, 1_1_5_7_7_4, 2_3, 8_0_5, 1_3_2_8, 4_6_8_7_6, 7, 1_3_6, 5_3_8_9_4, 1_9_4_0, 4_2_2_2_7, 4_1_1_5_9, 1_7_7_2_1, 8_2_3, 4_2_5, 4, 2_7_5_1_2, 9_8_7_2_2, 2_0_6, 1_3_6, 5_5_3_1, 4_9_7_0, 9_1_9, 1_7_3_3_6, 5, 2], [0, 2_0_0_8_0, 6_1_8, 8_3, 8_2_7_7_5, 4_7, 4_7_9, 9, 1_5_1_7, 7_3, 5_3_8_9_4, 3_3_3, 8_0_5_8_1, 1_1_0_1_1_7, 1_8_8_1_1, 5_2_5_6, 1_2_9_5, 5_1, 1_5_2_5_2_6, 2_9_7, 7_9_8_6, 3_9_0, 1_2_4_4_1_6, 5_3_8, 3_5_4_3_1, 2_1_4, 9_8, 1_5_0_4_4, 2_5_7_3_7, 1_3_6, 7_1_0_8, 4_3_7_0_1, 2_3, 7_5_6, 1_3_5_3_5_5, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 5_8_1, 6_3_7_7_3, 1_1_9_4_5_5, 6, 1_4_7_7_9_7, 8_8_2_0_3, 7, 6_4_5, 7_0, 2_1, 3_2_8_5, 1_0_2_6_9, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ , model_name='''xlm-roberta-base''' , revision='''d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3''' , )
661
0
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING _a = logging.get_logger(__name__) _a = { """Salesforce/instruct-blip-flan-t5""": """https://huggingface.co/Salesforce/instruct-blip-flan-t5/resolve/main/config.json""", } class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'instructblip_vision_model' def __init__( self , __a=14_08 , __a=61_44 , __a=39 , __a=16 , __a=2_24 , __a=14 , __a="gelu" , __a=1e-6 , __a=0.0 , __a=1e-10 , __a=True , **__a , ) -> int: '''simple docstring''' super().__init__(**__a) _UpperCamelCase = hidden_size _UpperCamelCase = intermediate_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = patch_size _UpperCamelCase = image_size _UpperCamelCase = initializer_range _UpperCamelCase = attention_dropout _UpperCamelCase = layer_norm_eps _UpperCamelCase = hidden_act _UpperCamelCase = qkv_bias @classmethod def UpperCAmelCase ( cls , __a , **__a) -> "PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(__a) _UpperCamelCase , _UpperCamelCase = cls.get_config_dict(__a , **__a) # get the vision config dict if we are loading from InstructBlipConfig if config_dict.get('''model_type''') == "instructblip": _UpperCamelCase = config_dict['''vision_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''') and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''') return cls.from_dict(__a , **__a) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'instructblip_qformer' def __init__( self , __a=3_05_22 , __a=7_68 , __a=12 , __a=12 , __a=30_72 , __a="gelu" , __a=0.1 , __a=0.1 , __a=5_12 , __a=0.02 , __a=1e-12 , __a=0 , __a="absolute" , __a=2 , __a=14_08 , **__a , ) -> Dict: '''simple docstring''' super().__init__(pad_token_id=__a , **__a) _UpperCamelCase = vocab_size _UpperCamelCase = hidden_size _UpperCamelCase = num_hidden_layers _UpperCamelCase = num_attention_heads _UpperCamelCase = hidden_act _UpperCamelCase = intermediate_size _UpperCamelCase = hidden_dropout_prob _UpperCamelCase = attention_probs_dropout_prob _UpperCamelCase = max_position_embeddings _UpperCamelCase = initializer_range _UpperCamelCase = layer_norm_eps _UpperCamelCase = position_embedding_type _UpperCamelCase = cross_attention_frequency _UpperCamelCase = encoder_hidden_size @classmethod def UpperCAmelCase ( cls , __a , **__a) -> "PretrainedConfig": '''simple docstring''' cls._set_token_in_kwargs(__a) _UpperCamelCase , _UpperCamelCase = cls.get_config_dict(__a , **__a) # get the qformer config dict if we are loading from InstructBlipConfig if config_dict.get('''model_type''') == "instructblip": _UpperCamelCase = config_dict['''qformer_config'''] if "model_type" in config_dict and hasattr(cls , '''model_type''') and config_dict["model_type"] != cls.model_type: logger.warning( F'''You are using a model of type {config_dict["model_type"]} to instantiate a model of type ''' F'''{cls.model_type}. This is not supported for all configurations of models and can yield errors.''') return cls.from_dict(__a , **__a) class _UpperCAmelCase( lowerCamelCase ): lowercase__ = 'instructblip' lowercase__ = True def __init__( self , __a=None , __a=None , __a=None , __a=32 , **__a) -> List[str]: '''simple docstring''' super().__init__(**__a) if vision_config is None: _UpperCamelCase = {} logger.info('''vision_config is None. initializing the InstructBlipVisionConfig with default values.''') if qformer_config is None: _UpperCamelCase = {} logger.info('''qformer_config is None. Initializing the InstructBlipQFormerConfig with default values.''') if text_config is None: _UpperCamelCase = {} logger.info('''text_config is None. Initializing the text config with default values (`OPTConfig`).''') _UpperCamelCase = InstructBlipVisionConfig(**__a) _UpperCamelCase = InstructBlipQFormerConfig(**__a) _UpperCamelCase = text_config['''model_type'''] if '''model_type''' in text_config else '''opt''' _UpperCamelCase = CONFIG_MAPPING[text_model_type](**__a) _UpperCamelCase = self.text_config.tie_word_embeddings _UpperCamelCase = self.text_config.is_encoder_decoder _UpperCamelCase = num_query_tokens _UpperCamelCase = self.vision_config.hidden_size _UpperCamelCase = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES _UpperCamelCase = 1.0 _UpperCamelCase = 0.02 @classmethod def UpperCAmelCase ( cls , __a , __a , __a , **__a , ) -> int: '''simple docstring''' return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **__a , ) def UpperCAmelCase ( self) -> Union[str, Any]: '''simple docstring''' _UpperCamelCase = copy.deepcopy(self.__dict__) _UpperCamelCase = self.vision_config.to_dict() _UpperCamelCase = self.qformer_config.to_dict() _UpperCamelCase = self.text_config.to_dict() _UpperCamelCase = self.__class__.model_type return output
19
"""simple docstring""" def lowerCAmelCase_( lowercase_ : int = 10 ) -> str: if not isinstance(lowercase_ , lowercase_ ) or n < 0: raise ValueError('''Invalid input''' ) _lowerCamelCase = 10**n _lowerCamelCase = 2_84_33 * (pow(2 , 7_83_04_57 , lowercase_ )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(F"""{solution(1_0) = }""")
661
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowerCAmelCase: Union[str, Any] = { 'configuration_roformer': ['ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoFormerConfig', 'RoFormerOnnxConfig'], 'tokenization_roformer': ['RoFormerTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase: Any = ['RoFormerTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase: Tuple = [ 'ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'RoFormerForCausalLM', 'RoFormerForMaskedLM', 'RoFormerForMultipleChoice', 'RoFormerForQuestionAnswering', 'RoFormerForSequenceClassification', 'RoFormerForTokenClassification', 'RoFormerLayer', 'RoFormerModel', 'RoFormerPreTrainedModel', 'load_tf_weights_in_roformer', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase: Optional[int] = [ 'TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFRoFormerForCausalLM', 'TFRoFormerForMaskedLM', 'TFRoFormerForMultipleChoice', 'TFRoFormerForQuestionAnswering', 'TFRoFormerForSequenceClassification', 'TFRoFormerForTokenClassification', 'TFRoFormerLayer', 'TFRoFormerModel', 'TFRoFormerPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowerCAmelCase: int = [ 'FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'FlaxRoFormerForMaskedLM', 'FlaxRoFormerForMultipleChoice', 'FlaxRoFormerForQuestionAnswering', 'FlaxRoFormerForSequenceClassification', 'FlaxRoFormerForTokenClassification', 'FlaxRoFormerModel', 'FlaxRoFormerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_roformer import ROFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, RoFormerConfig, RoFormerOnnxConfig from .tokenization_roformer import RoFormerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roformer_fast import RoFormerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roformer import ( ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, RoFormerForCausalLM, RoFormerForMaskedLM, RoFormerForMultipleChoice, RoFormerForQuestionAnswering, RoFormerForSequenceClassification, RoFormerForTokenClassification, RoFormerLayer, RoFormerModel, RoFormerPreTrainedModel, load_tf_weights_in_roformer, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roformer import ( TF_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFRoFormerForCausalLM, TFRoFormerForMaskedLM, TFRoFormerForMultipleChoice, TFRoFormerForQuestionAnswering, TFRoFormerForSequenceClassification, TFRoFormerForTokenClassification, TFRoFormerLayer, TFRoFormerModel, TFRoFormerPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roformer import ( FLAX_ROFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, FlaxRoFormerPreTrainedModel, ) else: import sys _lowerCAmelCase: Union[str, Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
20
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __SCREAMING_SNAKE_CASE : Optional[int] = { '''configuration_informer''': [ '''INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : int = [ '''INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InformerForPrediction''', '''InformerModel''', '''InformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
661
0
import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch UpperCAmelCase_ : Dict = random.Random() def lowerCAmelCase_ ( lowerCamelCase , lowerCamelCase=1.0 , lowerCamelCase=None , lowerCamelCase=None ): if rng is None: __magic_name__ : Dict =global_rng __magic_name__ : int =[] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __A ( unittest.TestCase ): def __init__( self :Tuple , __snake_case :List[str] , __snake_case :Union[str, Any]=7 , __snake_case :int=4_00 , __snake_case :Dict=20_00 , __snake_case :Optional[int]=10 , __snake_case :int=1_60 , __snake_case :Union[str, Any]=8 , __snake_case :Any=0.0 , __snake_case :str=40_00 , __snake_case :Dict=False , __snake_case :Optional[Any]=True , ): '''simple docstring''' __magic_name__ : Tuple =parent __magic_name__ : Optional[Any] =batch_size __magic_name__ : Optional[int] =min_seq_length __magic_name__ : Optional[int] =max_seq_length __magic_name__ : int =(self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) __magic_name__ : int =padding_value __magic_name__ : Any =sampling_rate __magic_name__ : Optional[Any] =return_attention_mask __magic_name__ : List[str] =do_normalize __magic_name__ : str =feature_size __magic_name__ : Optional[int] =chunk_length __magic_name__ : Tuple =hop_length def A__ ( self :Any ): '''simple docstring''' return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def A__ ( self :str , __snake_case :Dict=False , __snake_case :Any=False ): '''simple docstring''' def _flatten(__snake_case :List[str] ): return list(itertools.chain(*__snake_case ) ) if equal_length: __magic_name__ : Tuple =[floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size __magic_name__ : List[Any] =[ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: __magic_name__ : Optional[Any] =[np.asarray(__snake_case ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __A ( UpperCamelCase__ , unittest.TestCase ): UpperCamelCase = WhisperFeatureExtractor if is_speech_available() else None def A__ ( self :int ): '''simple docstring''' __magic_name__ : Optional[Any] =WhisperFeatureExtractionTester(self ) def A__ ( self :Dict ): '''simple docstring''' __magic_name__ : Optional[Any] =self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __magic_name__ : int =feat_extract_first.save_pretrained(__snake_case )[0] check_json_file_has_correct_format(__snake_case ) __magic_name__ : Dict =self.feature_extraction_class.from_pretrained(__snake_case ) __magic_name__ : str =feat_extract_first.to_dict() __magic_name__ : Union[str, Any] =feat_extract_second.to_dict() __magic_name__ : int =feat_extract_first.mel_filters __magic_name__ : List[Any] =feat_extract_second.mel_filters self.assertTrue(np.allclose(__snake_case , __snake_case ) ) self.assertEqual(__snake_case , __snake_case ) def A__ ( self :str ): '''simple docstring''' __magic_name__ : Optional[Any] =self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __magic_name__ : str =os.path.join(__snake_case , """feat_extract.json""" ) feat_extract_first.to_json_file(__snake_case ) __magic_name__ : Tuple =self.feature_extraction_class.from_json_file(__snake_case ) __magic_name__ : str =feat_extract_first.to_dict() __magic_name__ : Union[str, Any] =feat_extract_second.to_dict() __magic_name__ : Dict =feat_extract_first.mel_filters __magic_name__ : List[str] =feat_extract_second.mel_filters self.assertTrue(np.allclose(__snake_case , __snake_case ) ) self.assertEqual(__snake_case , __snake_case ) def A__ ( self :List[str] ): '''simple docstring''' __magic_name__ : Any =self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 __magic_name__ : Dict =[floats_list((1, x) )[0] for x in range(8_00 , 14_00 , 2_00 )] __magic_name__ : List[str] =[np.asarray(__snake_case ) for speech_input in speech_inputs] # Test feature size __magic_name__ : Dict =feature_extractor(__snake_case , padding="""max_length""" , return_tensors="""np""" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input __magic_name__ : Optional[int] =feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_features __magic_name__ : int =feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_features self.assertTrue(np.allclose(__snake_case , __snake_case , atol=1E-3 ) ) # Test batched __magic_name__ : Union[str, Any] =feature_extractor(__snake_case , return_tensors="""np""" ).input_features __magic_name__ : int =feature_extractor(__snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(__snake_case , __snake_case ): self.assertTrue(np.allclose(__snake_case , __snake_case , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. __magic_name__ : Optional[Any] =[floats_list((1, x) )[0] for x in (8_00, 8_00, 8_00)] __magic_name__ : Tuple =np.asarray(__snake_case ) __magic_name__ : List[str] =feature_extractor(__snake_case , return_tensors="""np""" ).input_features __magic_name__ : Dict =feature_extractor(__snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(__snake_case , __snake_case ): self.assertTrue(np.allclose(__snake_case , __snake_case , atol=1E-3 ) ) # Test truncation required __magic_name__ : Any =[floats_list((1, x) )[0] for x in range(2_00 , (feature_extractor.n_samples + 5_00) , 2_00 )] __magic_name__ : int =[np.asarray(__snake_case ) for speech_input in speech_inputs] __magic_name__ : List[Any] =[x[: feature_extractor.n_samples] for x in speech_inputs] __magic_name__ : str =[np.asarray(__snake_case ) for speech_input in speech_inputs_truncated] __magic_name__ : Optional[Any] =feature_extractor(__snake_case , return_tensors="""np""" ).input_features __magic_name__ : int =feature_extractor(__snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(__snake_case , __snake_case ): self.assertTrue(np.allclose(__snake_case , __snake_case , atol=1E-3 ) ) def A__ ( self :Any ): '''simple docstring''' import torch __magic_name__ : List[str] =self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __magic_name__ : str =np.random.rand(1_00 , 32 ).astype(np.floataa ) __magic_name__ : int =np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: __magic_name__ : Dict =feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) __magic_name__ : List[str] =feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def A__ ( self :Optional[int] , __snake_case :Tuple ): '''simple docstring''' __magic_name__ : str =load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech __magic_name__ : Any =ds.sort("""id""" ).select(range(__snake_case ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def A__ ( self :Any ): '''simple docstring''' __magic_name__ : int =torch.tensor( [ 0.1193, -0.0946, -0.1098, -0.0196, 0.0225, -0.0690, -0.1736, 0.0951, 0.0971, -0.0817, -0.0702, 0.0162, 0.0260, 0.0017, -0.0192, -0.1678, 0.0709, -0.1867, -0.0655, -0.0274, -0.0234, -0.1884, -0.0516, -0.0554, -0.0274, -0.1425, -0.1423, 0.0837, 0.0377, -0.0854 ] ) # fmt: on __magic_name__ : List[Any] =self._load_datasamples(1 ) __magic_name__ : int =WhisperFeatureExtractor() __magic_name__ : Optional[Any] =feature_extractor(__snake_case , return_tensors="""pt""" ).input_features self.assertEqual(input_features.shape , (1, 80, 30_00) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , __snake_case , atol=1E-4 ) ) def A__ ( self :Optional[Any] ): '''simple docstring''' __magic_name__ : Union[str, Any] =self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) __magic_name__ : Tuple =self._load_datasamples(1 )[0] __magic_name__ : List[str] =((audio - audio.min()) / (audio.max() - audio.min())) * 6_55_35 # Rescale to [0, 65535] to show issue __magic_name__ : List[str] =feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=__snake_case )[0] self.assertTrue(np.all(np.mean(__snake_case ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(__snake_case ) - 1 ) < 1E-3 ) )
21
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> float: def get_matched_characters(lowercase_ : str , lowercase_ : str ) -> str: _lowerCamelCase = [] _lowerCamelCase = min(len(_stra ) , len(_stra ) ) // 2 for i, l in enumerate(_stra ): _lowerCamelCase = int(max(0 , i - limit ) ) _lowerCamelCase = int(min(i + limit + 1 , len(_stra ) ) ) if l in _stra[left:right]: matched.append(lowercase_ ) _lowerCamelCase = F"""{_stra[0:_stra.index(lowercase_ )]} {_stra[_stra.index(lowercase_ ) + 1:]}""" return "".join(lowercase_ ) # matching characters _lowerCamelCase = get_matched_characters(lowercase_ , lowercase_ ) _lowerCamelCase = get_matched_characters(lowercase_ , lowercase_ ) _lowerCamelCase = len(lowercase_ ) # transposition _lowerCamelCase = ( len([(ca, ca) for ca, ca in zip(lowercase_ , lowercase_ ) if ca != ca] ) // 2 ) if not match_count: _lowerCamelCase = 0.0 else: _lowerCamelCase = ( 1 / 3 * ( match_count / len(lowercase_ ) + match_count / len(lowercase_ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters _lowerCamelCase = 0 for ca, ca in zip(stra[:4] , stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('''hello''', '''world'''))
661
0
'''simple docstring''' def snake_case_ (UpperCamelCase : int ): '''simple docstring''' if num <= 0: raise ValueError('''Input must be a positive integer''' ) _a = [True] * (num + 1) _a = 2 while p * p <= num: if primes[p]: for i in range(p * p , num + 1 , UpperCamelCase ): _a = False p += 1 return [prime for prime in range(2 , num + 1 ) if primes[prime]] if __name__ == "__main__": import doctest doctest.testmod() _snake_case : Optional[Any] = int(input('Enter a positive integer: ').strip()) print(prime_sieve_eratosthenes(user_num))
22
"""simple docstring""" from typing import Optional import numpy as np import torch from torch import nn from transformers import GPTaConfig, GPTaLMHeadModel from transformers.modeling_utils import ModuleUtilsMixin from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class lowerCamelCase_( A__, A__, A__ ): '''simple docstring''' lowercase__ : List[Any] = [r'h\.\d+\.attn\.bias', r'h\.\d+\.attn\.masked_bias'] @register_to_config def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = 5_0_2_5_7 , lowerCamelCase__ = 1_0_2_4 , lowerCamelCase__ = 7_6_8 , lowerCamelCase__ = 1_2 , lowerCamelCase__ = 1_2 , lowerCamelCase__ = None , lowerCamelCase__ = "gelu_new" , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 0.1 , lowerCamelCase__ = 1e-5 , lowerCamelCase__ = 0.0_2 , lowerCamelCase__ = True , lowerCamelCase__ = True , lowerCamelCase__ = False , lowerCamelCase__ = False , ): super().__init__() _lowerCamelCase = prefix_length if prefix_inner_dim != n_embd and prefix_hidden_dim is None: raise ValueError( F"""`prefix_hidden_dim` cannot be `None` when `prefix_inner_dim`: {prefix_hidden_dim} and""" F""" `n_embd`: {n_embd} are not equal.""" ) _lowerCamelCase = prefix_inner_dim _lowerCamelCase = prefix_hidden_dim _lowerCamelCase = ( nn.Linear(self.prefix_inner_dim , self.prefix_hidden_dim ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCamelCase = ( nn.Linear(self.prefix_hidden_dim , lowerCamelCase__ ) if self.prefix_hidden_dim is not None else nn.Identity() ) _lowerCamelCase = GPTaConfig( vocab_size=lowerCamelCase__ , n_positions=lowerCamelCase__ , n_embd=lowerCamelCase__ , n_layer=lowerCamelCase__ , n_head=lowerCamelCase__ , n_inner=lowerCamelCase__ , activation_function=lowerCamelCase__ , resid_pdrop=lowerCamelCase__ , embd_pdrop=lowerCamelCase__ , attn_pdrop=lowerCamelCase__ , layer_norm_epsilon=lowerCamelCase__ , initializer_range=lowerCamelCase__ , scale_attn_weights=lowerCamelCase__ , use_cache=lowerCamelCase__ , scale_attn_by_inverse_layer_idx=lowerCamelCase__ , reorder_and_upcast_attn=lowerCamelCase__ , ) _lowerCamelCase = GPTaLMHeadModel(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = None , ): _lowerCamelCase = self.transformer.transformer.wte(lowerCamelCase__ ) _lowerCamelCase = self.encode_prefix(lowerCamelCase__ ) _lowerCamelCase = self.decode_prefix(lowerCamelCase__ ) _lowerCamelCase = torch.cat((prefix_embeds, embedding_text) , dim=1 ) if labels is not None: _lowerCamelCase = self.get_dummy_token(input_ids.shape[0] , input_ids.device ) _lowerCamelCase = torch.cat((dummy_token, input_ids) , dim=1 ) _lowerCamelCase = self.transformer(inputs_embeds=lowerCamelCase__ , labels=lowerCamelCase__ , attention_mask=lowerCamelCase__ ) if self.prefix_hidden_dim is not None: return out, hidden else: return out def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): return torch.zeros(lowerCamelCase__ , self.prefix_length , dtype=torch.intaa , device=lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): return self.encode_prefix(lowerCamelCase__ ) @torch.no_grad() def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = torch.split(lowerCamelCase__ , 1 , dim=0 ) _lowerCamelCase = [] _lowerCamelCase = [] for feature in features: _lowerCamelCase = self.decode_prefix(feature.to(lowerCamelCase__ ) ) # back to the clip feature # Only support beam search for now _lowerCamelCase , _lowerCamelCase = self.generate_beam( input_embeds=lowerCamelCase__ , device=lowerCamelCase__ , eos_token_id=lowerCamelCase__ ) generated_tokens.append(output_tokens[0] ) generated_seq_lengths.append(seq_lengths[0] ) _lowerCamelCase = torch.stack(lowerCamelCase__ ) _lowerCamelCase = torch.stack(lowerCamelCase__ ) return generated_tokens, generated_seq_lengths @torch.no_grad() def snake_case__ ( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__ = 5 , lowerCamelCase__ = 6_7 , lowerCamelCase__ = 1.0 , lowerCamelCase__ = None , ): _lowerCamelCase = eos_token_id _lowerCamelCase = None _lowerCamelCase = None _lowerCamelCase = torch.ones(lowerCamelCase__ , device=lowerCamelCase__ , dtype=torch.int ) _lowerCamelCase = torch.zeros(lowerCamelCase__ , device=lowerCamelCase__ , dtype=torch.bool ) if input_embeds is not None: _lowerCamelCase = input_embeds else: _lowerCamelCase = self.transformer.transformer.wte(lowerCamelCase__ ) for i in range(lowerCamelCase__ ): _lowerCamelCase = self.transformer(inputs_embeds=lowerCamelCase__ ) _lowerCamelCase = outputs.logits _lowerCamelCase = logits[:, -1, :] / (temperature if temperature > 0 else 1.0) _lowerCamelCase = logits.softmax(-1 ).log() if scores is None: _lowerCamelCase , _lowerCamelCase = logits.topk(lowerCamelCase__ , -1 ) _lowerCamelCase = generated.expand(lowerCamelCase__ , *generated.shape[1:] ) _lowerCamelCase , _lowerCamelCase = next_tokens.permute(1 , 0 ), scores.squeeze(0 ) if tokens is None: _lowerCamelCase = next_tokens else: _lowerCamelCase = tokens.expand(lowerCamelCase__ , *tokens.shape[1:] ) _lowerCamelCase = torch.cat((tokens, next_tokens) , dim=1 ) else: _lowerCamelCase = -float(np.inf ) _lowerCamelCase = 0 _lowerCamelCase = scores[:, None] + logits seq_lengths[~is_stopped] += 1 _lowerCamelCase = scores_sum / seq_lengths[:, None] _lowerCamelCase , _lowerCamelCase = scores_sum_average.view(-1 ).topk(lowerCamelCase__ , -1 ) _lowerCamelCase = next_tokens // scores_sum.shape[1] _lowerCamelCase = seq_lengths[next_tokens_source] _lowerCamelCase = next_tokens % scores_sum.shape[1] _lowerCamelCase = next_tokens.unsqueeze(1 ) _lowerCamelCase = tokens[next_tokens_source] _lowerCamelCase = torch.cat((tokens, next_tokens) , dim=1 ) _lowerCamelCase = generated[next_tokens_source] _lowerCamelCase = scores_sum_average * seq_lengths _lowerCamelCase = is_stopped[next_tokens_source] _lowerCamelCase = self.transformer.transformer.wte(next_tokens.squeeze() ).view(generated.shape[0] , 1 , -1 ) _lowerCamelCase = torch.cat((generated, next_token_embed) , dim=1 ) _lowerCamelCase = is_stopped + next_tokens.eq(lowerCamelCase__ ).squeeze() if is_stopped.all(): break _lowerCamelCase = scores / seq_lengths _lowerCamelCase = scores.argsort(descending=lowerCamelCase__ ) # tokens tensors are already padded to max_seq_length _lowerCamelCase = [tokens[i] for i in order] _lowerCamelCase = torch.stack(lowerCamelCase__ , dim=0 ) _lowerCamelCase = torch.tensor([seq_lengths[i] for i in order] , dtype=seq_lengths.dtype ) return output_texts, seq_lengths
661
0
import gc import unittest from transformers import MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, FillMaskPipeline, pipeline from transformers.pipelines import PipelineException from transformers.testing_utils import ( is_pipeline_test, is_torch_available, nested_simplify, require_tf, require_torch, require_torch_gpu, slow, ) from .test_pipelines_common import ANY @is_pipeline_test class _a ( unittest.TestCase ): """simple docstring""" A_ = MODEL_FOR_MASKED_LM_MAPPING A_ = TF_MODEL_FOR_MASKED_LM_MAPPING def _UpperCAmelCase ( self ) -> List[str]: super().tearDown() # clean-up as much as possible GPU memory occupied by PyTorch gc.collect() if is_torch_available(): import torch torch.cuda.empty_cache() @require_tf def _UpperCAmelCase ( self ) -> str: UpperCamelCase_ = pipeline(task='fill-mask' , model='sshleifer/tiny-distilroberta-base' , top_k=2 , framework='tf' ) UpperCamelCase_ = unmasker('My name is <mask>' ) self.assertEqual( nested_simplify(_UpperCAmelCase , decimals=6 ) , [ {'sequence': 'My name is grouped', 'score': 2.1e-05, 'token': 38015, 'token_str': ' grouped'}, {'sequence': 'My name is accuser', 'score': 2.1e-05, 'token': 25506, 'token_str': ' accuser'}, ] , ) UpperCamelCase_ = unmasker('The largest city in France is <mask>' ) self.assertEqual( nested_simplify(_UpperCAmelCase , decimals=6 ) , [ { 'sequence': 'The largest city in France is grouped', 'score': 2.1e-05, 'token': 38015, 'token_str': ' grouped', }, { 'sequence': 'The largest city in France is accuser', 'score': 2.1e-05, 'token': 25506, 'token_str': ' accuser', }, ] , ) UpperCamelCase_ = unmasker('My name is <mask>' , targets=[' Patrick', ' Clara', ' Teven'] , top_k=3 ) self.assertEqual( nested_simplify(_UpperCAmelCase , decimals=6 ) , [ {'sequence': 'My name is Clara', 'score': 2e-05, 'token': 13606, 'token_str': ' Clara'}, {'sequence': 'My name is Patrick', 'score': 2e-05, 'token': 3499, 'token_str': ' Patrick'}, {'sequence': 'My name is Te', 'score': 1.9e-05, 'token': 2941, 'token_str': ' Te'}, ] , ) @require_torch def _UpperCAmelCase ( self ) -> Union[str, Any]: UpperCamelCase_ = pipeline(task='fill-mask' , model='sshleifer/tiny-distilroberta-base' , top_k=2 , framework='pt' ) UpperCamelCase_ = unmasker('My name is <mask>' ) self.assertEqual( nested_simplify(_UpperCAmelCase , decimals=6 ) , [ {'sequence': 'My name is Maul', 'score': 2.2e-05, 'token': 35676, 'token_str': ' Maul'}, {'sequence': 'My name isELS', 'score': 2.2e-05, 'token': 16416, 'token_str': 'ELS'}, ] , ) UpperCamelCase_ = unmasker('The largest city in France is <mask>' ) self.assertEqual( nested_simplify(_UpperCAmelCase , decimals=6 ) , [ { 'sequence': 'The largest city in France is Maul', 'score': 2.2e-05, 'token': 35676, 'token_str': ' Maul', }, {'sequence': 'The largest city in France isELS', 'score': 2.2e-05, 'token': 16416, 'token_str': 'ELS'}, ] , ) UpperCamelCase_ = unmasker('My name is <mask>' , targets=[' Patrick', ' Clara', ' Teven'] , top_k=3 ) self.assertEqual( nested_simplify(_UpperCAmelCase , decimals=6 ) , [ {'sequence': 'My name is Patrick', 'score': 2.1e-05, 'token': 3499, 'token_str': ' Patrick'}, {'sequence': 'My name is Te', 'score': 2e-05, 'token': 2941, 'token_str': ' Te'}, {'sequence': 'My name is Clara', 'score': 2e-05, 'token': 13606, 'token_str': ' Clara'}, ] , ) UpperCamelCase_ = unmasker('My name is <mask> <mask>' , top_k=2 ) self.assertEqual( nested_simplify(_UpperCAmelCase , decimals=6 ) , [ [ { 'score': 2.2e-05, 'token': 35676, 'token_str': ' Maul', 'sequence': '<s>My name is Maul<mask></s>', }, {'score': 2.2e-05, 'token': 16416, 'token_str': 'ELS', 'sequence': '<s>My name isELS<mask></s>'}, ], [ { 'score': 2.2e-05, 'token': 35676, 'token_str': ' Maul', 'sequence': '<s>My name is<mask> Maul</s>', }, {'score': 2.2e-05, 'token': 16416, 'token_str': 'ELS', 'sequence': '<s>My name is<mask>ELS</s>'}, ], ] , ) @require_torch_gpu def _UpperCAmelCase ( self ) -> Optional[Any]: UpperCamelCase_ = pipeline('fill-mask' , model='hf-internal-testing/tiny-random-distilbert' , device=0 , framework='pt' ) # convert model to fp16 pipe.model.half() UpperCamelCase_ = pipe('Paris is the [MASK] of France.' ) # We actually don't care about the result, we just want to make sure # it works, meaning the float16 tensor got casted back to float32 # for postprocessing. self.assertIsInstance(_UpperCAmelCase , _UpperCAmelCase ) @slow @require_torch def _UpperCAmelCase ( self ) -> List[Any]: UpperCamelCase_ = pipeline(task='fill-mask' , model='distilroberta-base' , top_k=2 , framework='pt' ) self.run_large_test(_UpperCAmelCase ) @slow @require_tf def _UpperCAmelCase ( self ) -> Union[str, Any]: UpperCamelCase_ = pipeline(task='fill-mask' , model='distilroberta-base' , top_k=2 , framework='tf' ) self.run_large_test(_UpperCAmelCase ) def _UpperCAmelCase ( self , _UpperCAmelCase ) -> Tuple: UpperCamelCase_ = unmasker('My name is <mask>' ) self.assertEqual( nested_simplify(_UpperCAmelCase ) , [ {'sequence': 'My name is John', 'score': 0.0_0_8, 'token': 610, 'token_str': ' John'}, {'sequence': 'My name is Chris', 'score': 0.0_0_7, 'token': 1573, 'token_str': ' Chris'}, ] , ) UpperCamelCase_ = unmasker('The largest city in France is <mask>' ) self.assertEqual( nested_simplify(_UpperCAmelCase ) , [ { 'sequence': 'The largest city in France is Paris', 'score': 0.2_5_1, 'token': 2201, 'token_str': ' Paris', }, { 'sequence': 'The largest city in France is Lyon', 'score': 0.2_1_4, 'token': 12790, 'token_str': ' Lyon', }, ] , ) UpperCamelCase_ = unmasker('My name is <mask>' , targets=[' Patrick', ' Clara', ' Teven'] , top_k=3 ) self.assertEqual( nested_simplify(_UpperCAmelCase ) , [ {'sequence': 'My name is Patrick', 'score': 0.0_0_5, 'token': 3499, 'token_str': ' Patrick'}, {'sequence': 'My name is Clara', 'score': 0.0_0_0, 'token': 13606, 'token_str': ' Clara'}, {'sequence': 'My name is Te', 'score': 0.0_0_0, 'token': 2941, 'token_str': ' Te'}, ] , ) @require_torch def _UpperCAmelCase ( self ) -> List[Any]: UpperCamelCase_ = pipeline(task='fill-mask' , model='sshleifer/tiny-distilroberta-base' , framework='pt' ) UpperCamelCase_ = None UpperCamelCase_ = None self.run_pipeline_test(_UpperCAmelCase , [] ) @require_tf def _UpperCAmelCase ( self ) -> Optional[int]: UpperCamelCase_ = pipeline(task='fill-mask' , model='sshleifer/tiny-distilroberta-base' , framework='tf' ) UpperCamelCase_ = None UpperCamelCase_ = None self.run_pipeline_test(_UpperCAmelCase , [] ) def _UpperCAmelCase ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[Any]: if tokenizer is None or tokenizer.mask_token_id is None: self.skipTest('The provided tokenizer has no mask token, (probably reformer or wav2vec2)' ) UpperCamelCase_ = FillMaskPipeline(model=_UpperCAmelCase , tokenizer=_UpperCAmelCase ) UpperCamelCase_ = [ f"""This is another {tokenizer.mask_token} test""", ] return fill_masker, examples def _UpperCAmelCase ( self , _UpperCAmelCase , _UpperCAmelCase ) -> Union[str, Any]: UpperCamelCase_ = fill_masker.tokenizer UpperCamelCase_ = fill_masker.model UpperCamelCase_ = fill_masker( f"""This is a {tokenizer.mask_token}""" , ) self.assertEqual( _UpperCAmelCase , [ {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, ] , ) UpperCamelCase_ = fill_masker([f"""This is a {tokenizer.mask_token}"""] ) self.assertEqual( _UpperCAmelCase , [ {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, ] , ) UpperCamelCase_ = fill_masker([f"""This is a {tokenizer.mask_token}""", f"""Another {tokenizer.mask_token} great test."""] ) self.assertEqual( _UpperCAmelCase , [ [ {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, ], [ {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, ], ] , ) with self.assertRaises(_UpperCAmelCase ): fill_masker([None] ) # No mask_token is not supported with self.assertRaises(_UpperCAmelCase ): fill_masker('This is' ) self.run_test_top_k(_UpperCAmelCase , _UpperCAmelCase ) self.run_test_targets(_UpperCAmelCase , _UpperCAmelCase ) self.run_test_top_k_targets(_UpperCAmelCase , _UpperCAmelCase ) self.fill_mask_with_duplicate_targets_and_top_k(_UpperCAmelCase , _UpperCAmelCase ) self.fill_mask_with_multiple_masks(_UpperCAmelCase , _UpperCAmelCase ) def _UpperCAmelCase ( self , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[Any]: UpperCamelCase_ = tokenizer.get_vocab() UpperCamelCase_ = sorted(vocab.keys() )[:2] # Pipeline argument UpperCamelCase_ = FillMaskPipeline(model=_UpperCAmelCase , tokenizer=_UpperCAmelCase , targets=_UpperCAmelCase ) UpperCamelCase_ = fill_masker(f"""This is a {tokenizer.mask_token}""" ) self.assertEqual( _UpperCAmelCase , [ {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, ] , ) UpperCamelCase_ = {vocab[el] for el in targets} self.assertEqual({el['token'] for el in outputs} , _UpperCAmelCase ) UpperCamelCase_ = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el['token_str'] for el in outputs} , set(_UpperCAmelCase ) ) # Call argument UpperCamelCase_ = FillMaskPipeline(model=_UpperCAmelCase , tokenizer=_UpperCAmelCase ) UpperCamelCase_ = fill_masker(f"""This is a {tokenizer.mask_token}""" , targets=_UpperCAmelCase ) self.assertEqual( _UpperCAmelCase , [ {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, ] , ) UpperCamelCase_ = {vocab[el] for el in targets} self.assertEqual({el['token'] for el in outputs} , _UpperCAmelCase ) UpperCamelCase_ = [tokenizer.decode([x] ) for x in target_ids] self.assertEqual({el['token_str'] for el in outputs} , set(_UpperCAmelCase ) ) # Score equivalence UpperCamelCase_ = fill_masker(f"""This is a {tokenizer.mask_token}""" , targets=_UpperCAmelCase ) UpperCamelCase_ = [top_mask['token_str'] for top_mask in outputs] UpperCamelCase_ = [top_mask['score'] for top_mask in outputs] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(_UpperCAmelCase ) == set(_UpperCAmelCase ): UpperCamelCase_ = fill_masker(f"""This is a {tokenizer.mask_token}""" , targets=_UpperCAmelCase ) UpperCamelCase_ = [top_mask['score'] for top_mask in unmasked_targets] self.assertEqual(nested_simplify(_UpperCAmelCase ) , nested_simplify(_UpperCAmelCase ) ) # Raises with invalid with self.assertRaises(_UpperCAmelCase ): UpperCamelCase_ = fill_masker(f"""This is a {tokenizer.mask_token}""" , targets=[] ) # For some tokenizers, `""` is actually in the vocabulary and the expected error won't raised if "" not in tokenizer.get_vocab(): with self.assertRaises(_UpperCAmelCase ): UpperCamelCase_ = fill_masker(f"""This is a {tokenizer.mask_token}""" , targets=[''] ) with self.assertRaises(_UpperCAmelCase ): UpperCamelCase_ = fill_masker(f"""This is a {tokenizer.mask_token}""" , targets='' ) def _UpperCAmelCase ( self , _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: UpperCamelCase_ = FillMaskPipeline(model=_UpperCAmelCase , tokenizer=_UpperCAmelCase , top_k=2 ) UpperCamelCase_ = fill_masker(f"""This is a {tokenizer.mask_token}""" ) self.assertEqual( _UpperCAmelCase , [ {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, ] , ) UpperCamelCase_ = FillMaskPipeline(model=_UpperCAmelCase , tokenizer=_UpperCAmelCase ) UpperCamelCase_ = fill_masker(f"""This is a {tokenizer.mask_token}""" , top_k=2 ) self.assertEqual( _UpperCAmelCase , [ {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, ] , ) self.assertEqual(nested_simplify(_UpperCAmelCase ) , nested_simplify(_UpperCAmelCase ) ) def _UpperCAmelCase ( self , _UpperCAmelCase , _UpperCAmelCase ) -> List[Any]: UpperCamelCase_ = tokenizer.get_vocab() UpperCamelCase_ = FillMaskPipeline(model=_UpperCAmelCase , tokenizer=_UpperCAmelCase ) # top_k=2, ntargets=3 UpperCamelCase_ = sorted(vocab.keys() )[:3] UpperCamelCase_ = fill_masker(f"""This is a {tokenizer.mask_token}""" , top_k=2 , targets=_UpperCAmelCase ) # If we use the most probably targets, and filter differently, we should still # have the same results UpperCamelCase_ = [el['token_str'] for el in sorted(_UpperCAmelCase , key=lambda _UpperCAmelCase : x["score"] , reverse=_UpperCAmelCase )] # For some BPE tokenizers, `</w>` is removed during decoding, so `token_str` won't be the same as in `targets`. if set(_UpperCAmelCase ).issubset(_UpperCAmelCase ): UpperCamelCase_ = fill_masker(f"""This is a {tokenizer.mask_token}""" , top_k=3 , targets=_UpperCAmelCase ) # They should yield exactly the same result self.assertEqual(nested_simplify(_UpperCAmelCase ) , nested_simplify(_UpperCAmelCase ) ) def _UpperCAmelCase ( self , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: UpperCamelCase_ = FillMaskPipeline(model=_UpperCAmelCase , tokenizer=_UpperCAmelCase ) UpperCamelCase_ = tokenizer.get_vocab() # String duplicates + id duplicates UpperCamelCase_ = sorted(vocab.keys() )[:3] UpperCamelCase_ = [targets[0], targets[1], targets[0], targets[2], targets[1]] UpperCamelCase_ = fill_masker(f"""My name is {tokenizer.mask_token}""" , targets=_UpperCAmelCase , top_k=10 ) # The target list contains duplicates, so we can't output more # than them self.assertEqual(len(_UpperCAmelCase ) , 3 ) def _UpperCAmelCase ( self , _UpperCAmelCase , _UpperCAmelCase ) -> List[str]: UpperCamelCase_ = FillMaskPipeline(model=_UpperCAmelCase , tokenizer=_UpperCAmelCase ) UpperCamelCase_ = fill_masker( f"""This is a {tokenizer.mask_token} {tokenizer.mask_token} {tokenizer.mask_token}""" , top_k=2 ) self.assertEqual( _UpperCAmelCase , [ [ {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, ], [ {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, ], [ {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, {'sequence': ANY(_UpperCAmelCase ), 'score': ANY(_UpperCAmelCase ), 'token': ANY(_UpperCAmelCase ), 'token_str': ANY(_UpperCAmelCase )}, ], ] , )
23
"""simple docstring""" import re import string from collections import Counter import sacrebleu import sacremoses from packaging import version import datasets __SCREAMING_SNAKE_CASE : Optional[int] = ''' @inproceedings{xu-etal-2016-optimizing, title = {Optimizing Statistical Machine Translation for Text Simplification}, authors={Xu, Wei and Napoles, Courtney and Pavlick, Ellie and Chen, Quanze and Callison-Burch, Chris}, journal = {Transactions of the Association for Computational Linguistics}, volume = {4}, year={2016}, url = {https://www.aclweb.org/anthology/Q16-1029}, pages = {401--415 }, @inproceedings{post-2018-call, title = "A Call for Clarity in Reporting {BLEU} Scores", author = "Post, Matt", booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers", month = oct, year = "2018", address = "Belgium, Brussels", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W18-6319", pages = "186--191", } ''' __SCREAMING_SNAKE_CASE : List[str] = '''\ WIKI_SPLIT is the combination of three metrics SARI, EXACT and SACREBLEU It can be used to evaluate the quality of machine-generated texts. ''' __SCREAMING_SNAKE_CASE : Optional[Any] = ''' Calculates sari score (between 0 and 100) given a list of source and predicted sentences, and a list of lists of reference sentences. It also computes the BLEU score as well as the exact match score. Args: sources: list of source sentences where each sentence should be a string. predictions: list of predicted sentences where each sentence should be a string. references: list of lists of reference sentences where each sentence should be a string. Returns: sari: sari score sacrebleu: sacrebleu score exact: exact score Examples: >>> sources=["About 95 species are currently accepted ."] >>> predictions=["About 95 you now get in ."] >>> references=[["About 95 species are currently known ."]] >>> wiki_split = datasets.load_metric("wiki_split") >>> results = wiki_split.compute(sources=sources, predictions=predictions, references=references) >>> print(results) {\'sari\': 21.805555555555557, \'sacrebleu\': 14.535768424205482, \'exact\': 0.0} ''' def lowerCAmelCase_( lowercase_ : Union[str, Any] ) -> str: def remove_articles(lowercase_ : int ): _lowerCamelCase = re.compile(r'''\b(a|an|the)\b''' , re.UNICODE ) return re.sub(lowercase_ , ''' ''' , lowercase_ ) def white_space_fix(lowercase_ : List[Any] ): return " ".join(text.split() ) def remove_punc(lowercase_ : Dict ): _lowerCamelCase = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(lowercase_ : Union[str, Any] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(lowercase_ ) ) ) ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[Any] ) -> Union[str, Any]: return int(normalize_answer(lowercase_ ) == normalize_answer(lowercase_ ) ) def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Tuple ) -> Tuple: _lowerCamelCase = [any(compute_exact(lowercase_ , lowercase_ ) for ref in refs ) for pred, refs in zip(lowercase_ , lowercase_ )] return (sum(lowercase_ ) / len(lowercase_ )) * 1_00 def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : str , lowercase_ : Optional[Any] , lowercase_ : str ) -> Optional[int]: _lowerCamelCase = [rgram for rgrams in rgramslist for rgram in rgrams] _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter() for sgram, scount in sgramcounter.items(): _lowerCamelCase = scount * numref _lowerCamelCase = Counter(lowercase_ ) _lowerCamelCase = Counter() for cgram, ccount in cgramcounter.items(): _lowerCamelCase = ccount * numref # KEEP _lowerCamelCase = sgramcounter_rep & cgramcounter_rep _lowerCamelCase = keepgramcounter_rep & rgramcounter _lowerCamelCase = sgramcounter_rep & rgramcounter _lowerCamelCase = 0 _lowerCamelCase = 0 for keepgram in keepgramcountergood_rep: keeptmpscorea += keepgramcountergood_rep[keepgram] / keepgramcounter_rep[keepgram] # Fix an alleged bug [2] in the keep score computation. # keeptmpscore2 += keepgramcountergood_rep[keepgram] / keepgramcounterall_rep[keepgram] keeptmpscorea += keepgramcountergood_rep[keepgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = keeptmpscorea / len(lowercase_ ) if len(lowercase_ ) > 0: # Fix an alleged bug [2] in the keep score computation. # keepscore_recall = keeptmpscore2 / len(keepgramcounterall_rep) _lowerCamelCase = keeptmpscorea / sum(keepgramcounterall_rep.values() ) _lowerCamelCase = 0 if keepscore_precision > 0 or keepscore_recall > 0: _lowerCamelCase = 2 * keepscore_precision * keepscore_recall / (keepscore_precision + keepscore_recall) # DELETION _lowerCamelCase = sgramcounter_rep - cgramcounter_rep _lowerCamelCase = delgramcounter_rep - rgramcounter _lowerCamelCase = sgramcounter_rep - rgramcounter _lowerCamelCase = 0 _lowerCamelCase = 0 for delgram in delgramcountergood_rep: deltmpscorea += delgramcountergood_rep[delgram] / delgramcounter_rep[delgram] deltmpscorea += delgramcountergood_rep[delgram] / delgramcounterall_rep[delgram] # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = deltmpscorea / len(lowercase_ ) # ADDITION _lowerCamelCase = set(lowercase_ ) - set(lowercase_ ) _lowerCamelCase = set(lowercase_ ) & set(lowercase_ ) _lowerCamelCase = set(lowercase_ ) - set(lowercase_ ) _lowerCamelCase = 0 for addgram in addgramcountergood: addtmpscore += 1 # Define 0/0=1 instead of 0 to give higher scores for predictions that match # a target exactly. _lowerCamelCase = 1 _lowerCamelCase = 1 if len(lowercase_ ) > 0: _lowerCamelCase = addtmpscore / len(lowercase_ ) if len(lowercase_ ) > 0: _lowerCamelCase = addtmpscore / len(lowercase_ ) _lowerCamelCase = 0 if addscore_precision > 0 or addscore_recall > 0: _lowerCamelCase = 2 * addscore_precision * addscore_recall / (addscore_precision + addscore_recall) return (keepscore, delscore_precision, addscore) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : Optional[Any] , lowercase_ : str ) -> List[str]: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = ssent.split(''' ''' ) _lowerCamelCase = csent.split(''' ''' ) _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] for rsent in rsents: _lowerCamelCase = rsent.split(''' ''' ) _lowerCamelCase = [] _lowerCamelCase = [] _lowerCamelCase = [] ragramslist.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] ragrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] + ''' ''' + ragrams[i + 2] ragrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = ragrams[i] + ''' ''' + ragrams[i + 1] + ''' ''' + ragrams[i + 2] + ''' ''' + ragrams[i + 3] ragrams.append(lowercase_ ) ragramslist.append(lowercase_ ) ragramslist.append(lowercase_ ) ragramslist.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] sagrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] + ''' ''' + sagrams[i + 2] sagrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = sagrams[i] + ''' ''' + sagrams[i + 1] + ''' ''' + sagrams[i + 2] + ''' ''' + sagrams[i + 3] sagrams.append(lowercase_ ) for i in range(0 , len(lowercase_ ) - 1 ): if i < len(lowercase_ ) - 1: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] cagrams.append(lowercase_ ) if i < len(lowercase_ ) - 2: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] + ''' ''' + cagrams[i + 2] cagrams.append(lowercase_ ) if i < len(lowercase_ ) - 3: _lowerCamelCase = cagrams[i] + ''' ''' + cagrams[i + 1] + ''' ''' + cagrams[i + 2] + ''' ''' + cagrams[i + 3] cagrams.append(lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) ((_lowerCamelCase) , (_lowerCamelCase) , (_lowerCamelCase)) = SARIngram(lowercase_ , lowercase_ , lowercase_ , lowercase_ ) _lowerCamelCase = sum([keepascore, keepascore, keepascore, keepascore] ) / 4 _lowerCamelCase = sum([delascore, delascore, delascore, delascore] ) / 4 _lowerCamelCase = sum([addascore, addascore, addascore, addascore] ) / 4 _lowerCamelCase = (avgkeepscore + avgdelscore + avgaddscore) / 3 return finalscore def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : bool = True , lowercase_ : str = "13a" , lowercase_ : bool = True ) -> int: # Normalization is requried for the ASSET dataset (one of the primary # datasets in sentence simplification) to allow using space # to split the sentence. Even though Wiki-Auto and TURK datasets, # do not require normalization, we do it for consistency. # Code adapted from the EASSE library [1] written by the authors of the ASSET dataset. # [1] https://github.com/feralvam/easse/blob/580bba7e1378fc8289c663f864e0487188fe8067/easse/utils/preprocessing.py#L7 if lowercase: _lowerCamelCase = sentence.lower() if tokenizer in ["13a", "intl"]: if version.parse(sacrebleu.__version__ ).major >= 2: _lowerCamelCase = sacrebleu.metrics.bleu._get_tokenizer(lowercase_ )()(lowercase_ ) else: _lowerCamelCase = sacrebleu.TOKENIZERS[tokenizer]()(lowercase_ ) elif tokenizer == "moses": _lowerCamelCase = sacremoses.MosesTokenizer().tokenize(lowercase_ , return_str=lowercase_ , escape=lowercase_ ) elif tokenizer == "penn": _lowerCamelCase = sacremoses.MosesTokenizer().penn_tokenize(lowercase_ , return_str=lowercase_ ) else: _lowerCamelCase = sentence if not return_str: _lowerCamelCase = normalized_sent.split() return normalized_sent def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Any , lowercase_ : List[Any] ) -> Optional[int]: if not (len(lowercase_ ) == len(lowercase_ ) == len(lowercase_ )): raise ValueError('''Sources length must match predictions and references lengths.''' ) _lowerCamelCase = 0 for src, pred, refs in zip(lowercase_ , lowercase_ , lowercase_ ): sari_score += SARIsent(normalize(lowercase_ ) , normalize(lowercase_ ) , [normalize(lowercase_ ) for sent in refs] ) _lowerCamelCase = sari_score / len(lowercase_ ) return 1_00 * sari_score def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Any , lowercase_ : List[Any]="exp" , lowercase_ : List[Any]=None , lowercase_ : Optional[Any]=False , lowercase_ : List[Any]=False , lowercase_ : List[Any]=False , ) -> Dict: _lowerCamelCase = len(references[0] ) if any(len(lowercase_ ) != references_per_prediction for refs in references ): raise ValueError('''Sacrebleu requires the same number of references for each prediction''' ) _lowerCamelCase = [[refs[i] for refs in references] for i in range(lowercase_ )] _lowerCamelCase = sacrebleu.corpus_bleu( lowercase_ , lowercase_ , smooth_method=lowercase_ , smooth_value=lowercase_ , force=lowercase_ , lowercase=lowercase_ , use_effective_order=lowercase_ , ) return output.score @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Sequence(datasets.Value('''string''' , id='''sequence''' ) , id='''references''' ), } ) , codebase_urls=[ '''https://github.com/huggingface/transformers/blob/master/src/transformers/data/metrics/squad_metrics.py''', '''https://github.com/cocoxu/simplification/blob/master/SARI.py''', '''https://github.com/tensorflow/tensor2tensor/blob/master/tensor2tensor/utils/sari_hook.py''', '''https://github.com/mjpost/sacreBLEU''', ] , reference_urls=[ '''https://www.aclweb.org/anthology/Q16-1029.pdf''', '''https://github.com/mjpost/sacreBLEU''', '''https://en.wikipedia.org/wiki/BLEU''', '''https://towardsdatascience.com/evaluating-text-output-in-nlp-bleu-at-your-own-risk-e8609665a213''', ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = {} result.update({'''sari''': compute_sari(sources=lowerCamelCase__ , predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) result.update({'''sacrebleu''': compute_sacrebleu(predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) result.update({'''exact''': compute_em(predictions=lowerCamelCase__ , references=lowerCamelCase__ )} ) return result
661
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_torch_available, ) UpperCAmelCase_ : int = { '''configuration_trocr''': ['''TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TrOCRConfig'''], '''processing_trocr''': ['''TrOCRProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase_ : Optional[Any] = [ '''TROCR_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TrOCRForCausalLM''', '''TrOCRPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_trocr import TROCR_PRETRAINED_CONFIG_ARCHIVE_MAP, TrOCRConfig from .processing_trocr import TrOCRProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_trocr import TROCR_PRETRAINED_MODEL_ARCHIVE_LIST, TrOCRForCausalLM, TrOCRPreTrainedModel else: import sys UpperCAmelCase_ : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
24
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''configuration_wav2vec2''': ['''WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Wav2Vec2Config'''], '''feature_extraction_wav2vec2''': ['''Wav2Vec2FeatureExtractor'''], '''processing_wav2vec2''': ['''Wav2Vec2Processor'''], '''tokenization_wav2vec2''': ['''Wav2Vec2CTCTokenizer''', '''Wav2Vec2Tokenizer'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : Tuple = [ '''WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Wav2Vec2ForAudioFrameClassification''', '''Wav2Vec2ForCTC''', '''Wav2Vec2ForMaskedLM''', '''Wav2Vec2ForPreTraining''', '''Wav2Vec2ForSequenceClassification''', '''Wav2Vec2ForXVector''', '''Wav2Vec2Model''', '''Wav2Vec2PreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : List[Any] = [ '''TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFWav2Vec2ForCTC''', '''TFWav2Vec2Model''', '''TFWav2Vec2PreTrainedModel''', '''TFWav2Vec2ForSequenceClassification''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __SCREAMING_SNAKE_CASE : str = [ '''FlaxWav2Vec2ForCTC''', '''FlaxWav2Vec2ForPreTraining''', '''FlaxWav2Vec2Model''', '''FlaxWav2Vec2PreTrainedModel''', ] if TYPE_CHECKING: from .configuration_wavaveca import WAV_2_VEC_2_PRETRAINED_CONFIG_ARCHIVE_MAP, WavaVecaConfig from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .processing_wavaveca import WavaVecaProcessor from .tokenization_wavaveca import WavaVecaCTCTokenizer, WavaVecaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_wavaveca import ( WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, WavaVecaForAudioFrameClassification, WavaVecaForCTC, WavaVecaForMaskedLM, WavaVecaForPreTraining, WavaVecaForSequenceClassification, WavaVecaForXVector, WavaVecaModel, WavaVecaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( TF_WAV_2_VEC_2_PRETRAINED_MODEL_ARCHIVE_LIST, TFWavaVecaForCTC, TFWavaVecaForSequenceClassification, TFWavaVecaModel, TFWavaVecaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_wavaveca import ( FlaxWavaVecaForCTC, FlaxWavaVecaForPreTraining, FlaxWavaVecaModel, FlaxWavaVecaPreTrainedModel, ) else: import sys __SCREAMING_SNAKE_CASE : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
661
0
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'edbeeching/decision-transformer-gym-hopper-medium': ( 'https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/config.json' ), # See all DecisionTransformer models at https://huggingface.co/models?filter=decision_transformer } class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ ='decision_transformer' lowerCamelCase__ =['past_key_values'] lowerCamelCase__ ={ 'max_position_embeddings': 'n_positions', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : str , a : List[str]=17 , a : Optional[int]=4 , a : List[str]=128 , a : Union[str, Any]=4096 , a : Union[str, Any]=True , a : Dict=1 , a : Optional[Any]=1024 , a : int=3 , a : Any=1 , a : str=None , a : List[Any]="relu" , a : Optional[Any]=0.1 , a : int=0.1 , a : Dict=0.1 , a : List[str]=1e-5 , a : List[str]=0.02 , a : str=True , a : Any=True , a : Tuple=5_0256 , a : List[str]=5_0256 , a : Dict=False , a : int=False , **a : Any , ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = state_dim SCREAMING_SNAKE_CASE : Optional[Any] = act_dim SCREAMING_SNAKE_CASE : List[Any] = hidden_size SCREAMING_SNAKE_CASE : int = max_ep_len SCREAMING_SNAKE_CASE : Any = action_tanh SCREAMING_SNAKE_CASE : List[Any] = vocab_size SCREAMING_SNAKE_CASE : Optional[int] = n_positions SCREAMING_SNAKE_CASE : Union[str, Any] = n_layer SCREAMING_SNAKE_CASE : Dict = n_head SCREAMING_SNAKE_CASE : List[Any] = n_inner SCREAMING_SNAKE_CASE : List[str] = activation_function SCREAMING_SNAKE_CASE : Optional[Any] = resid_pdrop SCREAMING_SNAKE_CASE : Dict = embd_pdrop SCREAMING_SNAKE_CASE : List[str] = attn_pdrop SCREAMING_SNAKE_CASE : Optional[int] = layer_norm_epsilon SCREAMING_SNAKE_CASE : Union[str, Any] = initializer_range SCREAMING_SNAKE_CASE : int = scale_attn_weights SCREAMING_SNAKE_CASE : List[str] = use_cache SCREAMING_SNAKE_CASE : Union[str, Any] = scale_attn_by_inverse_layer_idx SCREAMING_SNAKE_CASE : Optional[int] = reorder_and_upcast_attn SCREAMING_SNAKE_CASE : Union[str, Any] = bos_token_id SCREAMING_SNAKE_CASE : Optional[Any] = eos_token_id super().__init__(bos_token_id=a , eos_token_id=a , **a )
25
"""simple docstring""" __SCREAMING_SNAKE_CASE : Any = { '''A''': '''.-''', '''B''': '''-...''', '''C''': '''-.-.''', '''D''': '''-..''', '''E''': '''.''', '''F''': '''..-.''', '''G''': '''--.''', '''H''': '''....''', '''I''': '''..''', '''J''': '''.---''', '''K''': '''-.-''', '''L''': '''.-..''', '''M''': '''--''', '''N''': '''-.''', '''O''': '''---''', '''P''': '''.--.''', '''Q''': '''--.-''', '''R''': '''.-.''', '''S''': '''...''', '''T''': '''-''', '''U''': '''..-''', '''V''': '''...-''', '''W''': '''.--''', '''X''': '''-..-''', '''Y''': '''-.--''', '''Z''': '''--..''', '''1''': '''.----''', '''2''': '''..---''', '''3''': '''...--''', '''4''': '''....-''', '''5''': '''.....''', '''6''': '''-....''', '''7''': '''--...''', '''8''': '''---..''', '''9''': '''----.''', '''0''': '''-----''', '''&''': '''.-...''', '''@''': '''.--.-.''', ''':''': '''---...''', ''',''': '''--..--''', '''.''': '''.-.-.-''', '''\'''': '''.----.''', '''"''': '''.-..-.''', '''?''': '''..--..''', '''/''': '''-..-.''', '''=''': '''-...-''', '''+''': '''.-.-.''', '''-''': '''-....-''', '''(''': '''-.--.''', ''')''': '''-.--.-''', '''!''': '''-.-.--''', ''' ''': '''/''' } # Exclamation mark is not in ITU-R recommendation # fmt: on __SCREAMING_SNAKE_CASE : List[str] = {value: key for key, value in MORSE_CODE_DICT.items()} def lowerCAmelCase_( lowercase_ : str ) -> str: return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def lowerCAmelCase_( lowercase_ : str ) -> str: return "".join(REVERSE_DICT[char] for char in message.split() ) def lowerCAmelCase_( ) -> None: _lowerCamelCase = '''Morse code here!''' print(lowercase_ ) _lowerCamelCase = encrypt(lowercase_ ) print(lowercase_ ) _lowerCamelCase = decrypt(lowercase_ ) print(lowercase_ ) if __name__ == "__main__": main()
661
0
'''simple docstring''' from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging __UpperCamelCase = logging.get_logger(__name__) class _A ( __lowercase ): lowercase__: Any = ['''pixel_values'''] def __init__( self : Any , __magic_name__ : bool = True , __magic_name__ : Dict[str, int] = None , __magic_name__ : PILImageResampling = PILImageResampling.BICUBIC , __magic_name__ : bool = True , __magic_name__ : Dict[str, int] = None , __magic_name__ : bool = True , __magic_name__ : Union[int, float] = 1 / 2_55 , __magic_name__ : bool = True , __magic_name__ : Optional[Union[float, Iterable[float]]] = IMAGENET_DEFAULT_MEAN , __magic_name__ : Optional[Union[float, Iterable[float]]] = IMAGENET_DEFAULT_STD , **__magic_name__ : Union[str, Any] , ) -> None: """simple docstring""" super().__init__(**__magic_name__ ) __snake_case : Dict = size if size is not None else {"""shortest_edge""": 2_24} __snake_case : str = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) __snake_case : int = crop_size if crop_size is not None else {"""height""": 2_24, """width""": 2_24} __snake_case : Optional[Any] = get_size_dict(__magic_name__ , param_name="""crop_size""" ) __snake_case : List[Any] = do_resize __snake_case : Dict = size __snake_case : str = resample __snake_case : Any = do_center_crop __snake_case : Optional[Any] = crop_size __snake_case : Optional[Any] = do_rescale __snake_case : Optional[int] = rescale_factor __snake_case : Tuple = do_normalize __snake_case : str = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN __snake_case : Any = image_std if image_std is not None else IMAGENET_DEFAULT_STD def lowercase__ ( self : List[Any] , __magic_name__ : np.ndarray , __magic_name__ : Dict[str, int] , __magic_name__ : PILImageResampling = PILImageResampling.BICUBIC , __magic_name__ : Optional[Union[str, ChannelDimension]] = None , **__magic_name__ : str , ) -> np.ndarray: """simple docstring""" __snake_case : Union[str, Any] = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: __snake_case : List[Any] = int((2_56 / 2_24) * size["""shortest_edge"""] ) __snake_case : int = get_resize_output_image_size(__magic_name__ , size=__magic_name__ , default_to_square=__magic_name__ ) __snake_case : List[str] = {"""height""": output_size[0], """width""": output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( f'''Size dict must have keys \'height\' and \'width\' or \'shortest_edge\'. Got {size_dict.keys()}''' ) return resize( __magic_name__ , size=(size_dict["""height"""], size_dict["""width"""]) , resample=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def lowercase__ ( self : Union[str, Any] , __magic_name__ : np.ndarray , __magic_name__ : Dict[str, int] , __magic_name__ : Optional[Union[str, ChannelDimension]] = None , **__magic_name__ : str , ) -> np.ndarray: """simple docstring""" __snake_case : str = get_size_dict(__magic_name__ ) if "height" not in size or "width" not in size: raise ValueError(f'''Size dict must have keys \'height\' and \'width\'. Got {size.keys()}''' ) return center_crop(__magic_name__ , size=(size["""height"""], size["""width"""]) , data_format=__magic_name__ , **__magic_name__ ) def lowercase__ ( self : Tuple , __magic_name__ : np.ndarray , __magic_name__ : Union[int, float] , __magic_name__ : Optional[Union[str, ChannelDimension]] = None , **__magic_name__ : int , ) -> np.ndarray: """simple docstring""" return rescale(__magic_name__ , scale=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def lowercase__ ( self : Any , __magic_name__ : np.ndarray , __magic_name__ : Union[float, List[float]] , __magic_name__ : Union[float, List[float]] , __magic_name__ : Optional[Union[str, ChannelDimension]] = None , **__magic_name__ : int , ) -> np.ndarray: """simple docstring""" return normalize(__magic_name__ , mean=__magic_name__ , std=__magic_name__ , data_format=__magic_name__ , **__magic_name__ ) def lowercase__ ( self : Union[str, Any] , __magic_name__ : ImageInput , __magic_name__ : Optional[bool] = None , __magic_name__ : Optional[Dict[str, int]] = None , __magic_name__ : PILImageResampling = None , __magic_name__ : Optional[bool] = None , __magic_name__ : Optional[Dict[str, int]] = None , __magic_name__ : Optional[bool] = None , __magic_name__ : Optional[float] = None , __magic_name__ : Optional[bool] = None , __magic_name__ : Optional[Union[float, Iterable[float]]] = None , __magic_name__ : Optional[Union[float, Iterable[float]]] = None , __magic_name__ : Optional[TensorType] = None , __magic_name__ : ChannelDimension = ChannelDimension.FIRST , **__magic_name__ : str , ) -> BatchFeature: """simple docstring""" __snake_case : str = do_resize if do_resize is not None else self.do_resize __snake_case : List[str] = resample if resample is not None else self.resample __snake_case : Optional[int] = do_center_crop if do_center_crop is not None else self.do_center_crop __snake_case : Union[str, Any] = do_rescale if do_rescale is not None else self.do_rescale __snake_case : Optional[int] = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case : List[str] = do_normalize if do_normalize is not None else self.do_normalize __snake_case : List[str] = image_mean if image_mean is not None else self.image_mean __snake_case : Optional[Any] = image_std if image_std is not None else self.image_std __snake_case : Optional[int] = size if size is not None else self.size __snake_case : Optional[int] = get_size_dict(__magic_name__ , default_to_square=__magic_name__ ) __snake_case : Tuple = crop_size if crop_size is not None else self.crop_size __snake_case : Tuple = get_size_dict(__magic_name__ , param_name="""crop_size""" ) __snake_case : int = make_list_of_images(__magic_name__ ) if not valid_images(__magic_name__ ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # All transformations expect numpy arrays. __snake_case : Optional[int] = [to_numpy_array(__magic_name__ ) for image in images] if do_resize: __snake_case : Any = [self.resize(__magic_name__ , __magic_name__ , __magic_name__ ) for image in images] if do_center_crop: __snake_case : Any = [self.center_crop(__magic_name__ , __magic_name__ ) for image in images] if do_rescale: __snake_case : Any = [self.rescale(__magic_name__ , __magic_name__ ) for image in images] if do_normalize: __snake_case : List[Any] = [self.normalize(__magic_name__ , __magic_name__ , __magic_name__ ) for image in images] __snake_case : List[str] = [to_channel_dimension_format(__magic_name__ , __magic_name__ ) for image in images] __snake_case : Optional[Any] = {"""pixel_values""": images} return BatchFeature(data=__magic_name__ , tensor_type=__magic_name__ )
26
"""simple docstring""" from argparse import ArgumentParser, Namespace from typing import Any, List, Optional from ..pipelines import Pipeline, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand try: from fastapi import Body, FastAPI, HTTPException from fastapi.routing import APIRoute from pydantic import BaseModel from starlette.responses import JSONResponse from uvicorn import run __SCREAMING_SNAKE_CASE : List[Any] = True except (ImportError, AttributeError): __SCREAMING_SNAKE_CASE : List[Any] = object def lowerCAmelCase_( *lowercase_ : Dict , **lowercase_ : str ) -> str: pass __SCREAMING_SNAKE_CASE : Tuple = False __SCREAMING_SNAKE_CASE : Any = logging.get_logger('''transformers-cli/serving''') def lowerCAmelCase_( lowercase_ : Namespace ) -> List[Any]: _lowerCamelCase = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) return ServeCommand(lowercase_ , args.host , args.port , args.workers ) class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : dict class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] lowercase__ : Optional[List[int]] class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : str class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Any class lowerCamelCase_( A__ ): '''simple docstring''' @staticmethod def snake_case__ ( lowerCamelCase__ ): _lowerCamelCase = parser.add_parser( '''serve''' , help='''CLI tool to run inference requests through REST and GraphQL endpoints.''' ) serve_parser.add_argument( '''--task''' , type=lowerCamelCase__ , choices=get_supported_tasks() , help='''The task to run the pipeline on''' , ) serve_parser.add_argument('''--host''' , type=lowerCamelCase__ , default='''localhost''' , help='''Interface the server will listen on.''' ) serve_parser.add_argument('''--port''' , type=lowerCamelCase__ , default=8_8_8_8 , help='''Port the serving will listen to.''' ) serve_parser.add_argument('''--workers''' , type=lowerCamelCase__ , default=1 , help='''Number of http workers''' ) serve_parser.add_argument('''--model''' , type=lowerCamelCase__ , help='''Model\'s name or path to stored model.''' ) serve_parser.add_argument('''--config''' , type=lowerCamelCase__ , help='''Model\'s config name or path to stored model.''' ) serve_parser.add_argument('''--tokenizer''' , type=lowerCamelCase__ , help='''Tokenizer name to use.''' ) serve_parser.add_argument( '''--device''' , type=lowerCamelCase__ , default=-1 , help='''Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)''' , ) serve_parser.set_defaults(func=lowerCamelCase__ ) def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = pipeline _lowerCamelCase = host _lowerCamelCase = port _lowerCamelCase = workers if not _serve_dependencies_installed: raise RuntimeError( '''Using serve command requires FastAPI and uvicorn. ''' '''Please install transformers with [serving]: pip install "transformers[serving]".''' '''Or install FastAPI and uvicorn separately.''' ) else: logger.info(F"""Serving model over {host}:{port}""" ) _lowerCamelCase = FastAPI( routes=[ APIRoute( '''/''' , self.model_info , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''GET'''] , ), APIRoute( '''/tokenize''' , self.tokenize , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), APIRoute( '''/detokenize''' , self.detokenize , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), APIRoute( '''/forward''' , self.forward , response_model=lowerCamelCase__ , response_class=lowerCamelCase__ , methods=['''POST'''] , ), ] , timeout=6_0_0 , ) def snake_case__ ( self ): run(self._app , host=self.host , port=self.port , workers=self.workers ) def snake_case__ ( self ): return ServeModelInfoResult(infos=vars(self._pipeline.model.config ) ) def snake_case__ ( self , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) ): try: _lowerCamelCase = self._pipeline.tokenizer.tokenize(lowerCamelCase__ ) if return_ids: _lowerCamelCase = self._pipeline.tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) return ServeTokenizeResult(tokens=lowerCamelCase__ , tokens_ids=lowerCamelCase__ ) else: return ServeTokenizeResult(tokens=lowerCamelCase__ ) except Exception as e: raise HTTPException(status_code=5_0_0 , detail={'''model''': '''''', '''error''': str(lowerCamelCase__ )} ) def snake_case__ ( self , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , lowerCamelCase__ = Body(lowerCamelCase__ , embed=lowerCamelCase__ ) , ): try: _lowerCamelCase = self._pipeline.tokenizer.decode(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) return ServeDeTokenizeResult(model='''''' , text=lowerCamelCase__ ) except Exception as e: raise HTTPException(status_code=5_0_0 , detail={'''model''': '''''', '''error''': str(lowerCamelCase__ )} ) async def snake_case__ ( self , lowerCamelCase__=Body(lowerCamelCase__ , embed=lowerCamelCase__ ) ): # Check we don't have empty string if len(lowerCamelCase__ ) == 0: return ServeForwardResult(output=[] , attention=[] ) try: # Forward through the model _lowerCamelCase = self._pipeline(lowerCamelCase__ ) return ServeForwardResult(output=lowerCamelCase__ ) except Exception as e: raise HTTPException(5_0_0 , {'''error''': str(lowerCamelCase__ )} )
661
0
def __lowerCAmelCase( _SCREAMING_SNAKE_CASE ) -> bool: """simple docstring""" _A = n ** (1 / 3) return (val * val * val) == n if __name__ == "__main__": print(perfect_cube(27)) print(perfect_cube(4))
27
"""simple docstring""" import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def lowerCAmelCase_( lowercase_ : str , lowercase_ : Dict ) -> List[Any]: assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : str , lowercase_ : str ) -> List[Any]: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ , keep_in_memory=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : int ) -> Optional[int]: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = features.copy() if features else default_expected_features _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = JsonDatasetReader(lowercase_ , features=lowercase_ , cache_dir=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''}, ] , ) def lowerCAmelCase_( lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : List[Any] ) -> Tuple: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_3''': '''float64''', '''col_1''': '''string''', '''col_2''': '''int64'''} _lowerCamelCase = features.copy() if features else default_expected_features _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = JsonDatasetReader(lowercase_ , features=lowercase_ , cache_dir=lowercase_ ).read() assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def lowerCAmelCase_( lowercase_ : Any , lowercase_ : Optional[int] ) -> List[str]: # jsonl_312_path features are {"col_3": "float64", "col_1": "string", "col_2": "int64"} _lowerCamelCase = {'''col_2''': '''int64''', '''col_3''': '''float64''', '''col_1''': '''string'''} _lowerCamelCase = features.copy() _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = JsonDatasetReader(lowercase_ , features=lowercase_ , cache_dir=lowercase_ ).read() assert isinstance(lowercase_ , lowercase_ ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : Optional[Any] ) -> int: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ , split=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) assert dataset.split == split if split else "train" @pytest.mark.parametrize('''path_type''' , [str, list] ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[int] , lowercase_ : Tuple ) -> Optional[int]: if issubclass(lowercase_ , lowercase_ ): _lowerCamelCase = jsonl_path elif issubclass(lowercase_ , lowercase_ ): _lowerCamelCase = [jsonl_path] _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ ).read() _check_json_dataset(lowercase_ , lowercase_ ) def lowerCAmelCase_( lowercase_ : int , lowercase_ : List[str] , lowercase_ : str=("train",) ) -> List[str]: assert isinstance(lowercase_ , lowercase_ ) for split in splits: _lowerCamelCase = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize('''keep_in_memory''' , [False, True] ) def lowerCAmelCase_( lowercase_ : List[str] , lowercase_ : Optional[int] , lowercase_ : int ) -> Dict: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): _lowerCamelCase = JsonDatasetReader({'''train''': jsonl_path} , cache_dir=lowercase_ , keep_in_memory=lowercase_ ).read() _check_json_datasetdict(lowercase_ , lowercase_ ) @pytest.mark.parametrize( '''features''' , [ None, {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''}, {'''col_1''': '''string''', '''col_2''': '''string''', '''col_3''': '''string'''}, {'''col_1''': '''int32''', '''col_2''': '''int32''', '''col_3''': '''int32'''}, {'''col_1''': '''float32''', '''col_2''': '''float32''', '''col_3''': '''float32'''}, ] , ) def lowerCAmelCase_( lowercase_ : Tuple , lowercase_ : Union[str, Any] , lowercase_ : Optional[Any] ) -> List[Any]: _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = features.copy() if features else default_expected_features _lowerCamelCase = ( Features({feature: Value(lowercase_ ) for feature, dtype in features.items()} ) if features is not None else None ) _lowerCamelCase = JsonDatasetReader({'''train''': jsonl_path} , features=lowercase_ , cache_dir=lowercase_ ).read() _check_json_datasetdict(lowercase_ , lowercase_ ) @pytest.mark.parametrize('''split''' , [None, NamedSplit('''train''' ), '''train''', '''test'''] ) def lowerCAmelCase_( lowercase_ : Optional[Any] , lowercase_ : str , lowercase_ : List[str] ) -> Optional[Any]: if split: _lowerCamelCase = {split: jsonl_path} else: _lowerCamelCase = '''train''' _lowerCamelCase = {'''train''': jsonl_path, '''test''': jsonl_path} _lowerCamelCase = tmp_path / '''cache''' _lowerCamelCase = {'''col_1''': '''string''', '''col_2''': '''int64''', '''col_3''': '''float64'''} _lowerCamelCase = JsonDatasetReader(lowercase_ , cache_dir=lowercase_ ).read() _check_json_datasetdict(lowercase_ , lowercase_ , splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def lowerCAmelCase_( lowercase_ : Optional[int] ) -> Optional[Any]: return json.load(lowercase_ ) def lowerCAmelCase_( lowercase_ : Tuple ) -> Tuple: return [json.loads(lowercase_ ) for line in buffer] class lowerCamelCase_: '''simple docstring''' @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ ).write() buffer.seek(0 ) _lowerCamelCase = load_json_function(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) assert isinstance(exported_content[0] , lowerCamelCase__ ) assert len(lowerCamelCase__ ) == 1_0 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ , orient=lowerCamelCase__ ).write() buffer.seek(0 ) _lowerCamelCase = load_json(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowerCamelCase__ , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 1_0 else: assert len(lowerCamelCase__ ) == 1_0 @pytest.mark.parametrize('''lines, load_json_function''' , [(True, load_json_lines), (False, load_json)] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ , num_proc=2 ).write() buffer.seek(0 ) _lowerCamelCase = load_json_function(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) assert isinstance(exported_content[0] , lowerCamelCase__ ) assert len(lowerCamelCase__ ) == 1_0 @pytest.mark.parametrize( '''orient, container, keys, len_at''' , [ ('''records''', list, {'''tokens''', '''labels''', '''answers''', '''id'''}, None), ('''split''', dict, {'''columns''', '''data'''}, '''data'''), ('''index''', dict, set('''0123456789''' ), None), ('''columns''', dict, {'''tokens''', '''labels''', '''answers''', '''id'''}, '''tokens'''), ('''values''', list, None, None), ('''table''', dict, {'''schema''', '''data'''}, '''data'''), ] , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , lines=lowerCamelCase__ , orient=lowerCamelCase__ , num_proc=2 ).write() buffer.seek(0 ) _lowerCamelCase = load_json(lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowerCamelCase__ , '''keys''' ) and not hasattr(exported_content[0] , '''keys''' ) if len_at: assert len(exported_content[len_at] ) == 1_0 else: assert len(lowerCamelCase__ ) == 1_0 def snake_case__ ( self , lowerCamelCase__ ): with pytest.raises(lowerCamelCase__ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , num_proc=0 ) @pytest.mark.parametrize('''compression, extension''' , [('''gzip''', '''gz'''), ('''bz2''', '''bz2'''), ('''xz''', '''xz''')] ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = tmp_path_factory.mktemp('''data''' ) / F"""test.json.{extension}""" _lowerCamelCase = str(shared_datadir / F"""test_file.json.{extension}""" ) JsonDatasetWriter(lowerCamelCase__ , lowerCamelCase__ , compression=lowerCamelCase__ ).write() with fsspec.open(lowerCamelCase__ , '''rb''' , compression='''infer''' ) as f: _lowerCamelCase = f.read() with fsspec.open(lowerCamelCase__ , '''rb''' , compression='''infer''' ) as f: _lowerCamelCase = f.read() assert exported_content == original_content
661
0
'''simple docstring''' import unittest from transformers import BigBirdTokenizer, BigBirdTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin UpperCamelCase_ = "▁" UpperCamelCase_ = get_tests_dir("fixtures/test_sentencepiece.model") @require_sentencepiece @require_tokenizers class _a ( SCREAMING_SNAKE_CASE , unittest.TestCase ): '''simple docstring''' A : Dict = BigBirdTokenizer A : Any = BigBirdTokenizerFast A : Optional[int] = True A : Any = True def UpperCamelCase_ ( self ): '''simple docstring''' super().setUp() SCREAMING_SNAKE_CASE : Tuple = self.tokenizer_class(A, keep_accents=A ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[Any] = '<s>' SCREAMING_SNAKE_CASE : int = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(A ), A ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(A ), A ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0], '<unk>' ) self.assertEqual(vocab_keys[1], '<s>' ) self.assertEqual(vocab_keys[-1], '[MASK]' ) self.assertEqual(len(A ), 1_004 ) def UpperCamelCase_ ( self ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size, 1_000 ) def UpperCamelCase_ ( self ): '''simple docstring''' if not self.test_rust_tokenizer: return SCREAMING_SNAKE_CASE : Union[str, Any] = self.get_tokenizer() SCREAMING_SNAKE_CASE : Any = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE : List[str] = 'I was born in 92000, and this is falsé.' SCREAMING_SNAKE_CASE : List[str] = tokenizer.tokenize(A ) SCREAMING_SNAKE_CASE : Optional[Any] = rust_tokenizer.tokenize(A ) self.assertListEqual(A, A ) SCREAMING_SNAKE_CASE : Dict = tokenizer.encode(A, add_special_tokens=A ) SCREAMING_SNAKE_CASE : Optional[int] = rust_tokenizer.encode(A, add_special_tokens=A ) self.assertListEqual(A, A ) SCREAMING_SNAKE_CASE : Optional[Any] = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE : Tuple = tokenizer.encode(A ) SCREAMING_SNAKE_CASE : Dict = rust_tokenizer.encode(A ) self.assertListEqual(A, A ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : int = BigBirdTokenizer(A, keep_accents=A ) SCREAMING_SNAKE_CASE : int = tokenizer.tokenize('This is a test' ) self.assertListEqual(A, ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(A ), [285, 46, 10, 170, 382], ) SCREAMING_SNAKE_CASE : Optional[int] = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( A, [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ], ) SCREAMING_SNAKE_CASE : List[Any] = tokenizer.convert_tokens_to_ids(A ) self.assertListEqual( A, [8, 21, 84, 55, 24, 19, 7, 0, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 0, 4], ) SCREAMING_SNAKE_CASE : str = tokenizer.convert_ids_to_tokens(A ) self.assertListEqual( A, [ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ], ) @cached_property def UpperCamelCase_ ( self ): '''simple docstring''' return BigBirdTokenizer.from_pretrained('google/bigbird-roberta-base' ) @slow def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[Any] = 'Hello World!' SCREAMING_SNAKE_CASE : int = [65, 18_536, 2_260, 101, 66] self.assertListEqual(A, self.big_tokenizer.encode(A ) ) @slow def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) # fmt: off SCREAMING_SNAKE_CASE : int = [65, 871, 419, 358, 946, 991, 2_521, 452, 358, 1_357, 387, 7_751, 3_536, 112, 985, 456, 126, 865, 938, 5_400, 5_734, 458, 1_368, 467, 786, 2_462, 5_246, 1_159, 633, 865, 4_519, 457, 582, 852, 2_557, 427, 916, 508, 405, 34_324, 497, 391, 408, 11_342, 1_244, 385, 100, 938, 985, 456, 574, 362, 12_597, 3_200, 3_129, 1_172, 66] # noqa: E231 # fmt: on self.assertListEqual(A, self.big_tokenizer.encode(A ) ) @require_torch @slow def UpperCamelCase_ ( self ): '''simple docstring''' import torch from transformers import BigBirdConfig, BigBirdModel # Build sequence SCREAMING_SNAKE_CASE : str = list(self.big_tokenizer.get_vocab().keys() )[:10] SCREAMING_SNAKE_CASE : Any = ' '.join(A ) SCREAMING_SNAKE_CASE : str = self.big_tokenizer.encode_plus(A, return_tensors='pt', return_token_type_ids=A ) SCREAMING_SNAKE_CASE : List[str] = self.big_tokenizer.batch_encode_plus( [sequence + ' ' + sequence], return_tensors='pt', return_token_type_ids=A ) SCREAMING_SNAKE_CASE : Optional[int] = BigBirdConfig(attention_type='original_full' ) SCREAMING_SNAKE_CASE : Optional[int] = BigBirdModel(A ) assert model.get_input_embeddings().weight.shape[0] >= self.big_tokenizer.vocab_size with torch.no_grad(): model(**A ) model(**A ) @slow def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Tuple = BigBirdTokenizer.from_pretrained('google/bigbird-roberta-base' ) SCREAMING_SNAKE_CASE : List[Any] = tokenizer.decode(tokenizer('Paris is the [MASK].' ).input_ids ) self.assertTrue(decoded_text == '[CLS] Paris is the[MASK].[SEP]' ) @slow def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = {'input_ids': [[65, 39_286, 458, 36_335, 2_001, 456, 13_073, 13_266, 455, 113, 7_746, 1_741, 11_157, 391, 13_073, 13_266, 455, 113, 3_967, 35_412, 113, 4_936, 109, 3_870, 2_377, 113, 30_084, 45_720, 458, 134, 17_496, 112, 503, 11_672, 113, 118, 112, 5_665, 13_347, 38_687, 112, 1_496, 31_389, 112, 3_268, 47_264, 134, 962, 112, 16_377, 8_035, 23_130, 430, 12_169, 15_518, 28_592, 458, 146, 41_697, 109, 391, 12_169, 15_518, 16_689, 458, 146, 41_358, 109, 452, 726, 4_034, 111, 763, 35_412, 5_082, 388, 1_903, 111, 9_051, 391, 2_870, 48_918, 1_900, 1_123, 550, 998, 112, 9_586, 15_985, 455, 391, 410, 22_955, 37_636, 114, 66], [65, 448, 17_496, 419, 3_663, 385, 763, 113, 27_533, 2_870, 3_283, 13_043, 1_639, 24_713, 523, 656, 24_013, 18_550, 2_521, 517, 27_014, 21_244, 420, 1_212, 1_465, 391, 927, 4_833, 388, 578, 11_786, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [65, 484, 2_169, 7_687, 21_932, 18_146, 726, 363, 17_032, 3_391, 114, 66, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=A, model_name='google/bigbird-roberta-base', revision='215c99f1600e06f83acce68422f2035b2b5c3510', )
28
"""simple docstring""" import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=3 , lowerCamelCase__=3_2 , lowerCamelCase__=3 , lowerCamelCase__=1_0 , lowerCamelCase__=[8, 1_6, 3_2, 6_4] , lowerCamelCase__=[1, 1, 2, 1] , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__="relu" , lowerCamelCase__=3 , lowerCamelCase__=None , lowerCamelCase__=["stage2", "stage3", "stage4"] , lowerCamelCase__=[2, 3, 4] , lowerCamelCase__=1 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = image_size _lowerCamelCase = num_channels _lowerCamelCase = embeddings_size _lowerCamelCase = hidden_sizes _lowerCamelCase = depths _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_act _lowerCamelCase = num_labels _lowerCamelCase = scope _lowerCamelCase = len(lowerCamelCase__ ) _lowerCamelCase = out_features _lowerCamelCase = out_indices _lowerCamelCase = num_groups def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.num_labels ) _lowerCamelCase = self.get_config() return config, pixel_values, labels def snake_case__ ( self ): return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = BitModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = self.num_labels _lowerCamelCase = BitForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = BitBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None _lowerCamelCase = None _lowerCamelCase = BitBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = config_and_inputs _lowerCamelCase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Dict = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () lowercase__ : Any = ( {'feature-extraction': BitModel, 'image-classification': BitForImageClassification} if is_torch_available() else {} ) lowercase__ : Union[str, Any] = False lowercase__ : List[Any] = False lowercase__ : Any = False lowercase__ : List[str] = False lowercase__ : Any = False def snake_case__ ( self ): _lowerCamelCase = BitModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ ) def snake_case__ ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def snake_case__ ( self ): return @unittest.skip(reason='''Bit does not output attentions''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''Bit does not use inputs_embeds''' ) def snake_case__ ( self ): pass @unittest.skip(reason='''Bit does not support input and output embeddings''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(config=lowerCamelCase__ ) for name, module in model.named_modules(): if isinstance(lowerCamelCase__ , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) def snake_case__ ( self ): def check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): _lowerCamelCase = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) _lowerCamelCase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _lowerCamelCase = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase__ ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() _lowerCamelCase = ['''preactivation''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: _lowerCamelCase = layer_type _lowerCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _lowerCamelCase = True check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) @unittest.skip(reason='''Bit does not use feedforward chunking''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = BitModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> List[Any]: _lowerCamelCase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def snake_case__ ( self ): _lowerCamelCase = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowerCamelCase__ ) _lowerCamelCase = self.default_image_processor _lowerCamelCase = prepare_img() _lowerCamelCase = image_processor(images=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**lowerCamelCase__ ) # verify the logits _lowerCamelCase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([[-0.6_5_2_6, -0.5_2_6_3, -1.4_3_9_8]] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) ) @require_torch class lowerCamelCase_( A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Optional[Any] = (BitBackbone,) if is_torch_available() else () lowercase__ : Tuple = BitConfig lowercase__ : Any = False def snake_case__ ( self ): _lowerCamelCase = BitModelTester(self )
661
0
"""simple docstring""" import argparse import torch from huggingface_hub import hf_hub_download from transformers import AutoTokenizer, RobertaPreLayerNormConfig, RobertaPreLayerNormForMaskedLM from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def lowercase ( lowerCAmelCase__ ,lowerCAmelCase__ ): lowerCamelCase_ = RobertaPreLayerNormConfig.from_pretrained( lowerCAmelCase__ ,architectures=['''RobertaPreLayerNormForMaskedLM'''] ) # convert state_dict lowerCamelCase_ = torch.load(hf_hub_download(repo_id=lowerCAmelCase__ ,filename='''pytorch_model.bin''' ) ) lowerCamelCase_ = {} for tensor_key, tensor_value in original_state_dict.items(): # The transformer implementation gives the model a unique name, rather than overwiriting 'roberta' if tensor_key.startswith('''roberta.''' ): lowerCamelCase_ = '''roberta_prelayernorm.''' + tensor_key[len('''roberta.''' ) :] # The original implementation contains weights which are not used, remove them from the state_dict if tensor_key.endswith('''.self.LayerNorm.weight''' ) or tensor_key.endswith('''.self.LayerNorm.bias''' ): continue lowerCamelCase_ = tensor_value lowerCamelCase_ = RobertaPreLayerNormForMaskedLM.from_pretrained( pretrained_model_name_or_path=lowerCAmelCase__ ,config=lowerCAmelCase__ ,state_dict=lowerCAmelCase__ ) model.save_pretrained(lowerCAmelCase__ ) # convert tokenizer lowerCamelCase_ = AutoTokenizer.from_pretrained(lowerCAmelCase__ ) tokenizer.save_pretrained(lowerCAmelCase__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( """--checkpoint-repo""", default=None, type=str, required=True, help="""Path the official PyTorch dump, e.g. 'andreasmadsen/efficient_mlm_m0.40'.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) A_ = parser.parse_args() convert_roberta_prelayernorm_checkpoint_to_pytorch(args.checkpoint_repo, args.pytorch_dump_folder_path)
29
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__=1_3 , lowerCamelCase__=2 , lowerCamelCase__=2_4 , lowerCamelCase__=1_6 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=3_2 , lowerCamelCase__=5 , lowerCamelCase__=4 , lowerCamelCase__=3_7 , lowerCamelCase__="gelu" , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=1_0 , lowerCamelCase__=0.0_2 , lowerCamelCase__=None , lowerCamelCase__=2 , lowerCamelCase__=2 , ): _lowerCamelCase = parent _lowerCamelCase = batch_size _lowerCamelCase = patch_size _lowerCamelCase = max_length _lowerCamelCase = num_mel_bins _lowerCamelCase = is_training _lowerCamelCase = use_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = intermediate_size _lowerCamelCase = hidden_act _lowerCamelCase = hidden_dropout_prob _lowerCamelCase = attention_probs_dropout_prob _lowerCamelCase = type_sequence_label_size _lowerCamelCase = initializer_range _lowerCamelCase = scope _lowerCamelCase = frequency_stride _lowerCamelCase = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _lowerCamelCase = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _lowerCamelCase = (self.max_length - self.patch_size) // self.time_stride + 1 _lowerCamelCase = frequency_out_dimension * time_out_dimension _lowerCamelCase = num_patches + 2 def snake_case__ ( self ): _lowerCamelCase = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) _lowerCamelCase = None if self.use_labels: _lowerCamelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase = self.get_config() return config, input_values, labels def snake_case__ ( self ): return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCamelCase__ , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = ASTModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() _lowerCamelCase = model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case__ ( self ): _lowerCamelCase = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) = config_and_inputs _lowerCamelCase = {'''input_values''': input_values} return config, inputs_dict @require_torch class lowerCamelCase_( A__, A__, unittest.TestCase ): '''simple docstring''' lowercase__ : Any = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) lowercase__ : List[str] = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) lowercase__ : int = False lowercase__ : str = False lowercase__ : Union[str, Any] = False lowercase__ : List[str] = False def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def snake_case__ ( self ): _lowerCamelCase = ASTModelTester(self ) _lowerCamelCase = ConfigTester(self , config_class=lowerCamelCase__ , has_text_modality=lowerCamelCase__ , hidden_size=3_7 ) def snake_case__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def snake_case__ ( self ): pass def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowerCamelCase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , nn.Linear ) ) def snake_case__ ( self ): _lowerCamelCase , _lowerCamelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase = model_class(lowerCamelCase__ ) _lowerCamelCase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase = [*signature.parameters.keys()] _lowerCamelCase = ['''input_values'''] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) @slow def snake_case__ ( self ): for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase = ASTModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def lowerCAmelCase_( ) -> str: _lowerCamelCase = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) _lowerCamelCase , _lowerCamelCase = torchaudio.load(lowercase_ ) return audio, sampling_rate @require_torch @require_torchaudio class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' @cached_property def snake_case__ ( self ): return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def snake_case__ ( self ): _lowerCamelCase = self.default_feature_extractor _lowerCamelCase = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(lowerCamelCase__ ) _lowerCamelCase = self.default_feature_extractor _lowerCamelCase , _lowerCamelCase = prepare_audio() _lowerCamelCase = audio.squeeze().numpy() _lowerCamelCase = feature_extractor(lowerCamelCase__ , sampling_rate=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): _lowerCamelCase = model(**lowerCamelCase__ ) # verify the logits _lowerCamelCase = torch.Size((1, 5_2_7) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) _lowerCamelCase = torch.tensor([-0.8_7_6_0, -7.0_0_4_2, -8.6_6_0_2] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) )
661
0
import operator as op __a = 'scaler.pt' __a = 'pytorch_model' __a = 'random_states' __a = 'optimizer' __a = 'scheduler' __a = 'pytorch_model.bin' __a = 'pytorch_model.bin.index.json' __a = 'model.safetensors' __a = 'model.safetensors.index.json' __a = '1.10.2' __a = 'py38' __a = '4.17.0' __a = ['ml.p3.16xlarge', 'ml.p3dn.24xlarge', 'ml.p4dn.24xlarge'] __a = ['FULL_SHARD', 'SHARD_GRAD_OP', 'NO_SHARD', 'HYBRID_SHARD', 'HYBRID_SHARD_ZERO2'] __a = ['TRANSFORMER_BASED_WRAP', 'SIZE_BASED_WRAP', 'NO_WRAP'] __a = ['BACKWARD_PRE', 'BACKWARD_POST', 'NO_PREFETCH'] __a = ['FULL_STATE_DICT', 'LOCAL_STATE_DICT', 'SHARDED_STATE_DICT'] __a = '2.0.1' __a = ['pdsh', 'standard', 'openmpi', 'mvapich'] __a = ['default', 'reduce-overhead', 'max-autotune'] __a = {'>': op.gt, '>=': op.ge, '==': op.eq, '!=': op.ne, '<=': op.le, '<': op.lt} # These are the args for `torch.distributed.launch` for pytorch < 1.9 __a = [ 'nnodes', 'nproc_per_node', 'rdzv_backend', 'rdzv_endpoint', 'rdzv_id', 'rdzv_conf', 'standalone', 'max_restarts', 'monitor_interval', 'start_method', 'role', 'module', 'm', 'no_python', 'run_path', 'log_dir', 'r', 'redirects', 't', 'tee', 'node_rank', 'master_addr', 'master_port', ] __a = ['DEEPSPEED', 'MULTI_GPU', 'FSDP', 'MEGATRON_LM'] __a = ['DEEPSPEED', 'MULTI_XPU', 'FSDP']
30
"""simple docstring""" import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class lowerCamelCase_: '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return None class lowerCamelCase_: '''simple docstring''' def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): return None class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' lowercase__ : Tuple = [ # (model_name, model_kwargs) ('bert-base-cased', {}), ('gpt2', {'use_cache': False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase__ , '''tf''' , 1_2 , **lowerCamelCase__ ) @require_torch @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , **lowerCamelCase__ ) @require_torch @slow def snake_case__ ( self ): from transformers import BertModel _lowerCamelCase = ['''[UNK]''', '''[SEP]''', '''[CLS]''', '''[PAD]''', '''[MASK]''', '''some''', '''other''', '''words'''] with NamedTemporaryFile(mode='''w+t''' ) as vocab_file: vocab_file.write('''\n'''.join(lowerCamelCase__ ) ) vocab_file.flush() _lowerCamelCase = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: _lowerCamelCase = BertModel(BertConfig(vocab_size=len(lowerCamelCase__ ) ) ) model.save_pretrained(lowerCamelCase__ ) self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , lowerCamelCase__ ) @require_tf @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCamelCase = self._test_export(lowerCamelCase__ , '''tf''' , 1_2 , **lowerCamelCase__ ) _lowerCamelCase = quantize(Path(lowerCamelCase__ ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) @require_torch @slow def snake_case__ ( self ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: _lowerCamelCase = self._test_export(lowerCamelCase__ , '''pt''' , 1_2 , **lowerCamelCase__ ) _lowerCamelCase = quantize(lowerCamelCase__ ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(lowerCamelCase__ ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=None , **lowerCamelCase__ ): try: # Compute path with TemporaryDirectory() as tempdir: _lowerCamelCase = Path(lowerCamelCase__ ).joinpath('''model.onnx''' ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , **lowerCamelCase__ ) return path except Exception as e: self.fail(lowerCamelCase__ ) @require_torch @require_tokenizers @slow def snake_case__ ( self ): from transformers import BertModel _lowerCamelCase = BertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) _lowerCamelCase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(lowerCamelCase__ , lowerCamelCase__ , '''pt''' ) @require_tf @require_tokenizers @slow def snake_case__ ( self ): from transformers import TFBertModel _lowerCamelCase = TFBertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) _lowerCamelCase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(lowerCamelCase__ , lowerCamelCase__ , '''tf''' ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = FeatureExtractionPipeline(lowerCamelCase__ , lowerCamelCase__ ) _lowerCamelCase = ['''input_ids''', '''token_type_ids''', '''attention_mask''', '''output_0''', '''output_1'''] _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = infer_shapes(lowerCamelCase__ , lowerCamelCase__ ) # Assert all variables are present self.assertEqual(len(lowerCamelCase__ ) , len(lowerCamelCase__ ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] , lowerCamelCase__ ) self.assertSequenceEqual(variable_names[3:] , lowerCamelCase__ ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] , {0: '''batch''', 1: '''sequence'''} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes['''output_0'''] , {0: '''batch''', 1: '''sequence'''} ) self.assertDictEqual(shapes['''output_1'''] , {0: '''batch'''} ) def snake_case__ ( self ): _lowerCamelCase = ['''input_ids''', '''attention_mask''', '''token_type_ids'''] _lowerCamelCase = {'''input_ids''': [1, 2, 3, 4], '''attention_mask''': [0, 0, 0, 0], '''token_type_ids''': [1, 1, 1, 1]} _lowerCamelCase , _lowerCamelCase = ensure_valid_input(FuncContiguousArgs() , lowerCamelCase__ , lowerCamelCase__ ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(lowerCamelCase__ ) , 3 ) # Should have exactly the same input names self.assertEqual(set(lowerCamelCase__ ) , set(lowerCamelCase__ ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(lowerCamelCase__ , (tokens['''input_ids'''], tokens['''token_type_ids'''], tokens['''attention_mask''']) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) _lowerCamelCase , _lowerCamelCase = ensure_valid_input(FuncNonContiguousArgs() , lowerCamelCase__ , lowerCamelCase__ ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(lowerCamelCase__ ) , 1 ) self.assertEqual(len(lowerCamelCase__ ) , 1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] , tokens['''input_ids'''] ) self.assertEqual(ordered_input_names[0] , '''input_ids''' ) def snake_case__ ( self ): _lowerCamelCase = generate_identified_filename(Path('''/home/something/my_fake_model.onnx''' ) , '''-test''' ) self.assertEqual('''/home/something/my_fake_model-test.onnx''' , generated.as_posix() )
661
0
def UpperCAmelCase_ ( __UpperCAmelCase : int = 50 ) -> int: SCREAMING_SNAKE_CASE_ = [1] * (length + 1) for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): ways_number[row_length] += ways_number[ row_length - tile_start - tile_length ] return ways_number[length] if __name__ == "__main__": print(f'''{solution() = }''')
31
"""simple docstring""" __SCREAMING_SNAKE_CASE : Optional[Any] = {str(digit): digit**5 for digit in range(1_0)} def lowerCAmelCase_( lowercase_ : int ) -> int: return sum(DIGITS_FIFTH_POWER[digit] for digit in str(lowercase_ ) ) def lowerCAmelCase_( ) -> int: return sum( number for number in range(10_00 , 1_00_00_00 ) if number == digits_fifth_powers_sum(lowercase_ ) ) if __name__ == "__main__": print(solution())
661
0
def A__ ( SCREAMING_SNAKE_CASE_ : list ) -> list: """simple docstring""" _UpperCAmelCase = len(SCREAMING_SNAKE_CASE_ ) for i in range(1 , SCREAMING_SNAKE_CASE_ ): _UpperCAmelCase = collection[i] _UpperCAmelCase = 0 _UpperCAmelCase = i - 1 while low <= high: _UpperCAmelCase = (low + high) // 2 if val < collection[mid]: _UpperCAmelCase = mid - 1 else: _UpperCAmelCase = mid + 1 for j in range(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , -1 ): _UpperCAmelCase = collection[j - 1] _UpperCAmelCase = val return collection if __name__ == "__main__": UpperCAmelCase_ = input("Enter numbers separated by a comma:\n").strip() UpperCAmelCase_ = [int(item) for item in user_input.split(",")] print(binary_insertion_sort(unsorted))
32
"""simple docstring""" from __future__ import annotations import os from collections.abc import Mapping __SCREAMING_SNAKE_CASE : str = tuple[int, int] class lowerCamelCase_: '''simple docstring''' def __init__( self , lowerCamelCase__ , lowerCamelCase__ ): _lowerCamelCase = vertices _lowerCamelCase = { (min(lowerCamelCase__ ), max(lowerCamelCase__ )): weight for edge, weight in edges.items() } def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): self.vertices.add(edge[0] ) self.vertices.add(edge[1] ) _lowerCamelCase = weight def snake_case__ ( self ): _lowerCamelCase = Graph({min(self.vertices )} , {} ) _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 while len(subgraph.vertices ) < len(self.vertices ): _lowerCamelCase = max(self.edges.values() ) + 1 for edge, weight in self.edges.items(): if (edge[0] in subgraph.vertices) ^ (edge[1] in subgraph.vertices): if weight < min_weight: _lowerCamelCase = edge _lowerCamelCase = weight subgraph.add_edge(lowerCamelCase__ , lowerCamelCase__ ) return subgraph def lowerCAmelCase_( lowercase_ : str = "p107_network.txt" ) -> int: _lowerCamelCase = os.path.abspath(os.path.dirname(lowercase_ ) ) _lowerCamelCase = os.path.join(lowercase_ , lowercase_ ) _lowerCamelCase = {} _lowerCamelCase = 42 _lowerCamelCase = 42 _lowerCamelCase = 42 with open(lowercase_ ) as f: _lowerCamelCase = f.read().strip().split('''\n''' ) _lowerCamelCase = [line.split(''',''' ) for line in data] for edgea in range(1 , len(lowercase_ ) ): for edgea in range(lowercase_ ): if adjaceny_matrix[edgea][edgea] != "-": _lowerCamelCase = int(adjaceny_matrix[edgea][edgea] ) _lowerCamelCase = Graph(set(range(len(lowercase_ ) ) ) , lowercase_ ) _lowerCamelCase = graph.prims_algorithm() _lowerCamelCase = sum(graph.edges.values() ) _lowerCamelCase = sum(subgraph.edges.values() ) return initial_total - optimal_total if __name__ == "__main__": print(F"""{solution() = }""")
661
0
import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Union[str, Any]: # Construct model if openai_config_file == "": snake_case__ = OpenAIGPTConfig() else: snake_case__ = OpenAIGPTConfig.from_json_file(__lowerCAmelCase ) snake_case__ = OpenAIGPTModel(__lowerCAmelCase ) # Load weights from numpy load_tf_weights_in_openai_gpt(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # Save pytorch-model snake_case__ = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME snake_case__ = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(F"""Save PyTorch model to {pytorch_weights_dump_path}""" ) torch.save(model.state_dict() , __lowerCAmelCase ) print(F"""Save configuration file to {pytorch_config_dump_path}""" ) with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": lowerCamelCase__ : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( """--openai_checkpoint_folder_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--openai_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained OpenAI model. \n""" """This specifies the model architecture.""" ), ) lowerCamelCase__ : Any = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
33
"""simple docstring""" from __future__ import annotations from math import ceil, floor, sqrt def lowerCAmelCase_( lowercase_ : int = 2_00_00_00 ) -> int: _lowerCamelCase = [0] _lowerCamelCase = 42 for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target _lowerCamelCase = 0 # the area corresponding to the grid that gives the product closest to target _lowerCamelCase = 0 # an estimate of b, using the quadratic formula _lowerCamelCase = 42 # the largest integer less than b_estimate _lowerCamelCase = 42 # the largest integer less than b_estimate _lowerCamelCase = 42 # the triangle number corresponding to b_floor _lowerCamelCase = 42 # the triangle number corresponding to b_ceil _lowerCamelCase = 42 for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): _lowerCamelCase = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 _lowerCamelCase = floor(lowercase_ ) _lowerCamelCase = ceil(lowercase_ ) _lowerCamelCase = triangle_numbers[b_floor] _lowerCamelCase = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): _lowerCamelCase = triangle_b_first_guess * triangle_a _lowerCamelCase = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): _lowerCamelCase = triangle_b_second_guess * triangle_a _lowerCamelCase = idx_a * b_ceil return area if __name__ == "__main__": print(F"""{solution() = }""")
661
0
"""simple docstring""" import datasets from .evaluate import evaluate SCREAMING_SNAKE_CASE_ = '\\n@inproceedings{Rajpurkar2016SQuAD10,\n title={SQuAD: 100, 000+ Questions for Machine Comprehension of Text},\n author={Pranav Rajpurkar and Jian Zhang and Konstantin Lopyrev and Percy Liang},\n booktitle={EMNLP},\n year={2016}\n}\n' SCREAMING_SNAKE_CASE_ = '\nThis metric wrap the official scoring script for version 1 of the Stanford Question Answering Dataset (SQuAD).\n\nStanford Question Answering Dataset (SQuAD) is a reading comprehension dataset, consisting of questions posed by\ncrowdworkers on a set of Wikipedia articles, where the answer to every question is a segment of text, or span,\nfrom the corresponding reading passage, or the question might be unanswerable.\n' SCREAMING_SNAKE_CASE_ = '\nComputes SQuAD scores (F1 and EM).\nArgs:\n predictions: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair as given in the references (see below)\n - \'prediction_text\': the text of the answer\n references: List of question-answers dictionaries with the following key-values:\n - \'id\': id of the question-answer pair (see above),\n - \'answers\': a Dict in the SQuAD dataset format\n {\n \'text\': list of possible texts for the answer, as a list of strings\n \'answer_start\': list of start positions for the answer, as a list of ints\n }\n Note that answer_start values are not taken into account to compute the metric.\nReturns:\n \'exact_match\': Exact match (the normalized answer exactly match the gold answer)\n \'f1\': The F-score of predicted tokens versus the gold answer\nExamples:\n\n >>> predictions = [{\'prediction_text\': \'1976\', \'id\': \'56e10a3be3433e1400422b22\'}]\n >>> references = [{\'answers\': {\'answer_start\': [97], \'text\': [\'1976\']}, \'id\': \'56e10a3be3433e1400422b22\'}]\n >>> squad_metric = datasets.load_metric("squad")\n >>> results = squad_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'exact_match\': 100.0, \'f1\': 100.0}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class snake_case_ ( datasets.Metric ): """simple docstring""" def UpperCAmelCase__ ( self) -> int: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': {'''id''': datasets.Value('''string'''), '''prediction_text''': datasets.Value('''string''')}, '''references''': { '''id''': datasets.Value('''string'''), '''answers''': datasets.features.Sequence( { '''text''': datasets.Value('''string'''), '''answer_start''': datasets.Value('''int32'''), }), }, }) , codebase_urls=['''https://rajpurkar.github.io/SQuAD-explorer/'''] , reference_urls=['''https://rajpurkar.github.io/SQuAD-explorer/'''] , ) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_) -> int: UpperCamelCase = {prediction['''id''']: prediction['''prediction_text'''] for prediction in predictions} UpperCamelCase = [ { '''paragraphs''': [ { '''qas''': [ { '''answers''': [{'''text''': answer_text} for answer_text in ref['''answers''']['''text''']], '''id''': ref['''id'''], } for ref in references ] } ] } ] UpperCamelCase = evaluate(dataset=lowerCamelCase_ , predictions=lowerCamelCase_) return score
34
"""simple docstring""" import warnings from ...utils import logging from .image_processing_deit import DeiTImageProcessor __SCREAMING_SNAKE_CASE : List[Any] = logging.get_logger(__name__) class lowerCamelCase_( A__ ): '''simple docstring''' def __init__( self , *lowerCamelCase__ , **lowerCamelCase__ ): warnings.warn( '''The class DeiTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use DeiTImageProcessor instead.''' , lowerCamelCase__ , ) super().__init__(*lowerCamelCase__ , **lowerCamelCase__ )
661
0
import argparse import json import logging import os import shutil import sys import tempfile import unittest from unittest import mock import torch from accelerate.utils import write_basic_config from transformers.testing_utils import TestCasePlus, get_gpu_count, run_command, slow, torch_device from transformers.utils import is_apex_available logging.basicConfig(level=logging.DEBUG) a_ :Optional[int] = logging.getLogger() def a ( ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[int] = argparse.ArgumentParser() parser.add_argument('''-f''' ) SCREAMING_SNAKE_CASE__ : int = parser.parse_args() return args.f def a ( A__ ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = {} SCREAMING_SNAKE_CASE__ : List[Any] = os.path.join(A__ , '''all_results.json''' ) if os.path.exists(A__ ): with open(A__ , '''r''' ) as f: SCREAMING_SNAKE_CASE__ : List[str] = json.load(A__ ) else: raise ValueError(f"""can't find {path}""" ) return results def a ( ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = torch.cuda.is_available() and torch_device == '''cuda''' return is_using_cuda and is_apex_available() a_ :Any = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class lowercase ( _UpperCAmelCase ): @classmethod def lowercase__ ( cls : int ): # Write Accelerate config, will pick up on CPU, GPU, and multi-GPU SCREAMING_SNAKE_CASE__ : List[Any] = tempfile.mkdtemp() SCREAMING_SNAKE_CASE__ : List[str] = os.path.join(cls.tmpdir , '''default_config.yml''' ) write_basic_config(save_location=cls.configPath ) SCREAMING_SNAKE_CASE__ : Any = ['''accelerate''', '''launch''', '''--config_file''', cls.configPath] @classmethod def lowercase__ ( cls : Any ): shutil.rmtree(cls.tmpdir ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def lowercase__ ( self : Dict ): SCREAMING_SNAKE_CASE__ : Any = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE__ : str = f""" {self.examples_dir}/pytorch/text-classification/run_glue_no_trainer.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --seed=42 --checkpointing_steps epoch --with_tracking """.split() if is_cuda_and_apex_available(): testargs.append('''--fp16''' ) run_command(self._launch_args + testargs ) SCREAMING_SNAKE_CASE__ : Optional[Any] = get_results(_lowercase ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.75 ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''glue_no_trainer''' ) ) ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def lowercase__ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE__ : List[Any] = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE__ : int = f""" {self.examples_dir}/pytorch/language-modeling/run_clm_no_trainer.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --block_size 128 --per_device_train_batch_size 5 --per_device_eval_batch_size 5 --num_train_epochs 2 --output_dir {tmp_dir} --checkpointing_steps epoch --with_tracking """.split() if torch.cuda.device_count() > 1: # Skipping because there are not enough batches to train the model + would need a drop_last to work. return run_command(self._launch_args + testargs ) SCREAMING_SNAKE_CASE__ : List[Any] = get_results(_lowercase ) self.assertLess(result['''perplexity'''] , 1_00 ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''clm_no_trainer''' ) ) ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def lowercase__ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE__ : Tuple = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE__ : Union[str, Any] = f""" {self.examples_dir}/pytorch/language-modeling/run_mlm_no_trainer.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --num_train_epochs=1 --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = get_results(_lowercase ) self.assertLess(result['''perplexity'''] , 42 ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''mlm_no_trainer''' ) ) ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def lowercase__ ( self : List[Any] ): # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu SCREAMING_SNAKE_CASE__ : str = 7 if get_gpu_count() > 1 else 2 SCREAMING_SNAKE_CASE__ : Dict = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE__ : Optional[Any] = f""" {self.examples_dir}/pytorch/token-classification/run_ner_no_trainer.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) SCREAMING_SNAKE_CASE__ : List[Any] = get_results(_lowercase ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.75 ) self.assertLess(result['''train_loss'''] , 0.5 ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''ner_no_trainer''' ) ) ) @unittest.skip(reason='''Fix me @muellerzr''' ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def lowercase__ ( self : Any ): SCREAMING_SNAKE_CASE__ : Optional[Any] = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE__ : Dict = f""" {self.examples_dir}/pytorch/question-answering/run_qa_no_trainer.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --seed=42 --max_train_steps=10 --num_warmup_steps=2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) SCREAMING_SNAKE_CASE__ : Any = get_results(_lowercase ) # Because we use --version_2_with_negative the testing script uses SQuAD v2 metrics. self.assertGreaterEqual(result['''eval_f1'''] , 28 ) self.assertGreaterEqual(result['''eval_exact'''] , 28 ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''qa_no_trainer''' ) ) ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def lowercase__ ( self : Dict ): SCREAMING_SNAKE_CASE__ : str = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE__ : Any = f""" {self.examples_dir}/pytorch/multiple-choice/run_swag_no_trainer.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/swag/sample.json --validation_file tests/fixtures/tests_samples/swag/sample.json --output_dir {tmp_dir} --max_train_steps=20 --num_warmup_steps=2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --with_tracking """.split() run_command(self._launch_args + testargs ) SCREAMING_SNAKE_CASE__ : Tuple = get_results(_lowercase ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.8 ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''swag_no_trainer''' ) ) ) @slow @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def lowercase__ ( self : int ): SCREAMING_SNAKE_CASE__ : Optional[Any] = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE__ : List[Any] = f""" {self.examples_dir}/pytorch/summarization/run_summarization_no_trainer.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --max_train_steps=50 --num_warmup_steps=8 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = get_results(_lowercase ) self.assertGreaterEqual(result['''eval_rouge1'''] , 10 ) self.assertGreaterEqual(result['''eval_rouge2'''] , 2 ) self.assertGreaterEqual(result['''eval_rougeL'''] , 7 ) self.assertGreaterEqual(result['''eval_rougeLsum'''] , 7 ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''summarization_no_trainer''' ) ) ) @slow @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def lowercase__ ( self : Any ): SCREAMING_SNAKE_CASE__ : Optional[Any] = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE__ : str = f""" {self.examples_dir}/pytorch/translation/run_translation_no_trainer.py --model_name_or_path sshleifer/student_marian_en_ro_6_1 --source_lang en --target_lang ro --train_file tests/fixtures/tests_samples/wmt16/sample.json --validation_file tests/fixtures/tests_samples/wmt16/sample.json --output_dir {tmp_dir} --max_train_steps=50 --num_warmup_steps=8 --num_beams=6 --learning_rate=3e-3 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --source_lang en_XX --target_lang ro_RO --checkpointing_steps epoch --with_tracking """.split() run_command(self._launch_args + testargs ) SCREAMING_SNAKE_CASE__ : Any = get_results(_lowercase ) self.assertGreaterEqual(result['''eval_bleu'''] , 30 ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''translation_no_trainer''' ) ) ) @slow def lowercase__ ( self : int ): SCREAMING_SNAKE_CASE__ : int = logging.StreamHandler(sys.stdout ) logger.addHandler(_lowercase ) SCREAMING_SNAKE_CASE__ : List[Any] = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE__ : List[str] = f""" {self.examples_dir}/pytorch/semantic-segmentation/run_semantic_segmentation_no_trainer.py --dataset_name huggingface/semantic-segmentation-test-sample --output_dir {tmp_dir} --max_train_steps=10 --num_warmup_steps=2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --checkpointing_steps epoch """.split() run_command(self._launch_args + testargs ) SCREAMING_SNAKE_CASE__ : Optional[Any] = get_results(_lowercase ) self.assertGreaterEqual(result['''eval_overall_accuracy'''] , 0.10 ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def lowercase__ ( self : List[Any] ): SCREAMING_SNAKE_CASE__ : Any = self.get_auto_remove_tmp_dir() SCREAMING_SNAKE_CASE__ : Dict = f""" {self.examples_dir}/pytorch/image-classification/run_image_classification_no_trainer.py --model_name_or_path google/vit-base-patch16-224-in21k --dataset_name hf-internal-testing/cats_vs_dogs_sample --learning_rate 1e-4 --per_device_train_batch_size 2 --per_device_eval_batch_size 1 --max_train_steps 2 --train_val_split 0.1 --seed 42 --output_dir {tmp_dir} --with_tracking --checkpointing_steps 1 """.split() if is_cuda_and_apex_available(): testargs.append('''--fp16''' ) run_command(self._launch_args + testargs ) SCREAMING_SNAKE_CASE__ : Optional[int] = get_results(_lowercase ) # The base model scores a 25% self.assertGreaterEqual(result['''eval_accuracy'''] , 0.6 ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''step_1''' ) ) ) self.assertTrue(os.path.exists(os.path.join(_lowercase , '''image_classification_no_trainer''' ) ) )
35
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_distilbert import DistilBertTokenizer __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Optional[Any] = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} __SCREAMING_SNAKE_CASE : Union[str, Any] = { '''vocab_file''': { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/vocab.txt''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/vocab.txt''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/vocab.txt''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/vocab.txt''' ), '''distilbert-base-german-cased''': '''https://huggingface.co/distilbert-base-german-cased/resolve/main/vocab.txt''', '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''distilbert-base-uncased''': '''https://huggingface.co/distilbert-base-uncased/resolve/main/tokenizer.json''', '''distilbert-base-uncased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-uncased-distilled-squad/resolve/main/tokenizer.json''' ), '''distilbert-base-cased''': '''https://huggingface.co/distilbert-base-cased/resolve/main/tokenizer.json''', '''distilbert-base-cased-distilled-squad''': ( '''https://huggingface.co/distilbert-base-cased-distilled-squad/resolve/main/tokenizer.json''' ), '''distilbert-base-german-cased''': ( '''https://huggingface.co/distilbert-base-german-cased/resolve/main/tokenizer.json''' ), '''distilbert-base-multilingual-cased''': ( '''https://huggingface.co/distilbert-base-multilingual-cased/resolve/main/tokenizer.json''' ), }, } __SCREAMING_SNAKE_CASE : str = { '''distilbert-base-uncased''': 5_1_2, '''distilbert-base-uncased-distilled-squad''': 5_1_2, '''distilbert-base-cased''': 5_1_2, '''distilbert-base-cased-distilled-squad''': 5_1_2, '''distilbert-base-german-cased''': 5_1_2, '''distilbert-base-multilingual-cased''': 5_1_2, } __SCREAMING_SNAKE_CASE : str = { '''distilbert-base-uncased''': {'''do_lower_case''': True}, '''distilbert-base-uncased-distilled-squad''': {'''do_lower_case''': True}, '''distilbert-base-cased''': {'''do_lower_case''': False}, '''distilbert-base-cased-distilled-squad''': {'''do_lower_case''': False}, '''distilbert-base-german-cased''': {'''do_lower_case''': False}, '''distilbert-base-multilingual-cased''': {'''do_lower_case''': False}, } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[Any] = VOCAB_FILES_NAMES lowercase__ : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP lowercase__ : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : str = PRETRAINED_INIT_CONFIGURATION lowercase__ : int = ['input_ids', 'attention_mask'] lowercase__ : Tuple = DistilBertTokenizer def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__="[UNK]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ): super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) _lowerCamelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , lowerCamelCase__ ) != do_lower_case or normalizer_state.get('''strip_accents''' , lowerCamelCase__ ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , lowerCamelCase__ ) != tokenize_chinese_chars ): _lowerCamelCase = getattr(lowerCamelCase__ , normalizer_state.pop('''type''' ) ) _lowerCamelCase = do_lower_case _lowerCamelCase = strip_accents _lowerCamelCase = tokenize_chinese_chars _lowerCamelCase = normalizer_class(**lowerCamelCase__ ) _lowerCamelCase = do_lower_case def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=None ): _lowerCamelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
661
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowercase : str = logging.get_logger(__name__) __lowercase : Union[str, Any] = { '''abeja/gpt-neox-japanese-2.7b''': '''https://huggingface.co/abeja/gpt-neox-japanese-2.7b/resolve/main/config.json''', } class _A ( snake_case ): '''simple docstring''' __lowerCamelCase : int = '''gpt_neox_japanese''' def __init__( self ,SCREAMING_SNAKE_CASE_=32000 ,SCREAMING_SNAKE_CASE_=2560 ,SCREAMING_SNAKE_CASE_=32 ,SCREAMING_SNAKE_CASE_=32 ,SCREAMING_SNAKE_CASE_=4 ,SCREAMING_SNAKE_CASE_="gelu" ,SCREAMING_SNAKE_CASE_=1.00 ,SCREAMING_SNAKE_CASE_=10000 ,SCREAMING_SNAKE_CASE_=2048 ,SCREAMING_SNAKE_CASE_=0.02 ,SCREAMING_SNAKE_CASE_=1E-5 ,SCREAMING_SNAKE_CASE_=True ,SCREAMING_SNAKE_CASE_=31996 ,SCREAMING_SNAKE_CASE_=31999 ,SCREAMING_SNAKE_CASE_=0.1 ,SCREAMING_SNAKE_CASE_=0.0 ,**SCREAMING_SNAKE_CASE_ ,): '''simple docstring''' super().__init__(bos_token_id=SCREAMING_SNAKE_CASE_ ,eos_token_id=SCREAMING_SNAKE_CASE_ ,**SCREAMING_SNAKE_CASE_ ) snake_case : Optional[Any] = vocab_size snake_case : Optional[int] = max_position_embeddings snake_case : List[str] = hidden_size snake_case : Union[str, Any] = num_hidden_layers snake_case : Optional[Any] = num_attention_heads snake_case : Tuple = intermediate_multiple_size snake_case : Optional[int] = hidden_act snake_case : List[str] = rotary_pct snake_case : str = rotary_emb_base snake_case : Any = initializer_range snake_case : str = layer_norm_eps snake_case : Optional[Any] = use_cache snake_case : Dict = attention_dropout snake_case : List[str] = hidden_dropout
36
"""simple docstring""" from math import sqrt import numpy as np from sympy import symbols # Coefficient # Speed of light (m/s) __SCREAMING_SNAKE_CASE : str = 2_9_9_7_9_2_4_5_8 # Symbols __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE : Union[str, Any] = symbols('''ct x y z''') def lowerCAmelCase_( lowercase_ : float ) -> float: if velocity > c: raise ValueError('''Speed must not exceed light speed 299,792,458 [m/s]!''' ) elif velocity < 1: # Usually the speed should be much higher than 1 (c order of magnitude) raise ValueError('''Speed must be greater than or equal to 1!''' ) return velocity / c def lowerCAmelCase_( lowercase_ : float ) -> float: return 1 / sqrt(1 - beta(lowercase_ ) ** 2 ) def lowerCAmelCase_( lowercase_ : float ) -> np.ndarray: return np.array( [ [gamma(lowercase_ ), -gamma(lowercase_ ) * beta(lowercase_ ), 0, 0], [-gamma(lowercase_ ) * beta(lowercase_ ), gamma(lowercase_ ), 0, 0], [0, 0, 1, 0], [0, 0, 0, 1], ] ) def lowerCAmelCase_( lowercase_ : float , lowercase_ : np.ndarray | None = None ) -> np.ndarray: # Ensure event is not empty if event is None: _lowerCamelCase = np.array([ct, x, y, z] ) # Symbolic four vector else: event[0] *= c # x0 is ct (speed of light * time) return transformation_matrix(lowercase_ ) @ event if __name__ == "__main__": import doctest doctest.testmod() # Example of symbolic vector: __SCREAMING_SNAKE_CASE : List[str] = transform(2_9_9_7_9_2_4_5) print('''Example of four vector: ''') print(F"""ct' = {four_vector[0]}""") print(F"""x' = {four_vector[1]}""") print(F"""y' = {four_vector[2]}""") print(F"""z' = {four_vector[3]}""") # Substitute symbols with numerical values __SCREAMING_SNAKE_CASE : Tuple = {ct: c, x: 1, y: 1, z: 1} __SCREAMING_SNAKE_CASE : int = [four_vector[i].subs(sub_dict) for i in range(4)] print(F"""\n{numerical_vector}""")
661
0
import copy import random from transformers import CLIPTokenizer class A__ ( A__ ): """simple docstring""" def __init__( self : Dict , *lowerCamelCase__ : Tuple , **lowerCamelCase__ : Optional[Any] ): super().__init__(*lowerCamelCase__ , **lowerCamelCase__ ) a__ : Dict = {} def _UpperCamelCase( self : Optional[int] , lowerCamelCase__ : Dict , *lowerCamelCase__ : int , **lowerCamelCase__ : List[str] ): a__ : Optional[Any] = super().add_tokens(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) if num_added_tokens == 0: raise ValueError( f'''The tokenizer already contains the token {placeholder_token}. Please pass a different''' " `placeholder_token` that is not already in the tokenizer." ) def _UpperCamelCase( self : Dict , lowerCamelCase__ : Dict , *lowerCamelCase__ : Optional[Any] , lowerCamelCase__ : Any=1 , **lowerCamelCase__ : List[str] ): a__ : int = [] if num_vec_per_token == 1: self.try_adding_tokens(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) output.append(lowerCamelCase__ ) else: a__ : str = [] for i in range(lowerCamelCase__ ): a__ : List[str] = placeholder_token + f'''_{i}''' self.try_adding_tokens(lowerCamelCase__ , *lowerCamelCase__ , **lowerCamelCase__ ) output.append(lowerCamelCase__ ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( f'''The tokenizer already has placeholder token {token} that can get confused with''' f''' {placeholder_token}keep placeholder tokens independent''' ) a__ : List[Any] = output def _UpperCamelCase( self : Optional[Any] , lowerCamelCase__ : List[str] , lowerCamelCase__ : Tuple=False , lowerCamelCase__ : int=1.0 ): if isinstance(lowerCamelCase__ , lowerCamelCase__ ): a__ : Any = [] for i in range(len(lowerCamelCase__ ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=lowerCamelCase__ ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: a__ : Any = self.token_map[placeholder_token] a__ : Optional[Any] = tokens[: 1 + int(len(lowerCamelCase__ ) * prop_tokens_to_load )] if vector_shuffle: a__ : str = copy.copy(lowerCamelCase__ ) random.shuffle(lowerCamelCase__ ) a__ : Union[str, Any] = text.replace(lowerCamelCase__ , " ".join(lowerCamelCase__ ) ) return text def __call__( self : str , lowerCamelCase__ : Union[str, Any] , *lowerCamelCase__ : Union[str, Any] , lowerCamelCase__ : Optional[int]=False , lowerCamelCase__ : List[str]=1.0 , **lowerCamelCase__ : Tuple ): return super().__call__( self.replace_placeholder_tokens_in_text( lowerCamelCase__ , vector_shuffle=lowerCamelCase__ , prop_tokens_to_load=lowerCamelCase__ ) , *lowerCamelCase__ , **lowerCamelCase__ , ) def _UpperCamelCase( self : Any , lowerCamelCase__ : Optional[int] , *lowerCamelCase__ : int , lowerCamelCase__ : Union[str, Any]=False , lowerCamelCase__ : Any=1.0 , **lowerCamelCase__ : List[str] ): return super().encode( self.replace_placeholder_tokens_in_text( lowerCamelCase__ , vector_shuffle=lowerCamelCase__ , prop_tokens_to_load=lowerCamelCase__ ) , *lowerCamelCase__ , **lowerCamelCase__ , )
37
"""simple docstring""" from __future__ import annotations __SCREAMING_SNAKE_CASE : Dict = 1.6_0_2_1e-1_9 # units = C def lowerCAmelCase_( lowercase_ : float , lowercase_ : float , lowercase_ : float , ) -> tuple[str, float]: if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif conductivity < 0: raise ValueError('''Conductivity cannot be negative''' ) elif electron_conc < 0: raise ValueError('''Electron concentration cannot be negative''' ) elif mobility < 0: raise ValueError('''mobility cannot be negative''' ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
661
0
'''simple docstring''' import random def UpperCamelCase__ ( __magic_name__ : int ) -> bool: '''simple docstring''' snake_case__ : List[str] = num - 1 snake_case__ : List[Any] = 0 while s % 2 == 0: snake_case__ : Optional[Any] = s // 2 t += 1 for _ in range(5 ): snake_case__ : Any = random.randrange(2 , num - 1 ) snake_case__ : Optional[Any] = pow(__magic_name__ , __magic_name__ , __magic_name__ ) if v != 1: snake_case__ : int = 0 while v != (num - 1): if i == t - 1: return False else: snake_case__ : int = i + 1 snake_case__ : str = (v**2) % num return True def UpperCamelCase__ ( __magic_name__ : int ) -> bool: '''simple docstring''' if num < 2: return False snake_case__ : Dict = [ 2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41, 43, 47, 53, 59, 61, 67, 71, 73, 79, 83, 89, 97, 1_01, 1_03, 1_07, 1_09, 1_13, 1_27, 1_31, 1_37, 1_39, 1_49, 1_51, 1_57, 1_63, 1_67, 1_73, 1_79, 1_81, 1_91, 1_93, 1_97, 1_99, 2_11, 2_23, 2_27, 2_29, 2_33, 2_39, 2_41, 2_51, 2_57, 2_63, 2_69, 2_71, 2_77, 2_81, 2_83, 2_93, 3_07, 3_11, 3_13, 3_17, 3_31, 3_37, 3_47, 3_49, 3_53, 3_59, 3_67, 3_73, 3_79, 3_83, 3_89, 3_97, 4_01, 4_09, 4_19, 4_21, 4_31, 4_33, 4_39, 4_43, 4_49, 4_57, 4_61, 4_63, 4_67, 4_79, 4_87, 4_91, 4_99, 5_03, 5_09, 5_21, 5_23, 5_41, 5_47, 5_57, 5_63, 5_69, 5_71, 5_77, 5_87, 5_93, 5_99, 6_01, 6_07, 6_13, 6_17, 6_19, 6_31, 6_41, 6_43, 6_47, 6_53, 6_59, 6_61, 6_73, 6_77, 6_83, 6_91, 7_01, 7_09, 7_19, 7_27, 7_33, 7_39, 7_43, 7_51, 7_57, 7_61, 7_69, 7_73, 7_87, 7_97, 8_09, 8_11, 8_21, 8_23, 8_27, 8_29, 8_39, 8_53, 8_57, 8_59, 8_63, 8_77, 8_81, 8_83, 8_87, 9_07, 9_11, 9_19, 9_29, 9_37, 9_41, 9_47, 9_53, 9_67, 9_71, 9_77, 9_83, 9_91, 9_97, ] if num in low_primes: return True for prime in low_primes: if (num % prime) == 0: return False return rabin_miller(__magic_name__ ) def UpperCamelCase__ ( __magic_name__ : int = 10_24 ) -> int: '''simple docstring''' while True: snake_case__ : Optional[Any] = random.randrange(2 ** (keysize - 1) , 2 ** (keysize) ) if is_prime_low_num(__magic_name__ ): return num if __name__ == "__main__": A_ : Dict = generate_large_prime() print(("Prime number:", num)) print(("is_prime_low_num:", is_prime_low_num(num)))
38
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class lowerCamelCase_( unittest.TestCase ): '''simple docstring''' def snake_case__ ( self ): _lowerCamelCase = { '''task_specific_params''': { '''summarization''': {'''length_penalty''': 1.0, '''max_length''': 1_2_8, '''min_length''': 1_2, '''num_beams''': 4}, '''summarization_cnn''': {'''length_penalty''': 2.0, '''max_length''': 1_4_2, '''min_length''': 5_6, '''num_beams''': 4}, '''summarization_xsum''': {'''length_penalty''': 1.0, '''max_length''': 6_2, '''min_length''': 1_1, '''num_beams''': 6}, } } _lowerCamelCase = { '''task_specific_params.summarization.length_penalty''': 1.0, '''task_specific_params.summarization.max_length''': 1_2_8, '''task_specific_params.summarization.min_length''': 1_2, '''task_specific_params.summarization.num_beams''': 4, '''task_specific_params.summarization_cnn.length_penalty''': 2.0, '''task_specific_params.summarization_cnn.max_length''': 1_4_2, '''task_specific_params.summarization_cnn.min_length''': 5_6, '''task_specific_params.summarization_cnn.num_beams''': 4, '''task_specific_params.summarization_xsum.length_penalty''': 1.0, '''task_specific_params.summarization_xsum.max_length''': 6_2, '''task_specific_params.summarization_xsum.min_length''': 1_1, '''task_specific_params.summarization_xsum.num_beams''': 6, } self.assertEqual(flatten_dict(lowerCamelCase__ ) , lowerCamelCase__ ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , x.transpose() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , np.asarray(transpose(lowerCamelCase__ ) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , np.asarray(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.reshape(lowerCamelCase__ , (4, 3) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , np.reshape(lowerCamelCase__ , (1_2, 5) ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , reshape(lowerCamelCase__ , (1_2, 5) ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , reshape(lowerCamelCase__ , (1_2, 5) ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.asarray(reshape(lowerCamelCase__ , (4, 3) ) ) ) ) _lowerCamelCase = np.random.randn(3 , 4 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (1_2, 5) ) , np.asarray(reshape(lowerCamelCase__ , (1_2, 5) ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.squeeze(lowerCamelCase__ ) ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.squeeze(lowerCamelCase__ , axis=2 ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(1 , 3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.asarray(squeeze(lowerCamelCase__ ) ) ) ) _lowerCamelCase = np.random.randn(1 , 4 , 1 , 5 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.asarray(squeeze(lowerCamelCase__ , axis=2 ) ) ) ) def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.expand_dims(lowerCamelCase__ , axis=1 ) ) ) @require_torch def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_tf def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_flax def snake_case__ ( self ): _lowerCamelCase = np.random.randn(3 , 4 ) _lowerCamelCase = jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.asarray(expand_dims(lowerCamelCase__ , axis=1 ) ) ) )
661
0
from __future__ import annotations import unittest from transformers import MobileBertConfig, is_tf_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_MODEL_FOR_PRETRAINING_MAPPING, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertModel, ) @require_tf class snake_case_ ( __A , __A , unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = ( ( TFMobileBertModel, TFMobileBertForMaskedLM, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertForMultipleChoice, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : List[Any] = ( { "feature-extraction": TFMobileBertModel, "fill-mask": TFMobileBertForMaskedLM, "question-answering": TFMobileBertForQuestionAnswering, "text-classification": TFMobileBertForSequenceClassification, "token-classification": TFMobileBertForTokenClassification, "zero-shot": TFMobileBertForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[int] = False def snake_case__( self : List[Any] , _UpperCamelCase : Optional[Any] , _UpperCamelCase : Tuple , _UpperCamelCase : Dict=False ) ->List[str]: snake_case_ = super()._prepare_for_class(_UpperCamelCase , _UpperCamelCase , return_labels=_UpperCamelCase ) if return_labels: if model_class in get_values(_UpperCamelCase ): snake_case_ = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) return inputs_dict class snake_case_ ( __A ): '''simple docstring''' def __init__( self : str , _UpperCamelCase : Optional[int] , _UpperCamelCase : Dict=1_3 , _UpperCamelCase : Tuple=7 , _UpperCamelCase : Optional[int]=True , _UpperCamelCase : str=True , _UpperCamelCase : Any=True , _UpperCamelCase : Optional[Any]=True , _UpperCamelCase : str=9_9 , _UpperCamelCase : Tuple=3_2 , _UpperCamelCase : int=3_2 , _UpperCamelCase : int=2 , _UpperCamelCase : List[str]=4 , _UpperCamelCase : Optional[int]=3_7 , _UpperCamelCase : List[str]="gelu" , _UpperCamelCase : List[str]=0.1 , _UpperCamelCase : Dict=0.1 , _UpperCamelCase : int=5_1_2 , _UpperCamelCase : Any=1_6 , _UpperCamelCase : Tuple=2 , _UpperCamelCase : List[str]=0.02 , _UpperCamelCase : Any=3 , _UpperCamelCase : List[Any]=4 , _UpperCamelCase : str=None , ) ->List[Any]: snake_case_ = parent snake_case_ = batch_size snake_case_ = seq_length snake_case_ = is_training snake_case_ = use_input_mask snake_case_ = use_token_type_ids snake_case_ = use_labels snake_case_ = vocab_size snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = max_position_embeddings snake_case_ = type_vocab_size snake_case_ = type_sequence_label_size snake_case_ = initializer_range snake_case_ = num_labels snake_case_ = num_choices snake_case_ = scope snake_case_ = embedding_size def snake_case__( self : Optional[int] ) ->Optional[int]: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ = None if self.use_input_mask: snake_case_ = random_attention_mask([self.batch_size, self.seq_length] ) snake_case_ = None if self.use_token_type_ids: snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ = None snake_case_ = None snake_case_ = None if self.use_labels: snake_case_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ = MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , embedding_size=self.embedding_size , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case__( self : Union[str, Any] , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Optional[int] , _UpperCamelCase : str , _UpperCamelCase : Optional[Any] , _UpperCamelCase : Tuple , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : List[Any] ) ->Optional[Any]: snake_case_ = TFMobileBertModel(config=_UpperCamelCase ) snake_case_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} snake_case_ = model(_UpperCamelCase ) snake_case_ = [input_ids, input_mask] snake_case_ = model(_UpperCamelCase ) snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def snake_case__( self : Dict , _UpperCamelCase : int , _UpperCamelCase : int , _UpperCamelCase : List[Any] , _UpperCamelCase : str , _UpperCamelCase : List[str] , _UpperCamelCase : Any , _UpperCamelCase : Optional[int] ) ->List[str]: snake_case_ = TFMobileBertForMaskedLM(config=_UpperCamelCase ) snake_case_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case__( self : Any , _UpperCamelCase : List[str] , _UpperCamelCase : List[Any] , _UpperCamelCase : str , _UpperCamelCase : Dict , _UpperCamelCase : List[Any] , _UpperCamelCase : Any , _UpperCamelCase : Optional[Any] ) ->Optional[Any]: snake_case_ = TFMobileBertForNextSentencePrediction(config=_UpperCamelCase ) snake_case_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def snake_case__( self : Optional[Any] , _UpperCamelCase : Any , _UpperCamelCase : List[Any] , _UpperCamelCase : Any , _UpperCamelCase : str , _UpperCamelCase : int , _UpperCamelCase : Tuple , _UpperCamelCase : Any ) ->List[Any]: snake_case_ = TFMobileBertForPreTraining(config=_UpperCamelCase ) snake_case_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual( result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def snake_case__( self : int , _UpperCamelCase : Any , _UpperCamelCase : str , _UpperCamelCase : Optional[Any] , _UpperCamelCase : str , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Optional[int] , _UpperCamelCase : int ) ->Any: snake_case_ = self.num_labels snake_case_ = TFMobileBertForSequenceClassification(config=_UpperCamelCase ) snake_case_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case__( self : List[Any] , _UpperCamelCase : Optional[Any] , _UpperCamelCase : Any , _UpperCamelCase : Dict , _UpperCamelCase : Optional[Any] , _UpperCamelCase : str , _UpperCamelCase : Dict , _UpperCamelCase : str ) ->Union[str, Any]: snake_case_ = self.num_choices snake_case_ = TFMobileBertForMultipleChoice(config=_UpperCamelCase ) snake_case_ = tf.tile(tf.expand_dims(_UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) snake_case_ = tf.tile(tf.expand_dims(_UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) snake_case_ = tf.tile(tf.expand_dims(_UpperCamelCase , 1 ) , (1, self.num_choices, 1) ) snake_case_ = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def snake_case__( self : List[Any] , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : Optional[Any] , _UpperCamelCase : List[Any] , _UpperCamelCase : Any , _UpperCamelCase : List[str] , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : str ) ->str: snake_case_ = self.num_labels snake_case_ = TFMobileBertForTokenClassification(config=_UpperCamelCase ) snake_case_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case__( self : Optional[int] , _UpperCamelCase : Tuple , _UpperCamelCase : Union[str, Any] , _UpperCamelCase : int , _UpperCamelCase : Optional[int] , _UpperCamelCase : List[str] , _UpperCamelCase : Any , _UpperCamelCase : Optional[Any] ) ->str: snake_case_ = TFMobileBertForQuestionAnswering(config=_UpperCamelCase ) snake_case_ = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} snake_case_ = model(_UpperCamelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case__( self : Union[str, Any] ) ->List[str]: snake_case_ = self.prepare_config_and_inputs() ( ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ) = config_and_inputs snake_case_ = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict def snake_case__( self : Tuple ) ->Optional[Any]: snake_case_ = TFMobileBertModelTest.TFMobileBertModelTester(self ) snake_case_ = ConfigTester(self , config_class=_UpperCamelCase , hidden_size=3_7 ) def snake_case__( self : Optional[Any] ) ->Any: self.config_tester.run_common_tests() def snake_case__( self : str ) ->Dict: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*_UpperCamelCase ) def snake_case__( self : List[str] ) ->Optional[Any]: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*_UpperCamelCase ) def snake_case__( self : Optional[Any] ) ->List[str]: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*_UpperCamelCase ) def snake_case__( self : Tuple ) ->Optional[Any]: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*_UpperCamelCase ) def snake_case__( self : List[str] ) ->Union[str, Any]: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*_UpperCamelCase ) def snake_case__( self : Union[str, Any] ) ->Dict: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*_UpperCamelCase ) def snake_case__( self : Optional[Any] ) ->Union[str, Any]: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*_UpperCamelCase ) def snake_case__( self : Union[str, Any] ) ->Tuple: snake_case_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*_UpperCamelCase ) @slow def snake_case__( self : List[Any] ) ->int: # for model_name in TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["google/mobilebert-uncased"]: snake_case_ = TFMobileBertModel.from_pretrained(_UpperCamelCase ) self.assertIsNotNone(_UpperCamelCase ) @require_tf class snake_case_ ( unittest.TestCase ): '''simple docstring''' @slow def snake_case__( self : Tuple ) ->List[Any]: snake_case_ = TFMobileBertForPreTraining.from_pretrained('''google/mobilebert-uncased''' ) snake_case_ = tf.constant([[0, 1, 2, 3, 4, 5]] ) snake_case_ = model(_UpperCamelCase )[0] snake_case_ = [1, 6, 3_0_5_2_2] self.assertEqual(output.shape , _UpperCamelCase ) snake_case_ = tf.constant( [ [ [-4.5919547, -9.248295, -9.645256], [-6.7306175, -6.440284, -6.6052837], [-7.2743506, -6.7847915, -6.024673], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , _UpperCamelCase , atol=1e-4 )
39
"""simple docstring""" import json import os from functools import lru_cache from typing import TYPE_CHECKING, List, Optional, Tuple import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __SCREAMING_SNAKE_CASE : Union[str, Any] = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : int = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } __SCREAMING_SNAKE_CASE : Tuple = { '''vocab_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/vocab.json'''}, '''merges_file''': {'''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/merges.txt'''}, '''tokenizer_config_file''': { '''facebook/blenderbot-3B''': '''https://huggingface.co/facebook/blenderbot-3B/resolve/main/tokenizer_config.json''' }, } __SCREAMING_SNAKE_CASE : str = {'''facebook/blenderbot-3B''': 1_2_8} @lru_cache() # Copied from transformers.models.roberta.tokenization_roberta.bytes_to_unicode def lowerCAmelCase_( ) -> str: _lowerCamelCase = ( list(range(ord('''!''' ) , ord('''~''' ) + 1 ) ) + list(range(ord('''¡''' ) , ord('''¬''' ) + 1 ) ) + list(range(ord('''®''' ) , ord('''ÿ''' ) + 1 ) ) ) _lowerCamelCase = bs[:] _lowerCamelCase = 0 for b in range(2**8 ): if b not in bs: bs.append(lowercase_ ) cs.append(2**8 + n ) n += 1 _lowerCamelCase = [chr(lowercase_ ) for n in cs] return dict(zip(lowercase_ , lowercase_ ) ) def lowerCAmelCase_( lowercase_ : str ) -> Dict: _lowerCamelCase = set() _lowerCamelCase = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _lowerCamelCase = char return pairs class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : Union[str, Any] = VOCAB_FILES_NAMES lowercase__ : int = PRETRAINED_VOCAB_FILES_MAP lowercase__ : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase__ : Tuple = ['input_ids', 'attention_mask'] def __init__( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__="replace" , lowerCamelCase__="<s>" , lowerCamelCase__="</s>" , lowerCamelCase__="</s>" , lowerCamelCase__="<s>" , lowerCamelCase__="<unk>" , lowerCamelCase__="<pad>" , lowerCamelCase__="<mask>" , lowerCamelCase__=False , **lowerCamelCase__ , ): _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else bos_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else eos_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else sep_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else cls_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else unk_token _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else pad_token # Mask token behave like a normal word, i.e. include the space before it _lowerCamelCase = AddedToken(lowerCamelCase__ , lstrip=lowerCamelCase__ , rstrip=lowerCamelCase__ ) if isinstance(lowerCamelCase__ , lowerCamelCase__ ) else mask_token super().__init__( errors=lowerCamelCase__ , bos_token=lowerCamelCase__ , eos_token=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , add_prefix_space=lowerCamelCase__ , **lowerCamelCase__ , ) with open(lowerCamelCase__ , encoding='''utf-8''' ) as vocab_handle: _lowerCamelCase = json.load(lowerCamelCase__ ) _lowerCamelCase = {v: k for k, v in self.encoder.items()} _lowerCamelCase = errors # how to handle errors in decoding _lowerCamelCase = bytes_to_unicode() _lowerCamelCase = {v: k for k, v in self.byte_encoder.items()} with open(lowerCamelCase__ , encoding='''utf-8''' ) as merges_handle: _lowerCamelCase = merges_handle.read().split('''\n''' )[1:-1] _lowerCamelCase = [tuple(merge.split() ) for merge in bpe_merges] _lowerCamelCase = dict(zip(lowerCamelCase__ , range(len(lowerCamelCase__ ) ) ) ) _lowerCamelCase = {} _lowerCamelCase = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions _lowerCamelCase = re.compile(R'''\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+''' ) @property # Copied from transformers.models.roberta.tokenization_roberta.RobertaTokenizer.vocab_size with Roberta->Blenderbot, RoBERTa->Blenderbot def snake_case__ ( self ): return len(self.encoder ) def snake_case__ ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def snake_case__ ( self , lowerCamelCase__ ): if token in self.cache: return self.cache[token] _lowerCamelCase = tuple(lowerCamelCase__ ) _lowerCamelCase = get_pairs(lowerCamelCase__ ) if not pairs: return token while True: _lowerCamelCase = min(lowerCamelCase__ , key=lambda lowerCamelCase__ : self.bpe_ranks.get(lowerCamelCase__ , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break _lowerCamelCase , _lowerCamelCase = bigram _lowerCamelCase = [] _lowerCamelCase = 0 while i < len(lowerCamelCase__ ): try: _lowerCamelCase = word.index(lowerCamelCase__ , lowerCamelCase__ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _lowerCamelCase = j if word[i] == first and i < len(lowerCamelCase__ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _lowerCamelCase = tuple(lowerCamelCase__ ) _lowerCamelCase = new_word if len(lowerCamelCase__ ) == 1: break else: _lowerCamelCase = get_pairs(lowerCamelCase__ ) _lowerCamelCase = ''' '''.join(lowerCamelCase__ ) _lowerCamelCase = word return word def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] for token in re.findall(self.pat , lowerCamelCase__ ): _lowerCamelCase = ''''''.join( self.byte_encoder[b] for b in token.encode('''utf-8''' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(lowerCamelCase__ ).split(''' ''' ) ) return bpe_tokens def snake_case__ ( self , lowerCamelCase__ ): return self.encoder.get(lowerCamelCase__ , self.encoder.get(self.unk_token ) ) def snake_case__ ( self , lowerCamelCase__ ): return self.decoder.get(lowerCamelCase__ ) def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = ''''''.join(lowerCamelCase__ ) _lowerCamelCase = bytearray([self.byte_decoder[c] for c in text] ).decode('''utf-8''' , errors=self.errors ) return text def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): if not os.path.isdir(lowerCamelCase__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) _lowerCamelCase = os.path.join( lowerCamelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(lowerCamelCase__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=lowerCamelCase__ , ensure_ascii=lowerCamelCase__ ) + '''\n''' ) _lowerCamelCase = 0 with open(lowerCamelCase__ , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda lowerCamelCase__ : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ''' Please check that the tokenizer is not corrupted!''' ) _lowerCamelCase = token_index writer.write(''' '''.join(lowerCamelCase__ ) + '''\n''' ) index += 1 return vocab_file, merge_file def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None , lowerCamelCase__ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase__ , token_ids_a=lowerCamelCase__ , already_has_special_tokens=lowerCamelCase__ ) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase__ )) + [1] return [1] + ([0] * len(lowerCamelCase__ )) + [1, 1] + ([0] * len(lowerCamelCase__ )) + [1] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): _lowerCamelCase = [self.sep_token_id] _lowerCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__=False , **lowerCamelCase__ ): _lowerCamelCase = kwargs.pop('''add_prefix_space''' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(lowerCamelCase__ ) > 0 and not text[0].isspace()): _lowerCamelCase = ''' ''' + text return (text, kwargs) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ = None ): return token_ids_a + [self.eos_token_id] def snake_case__ ( self , lowerCamelCase__ ): _lowerCamelCase = [] for is_user, text in conversation.iter_texts(): if is_user: # We need to space prefix as it's being done within blenderbot inputs.append(''' ''' + text ) else: # Generated responses should contain them already. inputs.append(lowerCamelCase__ ) _lowerCamelCase = ''' '''.join(lowerCamelCase__ ) _lowerCamelCase = self.encode(lowerCamelCase__ ) if len(lowerCamelCase__ ) > self.model_max_length: _lowerCamelCase = input_ids[-self.model_max_length :] logger.warning(F"""Trimmed input from conversation as it was longer than {self.model_max_length} tokens.""" ) return input_ids
661
0