code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class _lowercase ( SCREAMING_SNAKE_CASE__ ): _UpperCAmelCase = (KDPMaDiscreteScheduler,) _UpperCAmelCase = 10 def UpperCamelCase ( self , **A__ ) -> Union[str, Any]: snake_case = { """num_train_timesteps""": 11_00, """beta_start""": 0.0_0_0_1, """beta_end""": 0.0_2, """beta_schedule""": """linear""", } config.update(**_lowercase ) return config def UpperCamelCase ( self ) -> int: for timesteps in [10, 50, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=_lowercase ) def UpperCamelCase ( self ) -> Union[str, Any]: for beta_start, beta_end in zip([0.0_0_0_0_1, 0.0_0_0_1, 0.0_0_1] , [0.0_0_0_2, 0.0_0_2, 0.0_2] ): self.check_over_configs(beta_start=_lowercase , beta_end=_lowercase ) def UpperCamelCase ( self ) -> List[Any]: for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=_lowercase ) def UpperCamelCase ( self ) -> Optional[Any]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=_lowercase ) def UpperCamelCase ( self ) -> Optional[int]: snake_case = self.scheduler_classes[0] snake_case = self.get_scheduler_config(prediction_type='''v_prediction''' ) snake_case = scheduler_class(**_lowercase ) scheduler.set_timesteps(self.num_inference_steps ) snake_case = self.dummy_model() snake_case = self.dummy_sample_deter * scheduler.init_noise_sigma snake_case = sample.to(_lowercase ) for i, t in enumerate(scheduler.timesteps ): snake_case = scheduler.scale_model_input(_lowercase , _lowercase ) snake_case = model(_lowercase , _lowercase ) snake_case = scheduler.step(_lowercase , _lowercase , _lowercase ) snake_case = output.prev_sample snake_case = torch.sum(torch.abs(_lowercase ) ) snake_case = torch.mean(torch.abs(_lowercase ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 4.6934e-07 ) < 1e-2 assert abs(result_mean.item() - 6.1112e-10 ) < 1e-3 else: # CUDA assert abs(result_sum.item() - 4.693428650170972e-07 ) < 1e-2 assert abs(result_mean.item() - 0.0_0_0_2 ) < 1e-3 def UpperCamelCase ( self ) -> Any: if torch_device == "mps": return snake_case = self.scheduler_classes[0] snake_case = self.get_scheduler_config() snake_case = scheduler_class(**_lowercase ) scheduler.set_timesteps(self.num_inference_steps ) snake_case = self.dummy_model() snake_case = self.dummy_sample_deter * scheduler.init_noise_sigma snake_case = sample.to(_lowercase ) for i, t in enumerate(scheduler.timesteps ): snake_case = scheduler.scale_model_input(_lowercase , _lowercase ) snake_case = model(_lowercase , _lowercase ) snake_case = scheduler.step(_lowercase , _lowercase , _lowercase ) snake_case = output.prev_sample snake_case = torch.sum(torch.abs(_lowercase ) ) snake_case = torch.mean(torch.abs(_lowercase ) ) if torch_device in ["cpu", "mps"]: assert abs(result_sum.item() - 2_0.4_1_2_5 ) < 1e-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1e-3 else: # CUDA assert abs(result_sum.item() - 2_0.4_1_2_5 ) < 1e-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1e-3 def UpperCamelCase ( self ) -> Optional[int]: if torch_device == "mps": return snake_case = self.scheduler_classes[0] snake_case = self.get_scheduler_config() snake_case = scheduler_class(**_lowercase ) scheduler.set_timesteps(self.num_inference_steps , device=_lowercase ) snake_case = self.dummy_model() snake_case = self.dummy_sample_deter.to(_lowercase ) * scheduler.init_noise_sigma for t in scheduler.timesteps: snake_case = scheduler.scale_model_input(_lowercase , _lowercase ) snake_case = model(_lowercase , _lowercase ) snake_case = scheduler.step(_lowercase , _lowercase , _lowercase ) snake_case = output.prev_sample snake_case = torch.sum(torch.abs(_lowercase ) ) snake_case = torch.mean(torch.abs(_lowercase ) ) if str(_lowercase ).startswith('''cpu''' ): # The following sum varies between 148 and 156 on mps. Why? assert abs(result_sum.item() - 2_0.4_1_2_5 ) < 1e-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1e-3 else: # CUDA assert abs(result_sum.item() - 2_0.4_1_2_5 ) < 1e-2 assert abs(result_mean.item() - 0.0_2_6_6 ) < 1e-3
721
'''simple docstring''' import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow _lowercase = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ 'text-classification', 'language-modeling', 'summarization', 'token-classification', 'question-answering', ] ] sys.path.extend(SRC_DIRS) if SRC_DIRS is not None: import run_clm_flax import run_flax_glue import run_flax_ner import run_mlm_flax import run_qa import run_summarization_flax import run_ta_mlm_flax logging.basicConfig(level=logging.DEBUG) _lowercase = logging.getLogger() def __UpperCamelCase ( ) ->Tuple: snake_case = argparse.ArgumentParser() parser.add_argument('''-f''' ) snake_case = parser.parse_args() return args.f def __UpperCamelCase ( a : Dict , a : Tuple="eval" ) ->List[Any]: snake_case = os.path.join(a , f"""{split}_results.json""" ) if os.path.exists(a ): with open(a , '''r''' ) as f: return json.load(a ) raise ValueError(f"""can't find {path}""" ) _lowercase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class _lowercase ( __a ): def UpperCamelCase ( self ) -> List[str]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --eval_steps=2 --warmup_steps=2 --seed=42 --max_seq_length=128 """.split() with patch.object(A__ , '''argv''' , A__ ): run_flax_glue.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.7_5 ) @slow def UpperCamelCase ( self ) -> List[Any]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_clm_flax.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --block_size 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(A__ , '''argv''' , A__ ): run_clm_flax.main() snake_case = get_results(A__ ) self.assertLess(result['''eval_perplexity'''] , 1_00 ) @slow def UpperCamelCase ( self ) -> int: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_summarization.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --test_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=8 --do_train --do_eval --do_predict --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --predict_with_generate """.split() with patch.object(A__ , '''argv''' , A__ ): run_summarization_flax.main() snake_case = get_results(A__ , split='''test''' ) self.assertGreaterEqual(result['''test_rouge1'''] , 10 ) self.assertGreaterEqual(result['''test_rouge2'''] , 2 ) self.assertGreaterEqual(result['''test_rougeL'''] , 7 ) self.assertGreaterEqual(result['''test_rougeLsum'''] , 7 ) @slow def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_mlm.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --overwrite_output_dir --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --logging_steps 2 --eval_steps 2 --do_train --do_eval --num_train_epochs=1 """.split() with patch.object(A__ , '''argv''' , A__ ): run_mlm_flax.main() snake_case = get_results(A__ ) self.assertLess(result['''eval_perplexity'''] , 42 ) @slow def UpperCamelCase ( self ) -> Dict: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_t5_mlm_flax.py --model_name_or_path t5-small --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(A__ , '''argv''' , A__ ): run_ta_mlm_flax.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.4_2 ) @slow def UpperCamelCase ( self ) -> int: # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu snake_case = 7 if get_gpu_count() > 1 else 2 snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_flax_ner.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --overwrite_output_dir --do_train --do_eval --warmup_steps=2 --learning_rate=2e-4 --logging_steps 2 --eval_steps 2 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 """.split() with patch.object(A__ , '''argv''' , A__ ): run_flax_ner.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.7_5 ) self.assertGreaterEqual(result['''eval_f1'''] , 0.3 ) @slow def UpperCamelCase ( self ) -> Any: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_qa.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=2 --do_train --do_eval --logging_steps 2 --eval_steps 2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 """.split() with patch.object(A__ , '''argv''' , A__ ): run_qa.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_f1'''] , 30 ) self.assertGreaterEqual(result['''eval_exact'''] , 30 )
44
0
'''simple docstring''' print((lambda quine: quine % quine)('print((lambda quine: quine %% quine)(%r))'))
700
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS _lowercase = logging.get_logger(__name__) _lowercase = { 'linear': get_linear_schedule_with_warmup, 'cosine': get_cosine_schedule_with_warmup, 'cosine_w_restarts': get_cosine_with_hard_restarts_schedule_with_warmup, 'polynomial': get_polynomial_decay_schedule_with_warmup, 'constant': get_constant_schedule, 'constant_w_warmup': get_constant_schedule_with_warmup, } class _lowercase ( __a ): def __init__( self , A__=None , A__=None , *A__ , **A__ ) -> Union[str, Any]: super().__init__(*A__ , **A__ ) if config is None: assert isinstance(self.model , A__ ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) snake_case = self.model.config else: snake_case = config snake_case = data_args snake_case = self.config.tgt_vocab_size if isinstance(self.config , A__ ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ''' padding..''' ) if self.args.label_smoothing == 0: snake_case = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss snake_case = label_smoothed_nll_loss def UpperCamelCase ( self , A__ ) -> Tuple: if self.optimizer is None: snake_case = ['''bias''', '''LayerNorm.weight'''] snake_case = [ { '''params''': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], '''weight_decay''': self.args.weight_decay, }, { '''params''': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0, }, ] snake_case = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: snake_case = Adafactor snake_case = {'''scale_parameter''': False, '''relative_step''': False} else: snake_case = AdamW snake_case = { '''betas''': (self.args.adam_betaa, self.args.adam_betaa), '''eps''': self.args.adam_epsilon, } snake_case = self.args.learning_rate if self.sharded_ddp: snake_case = OSS( params=A__ , optim=A__ , **A__ , ) else: snake_case = optimizer_cls(A__ , **A__ ) if self.lr_scheduler is None: snake_case = self._get_lr_scheduler(A__ ) else: # ignoring --lr_scheduler logger.warning('''scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.''' ) def UpperCamelCase ( self , A__ ) -> Tuple: snake_case = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": snake_case = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": snake_case = schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: snake_case = schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=A__ ) return scheduler def UpperCamelCase ( self ) -> Optional[torch.utils.data.Sampler]: if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> List[Any]: if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token snake_case = model(**A__ , use_cache=A__ )[0] snake_case = self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models snake_case , snake_case = model(**A__ , labels=A__ , use_cache=A__ )[:2] else: # compute label smoothed loss snake_case = model(**A__ , use_cache=A__ )[0] snake_case = torch.nn.functional.log_softmax(A__ , dim=-1 ) snake_case , snake_case = self.loss_fn(A__ , A__ , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def UpperCamelCase ( self , A__ , A__ ) -> Any: snake_case = inputs.pop('''labels''' ) snake_case , snake_case = self._compute_loss(A__ , A__ , A__ ) return loss def UpperCamelCase ( self , A__ , A__ , A__ , A__ = None , ) -> Tuple[Optional[float], Optional[torch.Tensor], Optional[torch.Tensor]]: snake_case = self._prepare_inputs(A__ ) snake_case = { '''max_length''': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, '''num_beams''': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: snake_case = self.model.generate( inputs['''input_ids'''] , attention_mask=inputs['''attention_mask'''] , **A__ , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: snake_case = self._pad_tensors_to_max_len(A__ , gen_kwargs['''max_length'''] ) snake_case = inputs.pop('''labels''' ) with torch.no_grad(): # compute loss on predict data snake_case , snake_case = self._compute_loss(A__ , A__ , A__ ) snake_case = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) snake_case = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: snake_case = self._pad_tensors_to_max_len(A__ , gen_kwargs['''max_length'''] ) return (loss, logits, labels) def UpperCamelCase ( self , A__ , A__ ) -> List[str]: # If PAD token is not defined at least EOS token has to be defined snake_case = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( '''Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be''' F""" padded to `max_length`={max_length}""" ) snake_case = pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) snake_case = tensor return padded_tensor
44
0
'''simple docstring''' import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder _lowercase = '__DUMMY_TRANSFORMERS_USER__' _lowercase = 'Dummy User' _lowercase = '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' _lowercase = 'https://hub-ci.huggingface.co' _lowercase = CI_HUB_ENDPOINT + '/datasets/{repo_id}/resolve/{revision}/{path}' _lowercase = CI_HUB_ENDPOINT + '/{repo_id}/resolve/{revision}/{filename}' _lowercase = Path('~/.huggingface/hub_ci_token').expanduser() @pytest.fixture def __UpperCamelCase ( a : Dict ) ->Dict: monkeypatch.setattr( '''huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE''' , __SCREAMING_SNAKE_CASE ) @pytest.fixture def __UpperCamelCase ( a : str ) ->Dict: monkeypatch.setattr('''datasets.config.HF_ENDPOINT''' , __SCREAMING_SNAKE_CASE ) monkeypatch.setattr('''datasets.config.HUB_DATASETS_URL''' , __SCREAMING_SNAKE_CASE ) @pytest.fixture def __UpperCamelCase ( a : List[Any] ) ->Any: monkeypatch.setattr('''huggingface_hub.hf_api.HfFolder.path_token''' , __SCREAMING_SNAKE_CASE ) @pytest.fixture def __UpperCamelCase ( a : Any , a : Union[str, Any] ) ->int: HfFolder.save_token(__SCREAMING_SNAKE_CASE ) yield HfFolder.delete_token() @pytest.fixture(scope='''session''' ) def __UpperCamelCase ( ) ->Union[str, Any]: return HfApi(endpoint=__SCREAMING_SNAKE_CASE ) @pytest.fixture(scope='''session''' ) def __UpperCamelCase ( a : Any ) ->Tuple: snake_case = HfFolder.get_token() HfFolder.save_token(__SCREAMING_SNAKE_CASE ) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(__SCREAMING_SNAKE_CASE ) @pytest.fixture def __UpperCamelCase ( a : Optional[int] ) ->Tuple: def _cleanup_repo(a : List[Any] ): hf_api.delete_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type='''dataset''' ) return _cleanup_repo @pytest.fixture def __UpperCamelCase ( a : Tuple ) ->Tuple: @contextmanager def _temporary_repo(a : List[Any] ): try: yield repo_id finally: cleanup_repo(__SCREAMING_SNAKE_CASE ) return _temporary_repo @pytest.fixture(scope='''session''' ) def __UpperCamelCase ( a : Union[str, Any] , a : int , a : str ) ->Any: snake_case = f"""repo_txt_data-{int(time.time() * 10e3 )}""" snake_case = f"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type='''dataset''' , private=__SCREAMING_SNAKE_CASE ) hf_api.upload_file( token=__SCREAMING_SNAKE_CASE , path_or_fileobj=str(__SCREAMING_SNAKE_CASE ) , path_in_repo='''data/text_data.txt''' , repo_id=__SCREAMING_SNAKE_CASE , repo_type='''dataset''' , ) yield repo_id try: hf_api.delete_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type='''dataset''' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __UpperCamelCase ( a : int , a : List[str] , a : List[Any] ) ->Optional[int]: return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope='''session''' ) def __UpperCamelCase ( a : Optional[Any] , a : List[Any] , a : Optional[Any] ) ->Any: snake_case = f"""repo_zipped_txt_data-{int(time.time() * 10e3 )}""" snake_case = f"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type='''dataset''' , private=__SCREAMING_SNAKE_CASE ) hf_api.upload_file( token=__SCREAMING_SNAKE_CASE , path_or_fileobj=str(__SCREAMING_SNAKE_CASE ) , path_in_repo='''data.zip''' , repo_id=__SCREAMING_SNAKE_CASE , repo_type='''dataset''' , ) yield repo_id try: hf_api.delete_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type='''dataset''' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __UpperCamelCase ( a : List[str] , a : Any , a : List[Any] ) ->Dict: return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope='''session''' ) def __UpperCamelCase ( a : Optional[Any] , a : Dict , a : Any ) ->str: snake_case = f"""repo_zipped_img_data-{int(time.time() * 10e3 )}""" snake_case = f"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type='''dataset''' , private=__SCREAMING_SNAKE_CASE ) hf_api.upload_file( token=__SCREAMING_SNAKE_CASE , path_or_fileobj=str(__SCREAMING_SNAKE_CASE ) , path_in_repo='''data.zip''' , repo_id=__SCREAMING_SNAKE_CASE , repo_type='''dataset''' , ) yield repo_id try: hf_api.delete_repo(__SCREAMING_SNAKE_CASE , token=__SCREAMING_SNAKE_CASE , repo_type='''dataset''' ) except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def __UpperCamelCase ( a : Tuple , a : Tuple , a : Dict ) ->List[Any]: return hf_private_dataset_repo_zipped_img_data_
701
'''simple docstring''' import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def __UpperCamelCase ( a : List[str] ) ->str: snake_case = [] for line in lines: snake_case = re.sub(R'''#.*''' , '''''' , a ) # remove comments if line: filtered_lines.append(a ) snake_case = '''\n'''.join(a ) # Make a hash from all this code snake_case = full_str.encode('''utf-8''' ) return shaaaa(a ).hexdigest() # get importable module names and hash for caching _lowercase = { 'csv': (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), 'json': (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), 'pandas': (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), 'parquet': (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), 'arrow': (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), 'text': (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), 'imagefolder': (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), 'audiofolder': (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions _lowercase = { '.csv': ('csv', {}), '.tsv': ('csv', {'sep': '\t'}), '.json': ('json', {}), '.jsonl': ('json', {}), '.parquet': ('parquet', {}), '.arrow': ('arrow', {}), '.txt': ('text', {}), } _EXTENSION_TO_MODULE.update({ext: ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _lowercase = {'imagefolder', 'audiofolder'} # Used to filter data files based on extensions given a module name _lowercase = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append('.zip') _MODULE_TO_EXTENSIONS["audiofolder"].append('.zip')
44
0
'''simple docstring''' from scipy.stats import pearsonr import datasets _lowercase = '\nPearson correlation coefficient and p-value for testing non-correlation.\nThe Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases.\nThe p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets.\n' _lowercase = '\nArgs:\n predictions (`list` of `int`): Predicted class labels, as returned by a model.\n references (`list` of `int`): Ground truth labels.\n return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`.\n\nReturns:\n pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation.\n p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities.\n\nExamples:\n\n Example 1-A simple example using only predictions and references.\n >>> pearsonr_metric = datasets.load_metric("pearsonr")\n >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5])\n >>> print(round(results[\'pearsonr\'], 2))\n -0.74\n\n Example 2-The same as Example 1, but that also returns the `p-value`.\n >>> pearsonr_metric = datasets.load_metric("pearsonr")\n >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True)\n >>> print(sorted(list(results.keys())))\n [\'p-value\', \'pearsonr\']\n >>> print(round(results[\'pearsonr\'], 2))\n -0.74\n >>> print(round(results[\'p-value\'], 2))\n 0.15\n' _lowercase = '\n@article{2020SciPy-NMeth,\nauthor = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, Ilhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Antonio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\ntitle = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\njournal = {Nature Methods},\nyear = {2020},\nvolume = {17},\npages = {261--272},\nadsurl = {https://rdcu.be/b08Wh},\ndoi = {10.1038/s41592-019-0686-2},\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowercase ( datasets.Metric ): def UpperCamelCase ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) , reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html'''] , ) def UpperCamelCase ( self , A__ , A__ , A__=False ): if return_pvalue: snake_case = pearsonr(A__ , A__ ) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(A__ , A__ )[0] )}
702
'''simple docstring''' _lowercase = { 'Pillow': 'Pillow', 'accelerate': 'accelerate>=0.11.0', 'compel': 'compel==0.1.8', 'black': 'black~=23.1', 'datasets': 'datasets', 'filelock': 'filelock', 'flax': 'flax>=0.4.1', 'hf-doc-builder': 'hf-doc-builder>=0.3.0', 'huggingface-hub': 'huggingface-hub>=0.13.2', 'requests-mock': 'requests-mock==1.10.0', 'importlib_metadata': 'importlib_metadata', 'invisible-watermark': 'invisible-watermark', 'isort': 'isort>=5.5.4', 'jax': 'jax>=0.2.8,!=0.3.2', 'jaxlib': 'jaxlib>=0.1.65', 'Jinja2': 'Jinja2', 'k-diffusion': 'k-diffusion>=0.0.12', 'torchsde': 'torchsde', 'note_seq': 'note_seq', 'librosa': 'librosa', 'numpy': 'numpy', 'omegaconf': 'omegaconf', 'parameterized': 'parameterized', 'protobuf': 'protobuf>=3.20.3,<4', 'pytest': 'pytest', 'pytest-timeout': 'pytest-timeout', 'pytest-xdist': 'pytest-xdist', 'ruff': 'ruff>=0.0.241', 'safetensors': 'safetensors', 'sentencepiece': 'sentencepiece>=0.1.91,!=0.1.92', 'scipy': 'scipy', 'onnx': 'onnx', 'regex': 'regex!=2019.12.17', 'requests': 'requests', 'tensorboard': 'tensorboard', 'torch': 'torch>=1.4', 'torchvision': 'torchvision', 'transformers': 'transformers>=4.25.1', 'urllib3': 'urllib3<=2.0.0', }
44
0
'''simple docstring''' from math import sqrt def __UpperCamelCase ( a : int ) ->List[Any]: if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(sqrt(snake_case_ ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def __UpperCamelCase ( a : int = 1_0001 ) ->Any: snake_case = 0 snake_case = 1 while count != nth and number < 3: number += 1 if is_prime(snake_case_ ): count += 1 while count != nth: number += 2 if is_prime(snake_case_ ): count += 1 return number if __name__ == "__main__": print(f'{solution() = }')
703
'''simple docstring''' import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _lowercase ( __a , __a , unittest.TestCase ): _UpperCAmelCase = IFInpaintingSuperResolutionPipeline _UpperCAmelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} _UpperCAmelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({'''original_image'''} ) _UpperCAmelCase = PipelineTesterMixin.required_optional_params - {'''latents'''} def UpperCamelCase ( self ) -> int: return self._get_superresolution_dummy_components() def UpperCamelCase ( self , A__ , A__=0 ) -> Union[str, Any]: if str(A__ ).startswith('''mps''' ): snake_case = torch.manual_seed(A__ ) else: snake_case = torch.Generator(device=A__ ).manual_seed(A__ ) snake_case = floats_tensor((1, 3, 16, 16) , rng=random.Random(A__ ) ).to(A__ ) snake_case = floats_tensor((1, 3, 32, 32) , rng=random.Random(A__ ) ).to(A__ ) snake_case = floats_tensor((1, 3, 32, 32) , rng=random.Random(A__ ) ).to(A__ ) snake_case = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase ( self ) -> List[Any]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def UpperCamelCase ( self ) -> Optional[Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def UpperCamelCase ( self ) -> List[str]: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1 ) def UpperCamelCase ( self ) -> int: self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def UpperCamelCase ( self ) -> Optional[Any]: self._test_save_load_local() def UpperCamelCase ( self ) -> Dict: self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
44
0
'''simple docstring''' from __future__ import annotations def __UpperCamelCase ( a : list , a : int | None = None , a : int | None = None ) ->int: if start is None: snake_case = 0 if end is None: snake_case = len(_lowerCamelCase ) - 1 if start >= end: return snake_case = (start + end) // 2 slowsort(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) slowsort(_lowerCamelCase , mid + 1 , _lowerCamelCase ) if sequence[end] < sequence[mid]: snake_case = sequence[mid], sequence[end] slowsort(_lowerCamelCase , _lowerCamelCase , end - 1 ) if __name__ == "__main__": from doctest import testmod testmod()
704
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy _lowercase = logging.get_logger(__name__) class _lowercase ( __a ): def __init__( self , A__ , A__ , A__ , **A__ ) -> Union[str, Any]: snake_case = feature_size snake_case = sampling_rate snake_case = padding_value snake_case = kwargs.pop('''padding_side''' , '''right''' ) snake_case = kwargs.pop('''return_attention_mask''' , A__ ) super().__init__(**A__ ) def UpperCamelCase ( self , A__ , A__ = True , A__ = None , A__ = False , A__ = None , A__ = None , A__ = None , ) -> BatchFeature: # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(A__ , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): snake_case = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( '''You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`''' F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) snake_case = processed_features[self.model_input_names[0]] snake_case = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(A__ ) == 0: if return_attention_mask: snake_case = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch snake_case = required_input[0] if isinstance(A__ , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. snake_case = 0 while len(required_input[index] ) == 0: index += 1 if index < len(A__ ): snake_case = required_input[index][0] if return_tensors is None: if is_tf_tensor(A__ ): snake_case = '''tf''' elif is_torch_tensor(A__ ): snake_case = '''pt''' elif isinstance(A__ , (int, float, list, tuple, np.ndarray) ): snake_case = '''np''' else: raise ValueError( F"""type of {first_element} unknown: {type(A__ )}. """ '''Should be one of a python, numpy, pytorch or tensorflow object.''' ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): snake_case = to_numpy(A__ ) else: snake_case = [to_numpy(A__ ) for v in value] # Convert padding_strategy in PaddingStrategy snake_case = self._get_padding_strategies(padding=A__ , max_length=A__ ) snake_case = processed_features[self.model_input_names[0]] snake_case = len(A__ ) if not all(len(A__ ) == batch_size for v in processed_features.values() ): raise ValueError('''Some items in the output dictionary have a different batch size than others.''' ) snake_case = [] for i in range(A__ ): snake_case = {k: v[i] for k, v in processed_features.items()} # truncation snake_case = self._truncate( A__ , max_length=A__ , pad_to_multiple_of=A__ , truncation=A__ , ) truncated_inputs.append(A__ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length snake_case = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) snake_case = PaddingStrategy.MAX_LENGTH snake_case = {} for i in range(A__ ): # padding snake_case = self._pad( truncated_inputs[i] , max_length=A__ , padding_strategy=A__ , pad_to_multiple_of=A__ , return_attention_mask=A__ , ) for key, value in outputs.items(): if key not in batch_outputs: snake_case = [] if value.dtype is np.dtype(np.floataa ): snake_case = value.astype(np.floataa ) batch_outputs[key].append(A__ ) return BatchFeature(A__ , tensor_type=A__ ) def UpperCamelCase ( self , A__ , A__ = None , A__ = PaddingStrategy.DO_NOT_PAD , A__ = None , A__ = None , ) -> dict: snake_case = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: snake_case = len(A__ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): snake_case = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of snake_case = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(A__ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: snake_case = np.ones(len(A__ ) , dtype=np.intaa ) if needs_to_be_padded: snake_case = max_length - len(A__ ) if self.padding_side == "right": if return_attention_mask: snake_case = np.pad( processed_features['''attention_mask'''] , (0, difference) ) snake_case = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) snake_case = np.pad( A__ , A__ , '''constant''' , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: snake_case = np.pad( processed_features['''attention_mask'''] , (difference, 0) ) snake_case = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) snake_case = np.pad( A__ , A__ , '''constant''' , constant_values=self.padding_value ) else: raise ValueError('''Invalid padding strategy:''' + str(self.padding_side ) ) return processed_features def UpperCamelCase ( self , A__ , A__ = None , A__ = None , A__ = None , ) -> Union[str, Any]: if not truncation: return processed_features elif truncation and max_length is None: raise ValueError('''When setting ``truncation=True``, make sure that ``max_length`` is defined.''' ) snake_case = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): snake_case = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of snake_case = len(A__ ) > max_length if needs_to_be_truncated: snake_case = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: snake_case = processed_features['''attention_mask'''][:max_length] return processed_features def UpperCamelCase ( self , A__=False , A__=None ) -> Union[str, Any]: # Get padding strategy if padding is not False: if padding is True: snake_case = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(A__ , A__ ): snake_case = PaddingStrategy(A__ ) elif isinstance(A__ , A__ ): snake_case = padding else: snake_case = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( '''Asking to pad but the feature_extractor does not have a padding value. Please select a value to use''' ''' as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.''' ) return padding_strategy
44
0
'''simple docstring''' from __future__ import annotations def __UpperCamelCase ( a : list , a : int ) ->Any: # Checks if the entire collection has been sorted if len(a ) <= 1 or n <= 1: return insert_next(a , n - 1 ) rec_insertion_sort(a , n - 1 ) def __UpperCamelCase ( a : list , a : int ) ->Dict: # Checks order between adjacent elements if index >= len(a ) or collection[index - 1] <= collection[index]: return # Swaps adjacent elements since they are not in ascending order snake_case = ( collection[index], collection[index - 1], ) insert_next(a , index + 1 ) if __name__ == "__main__": _lowercase = input('Enter integers separated by spaces: ') _lowercase = [int(num) for num in numbers.split()] rec_insertion_sort(number_list, len(number_list)) print(number_list)
705
'''simple docstring''' from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class _lowercase ( yaml.SafeLoader ): def UpperCamelCase ( self , A__ ) -> List[str]: snake_case = [self.constructed_objects[key_node] for key_node, _ in node.value] snake_case = [tuple(A__ ) if isinstance(A__ , A__ ) else key for key in keys] snake_case = Counter(A__ ) snake_case = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F"""Got duplicate yaml keys: {duplicate_keys}""" ) def UpperCamelCase ( self , A__ , A__=False ) -> List[Any]: snake_case = super().construct_mapping(A__ , deep=A__ ) self._check_no_duplicates_on_constructed_node(A__ ) return mapping def __UpperCamelCase ( a : str ) ->Tuple[Optional[str], str]: snake_case = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: snake_case = full_content[1:].index('''---''' ) + 1 snake_case = '''\n'''.join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(a ) class _lowercase ( __a ): # class attributes _UpperCAmelCase = {'''train_eval_index'''} # train-eval-index in the YAML metadata @classmethod def UpperCamelCase ( cls , A__ ) -> "DatasetMetadata": with open(A__ , encoding='''utf-8''' ) as readme_file: snake_case , snake_case = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(A__ ) else: return cls() def UpperCamelCase ( self , A__ ) -> str: if path.exists(): with open(A__ , encoding='''utf-8''' ) as readme_file: snake_case = readme_file.read() else: snake_case = None snake_case = self._to_readme(A__ ) with open(A__ , '''w''' , encoding='''utf-8''' ) as readme_file: readme_file.write(A__ ) def UpperCamelCase ( self , A__ = None ) -> str: if readme_content is not None: snake_case , snake_case = _split_yaml_from_readme(A__ ) snake_case = '''---\n''' + self.to_yaml_string() + '''---\n''' + content else: snake_case = '''---\n''' + self.to_yaml_string() + '''---\n''' return full_content @classmethod def UpperCamelCase ( cls , A__ ) -> "DatasetMetadata": snake_case = yaml.load(A__ , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields snake_case = { (key.replace('''-''' , '''_''' ) if key.replace('''-''' , '''_''' ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**A__ ) def UpperCamelCase ( self ) -> str: return yaml.safe_dump( { (key.replace('''_''' , '''-''' ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=A__ , allow_unicode=A__ , encoding='''utf-8''' , ).decode('''utf-8''' ) _lowercase = { 'image-classification': [], 'translation': [], 'image-segmentation': [], 'fill-mask': [], 'automatic-speech-recognition': [], 'token-classification': [], 'sentence-similarity': [], 'audio-classification': [], 'question-answering': [], 'summarization': [], 'zero-shot-classification': [], 'table-to-text': [], 'feature-extraction': [], 'other': [], 'multiple-choice': [], 'text-classification': [], 'text-to-image': [], 'text2text-generation': [], 'zero-shot-image-classification': [], 'tabular-classification': [], 'tabular-regression': [], 'image-to-image': [], 'tabular-to-text': [], 'unconditional-image-generation': [], 'text-retrieval': [], 'text-to-speech': [], 'object-detection': [], 'audio-to-audio': [], 'text-generation': [], 'conversational': [], 'table-question-answering': [], 'visual-question-answering': [], 'image-to-text': [], 'reinforcement-learning': [], 'voice-activity-detection': [], 'time-series-forecasting': [], 'document-question-answering': [], } if __name__ == "__main__": from argparse import ArgumentParser _lowercase = ArgumentParser(usage='Validate the yaml metadata block of a README.md file.') ap.add_argument('readme_filepath') _lowercase = ap.parse_args() _lowercase = Path(args.readme_filepath) _lowercase = DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
44
0
'''simple docstring''' import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() _lowercase = logging.get_logger('transformers.models.speecht5') def __UpperCamelCase ( a : int , a : int , a : Tuple ) ->List[Any]: hf_model.apply_weight_norm() snake_case = checkpoint["""input_conv.weight_g"""] snake_case = checkpoint["""input_conv.weight_v"""] snake_case = checkpoint["""input_conv.bias"""] for i in range(len(config.upsample_rates ) ): snake_case = checkpoint[f"""upsamples.{i}.1.weight_g"""] snake_case = checkpoint[f"""upsamples.{i}.1.weight_v"""] snake_case = checkpoint[f"""upsamples.{i}.1.bias"""] for i in range(len(config.upsample_rates ) * len(config.resblock_kernel_sizes ) ): for j in range(len(config.resblock_dilation_sizes ) ): snake_case = checkpoint[f"""blocks.{i}.convs1.{j}.1.weight_g"""] snake_case = checkpoint[f"""blocks.{i}.convs1.{j}.1.weight_v"""] snake_case = checkpoint[f"""blocks.{i}.convs1.{j}.1.bias"""] snake_case = checkpoint[f"""blocks.{i}.convs2.{j}.1.weight_g"""] snake_case = checkpoint[f"""blocks.{i}.convs2.{j}.1.weight_v"""] snake_case = checkpoint[f"""blocks.{i}.convs2.{j}.1.bias"""] snake_case = checkpoint["""output_conv.1.weight_g"""] snake_case = checkpoint["""output_conv.1.weight_v"""] snake_case = checkpoint["""output_conv.1.bias"""] hf_model.remove_weight_norm() @torch.no_grad() def __UpperCamelCase ( a : List[Any] , a : Union[str, Any] , a : int , a : str=None , a : Dict=None , ) ->Any: if config_path is not None: snake_case = SpeechTaHifiGanConfig.from_pretrained(snake_case__ ) else: snake_case = SpeechTaHifiGanConfig() snake_case = SpeechTaHifiGan(snake_case__ ) snake_case = torch.load(snake_case__ ) load_weights(orig_checkpoint['''model''']['''generator'''] , snake_case__ , snake_case__ ) snake_case = np.load(snake_case__ ) snake_case = stats[0].reshape(-1 ) snake_case = stats[1].reshape(-1 ) snake_case = torch.from_numpy(snake_case__ ).float() snake_case = torch.from_numpy(snake_case__ ).float() model.save_pretrained(snake_case__ ) if repo_id: print('''Pushing to the hub...''' ) model.push_to_hub(snake_case__ ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument('--checkpoint_path', required=True, default=None, type=str, help='Path to original checkpoint') parser.add_argument('--stats_path', required=True, default=None, type=str, help='Path to stats.npy file') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--pytorch_dump_folder_path', required=True, default=None, type=str, help='Path to the output PyTorch model.' ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) _lowercase = parser.parse_args() convert_hifigan_checkpoint( args.checkpoint_path, args.stats_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
706
'''simple docstring''' import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowercase ( __a , unittest.TestCase ): _UpperCAmelCase = CodeGenTokenizer _UpperCAmelCase = CodeGenTokenizerFast _UpperCAmelCase = True _UpperCAmelCase = {'''add_prefix_space''': True} _UpperCAmelCase = False def UpperCamelCase ( self ) -> Tuple: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt snake_case = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', '''<|endoftext|>''', ] snake_case = dict(zip(A__ , range(len(A__ ) ) ) ) snake_case = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] snake_case = {'''unk_token''': '''<unk>'''} snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(A__ ) ) def UpperCamelCase ( self , **A__ ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **A__ ) def UpperCamelCase ( self , **A__ ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **A__ ) def UpperCamelCase ( self , A__ ) -> Tuple: snake_case = '''lower newer''' snake_case = '''lower newer''' return input_text, output_text def UpperCamelCase ( self ) -> List[Any]: snake_case = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) snake_case = '''lower newer''' snake_case = ['''\u0120low''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] snake_case = tokenizer.tokenize(A__ , add_prefix_space=A__ ) self.assertListEqual(A__ , A__ ) snake_case = tokens + [tokenizer.unk_token] snake_case = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(A__ ) , A__ ) def UpperCamelCase ( self ) -> Optional[int]: if not self.test_rust_tokenizer: return snake_case = self.get_tokenizer() snake_case = self.get_rust_tokenizer(add_prefix_space=A__ ) snake_case = '''lower newer''' # Testing tokenization snake_case = tokenizer.tokenize(A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.tokenize(A__ ) self.assertListEqual(A__ , A__ ) # Testing conversion to ids without special tokens snake_case = tokenizer.encode(A__ , add_special_tokens=A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.encode(A__ , add_special_tokens=A__ ) self.assertListEqual(A__ , A__ ) # Testing conversion to ids with special tokens snake_case = self.get_rust_tokenizer(add_prefix_space=A__ ) snake_case = tokenizer.encode(A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.encode(A__ ) self.assertListEqual(A__ , A__ ) # Testing the unknown token snake_case = tokens + [rust_tokenizer.unk_token] snake_case = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(A__ ) , A__ ) def UpperCamelCase ( self , *A__ , **A__ ) -> List[str]: # It's very difficult to mix/test pretokenization with byte-level # And get both CodeGen and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def UpperCamelCase ( self , A__=15 ) -> Tuple: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): snake_case = self.rust_tokenizer_class.from_pretrained(A__ , **A__ ) # Simple input snake_case = '''This is a simple input''' snake_case = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case = ('''This is a simple input''', '''This is a pair''') snake_case = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(A__ , tokenizer_r.encode , A__ , max_length=A__ , padding='''max_length''' ) # Simple input self.assertRaises(A__ , tokenizer_r.encode_plus , A__ , max_length=A__ , padding='''max_length''' ) # Simple input self.assertRaises( A__ , tokenizer_r.batch_encode_plus , A__ , max_length=A__ , padding='''max_length''' , ) # Pair input self.assertRaises(A__ , tokenizer_r.encode , A__ , max_length=A__ , padding='''max_length''' ) # Pair input self.assertRaises(A__ , tokenizer_r.encode_plus , A__ , max_length=A__ , padding='''max_length''' ) # Pair input self.assertRaises( A__ , tokenizer_r.batch_encode_plus , A__ , max_length=A__ , padding='''max_length''' , ) def UpperCamelCase ( self ) -> Tuple: snake_case = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token='''<pad>''' ) # Simple input snake_case = '''This is a simple input''' snake_case = ['''This is a simple input looooooooong''', '''This is a simple input'''] snake_case = ('''This is a simple input''', '''This is a pair''') snake_case = [ ('''This is a simple input loooooong''', '''This is a simple input'''), ('''This is a simple pair loooooong''', '''This is a simple pair'''), ] snake_case = tokenizer.pad_token_id snake_case = tokenizer(A__ , padding='''max_length''' , max_length=30 , return_tensors='''np''' ) snake_case = tokenizer(A__ , padding=A__ , truncate=A__ , return_tensors='''np''' ) snake_case = tokenizer(*A__ , padding='''max_length''' , max_length=60 , return_tensors='''np''' ) snake_case = tokenizer(A__ , padding=A__ , truncate=A__ , return_tensors='''np''' ) # s # test single string max_length padding self.assertEqual(out_s['''input_ids'''].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s['''input_ids'''] ) self.assertTrue(0 in out_s['''attention_mask'''] ) # s2 # test automatic padding self.assertEqual(out_sa['''input_ids'''].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['''input_ids'''][0] ) self.assertFalse(0 in out_sa['''attention_mask'''][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['''input_ids'''][1] ) self.assertTrue(0 in out_sa['''attention_mask'''][1] ) # p # test single pair max_length padding self.assertEqual(out_p['''input_ids'''].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p['''input_ids'''] ) self.assertTrue(0 in out_p['''attention_mask'''] ) # p2 # test automatic padding pair self.assertEqual(out_pa['''input_ids'''].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['''input_ids'''][0] ) self.assertFalse(0 in out_pa['''attention_mask'''][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['''input_ids'''][1] ) self.assertTrue(0 in out_pa['''attention_mask'''][1] ) def UpperCamelCase ( self ) -> str: snake_case = '''$$$''' snake_case = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=A__ , add_bos_token=A__ ) snake_case = '''This is a simple input''' snake_case = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case = tokenizer.bos_token_id snake_case = tokenizer(A__ ) snake_case = tokenizer(A__ ) self.assertEqual(out_s.input_ids[0] , A__ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) snake_case = tokenizer.decode(out_s.input_ids ) snake_case = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , A__ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def UpperCamelCase ( self ) -> Any: snake_case = CodeGenTokenizer.from_pretrained('''Salesforce/codegen-350M-mono''' ) snake_case = '''\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#''' snake_case = '''\nif len_a > len_b: result = a\nelse: result = b''' snake_case = tokenizer.encode(A__ ) snake_case = ['''^#''', re.escape('''<|endoftext|>''' ), '''^\'\'\'''', '''^"""''', '''\n\n\n'''] snake_case = tokenizer.decode(A__ , truncate_before_pattern=A__ ) self.assertEqual(A__ , A__ ) def UpperCamelCase ( self ) -> Union[str, Any]: pass
44
0
'''simple docstring''' import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = '▁' _lowercase = {'vocab_file': 'vocab.txt', 'sentencepiece_model_ckpt': 'sentencepiece.bpe.model'} _lowercase = { 'sentencepiece_model_file': 'sentencepiece.bpe.model', 'vocab_file': 'vocab.txt', } _lowercase = { 'vocab_file': { 'ernie-m-base': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt', 'ernie-m-large': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt', }, 'sentencepiece_model_file': { 'ernie-m-base': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model', 'ernie-m-large': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model', }, } _lowercase = { 'ernie-m-base': 514, 'ernie-m-large': 514, } _lowercase = { 'ernie-m-base': {'do_lower_case': False}, 'ernie-m-large': {'do_lower_case': False}, } class _lowercase ( __a ): _UpperCAmelCase = ['''input_ids'''] _UpperCAmelCase = VOCAB_FILES_NAMES _UpperCAmelCase = PRETRAINED_INIT_CONFIGURATION _UpperCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase = RESOURCE_FILES_NAMES def __init__( self , A__ , A__=None , A__=False , A__="utf8" , A__="[UNK]" , A__="[SEP]" , A__="[PAD]" , A__="[CLS]" , A__="[MASK]" , A__ = None , **A__ , ) -> str: snake_case = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__lowerCAmelCase , unk_token=__lowerCAmelCase , sep_token=__lowerCAmelCase , pad_token=__lowerCAmelCase , cls_token=__lowerCAmelCase , mask_token=__lowerCAmelCase , vocab_file=__lowerCAmelCase , encoding=__lowerCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__lowerCAmelCase , ) snake_case = do_lower_case snake_case = sentencepiece_model_ckpt snake_case = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__lowerCAmelCase ) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: snake_case = self.load_vocab(filepath=__lowerCAmelCase ) else: snake_case = {self.sp_model.id_to_piece(__lowerCAmelCase ): id for id in range(self.sp_model.get_piece_size() )} snake_case = {v: k for k, v in self.vocab.items()} def UpperCamelCase ( self , A__ ) -> List[str]: if text is None: return None snake_case = self.tokenize(__lowerCAmelCase ) snake_case , snake_case = '''''', [] for i, ch in enumerate(__lowerCAmelCase ): if ch in self.SP_CHAR_MAPPING: snake_case = self.SP_CHAR_MAPPING.get(__lowerCAmelCase ) else: snake_case = unicodedata.normalize('''NFKC''' , __lowerCAmelCase ) if self.is_whitespace(__lowerCAmelCase ): continue normalized_text += ch char_mapping.extend([i] * len(__lowerCAmelCase ) ) snake_case , snake_case , snake_case = normalized_text, [], 0 if self.do_lower_case: snake_case = text.lower() for token in split_tokens: if token[:1] == "▁": snake_case = token[1:] snake_case = text[offset:].index(__lowerCAmelCase ) + offset snake_case = start + len(__lowerCAmelCase ) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1) ) snake_case = end return token_mapping @property def UpperCamelCase ( self ) -> Any: return len(self.vocab ) def UpperCamelCase ( self ) -> List[str]: return dict(self.vocab , **self.added_tokens_encoder ) def __getstate__( self ) -> List[Any]: snake_case = self.__dict__.copy() snake_case = None return state def __setstate__( self , A__ ) -> Optional[int]: snake_case = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): snake_case = {} snake_case = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.sentencepiece_model_ckpt ) def UpperCamelCase ( self , A__ ) -> List[str]: return "".join((self.SP_CHAR_MAPPING.get(__lowerCAmelCase , __lowerCAmelCase ) for c in text) ) def UpperCamelCase ( self , A__ , A__=False , A__=64 , A__=0.1 ) -> List[Any]: if self.sp_model_kwargs.get('''enable_sampling''' ) is True: snake_case = True if self.sp_model_kwargs.get('''alpha''' ) is not None: snake_case = self.sp_model_kwargs.get('''alpha''' ) if self.sp_model_kwargs.get('''nbest_size''' ) is not None: snake_case = self.sp_model_kwargs.get('''nbest_size''' ) if not enable_sampling: snake_case = self.sp_model.EncodeAsPieces(__lowerCAmelCase ) else: snake_case = self.sp_model.SampleEncodeAsPieces(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) snake_case = [] for pi, piece in enumerate(__lowerCAmelCase ): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(__lowerCAmelCase ) and pi != 0: new_pieces.append(__lowerCAmelCase ) continue else: continue snake_case = 0 for i, chunk in enumerate(__lowerCAmelCase ): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(__lowerCAmelCase ) or self.is_punct(__lowerCAmelCase ): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) new_pieces.append(__lowerCAmelCase ) snake_case = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) snake_case = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) snake_case = i if len(__lowerCAmelCase ) > lst_i: new_pieces.append(piece[lst_i:] ) return new_pieces def UpperCamelCase ( self , A__ ) -> List[str]: snake_case = ''''''.join(__lowerCAmelCase ).replace(__lowerCAmelCase , ''' ''' ).strip() return out_string def UpperCamelCase ( self , A__ ) -> Dict: snake_case = self.convert_ids_to_tokens(__lowerCAmelCase ) snake_case = ''''''.join(__lowerCAmelCase ).replace(__lowerCAmelCase , ''' ''' ).strip() return out_string def UpperCamelCase ( self , A__ ) -> Any: return self.vocab.get(__lowerCAmelCase , self.vocab.get(self.unk_token ) ) def UpperCamelCase ( self , A__ ) -> Dict: return self.reverse_vocab.get(__lowerCAmelCase , self.unk_token ) def UpperCamelCase ( self , A__ , A__=None ) -> Optional[Any]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] snake_case = [self.cls_token_id] snake_case = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def UpperCamelCase ( self , A__ , A__=None ) -> Tuple: if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def UpperCamelCase ( self , A__ , A__=None , A__=False ) -> Tuple: if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(__lowerCAmelCase )) + [1, 1] + ([0] * len(__lowerCAmelCase )) + [1] return [1] + ([0] * len(__lowerCAmelCase )) + [1] def UpperCamelCase ( self , A__ , A__ = None ) -> Dict: if token_ids_a is None: # [CLS] X [SEP] return (len(__lowerCAmelCase ) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(__lowerCAmelCase ) + 1) + [1] * (len(__lowerCAmelCase ) + 3) def UpperCamelCase ( self , A__ ) -> Optional[int]: if "\u4e00" <= char <= "\u9fff": return True return False def UpperCamelCase ( self , A__ ) -> Dict: if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def UpperCamelCase ( self , A__ ) -> List[str]: if char in ",;:.?!~,;:。?!《》【】": return True return False def UpperCamelCase ( self , A__ ) -> Union[str, Any]: if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(__lowerCAmelCase ) == 1: snake_case = unicodedata.category(__lowerCAmelCase ) if cat == "Zs": return True return False def UpperCamelCase ( self , A__ ) -> List[Any]: snake_case = {} with io.open(__lowerCAmelCase , '''r''' , encoding='''utf-8''' ) as f: for index, line in enumerate(__lowerCAmelCase ): snake_case = line.rstrip('''\n''' ) snake_case = int(__lowerCAmelCase ) return token_to_idx def UpperCamelCase ( self , A__ , A__ = None ) -> Union[str, Any]: snake_case = 0 if os.path.isdir(__lowerCAmelCase ): snake_case = os.path.join( __lowerCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) else: snake_case = (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as writer: for token, token_index in sorted(self.vocab.items() , key=lambda A__ : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" ''' Please check that the vocabulary is not corrupted!''' ) snake_case = token_index writer.write(token + '''\n''' ) index += 1 snake_case = os.path.join(__lowerCAmelCase , '''sentencepiece.bpe.model''' ) with open(__lowerCAmelCase , '''wb''' ) as fi: snake_case = self.sp_model.serialized_model_proto() fi.write(__lowerCAmelCase ) return (vocab_file,)
707
'''simple docstring''' from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowercase : def __init__( self , A__ , A__=13 , A__=30 , A__=2 , A__=3 , A__=True , A__=True , A__=32 , A__=2 , A__=4 , A__=37 , A__="gelu" , A__=0.1 , A__=0.1 , A__=10 , A__=0.0_2 , A__=3 , A__=None , ) -> List[Any]: snake_case = parent snake_case = batch_size snake_case = image_size snake_case = patch_size snake_case = num_channels snake_case = is_training snake_case = use_labels snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_size snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = type_sequence_label_size snake_case = initializer_range snake_case = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) snake_case = (image_size // patch_size) ** 2 snake_case = num_patches + 1 def UpperCamelCase ( self ) -> int: snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case = None if self.use_labels: snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ) -> int: return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A__ , initializer_range=self.initializer_range , ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Union[str, Any]: snake_case = TFViTModel(config=A__ ) snake_case = model(A__ , training=A__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. snake_case = self.image_size // 2 snake_case = pixel_values[:, :, :image_size, :image_size] snake_case = model(A__ , interpolate_pos_encoding=A__ , training=A__ ) snake_case = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Optional[int]: snake_case = self.type_sequence_label_size snake_case = TFViTForImageClassification(A__ ) snake_case = model(A__ , labels=A__ , training=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. snake_case = self.image_size // 2 snake_case = pixel_values[:, :, :image_size, :image_size] snake_case = model(A__ , interpolate_pos_encoding=A__ , training=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images snake_case = 1 snake_case = TFViTForImageClassification(A__ ) snake_case = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case = model(A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.prepare_config_and_inputs() snake_case , snake_case , snake_case = config_and_inputs snake_case = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class _lowercase ( __a , __a , unittest.TestCase ): _UpperCAmelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () _UpperCAmelCase = ( {'''feature-extraction''': TFViTModel, '''image-classification''': TFViTForImageClassification} if is_tf_available() else {} ) _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False def UpperCamelCase ( self ) -> List[Any]: snake_case = TFViTModelTester(self ) snake_case = ConfigTester(self , config_class=A__ , has_text_modality=A__ , hidden_size=37 ) def UpperCamelCase ( self ) -> int: self.config_tester.run_common_tests() @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def UpperCamelCase ( self ) -> int: pass @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def UpperCamelCase ( self ) -> str: pass def UpperCamelCase ( self ) -> Union[str, Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(A__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) snake_case = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A__ , tf.keras.layers.Layer ) ) def UpperCamelCase ( self ) -> List[Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(A__ ) snake_case = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case = [*signature.parameters.keys()] snake_case = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A__ ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A__ ) def UpperCamelCase ( self ) -> Optional[Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A__ ) @slow def UpperCamelCase ( self ) -> Any: snake_case = TFViTModel.from_pretrained('''google/vit-base-patch16-224''' ) self.assertIsNotNone(A__ ) def __UpperCamelCase ( ) ->Any: snake_case = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class _lowercase ( unittest.TestCase ): @cached_property def UpperCamelCase ( self ) -> Optional[int]: return ViTImageProcessor.from_pretrained('''google/vit-base-patch16-224''' ) if is_vision_available() else None @slow def UpperCamelCase ( self ) -> Dict: snake_case = TFViTForImageClassification.from_pretrained('''google/vit-base-patch16-224''' ) snake_case = self.default_image_processor snake_case = prepare_img() snake_case = image_processor(images=A__ , return_tensors='''tf''' ) # forward pass snake_case = model(**A__ ) # verify the logits snake_case = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , A__ ) snake_case = tf.constant([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ) tf.debugging.assert_near(outputs.logits[0, :3] , A__ , atol=1e-4 )
44
0
'''simple docstring''' _lowercase = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/" def __UpperCamelCase ( a : bytes ) ->bytes: # Make sure the supplied data is a bytes-like object if not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case = f"""a bytes-like object is required, not \'{data.__class__.__name__}\'""" raise TypeError(UpperCAmelCase__ ) snake_case = """""".join(bin(UpperCAmelCase__ )[2:].zfill(8 ) for byte in data ) snake_case = len(UpperCAmelCase__ ) % 6 != 0 if padding_needed: # The padding that will be added later snake_case = b"""=""" * ((6 - len(UpperCAmelCase__ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(UpperCAmelCase__ ) % 6) else: snake_case = b"""""" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(UpperCAmelCase__ ) , 6 ) ).encode() + padding ) def __UpperCamelCase ( a : str ) ->bytes: # Make sure encoded_data is either a string or a bytes-like object if not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) and not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): snake_case = ( """argument should be a bytes-like object or ASCII string, """ f"""not \'{encoded_data.__class__.__name__}\'""" ) raise TypeError(UpperCAmelCase__ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): try: snake_case = encoded_data.decode('''utf-8''' ) except UnicodeDecodeError: raise ValueError('''base64 encoded data should only contain ASCII characters''' ) snake_case = encoded_data.count('''=''' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(UpperCAmelCase__ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one snake_case = encoded_data[:-padding] snake_case = """""".join( bin(B64_CHARSET.index(UpperCAmelCase__ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: snake_case = """""".join( bin(B64_CHARSET.index(UpperCAmelCase__ ) )[2:].zfill(6 ) for char in encoded_data ) snake_case = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(UpperCAmelCase__ ) , 8 ) ] return bytes(UpperCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod()
708
'''simple docstring''' import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from datasets.iterable_dataset import IterableDataset from datasets.load import dataset_module_factory, import_main_class from datasets.utils.file_utils import cached_path _lowercase = [ {'dataset': 'wikipedia', 'config_name': '20220301.de'}, {'dataset': 'wikipedia', 'config_name': '20220301.en'}, {'dataset': 'wikipedia', 'config_name': '20220301.fr'}, {'dataset': 'wikipedia', 'config_name': '20220301.frr'}, {'dataset': 'wikipedia', 'config_name': '20220301.it'}, {'dataset': 'wikipedia', 'config_name': '20220301.simple'}, {'dataset': 'snli', 'config_name': 'plain_text'}, {'dataset': 'eli5', 'config_name': 'LFQA_reddit'}, {'dataset': 'wiki40b', 'config_name': 'en'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.compressed'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.no_index'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.multiset.no_index'}, {'dataset': 'natural_questions', 'config_name': 'default'}, ] def __UpperCamelCase ( a : Dict=True ) ->str: if with_config: return [ { "testcase_name": d["dataset"] + "/" + d["config_name"], "dataset": d["dataset"], "config_name": d["config_name"], } for d in DATASETS_ON_HF_GCP ] else: return [ {"testcase_name": dataset, "dataset": dataset} for dataset in {d["dataset"] for d in DATASETS_ON_HF_GCP} ] @parameterized.named_parameters(list_datasets_on_hf_gcp_parameters(with_config=__a ) ) class _lowercase ( __a ): _UpperCAmelCase = None _UpperCAmelCase = None def UpperCamelCase ( self , A__ , A__ ) -> str: with TemporaryDirectory() as tmp_dir: snake_case = dataset_module_factory(A__ , cache_dir=A__ ) snake_case = import_main_class(dataset_module.module_path , dataset=A__ ) snake_case = builder_cls( cache_dir=A__ , config_name=A__ , hash=dataset_module.hash , ) snake_case = '''/'''.join( [ HF_GCP_BASE_URL, builder_instance._relative_data_dir(with_hash=A__ ).replace(os.sep , '''/''' ), config.DATASET_INFO_FILENAME, ] ) snake_case = cached_path(A__ , cache_dir=A__ ) self.assertTrue(os.path.exists(A__ ) ) @pytest.mark.integration def __UpperCamelCase ( a : List[str] ) ->Any: snake_case = tmp_path_factory.mktemp('''test_hf_gcp''' ) / '''test_wikipedia_simple''' snake_case = dataset_module_factory('''wikipedia''' , cache_dir=a ) snake_case = import_main_class(dataset_module.module_path ) snake_case = builder_cls( cache_dir=a , config_name='''20220301.frr''' , hash=dataset_module.hash , ) # use the HF cloud storage, not the original download_and_prepare that uses apache-beam snake_case = None builder_instance.download_and_prepare() snake_case = builder_instance.as_dataset() assert ds @pytest.mark.integration def __UpperCamelCase ( a : Any ) ->Union[str, Any]: snake_case = dataset_module_factory('''wikipedia''' , cache_dir=a ) snake_case = import_main_class(dataset_module.module_path , dataset=a ) snake_case = builder_cls( cache_dir=a , config_name='''20220301.frr''' , hash=dataset_module.hash , ) snake_case = builder_instance.as_streaming_dataset() assert ds assert isinstance(a , a ) assert "train" in ds assert isinstance(ds['''train'''] , a ) assert next(iter(ds['''train'''] ) )
44
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowercase = { '''configuration_longformer''': [ '''LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LongformerConfig''', '''LongformerOnnxConfig''', ], '''tokenization_longformer''': ['''LongformerTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ['''LongformerTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ '''LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''LongformerForMaskedLM''', '''LongformerForMultipleChoice''', '''LongformerForQuestionAnswering''', '''LongformerForSequenceClassification''', '''LongformerForTokenClassification''', '''LongformerModel''', '''LongformerPreTrainedModel''', '''LongformerSelfAttention''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ '''TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFLongformerForMaskedLM''', '''TFLongformerForMultipleChoice''', '''TFLongformerForQuestionAnswering''', '''TFLongformerForSequenceClassification''', '''TFLongformerForTokenClassification''', '''TFLongformerModel''', '''TFLongformerPreTrainedModel''', '''TFLongformerSelfAttention''', ] if TYPE_CHECKING: from .configuration_longformer import ( LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, LongformerConfig, LongformerOnnxConfig, ) from .tokenization_longformer import LongformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_longformer_fast import LongformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longformer import ( LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, LongformerForMaskedLM, LongformerForMultipleChoice, LongformerForQuestionAnswering, LongformerForSequenceClassification, LongformerForTokenClassification, LongformerModel, LongformerPreTrainedModel, LongformerSelfAttention, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_longformer import ( TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFLongformerForMaskedLM, TFLongformerForMultipleChoice, TFLongformerForQuestionAnswering, TFLongformerForSequenceClassification, TFLongformerForTokenClassification, TFLongformerModel, TFLongformerPreTrainedModel, TFLongformerSelfAttention, ) else: import sys _lowercase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
709
'''simple docstring''' def __UpperCamelCase ( a : int , a : int ) ->int: while b: snake_case , snake_case = b, a % b return a def __UpperCamelCase ( a : int , a : int ) ->int: return a if b == 0 else euclidean_gcd_recursive(a , a % b ) def __UpperCamelCase ( ) ->Optional[Any]: print(f"""euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}""" ) print(f"""euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}""" ) print(f"""euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}""" ) print(f"""euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}""" ) print(f"""euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}""" ) print(f"""euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}""" ) print(f"""euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}""" ) print(f"""euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}""" ) if __name__ == "__main__": main()
44
0
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPanoramaPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() @skip_mps class _lowercase ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): _UpperCAmelCase = StableDiffusionPanoramaPipeline _UpperCAmelCase = TEXT_TO_IMAGE_PARAMS _UpperCAmelCase = TEXT_TO_IMAGE_BATCH_PARAMS _UpperCAmelCase = TEXT_TO_IMAGE_IMAGE_PARAMS _UpperCAmelCase = TEXT_TO_IMAGE_IMAGE_PARAMS def UpperCamelCase ( self ) -> List[str]: torch.manual_seed(0 ) snake_case = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=1 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) snake_case = DDIMScheduler() torch.manual_seed(0 ) snake_case = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) torch.manual_seed(0 ) snake_case = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) snake_case = CLIPTextModel(__lowerCAmelCase ) snake_case = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) snake_case = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def UpperCamelCase ( self , A__ , A__=0 ) -> Optional[int]: snake_case = torch.manual_seed(__lowerCAmelCase ) snake_case = { '''prompt''': '''a photo of the dolomites''', '''generator''': generator, # Setting height and width to None to prevent OOMs on CPU. '''height''': None, '''width''': None, '''num_inference_steps''': 1, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def UpperCamelCase ( self ) -> str: snake_case = '''cpu''' # ensure determinism for the device-dependent torch.Generator snake_case = self.get_dummy_components() snake_case = StableDiffusionPanoramaPipeline(**__lowerCAmelCase ) snake_case = sd_pipe.to(__lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=__lowerCAmelCase ) snake_case = self.get_dummy_inputs(__lowerCAmelCase ) snake_case = sd_pipe(**__lowerCAmelCase ).images snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) snake_case = np.array([0.6_1_8_6, 0.5_3_7_4, 0.4_9_1_5, 0.4_1_3_5, 0.4_1_1_4, 0.4_5_6_3, 0.5_1_2_8, 0.4_9_7_7, 0.4_7_5_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase ( self ) -> Dict: super().test_inference_batch_consistent(batch_sizes=[1, 2] ) def UpperCamelCase ( self ) -> Optional[Any]: super().test_inference_batch_single_identical(batch_size=2 , expected_max_diff=3.25e-3 ) def UpperCamelCase ( self ) -> Optional[int]: snake_case = '''cpu''' # ensure determinism for the device-dependent torch.Generator snake_case = self.get_dummy_components() snake_case = StableDiffusionPanoramaPipeline(**__lowerCAmelCase ) snake_case = sd_pipe.to(__lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=__lowerCAmelCase ) snake_case = self.get_dummy_inputs(__lowerCAmelCase ) snake_case = '''french fries''' snake_case = sd_pipe(**__lowerCAmelCase , negative_prompt=__lowerCAmelCase ) snake_case = output.images snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) snake_case = np.array([0.6_1_8_7, 0.5_3_7_5, 0.4_9_1_5, 0.4_1_3_6, 0.4_1_1_4, 0.4_5_6_3, 0.5_1_2_8, 0.4_9_7_6, 0.4_7_5_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase ( self ) -> str: snake_case = '''cpu''' # ensure determinism for the device-dependent torch.Generator snake_case = self.get_dummy_components() snake_case = StableDiffusionPanoramaPipeline(**__lowerCAmelCase ) snake_case = sd_pipe.to(__lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=__lowerCAmelCase ) snake_case = self.get_dummy_inputs(__lowerCAmelCase ) snake_case = sd_pipe(**__lowerCAmelCase , view_batch_size=2 ) snake_case = output.images snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) snake_case = np.array([0.6_1_8_7, 0.5_3_7_5, 0.4_9_1_5, 0.4_1_3_6, 0.4_1_1_4, 0.4_5_6_3, 0.5_1_2_8, 0.4_9_7_6, 0.4_7_5_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase ( self ) -> Any: snake_case = '''cpu''' # ensure determinism for the device-dependent torch.Generator snake_case = self.get_dummy_components() snake_case = EulerAncestralDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule='''scaled_linear''' ) snake_case = StableDiffusionPanoramaPipeline(**__lowerCAmelCase ) snake_case = sd_pipe.to(__lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=__lowerCAmelCase ) snake_case = self.get_dummy_inputs(__lowerCAmelCase ) snake_case = sd_pipe(**__lowerCAmelCase ).images snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) snake_case = np.array([0.4_0_2_4, 0.6_5_1_0, 0.4_9_0_1, 0.5_3_7_8, 0.5_8_1_3, 0.5_6_2_2, 0.4_7_9_5, 0.4_4_6_7, 0.4_9_5_2] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCamelCase ( self ) -> List[Any]: snake_case = '''cpu''' # ensure determinism for the device-dependent torch.Generator snake_case = self.get_dummy_components() snake_case = PNDMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule='''scaled_linear''' , skip_prk_steps=__lowerCAmelCase ) snake_case = StableDiffusionPanoramaPipeline(**__lowerCAmelCase ) snake_case = sd_pipe.to(__lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=__lowerCAmelCase ) snake_case = self.get_dummy_inputs(__lowerCAmelCase ) snake_case = sd_pipe(**__lowerCAmelCase ).images snake_case = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) snake_case = np.array([0.6_3_9_1, 0.6_2_9_1, 0.4_8_6_1, 0.5_1_3_4, 0.5_5_5_2, 0.4_5_7_8, 0.5_0_3_2, 0.5_0_2_3, 0.4_5_3_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class _lowercase ( unittest.TestCase ): def UpperCamelCase ( self ) -> Union[str, Any]: super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self , A__=0 ) -> Tuple: snake_case = torch.manual_seed(__lowerCAmelCase ) snake_case = { '''prompt''': '''a photo of the dolomites''', '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = '''stabilityai/stable-diffusion-2-base''' snake_case = DDIMScheduler.from_pretrained(__lowerCAmelCase , subfolder='''scheduler''' ) snake_case = StableDiffusionPanoramaPipeline.from_pretrained(__lowerCAmelCase , scheduler=__lowerCAmelCase , safety_checker=__lowerCAmelCase ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) pipe.enable_attention_slicing() snake_case = self.get_inputs() snake_case = pipe(**__lowerCAmelCase ).images snake_case = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 20_48, 3) snake_case = np.array( [ 0.3_6_9_6_8_3_9_2, 0.2_7_0_2_5_3_7_2, 0.3_2_4_4_6_7_6_6, 0.2_8_3_7_9_3_8_7, 0.3_6_3_6_3_2_7_4, 0.3_0_7_3_3_3_4_7, 0.2_7_1_0_0_0_2_7, 0.2_7_0_5_4_1_2_5, 0.2_5_5_3_6_0_9_6, ] ) assert np.abs(expected_slice - image_slice ).max() < 1e-2 def UpperCamelCase ( self ) -> List[Any]: snake_case = StableDiffusionPanoramaPipeline.from_pretrained( '''stabilityai/stable-diffusion-2-base''' , safety_checker=__lowerCAmelCase ) snake_case = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) pipe.enable_attention_slicing() snake_case = self.get_inputs() snake_case = pipe(**__lowerCAmelCase ).images snake_case = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 5_12, 20_48, 3) snake_case = np.array( [ [ 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, ] ] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def UpperCamelCase ( self ) -> int: snake_case = 0 def callback_fn(A__ , A__ , A__ ) -> None: snake_case = True nonlocal number_of_steps number_of_steps += 1 if step == 1: snake_case = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 2_56) snake_case = latents[0, -3:, -3:, -1] snake_case = np.array( [ 0.1_8_6_8_1_8_6_9, 0.3_3_9_0_7_8_1_6, 0.5_3_6_1_2_7_6, 0.1_4_4_3_2_8_6_5, -0.0_2_8_5_6_6_1_1, -0.7_3_9_4_1_1_2_3, 0.2_3_3_9_7_9_8_7, 0.4_7_3_2_2_6_8_2, -0.3_7_8_2_3_1_6_4, ] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: snake_case = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 2_56) snake_case = latents[0, -3:, -3:, -1] snake_case = np.array( [ 0.1_8_5_3_9_6_4_5, 0.3_3_9_8_7_2_4_8, 0.5_3_7_8_5_5_9, 0.1_4_4_3_7_1_4_2, -0.0_2_4_5_5_2_6_1, -0.7_3_3_8_3_1_7, 0.2_3_9_9_0_7_5_5, 0.4_7_3_5_6_2_7_2, -0.3_7_8_6_5_0_5, ] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 snake_case = False snake_case = '''stabilityai/stable-diffusion-2-base''' snake_case = DDIMScheduler.from_pretrained(__lowerCAmelCase , subfolder='''scheduler''' ) snake_case = StableDiffusionPanoramaPipeline.from_pretrained(__lowerCAmelCase , scheduler=__lowerCAmelCase , safety_checker=__lowerCAmelCase ) snake_case = pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) pipe.enable_attention_slicing() snake_case = self.get_inputs() pipe(**__lowerCAmelCase , callback=__lowerCAmelCase , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def UpperCamelCase ( self ) -> Optional[Any]: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() snake_case = '''stabilityai/stable-diffusion-2-base''' snake_case = DDIMScheduler.from_pretrained(__lowerCAmelCase , subfolder='''scheduler''' ) snake_case = StableDiffusionPanoramaPipeline.from_pretrained(__lowerCAmelCase , scheduler=__lowerCAmelCase , safety_checker=__lowerCAmelCase ) snake_case = pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() snake_case = self.get_inputs() snake_case = pipe(**__lowerCAmelCase ) snake_case = torch.cuda.max_memory_allocated() # make sure that less than 5.2 GB is allocated assert mem_bytes < 5.5 * 10**9
710
'''simple docstring''' import argparse import copy def __UpperCamelCase ( a : Union[str, Any] ) ->Tuple: snake_case = {} with open(a ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: snake_case = [] _list.append([line.split()[1], line.split()[2]] ) snake_case = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: snake_case = [] _list.append([line.split()[0], line.split()[2]] ) snake_case = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def __UpperCamelCase ( a : Dict , a : Tuple ) ->int: with open(a ) as f: snake_case = f.read(1 ) snake_case = start_node snake_case = [] snake_case = start_node snake_case = 0 while visiting not in first_solution: snake_case = 1_0000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(a ) and k[0] not in first_solution: snake_case = k[1] snake_case = k[0] first_solution.append(a ) snake_case = distance_of_first_solution + int(a ) snake_case = best_node first_solution.append(a ) snake_case = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 snake_case = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_0000 ) return first_solution, distance_of_first_solution def __UpperCamelCase ( a : Optional[int] , a : str ) ->str: snake_case = [] for n in solution[1:-1]: snake_case = solution.index(a ) for kn in solution[1:-1]: snake_case = solution.index(a ) if n == kn: continue snake_case = copy.deepcopy(a ) snake_case = kn snake_case = n snake_case = 0 for k in _tmp[:-1]: snake_case = _tmp[_tmp.index(a ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: snake_case = distance + int(i[1] ) _tmp.append(a ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) snake_case = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda a : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def __UpperCamelCase ( a : Any , a : Optional[Any] , a : int , a : Optional[int] , a : Union[str, Any] ) ->List[Any]: snake_case = 1 snake_case = first_solution snake_case = [] snake_case = distance_of_first_solution snake_case = solution while count <= iters: snake_case = find_neighborhood(a , a ) snake_case = 0 snake_case = neighborhood[index_of_best_solution] snake_case = len(a ) - 1 snake_case = False while not found: snake_case = 0 while i < len(a ): if best_solution[i] != solution[i]: snake_case = best_solution[i] snake_case = solution[i] break snake_case = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) snake_case = True snake_case = best_solution[:-1] snake_case = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: snake_case = cost snake_case = solution else: snake_case = index_of_best_solution + 1 snake_case = neighborhood[index_of_best_solution] if len(a ) >= size: tabu_list.pop(0 ) snake_case = count + 1 return best_solution_ever, best_cost def __UpperCamelCase ( a : Union[str, Any]=None ) ->Optional[Any]: snake_case = generate_neighbours(args.File ) snake_case , snake_case = generate_first_solution( args.File , a ) snake_case , snake_case = tabu_search( a , a , a , args.Iterations , args.Size , ) print(f"""Best solution: {best_sol}, with total distance: {best_cost}.""" ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser(description='Tabu Search') parser.add_argument( '-f', '--File', type=str, help='Path to the file containing the data', required=True, ) parser.add_argument( '-i', '--Iterations', type=int, help='How many iterations the algorithm should perform', required=True, ) parser.add_argument( '-s', '--Size', type=int, help='Size of the tabu list', required=True ) # Pass the arguments to main method main(parser.parse_args())
44
0
'''simple docstring''' import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { 'asapp/sew-tiny-100k': 'https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json', # See all SEW models at https://huggingface.co/models?filter=sew } class _lowercase ( __UpperCAmelCase ): _UpperCAmelCase = """sew""" def __init__( self , A__=32 , A__=7_68 , A__=12 , A__=12 , A__=30_72 , A__=2 , A__="gelu" , A__=0.1 , A__=0.1 , A__=0.1 , A__=0.0 , A__=0.1 , A__=0.1 , A__=0.0_2 , A__=1e-5 , A__="group" , A__="gelu" , A__=(64, 1_28, 1_28, 1_28, 1_28, 2_56, 2_56, 2_56, 2_56, 5_12, 5_12, 5_12, 5_12) , A__=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , A__=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , A__=False , A__=1_28 , A__=16 , A__=True , A__=0.0_5 , A__=10 , A__=2 , A__=0.0 , A__=10 , A__=0 , A__="mean" , A__=False , A__=False , A__=2_56 , A__=0 , A__=1 , A__=2 , **A__ , ) -> Dict: super().__init__(**UpperCAmelCase_ , pad_token_id=UpperCAmelCase_ , bos_token_id=UpperCAmelCase_ , eos_token_id=UpperCAmelCase_ ) snake_case = hidden_size snake_case = feat_extract_norm snake_case = feat_extract_activation snake_case = list(UpperCAmelCase_ ) snake_case = list(UpperCAmelCase_ ) snake_case = list(UpperCAmelCase_ ) snake_case = conv_bias snake_case = num_conv_pos_embeddings snake_case = num_conv_pos_embedding_groups snake_case = len(self.conv_dim ) snake_case = num_hidden_layers snake_case = intermediate_size snake_case = squeeze_factor snake_case = hidden_act snake_case = num_attention_heads snake_case = hidden_dropout snake_case = attention_dropout snake_case = activation_dropout snake_case = feat_proj_dropout snake_case = final_dropout snake_case = layerdrop snake_case = layer_norm_eps snake_case = initializer_range snake_case = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect.''' '''It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,''' F"""but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)""" F"""= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 snake_case = apply_spec_augment snake_case = mask_time_prob snake_case = mask_time_length snake_case = mask_time_min_masks snake_case = mask_feature_prob snake_case = mask_feature_length snake_case = mask_feature_min_masks # ctc loss snake_case = ctc_loss_reduction snake_case = ctc_zero_infinity # sequence classification snake_case = use_weighted_layer_sum snake_case = classifier_proj_size @property def UpperCamelCase ( self ) -> List[str]: return functools.reduce(operator.mul , self.conv_stride , 1 )
711
'''simple docstring''' from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .notes_encoder import SpectrogramNotesEncoder from .continous_encoder import SpectrogramContEncoder from .pipeline_spectrogram_diffusion import ( SpectrogramContEncoder, SpectrogramDiffusionPipeline, TaFilmDecoder, ) try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .midi_utils import MidiProcessor
44
0
import copy import fnmatch import json import os import pickle as pkl import shutil import sys import tarfile import tempfile from collections import OrderedDict from contextlib import contextmanager from functools import partial from hashlib import shaaaa from io import BytesIO from pathlib import Path from urllib.parse import urlparse from zipfile import ZipFile, is_zipfile import cva import numpy as np import requests import wget from filelock import FileLock from PIL import Image from tqdm.auto import tqdm from yaml import Loader, dump, load try: import torch _lowercase = True except ImportError: _lowercase = False try: from torch.hub import _get_torch_home _lowercase = _get_torch_home() except ImportError: _lowercase = os.path.expanduser( os.getenv('TORCH_HOME', os.path.join(os.getenv('XDG_CACHE_HOME', '~/.cache'), 'torch')) ) _lowercase = os.path.join(torch_cache_home, 'transformers') _lowercase = 'https://cdn.huggingface.co' _lowercase = 'https://s3.amazonaws.com/models.huggingface.co/bert' _lowercase = '/'.join(str(Path(__file__).resolve()).split('/')[:-1]) _lowercase = os.path.join(PATH, 'config.yaml') _lowercase = os.path.join(PATH, 'attributes.txt') _lowercase = os.path.join(PATH, 'objects.txt') _lowercase = os.getenv('PYTORCH_PRETRAINED_BERT_CACHE', default_cache_path) _lowercase = os.getenv('PYTORCH_TRANSFORMERS_CACHE', PYTORCH_PRETRAINED_BERT_CACHE) _lowercase = os.getenv('TRANSFORMERS_CACHE', PYTORCH_TRANSFORMERS_CACHE) _lowercase = 'pytorch_model.bin' _lowercase = 'config.yaml' def __UpperCamelCase ( a : Tuple=OBJECTS , a : List[Any]=ATTRIBUTES ) ->Dict: snake_case = [] with open(_lowercase ) as f: for object in f.readlines(): vg_classes.append(object.split(''',''' )[0].lower().strip() ) snake_case = [] with open(_lowercase ) as f: for object in f.readlines(): vg_attrs.append(object.split(''',''' )[0].lower().strip() ) return vg_classes, vg_attrs def __UpperCamelCase ( a : int ) ->Dict: snake_case = OrderedDict() with open(_lowercase , '''rb''' ) as f: snake_case = pkl.load(_lowercase )['''model'''] for k in copy.deepcopy(list(ckp.keys() ) ): snake_case = ckp.pop(_lowercase ) if isinstance(_lowercase , np.ndarray ): snake_case = torch.tensor(_lowercase ) else: assert isinstance(_lowercase , torch.tensor ), type(_lowercase ) snake_case = v return r class _lowercase : _UpperCAmelCase = {} def __init__( self , A__ , A__ = "root" , A__=0 ) -> List[Any]: snake_case = name snake_case = level snake_case = {} for k, v in dictionary.items(): if v is None: raise ValueError() snake_case = copy.deepcopy(UpperCamelCase__ ) snake_case = copy.deepcopy(UpperCamelCase__ ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ): snake_case = Config(UpperCamelCase__ , name=UpperCamelCase__ , level=level + 1 ) snake_case = v setattr(self , UpperCamelCase__ , UpperCamelCase__ ) snake_case = d def __repr__( self ) -> str: return str(list((self._pointer.keys()) ) ) def __setattr__( self , A__ , A__ ) -> List[str]: snake_case = val snake_case = val snake_case = key.split('''.''' ) snake_case = len(UpperCamelCase__ ) - 1 snake_case = self._pointer if len(UpperCamelCase__ ) > 1: for i, l in enumerate(UpperCamelCase__ ): if hasattr(self , UpperCamelCase__ ) and isinstance(getattr(self , UpperCamelCase__ ) , UpperCamelCase__ ): setattr(getattr(self , UpperCamelCase__ ) , '''.'''.join(levels[i:] ) , UpperCamelCase__ ) if l == last_level: snake_case = val else: snake_case = pointer[l] def UpperCamelCase ( self ) -> List[Any]: return self._pointer def UpperCamelCase ( self , A__ , A__ ) -> Tuple: with open(F"""{file_name}""" , '''w''' ) as stream: dump(UpperCamelCase__ , UpperCamelCase__ ) def UpperCamelCase ( self , A__ , A__ ) -> Any: with open(F"""{file_name}""" , '''w''' ) as stream: json.dump(UpperCamelCase__ , UpperCamelCase__ ) @staticmethod def UpperCamelCase ( A__ ) -> Any: with open(UpperCamelCase__ ) as stream: snake_case = load(UpperCamelCase__ , Loader=UpperCamelCase__ ) return data def __str__( self ) -> List[Any]: snake_case = ''' ''' if self._name != "root": snake_case = F"""{t * (self._level-1)}{self._name}:\n""" else: snake_case = '''''' snake_case = self._level for i, (k, v) in enumerate(self._pointer.items() ): if isinstance(UpperCamelCase__ , UpperCamelCase__ ): r += F"""{t * (self._level)}{v}\n""" self._level += 1 else: r += F"""{t * (self._level)}{k}: {v} ({type(UpperCamelCase__ ).__name__})\n""" snake_case = level return r[:-1] @classmethod def UpperCamelCase ( cls , A__ , **A__ ) -> List[Any]: snake_case = cls.get_config_dict(UpperCamelCase__ , **UpperCamelCase__ ) return cls(UpperCamelCase__ ) @classmethod def UpperCamelCase ( cls , A__ , **A__ ) -> Union[str, Any]: snake_case = kwargs.pop('''cache_dir''' , UpperCamelCase__ ) snake_case = kwargs.pop('''force_download''' , UpperCamelCase__ ) snake_case = kwargs.pop('''resume_download''' , UpperCamelCase__ ) snake_case = kwargs.pop('''proxies''' , UpperCamelCase__ ) snake_case = kwargs.pop('''local_files_only''' , UpperCamelCase__ ) if os.path.isdir(UpperCamelCase__ ): snake_case = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) elif os.path.isfile(UpperCamelCase__ ) or is_remote_url(UpperCamelCase__ ): snake_case = pretrained_model_name_or_path else: snake_case = hf_bucket_url(UpperCamelCase__ , filename=UpperCamelCase__ , use_cdn=UpperCamelCase__ ) try: # Load from URL or cache if already cached snake_case = cached_path( UpperCamelCase__ , cache_dir=UpperCamelCase__ , force_download=UpperCamelCase__ , proxies=UpperCamelCase__ , resume_download=UpperCamelCase__ , local_files_only=UpperCamelCase__ , ) # Load config dict if resolved_config_file is None: raise EnvironmentError snake_case = Config.load_yaml(UpperCamelCase__ ) except EnvironmentError: snake_case = '''Can\'t load config for''' raise EnvironmentError(UpperCamelCase__ ) if resolved_config_file == config_file: print('''loading configuration file from path''' ) else: print('''loading configuration file cache''' ) return Config.load_yaml(UpperCamelCase__ ), kwargs def __UpperCamelCase ( a : Optional[Any] ) ->Any: snake_case = torch.load('''dump.pt''' , map_location=in_tensor.device ) snake_case = in_tensor.numpy() snake_case = out_tensor.numpy()[0] print(na.shape , na[0, 0, :5] ) print(na.shape , na[0, 0, :5] ) assert np.allclose(_lowercase , _lowercase , rtol=0.01 , atol=0.1 ), ( f"""{sum([1 for x in np.isclose(_lowercase , _lowercase , rtol=0.01 , atol=0.1 ).flatten() if x is False] )/len(na.flatten() )*100:.4f} %""" " element-wise mismatch" ) raise Exception('''tensors are all good''' ) # Hugging face functions below def __UpperCamelCase ( a : Optional[Any] ) ->int: snake_case = urlparse(_lowercase ) return parsed.scheme in ("http", "https") def __UpperCamelCase ( a : Union[str, Any] , a : Tuple , a : Optional[int]=True ) ->List[str]: snake_case = CLOUDFRONT_DISTRIB_PREFIX if use_cdn else S3_BUCKET_PREFIX snake_case = '''/''' not in model_id if legacy_format: return f"""{endpoint}/{model_id}-{filename}""" else: return f"""{endpoint}/{model_id}/{filename}""" def __UpperCamelCase ( a : Any , a : Tuple , a : List[Any]=None , a : Optional[int]=0 , a : Any=None , ) ->Dict: snake_case = '''python/{}'''.format(sys.version.split()[0] ) if _torch_available: ua += "; torch/{}".format(torch.__version__ ) if isinstance(_lowercase , _lowercase ): ua += "; " + "; ".join('''{}/{}'''.format(_lowercase , _lowercase ) for k, v in user_agent.items() ) elif isinstance(_lowercase , _lowercase ): ua += "; " + user_agent snake_case = {'''user-agent''': ua} if resume_size > 0: snake_case = '''bytes=%d-''' % (resume_size,) snake_case = requests.get(_lowercase , stream=_lowercase , proxies=_lowercase , headers=_lowercase ) if response.status_code == 416: # Range not satisfiable return snake_case = response.headers.get('''Content-Length''' ) snake_case = resume_size + int(_lowercase ) if content_length is not None else None snake_case = tqdm( unit='''B''' , unit_scale=_lowercase , total=_lowercase , initial=_lowercase , desc='''Downloading''' , ) for chunk in response.iter_content(chunk_size=1024 ): if chunk: # filter out keep-alive new chunks progress.update(len(_lowercase ) ) temp_file.write(_lowercase ) progress.close() def __UpperCamelCase ( a : Dict , a : Tuple=None , a : Tuple=False , a : Tuple=None , a : List[str]=10 , a : int=False , a : str=None , a : str=False , ) ->Optional[int]: if cache_dir is None: snake_case = TRANSFORMERS_CACHE if isinstance(_lowercase , _lowercase ): snake_case = str(_lowercase ) os.makedirs(_lowercase , exist_ok=_lowercase ) snake_case = None if not local_files_only: try: snake_case = requests.head(_lowercase , allow_redirects=_lowercase , proxies=_lowercase , timeout=_lowercase ) if response.status_code == 200: snake_case = response.headers.get('''ETag''' ) except (EnvironmentError, requests.exceptions.Timeout): # etag is already None pass snake_case = url_to_filename(_lowercase , _lowercase ) # get cache path to put the file snake_case = os.path.join(_lowercase , _lowercase ) # etag is None = we don't have a connection, or url doesn't exist, or is otherwise inaccessible. # try to get the last downloaded one if etag is None: if os.path.exists(_lowercase ): return cache_path else: snake_case = [ file for file in fnmatch.filter(os.listdir(_lowercase ) , filename + '''.*''' ) if not file.endswith('''.json''' ) and not file.endswith('''.lock''' ) ] if len(_lowercase ) > 0: return os.path.join(_lowercase , matching_files[-1] ) else: # If files cannot be found and local_files_only=True, # the models might've been found if local_files_only=False # Notify the user about that if local_files_only: raise ValueError( '''Cannot find the requested files in the cached path and outgoing traffic has been''' ''' disabled. To enable model look-ups and downloads online, set \'local_files_only\'''' ''' to False.''' ) return None # From now on, etag is not None. if os.path.exists(_lowercase ) and not force_download: return cache_path # Prevent parallel downloads of the same file with a lock. snake_case = cache_path + '''.lock''' with FileLock(_lowercase ): # If the download just completed while the lock was activated. if os.path.exists(_lowercase ) and not force_download: # Even if returning early like here, the lock will be released. return cache_path if resume_download: snake_case = cache_path + '''.incomplete''' @contextmanager def _resumable_file_manager(): with open(_lowercase , '''a+b''' ) as f: yield f snake_case = _resumable_file_manager if os.path.exists(_lowercase ): snake_case = os.stat(_lowercase ).st_size else: snake_case = 0 else: snake_case = partial(tempfile.NamedTemporaryFile , dir=_lowercase , delete=_lowercase ) snake_case = 0 # Download to temporary file, then copy to cache dir once finished. # Otherwise you get corrupt cache entries if the download gets interrupted. with temp_file_manager() as temp_file: print( '''%s not found in cache or force_download set to True, downloading to %s''' , _lowercase , temp_file.name , ) http_get( _lowercase , _lowercase , proxies=_lowercase , resume_size=_lowercase , user_agent=_lowercase , ) os.replace(temp_file.name , _lowercase ) snake_case = {'''url''': url, '''etag''': etag} snake_case = cache_path + '''.json''' with open(_lowercase , '''w''' ) as meta_file: json.dump(_lowercase , _lowercase ) return cache_path def __UpperCamelCase ( a : int , a : Dict=None ) ->Tuple: snake_case = url.encode('''utf-8''' ) snake_case = shaaaa(_lowercase ) snake_case = url_hash.hexdigest() if etag: snake_case = etag.encode('''utf-8''' ) snake_case = shaaaa(_lowercase ) filename += "." + etag_hash.hexdigest() if url.endswith('''.h5''' ): filename += ".h5" return filename def __UpperCamelCase ( a : List[str] , a : int=None , a : Optional[Any]=False , a : Union[str, Any]=None , a : Optional[int]=False , a : List[str]=None , a : str=False , a : Optional[Any]=False , a : Optional[Any]=False , ) ->Union[str, Any]: if cache_dir is None: snake_case = TRANSFORMERS_CACHE if isinstance(_lowercase , _lowercase ): snake_case = str(_lowercase ) if isinstance(_lowercase , _lowercase ): snake_case = str(_lowercase ) if is_remote_url(_lowercase ): # URL, so get it from the cache (downloading if necessary) snake_case = get_from_cache( _lowercase , cache_dir=_lowercase , force_download=_lowercase , proxies=_lowercase , resume_download=_lowercase , user_agent=_lowercase , local_files_only=_lowercase , ) elif os.path.exists(_lowercase ): # File, and it exists. snake_case = url_or_filename elif urlparse(_lowercase ).scheme == "": # File, but it doesn't exist. raise EnvironmentError('''file {} not found'''.format(_lowercase ) ) else: # Something unknown raise ValueError('''unable to parse {} as a URL or as a local path'''.format(_lowercase ) ) if extract_compressed_file: if not is_zipfile(_lowercase ) and not tarfile.is_tarfile(_lowercase ): return output_path # Path where we extract compressed archives # We avoid '.' in dir name and add "-extracted" at the end: "./model.zip" => "./model-zip-extracted/" snake_case = os.path.split(_lowercase ) snake_case = output_file.replace('''.''' , '''-''' ) + '''-extracted''' snake_case = os.path.join(_lowercase , _lowercase ) if os.path.isdir(_lowercase ) and os.listdir(_lowercase ) and not force_extract: return output_path_extracted # Prevent parallel extractions snake_case = output_path + '''.lock''' with FileLock(_lowercase ): shutil.rmtree(_lowercase , ignore_errors=_lowercase ) os.makedirs(_lowercase ) if is_zipfile(_lowercase ): with ZipFile(_lowercase , '''r''' ) as zip_file: zip_file.extractall(_lowercase ) zip_file.close() elif tarfile.is_tarfile(_lowercase ): snake_case = tarfile.open(_lowercase ) tar_file.extractall(_lowercase ) tar_file.close() else: raise EnvironmentError('''Archive format of {} could not be identified'''.format(_lowercase ) ) return output_path_extracted return output_path def __UpperCamelCase ( a : str , a : List[str]="," ) ->Union[str, Any]: assert isinstance(_lowercase , _lowercase ) if os.path.isfile(_lowercase ): with open(_lowercase ) as f: snake_case = eval(f.read() ) else: snake_case = requests.get(_lowercase ) try: snake_case = requests.json() except Exception: snake_case = req.content.decode() assert data is not None, "could not connect" try: snake_case = eval(_lowercase ) except Exception: snake_case = data.split('''\n''' ) req.close() return data def __UpperCamelCase ( a : Any ) ->List[str]: snake_case = requests.get(_lowercase ) snake_case = np.array(Image.open(BytesIO(response.content ) ) ) return img def __UpperCamelCase ( a : Any ) ->Optional[int]: snake_case = url.split('''/''' )[-1] if fn not in os.listdir(os.getcwd() ): wget.download(_lowercase ) with open(_lowercase , '''rb''' ) as stream: snake_case = pkl.load(_lowercase ) snake_case = weights.pop('''model''' ) snake_case = {} for k, v in model.items(): snake_case = torch.from_numpy(_lowercase ) if "running_var" in k: snake_case = torch.tensor([0] ) snake_case = k.replace('''running_var''' , '''num_batches_tracked''' ) snake_case = zero return new def __UpperCamelCase ( ) ->int: print(f"""{os.path.abspath(os.path.join(_lowercase , os.pardir ) )}/demo.ipynb""" ) def __UpperCamelCase ( a : Tuple , a : Optional[int]="RGB" ) ->Any: assert isinstance(_lowercase , _lowercase ) if os.path.isfile(_lowercase ): snake_case = cva.imread(_lowercase ) else: snake_case = get_image_from_url(_lowercase ) assert img is not None, f"""could not connect to: {im}""" snake_case = cva.cvtColor(_lowercase , cva.COLOR_BGR2RGB ) if input_format == "RGB": snake_case = img[:, :, ::-1] return img def __UpperCamelCase ( a : Any , a : int=1 ) ->List[str]: return (images[i : i + batch] for i in range(0 , len(_lowercase ) , _lowercase ))
712
'''simple docstring''' from ...processing_utils import ProcessorMixin class _lowercase ( __a ): _UpperCAmelCase = '''WhisperFeatureExtractor''' _UpperCAmelCase = '''WhisperTokenizer''' def __init__( self , A__ , A__ ) -> Optional[Any]: super().__init__(A__ , A__ ) snake_case = self.feature_extractor snake_case = False def UpperCamelCase ( self , A__=None , A__=None , A__=True ) -> Union[str, Any]: return self.tokenizer.get_decoder_prompt_ids(task=A__ , language=A__ , no_timestamps=A__ ) def __call__( self , *A__ , **A__ ) -> Dict: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*A__ , **A__ ) snake_case = kwargs.pop('''audio''' , A__ ) snake_case = kwargs.pop('''sampling_rate''' , A__ ) snake_case = kwargs.pop('''text''' , A__ ) if len(A__ ) > 0: snake_case = args[0] snake_case = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if audio is not None: snake_case = self.feature_extractor(A__ , *A__ , sampling_rate=A__ , **A__ ) if text is not None: snake_case = self.tokenizer(A__ , **A__ ) if text is None: return inputs elif audio is None: return encodings else: snake_case = encodings['''input_ids'''] return inputs def UpperCamelCase ( self , *A__ , **A__ ) -> Optional[Any]: return self.tokenizer.batch_decode(*A__ , **A__ ) def UpperCamelCase ( self , *A__ , **A__ ) -> str: return self.tokenizer.decode(*A__ , **A__ ) def UpperCamelCase ( self , A__ , A__="np" ) -> Optional[Any]: return self.tokenizer.get_prompt_ids(A__ , return_tensors=A__ )
44
0
'''simple docstring''' from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class _lowercase : _UpperCAmelCase = 42 _UpperCAmelCase = None # Automatically constructed _UpperCAmelCase = '''dict''' _UpperCAmelCase = None _UpperCAmelCase = field(default='''Translation''' , init=__a , repr=__a ) def __call__( self ) -> List[Any]: return pa.struct({lang: pa.string() for lang in sorted(self.languages )} ) def UpperCamelCase ( self ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value return {k: Value('''string''' ) for k in sorted(self.languages )} @dataclass class _lowercase : _UpperCAmelCase = None _UpperCAmelCase = None _UpperCAmelCase = None # Automatically constructed _UpperCAmelCase = '''dict''' _UpperCAmelCase = None _UpperCAmelCase = field(default='''TranslationVariableLanguages''' , init=__a , repr=__a ) def UpperCamelCase ( self ) -> Dict: snake_case = sorted(set(self.languages ) ) if self.languages else None snake_case = len(self.languages ) if self.languages else None def __call__( self ) -> List[Any]: return pa.struct({'''language''': pa.list_(pa.string() ), '''translation''': pa.list_(pa.string() )} ) def UpperCamelCase ( self , A__ ) -> str: snake_case = set(self.languages ) if self.languages and set(_SCREAMING_SNAKE_CASE ) - lang_set: raise ValueError( F"""Some languages in example ({', '.join(sorted(set(_SCREAMING_SNAKE_CASE ) - lang_set ) )}) are not in valid set ({', '.join(_SCREAMING_SNAKE_CASE )}).""" ) # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. snake_case = [] for lang, text in translation_dict.items(): if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): translation_tuples.append((lang, text) ) else: translation_tuples.extend([(lang, el) for el in text] ) # Ensure translations are in ascending order by language code. snake_case , snake_case = zip(*sorted(_SCREAMING_SNAKE_CASE ) ) return {"language": languages, "translation": translations} def UpperCamelCase ( self ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Sequence, Value return { "language": Sequence(Value('''string''' ) ), "translation": Sequence(Value('''string''' ) ), }
713
'''simple docstring''' import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class _lowercase ( __a ): _UpperCAmelCase = '''char''' _UpperCAmelCase = '''bpe''' _UpperCAmelCase = '''wp''' _lowercase = (DecodeType.CHARACTER, DecodeType.BPE, DecodeType.WORDPIECE) class _lowercase ( __a ): _UpperCAmelCase = ['''image_processor''', '''char_tokenizer'''] _UpperCAmelCase = '''ViTImageProcessor''' _UpperCAmelCase = '''MgpstrTokenizer''' def __init__( self , A__=None , A__=None , **A__ ) -> List[Any]: snake_case = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , A__ , ) snake_case = kwargs.pop('''feature_extractor''' ) snake_case = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) snake_case = tokenizer snake_case = AutoTokenizer.from_pretrained('''gpt2''' ) snake_case = AutoTokenizer.from_pretrained('''bert-base-uncased''' ) super().__init__(A__ , A__ ) def __call__( self , A__=None , A__=None , A__=None , **A__ ) -> List[str]: if images is None and text is None: raise ValueError('''You need to specify either an `images` or `text` input to process.''' ) if images is not None: snake_case = self.image_processor(A__ , return_tensors=A__ , **A__ ) if text is not None: snake_case = self.char_tokenizer(A__ , return_tensors=A__ , **A__ ) if text is None: return inputs elif images is None: return encodings else: snake_case = encodings['''input_ids'''] return inputs def UpperCamelCase ( self , A__ ) -> Dict: snake_case , snake_case , snake_case = sequences snake_case = char_preds.size(0 ) snake_case , snake_case = self._decode_helper(A__ , '''char''' ) snake_case , snake_case = self._decode_helper(A__ , '''bpe''' ) snake_case , snake_case = self._decode_helper(A__ , '''wp''' ) snake_case = [] snake_case = [] for i in range(A__ ): snake_case = [char_scores[i], bpe_scores[i], wp_scores[i]] snake_case = [char_strs[i], bpe_strs[i], wp_strs[i]] snake_case = scores.index(max(A__ ) ) final_strs.append(strs[max_score_index] ) final_scores.append(scores[max_score_index] ) snake_case = {} snake_case = final_strs snake_case = final_scores snake_case = char_strs snake_case = bpe_strs snake_case = wp_strs return out def UpperCamelCase ( self , A__ , A__ ) -> Optional[Any]: if format == DecodeType.CHARACTER: snake_case = self.char_decode snake_case = 1 snake_case = '''[s]''' elif format == DecodeType.BPE: snake_case = self.bpe_decode snake_case = 2 snake_case = '''#''' elif format == DecodeType.WORDPIECE: snake_case = self.wp_decode snake_case = 1_02 snake_case = '''[SEP]''' else: raise ValueError(F"""Format {format} is not supported.""" ) snake_case , snake_case = [], [] snake_case = pred_logits.size(0 ) snake_case = pred_logits.size(1 ) snake_case , snake_case = pred_logits.topk(1 , dim=-1 , largest=A__ , sorted=A__ ) snake_case = preds_index.view(-1 , A__ )[:, 1:] snake_case = decoder(A__ ) snake_case , snake_case = torch.nn.functional.softmax(A__ , dim=2 ).max(dim=2 ) snake_case = preds_max_prob[:, 1:] for index in range(A__ ): snake_case = preds_str[index].find(A__ ) snake_case = preds_str[index][:pred_eos] snake_case = preds_index[index].cpu().tolist() snake_case = pred_index.index(A__ ) if eos_token in pred_index else -1 snake_case = preds_max_prob[index][: pred_eos_index + 1] snake_case = pred_max_prob.cumprod(dim=0 )[-1] if pred_max_prob.nelement() != 0 else 0.0 dec_strs.append(A__ ) conf_scores.append(A__ ) return dec_strs, conf_scores def UpperCamelCase ( self , A__ ) -> int: snake_case = [seq.replace(''' ''' , '''''' ) for seq in self.char_tokenizer.batch_decode(A__ )] return decode_strs def UpperCamelCase ( self , A__ ) -> List[str]: return self.bpe_tokenizer.batch_decode(A__ ) def UpperCamelCase ( self , A__ ) -> Union[str, Any]: snake_case = [seq.replace(''' ''' , '''''' ) for seq in self.wp_tokenizer.batch_decode(A__ )] return decode_strs
44
0
'''simple docstring''' import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class _lowercase : @staticmethod def UpperCamelCase ( *A__ , **A__ ) -> Optional[int]: pass @is_pipeline_test @require_vision @require_torch class _lowercase ( unittest.TestCase ): _UpperCAmelCase = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def UpperCamelCase ( self , A__ , A__ , A__ ) -> List[str]: snake_case = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) snake_case = [ { """image""": """./tests/fixtures/tests_samples/COCO/000000039769.png""", """candidate_labels""": ["""cat""", """remote""", """couch"""], } ] return object_detector, examples def UpperCamelCase ( self , A__ , A__ ) -> Any: snake_case = object_detector(examples[0] , threshold=0.0 ) snake_case = len(A__ ) self.assertGreater(A__ , 0 ) self.assertEqual( A__ , [ { '''score''': ANY(A__ ), '''label''': ANY(A__ ), '''box''': {'''xmin''': ANY(A__ ), '''ymin''': ANY(A__ ), '''xmax''': ANY(A__ ), '''ymax''': ANY(A__ )}, } for i in range(A__ ) ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def UpperCamelCase ( self ) -> List[str]: pass @require_torch def UpperCamelCase ( self ) -> List[Any]: snake_case = pipeline( '''zero-shot-object-detection''' , model='''hf-internal-testing/tiny-random-owlvit-object-detection''' ) snake_case = object_detector( '''./tests/fixtures/tests_samples/COCO/000000039769.png''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=0.6_4 , ) self.assertEqual( nested_simplify(A__ , decimals=4 ) , [ {'''score''': 0.7_2_3_5, '''label''': '''cat''', '''box''': {'''xmin''': 2_04, '''ymin''': 1_67, '''xmax''': 2_32, '''ymax''': 1_90}}, {'''score''': 0.7_2_1_8, '''label''': '''remote''', '''box''': {'''xmin''': 2_04, '''ymin''': 1_67, '''xmax''': 2_32, '''ymax''': 1_90}}, {'''score''': 0.7_1_8_4, '''label''': '''couch''', '''box''': {'''xmin''': 2_04, '''ymin''': 1_67, '''xmax''': 2_32, '''ymax''': 1_90}}, {'''score''': 0.6_7_4_8, '''label''': '''remote''', '''box''': {'''xmin''': 5_71, '''ymin''': 83, '''xmax''': 5_98, '''ymax''': 1_03}}, {'''score''': 0.6_6_5_6, '''label''': '''cat''', '''box''': {'''xmin''': 5_71, '''ymin''': 83, '''xmax''': 5_98, '''ymax''': 1_03}}, {'''score''': 0.6_6_1_4, '''label''': '''couch''', '''box''': {'''xmin''': 5_71, '''ymin''': 83, '''xmax''': 5_98, '''ymax''': 1_03}}, {'''score''': 0.6_4_5_6, '''label''': '''remote''', '''box''': {'''xmin''': 4_94, '''ymin''': 1_05, '''xmax''': 5_21, '''ymax''': 1_27}}, {'''score''': 0.6_4_2, '''label''': '''remote''', '''box''': {'''xmin''': 67, '''ymin''': 2_74, '''xmax''': 93, '''ymax''': 2_97}}, {'''score''': 0.6_4_1_9, '''label''': '''cat''', '''box''': {'''xmin''': 4_94, '''ymin''': 1_05, '''xmax''': 5_21, '''ymax''': 1_27}}, ] , ) snake_case = object_detector( [ { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], } ] , threshold=0.6_4 , ) self.assertEqual( nested_simplify(A__ , decimals=4 ) , [ [ {'''score''': 0.7_2_3_5, '''label''': '''cat''', '''box''': {'''xmin''': 2_04, '''ymin''': 1_67, '''xmax''': 2_32, '''ymax''': 1_90}}, {'''score''': 0.7_2_1_8, '''label''': '''remote''', '''box''': {'''xmin''': 2_04, '''ymin''': 1_67, '''xmax''': 2_32, '''ymax''': 1_90}}, {'''score''': 0.7_1_8_4, '''label''': '''couch''', '''box''': {'''xmin''': 2_04, '''ymin''': 1_67, '''xmax''': 2_32, '''ymax''': 1_90}}, {'''score''': 0.6_7_4_8, '''label''': '''remote''', '''box''': {'''xmin''': 5_71, '''ymin''': 83, '''xmax''': 5_98, '''ymax''': 1_03}}, {'''score''': 0.6_6_5_6, '''label''': '''cat''', '''box''': {'''xmin''': 5_71, '''ymin''': 83, '''xmax''': 5_98, '''ymax''': 1_03}}, {'''score''': 0.6_6_1_4, '''label''': '''couch''', '''box''': {'''xmin''': 5_71, '''ymin''': 83, '''xmax''': 5_98, '''ymax''': 1_03}}, {'''score''': 0.6_4_5_6, '''label''': '''remote''', '''box''': {'''xmin''': 4_94, '''ymin''': 1_05, '''xmax''': 5_21, '''ymax''': 1_27}}, {'''score''': 0.6_4_2, '''label''': '''remote''', '''box''': {'''xmin''': 67, '''ymin''': 2_74, '''xmax''': 93, '''ymax''': 2_97}}, {'''score''': 0.6_4_1_9, '''label''': '''cat''', '''box''': {'''xmin''': 4_94, '''ymin''': 1_05, '''xmax''': 5_21, '''ymax''': 1_27}}, ] ] , ) @require_torch @slow def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = pipeline('''zero-shot-object-detection''' ) snake_case = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , ) self.assertEqual( nested_simplify(A__ , decimals=4 ) , [ {'''score''': 0.2_8_6_8, '''label''': '''cat''', '''box''': {'''xmin''': 3_24, '''ymin''': 20, '''xmax''': 6_40, '''ymax''': 3_73}}, {'''score''': 0.2_7_7, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 72, '''xmax''': 1_77, '''ymax''': 1_15}}, {'''score''': 0.2_5_3_7, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 55, '''xmax''': 3_15, '''ymax''': 4_72}}, {'''score''': 0.1_4_7_4, '''label''': '''remote''', '''box''': {'''xmin''': 3_35, '''ymin''': 74, '''xmax''': 3_71, '''ymax''': 1_87}}, {'''score''': 0.1_2_0_8, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_42, '''ymax''': 4_76}}, ] , ) snake_case = object_detector( [ { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, { '''image''': '''http://images.cocodataset.org/val2017/000000039769.jpg''', '''candidate_labels''': ['''cat''', '''remote''', '''couch'''], }, ] , ) self.assertEqual( nested_simplify(A__ , decimals=4 ) , [ [ {'''score''': 0.2_8_6_8, '''label''': '''cat''', '''box''': {'''xmin''': 3_24, '''ymin''': 20, '''xmax''': 6_40, '''ymax''': 3_73}}, {'''score''': 0.2_7_7, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 72, '''xmax''': 1_77, '''ymax''': 1_15}}, {'''score''': 0.2_5_3_7, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 55, '''xmax''': 3_15, '''ymax''': 4_72}}, {'''score''': 0.1_4_7_4, '''label''': '''remote''', '''box''': {'''xmin''': 3_35, '''ymin''': 74, '''xmax''': 3_71, '''ymax''': 1_87}}, {'''score''': 0.1_2_0_8, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_42, '''ymax''': 4_76}}, ], [ {'''score''': 0.2_8_6_8, '''label''': '''cat''', '''box''': {'''xmin''': 3_24, '''ymin''': 20, '''xmax''': 6_40, '''ymax''': 3_73}}, {'''score''': 0.2_7_7, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 72, '''xmax''': 1_77, '''ymax''': 1_15}}, {'''score''': 0.2_5_3_7, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 55, '''xmax''': 3_15, '''ymax''': 4_72}}, {'''score''': 0.1_4_7_4, '''label''': '''remote''', '''box''': {'''xmin''': 3_35, '''ymin''': 74, '''xmax''': 3_71, '''ymax''': 1_87}}, {'''score''': 0.1_2_0_8, '''label''': '''couch''', '''box''': {'''xmin''': 4, '''ymin''': 0, '''xmax''': 6_42, '''ymax''': 4_76}}, ], ] , ) @require_tf @unittest.skip('''Zero Shot Object Detection not implemented in TF''' ) def UpperCamelCase ( self ) -> Any: pass @require_torch @slow def UpperCamelCase ( self ) -> Optional[int]: snake_case = 0.2 snake_case = pipeline('''zero-shot-object-detection''' ) snake_case = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , threshold=A__ , ) self.assertEqual( nested_simplify(A__ , decimals=4 ) , [ {'''score''': 0.2_8_6_8, '''label''': '''cat''', '''box''': {'''xmin''': 3_24, '''ymin''': 20, '''xmax''': 6_40, '''ymax''': 3_73}}, {'''score''': 0.2_7_7, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 72, '''xmax''': 1_77, '''ymax''': 1_15}}, {'''score''': 0.2_5_3_7, '''label''': '''cat''', '''box''': {'''xmin''': 1, '''ymin''': 55, '''xmax''': 3_15, '''ymax''': 4_72}}, ] , ) @require_torch @slow def UpperCamelCase ( self ) -> int: snake_case = 2 snake_case = pipeline('''zero-shot-object-detection''' ) snake_case = object_detector( '''http://images.cocodataset.org/val2017/000000039769.jpg''' , candidate_labels=['''cat''', '''remote''', '''couch'''] , top_k=A__ , ) self.assertEqual( nested_simplify(A__ , decimals=4 ) , [ {'''score''': 0.2_8_6_8, '''label''': '''cat''', '''box''': {'''xmin''': 3_24, '''ymin''': 20, '''xmax''': 6_40, '''ymax''': 3_73}}, {'''score''': 0.2_7_7, '''label''': '''remote''', '''box''': {'''xmin''': 40, '''ymin''': 72, '''xmax''': 1_77, '''ymax''': 1_15}}, ] , )
714
'''simple docstring''' import os from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen, xsplitext from ..table import array_cast from ..utils.py_utils import no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: from .features import FeatureType _lowercase , _lowercase , _lowercase = False, False, False @dataclass class _lowercase : _UpperCAmelCase = None _UpperCAmelCase = True _UpperCAmelCase = True _UpperCAmelCase = None # Automatically constructed _UpperCAmelCase = "dict" _UpperCAmelCase = pa.struct({'''bytes''': pa.binary(), '''path''': pa.string()} ) _UpperCAmelCase = field(default='''Audio''' , init=__a , repr=__a ) def __call__( self ) -> Optional[Any]: return self.pa_type def UpperCamelCase ( self , A__ ) -> dict: try: import soundfile as sf # soundfile is a dependency of librosa, needed to decode audio files. except ImportError as err: raise ImportError('''To support encoding audio data, please install \'soundfile\'.''' ) from err if isinstance(A__ , A__ ): return {"bytes": None, "path": value} elif isinstance(A__ , A__ ): return {"bytes": value, "path": None} elif "array" in value: # convert the audio array to wav bytes snake_case = BytesIO() sf.write(A__ , value['''array'''] , value['''sampling_rate'''] , format='''wav''' ) return {"bytes": buffer.getvalue(), "path": None} elif value.get('''path''' ) is not None and os.path.isfile(value['''path'''] ): # we set "bytes": None to not duplicate the data if they're already available locally if value["path"].endswith('''pcm''' ): # "PCM" only has raw audio bytes if value.get('''sampling_rate''' ) is None: # At least, If you want to convert "PCM-byte" to "WAV-byte", you have to know sampling rate raise KeyError('''To use PCM files, please specify a \'sampling_rate\' in Audio object''' ) if value.get('''bytes''' ): # If we already had PCM-byte, we don`t have to make "read file, make bytes" (just use it!) snake_case = np.frombuffer(value['''bytes'''] , dtype=np.intaa ).astype(np.floataa ) / 3_27_67 else: snake_case = np.memmap(value['''path'''] , dtype='''h''' , mode='''r''' ).astype(np.floataa ) / 3_27_67 snake_case = BytesIO(bytes() ) sf.write(A__ , A__ , value['''sampling_rate'''] , format='''wav''' ) return {"bytes": buffer.getvalue(), "path": None} else: return {"bytes": None, "path": value.get('''path''' )} elif value.get('''bytes''' ) is not None or value.get('''path''' ) is not None: # store the audio bytes, and path is used to infer the audio format using the file extension return {"bytes": value.get('''bytes''' ), "path": value.get('''path''' )} else: raise ValueError( F"""An audio sample should have one of 'path' or 'bytes' but they are missing or None in {value}.""" ) def UpperCamelCase ( self , A__ , A__ = None ) -> dict: if not self.decode: raise RuntimeError('''Decoding is disabled for this feature. Please use Audio(decode=True) instead.''' ) snake_case , snake_case = (value['''path'''], BytesIO(value['''bytes'''] )) if value['''bytes'''] is not None else (value['''path'''], None) if path is None and file is None: raise ValueError(F"""An audio sample should have one of 'path' or 'bytes' but both are None in {value}.""" ) try: import librosa import soundfile as sf except ImportError as err: raise ImportError('''To support decoding audio files, please install \'librosa\' and \'soundfile\'.''' ) from err snake_case = xsplitext(A__ )[1][1:].lower() if path is not None else None if not config.IS_OPUS_SUPPORTED and audio_format == "opus": raise RuntimeError( '''Decoding \'opus\' files requires system library \'libsndfile\'>=1.0.31, ''' '''You can try to update `soundfile` python library: `pip install "soundfile>=0.12.1"`. ''' ) elif not config.IS_MP3_SUPPORTED and audio_format == "mp3": raise RuntimeError( '''Decoding \'mp3\' files requires system library \'libsndfile\'>=1.1.0, ''' '''You can try to update `soundfile` python library: `pip install "soundfile>=0.12.1"`. ''' ) if file is None: snake_case = token_per_repo_id or {} snake_case = path.split('''::''' )[-1] try: snake_case = string_to_dict(A__ , config.HUB_DATASETS_URL )['''repo_id'''] snake_case = token_per_repo_id[repo_id] except (ValueError, KeyError): snake_case = None with xopen(A__ , '''rb''' , use_auth_token=A__ ) as f: snake_case , snake_case = sf.read(A__ ) else: snake_case , snake_case = sf.read(A__ ) snake_case = array.T if self.mono: snake_case = librosa.to_mono(A__ ) if self.sampling_rate and self.sampling_rate != sampling_rate: snake_case = librosa.resample(A__ , orig_sr=A__ , target_sr=self.sampling_rate ) snake_case = self.sampling_rate return {"path": path, "array": array, "sampling_rate": sampling_rate} def UpperCamelCase ( self ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value if self.decode: raise ValueError('''Cannot flatten a decoded Audio feature.''' ) return { "bytes": Value('''binary''' ), "path": Value('''string''' ), } def UpperCamelCase ( self , A__ ) -> pa.StructArray: if pa.types.is_string(storage.type ): snake_case = pa.array([None] * len(A__ ) , type=pa.binary() ) snake_case = pa.StructArray.from_arrays([bytes_array, storage] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): snake_case = pa.array([None] * len(A__ ) , type=pa.string() ) snake_case = pa.StructArray.from_arrays([storage, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ) and storage.type.get_all_field_indices('''array''' ): snake_case = pa.array([Audio().encode_example(A__ ) if x is not None else None for x in storage.to_pylist()] ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index('''bytes''' ) >= 0: snake_case = storage.field('''bytes''' ) else: snake_case = pa.array([None] * len(A__ ) , type=pa.binary() ) if storage.type.get_field_index('''path''' ) >= 0: snake_case = storage.field('''path''' ) else: snake_case = pa.array([None] * len(A__ ) , type=pa.string() ) snake_case = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null() ) return array_cast(A__ , self.pa_type ) def UpperCamelCase ( self , A__ ) -> pa.StructArray: @no_op_if_value_is_null def path_to_bytes(A__ ): with xopen(A__ , '''rb''' ) as f: snake_case = f.read() return bytes_ snake_case = pa.array( [ (path_to_bytes(x['''path'''] ) if x['''bytes'''] is None else x['''bytes''']) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) snake_case = pa.array( [os.path.basename(A__ ) if path is not None else None for path in storage.field('''path''' ).to_pylist()] , type=pa.string() , ) snake_case = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=bytes_array.is_null() ) return array_cast(A__ , self.pa_type )
44
0
'''simple docstring''' import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING, MODEL_MAPPING, LevitForImageClassification, LevitForImageClassificationWithTeacher, LevitModel, ) from transformers.models.levit.modeling_levit import LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class _lowercase ( lowercase__ ): def UpperCamelCase ( self ) -> Optional[Any]: '''simple docstring''' snake_case = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(__lowercase , '''hidden_sizes''' ) ) self.parent.assertTrue(hasattr(__lowercase , '''num_attention_heads''' ) ) class _lowercase : def __init__( self , A__ , A__=13 , A__=64 , A__=3 , A__=3 , A__=2 , A__=1 , A__=16 , A__=[1_28, 2_56, 3_84] , A__=[4, 6, 8] , A__=[2, 3, 4] , A__=[16, 16, 16] , A__=0 , A__=[2, 2, 2] , A__=[2, 2, 2] , A__=0.0_2 , A__=True , A__=True , A__=2 , ) -> int: '''simple docstring''' snake_case = parent snake_case = batch_size snake_case = image_size snake_case = num_channels snake_case = kernel_size snake_case = stride snake_case = padding snake_case = hidden_sizes snake_case = num_attention_heads snake_case = depths snake_case = key_dim snake_case = drop_path_rate snake_case = patch_size snake_case = attention_ratio snake_case = mlp_ratio snake_case = initializer_range snake_case = [ ['''Subsample''', key_dim[0], hidden_sizes[0] // key_dim[0], 4, 2, 2], ['''Subsample''', key_dim[0], hidden_sizes[1] // key_dim[0], 4, 2, 2], ] snake_case = is_training snake_case = use_labels snake_case = num_labels snake_case = initializer_range def UpperCamelCase ( self ) -> Optional[int]: '''simple docstring''' snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case = None if self.use_labels: snake_case = ids_tensor([self.batch_size] , self.num_labels ) snake_case = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ) -> List[str]: '''simple docstring''' return LevitConfig( image_size=self.image_size , num_channels=self.num_channels , kernel_size=self.kernel_size , stride=self.stride , padding=self.padding , patch_size=self.patch_size , hidden_sizes=self.hidden_sizes , num_attention_heads=self.num_attention_heads , depths=self.depths , key_dim=self.key_dim , drop_path_rate=self.drop_path_rate , mlp_ratio=self.mlp_ratio , attention_ratio=self.attention_ratio , initializer_range=self.initializer_range , down_ops=self.down_ops , ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Optional[Any]: '''simple docstring''' snake_case = LevitModel(config=__lowercase ) model.to(__lowercase ) model.eval() snake_case = model(__lowercase ) snake_case = (self.image_size, self.image_size) snake_case , snake_case = image_size[0], image_size[1] for _ in range(4 ): snake_case = floor(((height + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) snake_case = floor(((width + 2 * self.padding - self.kernel_size) / self.stride) + 1 ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, ceil(height / 4 ) * ceil(width / 4 ), self.hidden_sizes[-1]) , ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> str: '''simple docstring''' snake_case = self.num_labels snake_case = LevitForImageClassification(__lowercase ) model.to(__lowercase ) model.eval() snake_case = model(__lowercase , labels=__lowercase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self ) -> List[str]: '''simple docstring''' snake_case = self.prepare_config_and_inputs() snake_case , snake_case , snake_case = config_and_inputs snake_case = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class _lowercase ( lowercase__ , lowercase__ , unittest.TestCase ): _UpperCAmelCase = ( (LevitModel, LevitForImageClassification, LevitForImageClassificationWithTeacher) if is_torch_available() else () ) _UpperCAmelCase = ( { """feature-extraction""": LevitModel, """image-classification""": (LevitForImageClassification, LevitForImageClassificationWithTeacher), } if is_torch_available() else {} ) _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False def UpperCamelCase ( self ) -> Tuple: '''simple docstring''' snake_case = LevitModelTester(self ) snake_case = ConfigTester(self , config_class=__lowercase , has_text_modality=__lowercase , hidden_size=37 ) def UpperCamelCase ( self ) -> List[str]: '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ) -> Union[str, Any]: '''simple docstring''' return @unittest.skip(reason='''Levit does not use inputs_embeds''' ) def UpperCamelCase ( self ) -> Optional[Any]: '''simple docstring''' pass @unittest.skip(reason='''Levit does not support input and output embeddings''' ) def UpperCamelCase ( self ) -> List[str]: '''simple docstring''' pass @unittest.skip(reason='''Levit does not output attentions''' ) def UpperCamelCase ( self ) -> Dict: '''simple docstring''' pass def UpperCamelCase ( self ) -> str: '''simple docstring''' snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(__lowercase ) snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case = [*signature.parameters.keys()] snake_case = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , __lowercase ) def UpperCamelCase ( self ) -> int: '''simple docstring''' def check_hidden_states_output(A__ , A__ , A__ ): snake_case = model_class(__lowercase ) model.to(__lowercase ) model.eval() with torch.no_grad(): snake_case = model(**self._prepare_for_class(__lowercase , __lowercase ) ) snake_case = outputs.hidden_states snake_case = len(self.model_tester.depths ) + 1 self.assertEqual(len(__lowercase ) , __lowercase ) snake_case = (self.model_tester.image_size, self.model_tester.image_size) snake_case , snake_case = image_size[0], image_size[1] for _ in range(4 ): snake_case = floor( ( (height + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) snake_case = floor( ( (width + 2 * self.model_tester.padding - self.model_tester.kernel_size) / self.model_tester.stride ) + 1 ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [ height * width, self.model_tester.hidden_sizes[0], ] , ) snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = True check_hidden_states_output(__lowercase , __lowercase , __lowercase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case = True check_hidden_states_output(__lowercase , __lowercase , __lowercase ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def UpperCamelCase ( self ) -> Union[str, Any]: '''simple docstring''' pass def UpperCamelCase ( self , A__ , A__ , A__=False ) -> List[Any]: '''simple docstring''' snake_case = super()._prepare_for_class(__lowercase , __lowercase , return_labels=__lowercase ) if return_labels: if model_class.__name__ == "LevitForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def UpperCamelCase ( self ) -> List[str]: '''simple docstring''' snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowercase ) def UpperCamelCase ( self ) -> Any: '''simple docstring''' snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowercase ) def UpperCamelCase ( self ) -> Dict: '''simple docstring''' if not self.model_tester.is_training: return snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() snake_case = True for model_class in self.all_model_classes: # LevitForImageClassificationWithTeacher supports inference-only if ( model_class in get_values(__lowercase ) or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue snake_case = model_class(__lowercase ) model.to(__lowercase ) model.train() snake_case = self._prepare_for_class(__lowercase , __lowercase , return_labels=__lowercase ) snake_case = model(**__lowercase ).loss loss.backward() def UpperCamelCase ( self ) -> int: '''simple docstring''' snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() if not self.model_tester.is_training: return snake_case = False snake_case = True for model_class in self.all_model_classes: if model_class in get_values(__lowercase ) or not model_class.supports_gradient_checkpointing: continue # LevitForImageClassificationWithTeacher supports inference-only if model_class.__name__ == "LevitForImageClassificationWithTeacher": continue snake_case = model_class(__lowercase ) model.gradient_checkpointing_enable() model.to(__lowercase ) model.train() snake_case = self._prepare_for_class(__lowercase , __lowercase , return_labels=__lowercase ) snake_case = model(**__lowercase ).loss loss.backward() def UpperCamelCase ( self ) -> Any: '''simple docstring''' snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() snake_case = [ {'''title''': '''multi_label_classification''', '''num_labels''': 2, '''dtype''': torch.float}, {'''title''': '''single_label_classification''', '''num_labels''': 1, '''dtype''': torch.long}, {'''title''': '''regression''', '''num_labels''': 1, '''dtype''': torch.float}, ] for model_class in self.all_model_classes: if ( model_class not in [ *get_values(__lowercase ), ] or model_class.__name__ == "LevitForImageClassificationWithTeacher" ): continue for problem_type in problem_types: with self.subTest(msg=F"""Testing {model_class} with {problem_type['title']}""" ): snake_case = problem_type['''title'''] snake_case = problem_type['''num_labels'''] snake_case = model_class(__lowercase ) model.to(__lowercase ) model.train() snake_case = self._prepare_for_class(__lowercase , __lowercase , return_labels=__lowercase ) if problem_type["num_labels"] > 1: snake_case = inputs['''labels'''].unsqueeze(1 ).repeat(1 , problem_type['''num_labels'''] ) snake_case = inputs['''labels'''].to(problem_type['''dtype'''] ) # This tests that we do not trigger the warning form PyTorch "Using a target size that is different # to the input size. This will likely lead to incorrect results due to broadcasting. Please ensure # they have the same size." which is a symptom something in wrong for the regression problem. # See https://github.com/huggingface/transformers/issues/11780 with warnings.catch_warnings(record=__lowercase ) as warning_list: snake_case = model(**__lowercase ).loss for w in warning_list: if "Using a target size that is different to the input size" in str(w.message ): raise ValueError( F"""Something is going wrong in the regression problem: intercepted {w.message}""" ) loss.backward() @slow def UpperCamelCase ( self ) -> Tuple: '''simple docstring''' for model_name in LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case = LevitModel.from_pretrained(__lowercase ) self.assertIsNotNone(__lowercase ) def __UpperCamelCase ( ) ->Tuple: snake_case = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class _lowercase ( unittest.TestCase ): @cached_property def UpperCamelCase ( self ) -> str: '''simple docstring''' return LevitImageProcessor.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def UpperCamelCase ( self ) -> int: '''simple docstring''' snake_case = LevitForImageClassificationWithTeacher.from_pretrained(LEVIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to( __lowercase ) snake_case = self.default_image_processor snake_case = prepare_img() snake_case = image_processor(images=__lowercase , return_tensors='''pt''' ).to(__lowercase ) # forward pass with torch.no_grad(): snake_case = model(**__lowercase ) # verify the logits snake_case = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __lowercase ) snake_case = torch.tensor([1.0_4_4_8, -0.3_7_4_5, -1.8_3_1_7] ).to(__lowercase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowercase , atol=1e-4 ) )
715
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class _lowercase : @staticmethod def UpperCamelCase ( *A__ , **A__ ) -> List[Any]: pass def __UpperCamelCase ( a : Image ) ->str: snake_case = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class _lowercase ( unittest.TestCase ): _UpperCAmelCase = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def UpperCamelCase ( self , A__ , A__ , A__ ) -> Union[str, Any]: snake_case = DepthEstimationPipeline(model=A__ , image_processor=A__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def UpperCamelCase ( self , A__ , A__ ) -> List[Any]: snake_case = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} , A__ ) import datasets snake_case = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' , '''image''' , split='''test''' ) snake_case = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] , A__ , ) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def UpperCamelCase ( self ) -> Optional[Any]: pass @slow @require_torch def UpperCamelCase ( self ) -> Dict: snake_case = '''Intel/dpt-large''' snake_case = pipeline('''depth-estimation''' , model=A__ ) snake_case = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) snake_case = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) , 2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) , 2.6_6_2 ) @require_torch def UpperCamelCase ( self ) -> Any: # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
44
0
'''simple docstring''' from __future__ import annotations from bisect import bisect_left from functools import total_ordering from heapq import merge @total_ordering class _lowercase ( __snake_case ): def __lt__( self , A__ ) -> List[Any]: return self[-1] < other[-1] def __eq__( self , A__ ) -> str: return self[-1] == other[-1] def __UpperCamelCase ( a : Any ) ->list: snake_case = [] # sort into stacks for element in collection: snake_case = Stack([element] ) snake_case = bisect_left(_lowerCAmelCase , _lowerCAmelCase ) if i != len(_lowerCAmelCase ): stacks[i].append(_lowerCAmelCase ) else: stacks.append(_lowerCAmelCase ) # use a heap-based merge to merge stack efficiently snake_case = merge(*(reversed(_lowerCAmelCase ) for stack in stacks) ) return collection if __name__ == "__main__": _lowercase = input('Enter numbers separated by a comma:\n').strip() _lowercase = [int(item) for item in user_input.split(',')] print(patience_sort(unsorted))
716
'''simple docstring''' import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def __UpperCamelCase ( a : Optional[int] ) ->Dict: snake_case = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(a , a ) def __UpperCamelCase ( a : Optional[Any] ) ->int: snake_case = list(s_dict.keys() ) for key in keys: if "transformer_layers" in key: snake_case = s_dict.pop(a ) elif "subsample" in key: snake_case = s_dict.pop(a ) def __UpperCamelCase ( a : Optional[int] ) ->Optional[int]: snake_case , snake_case = emb.weight.shape snake_case = nn.Linear(a , a , bias=a ) snake_case = emb.weight.data return lin_layer def __UpperCamelCase ( a : Any , a : Tuple ) ->Tuple: snake_case = torch.load(a , map_location='''cpu''' ) snake_case = mam_aaa['''args'''] snake_case = mam_aaa['''model'''] snake_case = state_dict['''decoder.output_projection.weight'''] remove_ignore_keys_(a ) rename_keys(a ) snake_case = state_dict['''decoder.embed_tokens.weight'''].shape[0] snake_case = args.share_decoder_input_output_embed snake_case = [int(a ) for i in args.conv_kernel_sizes.split(''',''' )] snake_case = SpeechaTextConfig( vocab_size=a , max_source_positions=args.max_source_positions , max_target_positions=args.max_target_positions , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='''relu''' , num_conv_layers=len(a ) , conv_channels=args.conv_channels , conv_kernel_sizes=a , input_feat_per_channel=args.input_feat_per_channel , input_channels=args.input_channels , tie_word_embeddings=a , num_beams=5 , max_length=200 , use_cache=a , decoder_start_token_id=2 , early_stopping=a , ) snake_case = SpeechaTextForConditionalGeneration(a ) snake_case , snake_case = model.model.load_state_dict(a , strict=a ) if len(a ) > 0 and not set(a ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( '''Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,''' f""" but all the following weights are missing {missing}""" ) if tie_embeds: snake_case = make_linear_from_emb(model.model.decoder.embed_tokens ) else: snake_case = lm_head_weights model.save_pretrained(a ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument('--fairseq_path', type=str, help='Path to the fairseq model (.pt) file.') parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') _lowercase = parser.parse_args() convert_fairseq_sat_checkpoint_to_tfms(args.fairseq_path, args.pytorch_dump_folder_path)
44
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { '''facebook/data2vec-vision-base-ft''': ( '''https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json''' ), } class _lowercase ( UpperCamelCase__ ): _UpperCAmelCase = """data2vec-vision""" def __init__( self , A__=7_68 , A__=12 , A__=12 , A__=30_72 , A__="gelu" , A__=0.0 , A__=0.0 , A__=0.0_2 , A__=1e-12 , A__=2_24 , A__=16 , A__=3 , A__=False , A__=False , A__=False , A__=False , A__=0.1 , A__=0.1 , A__=True , A__=[3, 5, 7, 11] , A__=[1, 2, 3, 6] , A__=True , A__=0.4 , A__=2_56 , A__=1 , A__=False , A__=2_55 , **A__ , ) -> Any: super().__init__(**A__ ) snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_size snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = initializer_range snake_case = layer_norm_eps snake_case = image_size snake_case = patch_size snake_case = num_channels snake_case = use_mask_token snake_case = use_absolute_position_embeddings snake_case = use_relative_position_bias snake_case = use_shared_relative_position_bias snake_case = layer_scale_init_value snake_case = drop_path_rate snake_case = use_mean_pooling # decode head attributes (semantic segmentation) snake_case = out_indices snake_case = pool_scales # auxiliary head attributes (semantic segmentation) snake_case = use_auxiliary_head snake_case = auxiliary_loss_weight snake_case = auxiliary_channels snake_case = auxiliary_num_convs snake_case = auxiliary_concat_input snake_case = semantic_loss_ignore_index class _lowercase ( UpperCamelCase__ ): _UpperCAmelCase = version.parse('''1.11''' ) @property def UpperCamelCase ( self ) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def UpperCamelCase ( self ) -> float: return 1e-4
717
'''simple docstring''' from ..utils import DummyObject, requires_backends class _lowercase ( metaclass=__a ): _UpperCAmelCase = ['''transformers''', '''torch''', '''note_seq'''] def __init__( self , *A__ , **A__ ) -> Union[str, Any]: requires_backends(self , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ) -> Optional[Any]: requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ) -> Any: requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] )
44
0
'''simple docstring''' import collections import inspect import unittest from transformers import FocalNetConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ( FocalNetBackbone, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetModel, ) from transformers.models.focalnet.modeling_focalnet import FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowercase : def __init__( self , A__ , A__=13 , A__=32 , A__=2 , A__=3 , A__=16 , A__=[32, 64, 1_28] , A__=[1, 2, 1] , A__=[2, 2, 4] , A__=2 , A__=2.0 , A__=True , A__=0.0 , A__=0.0 , A__=0.1 , A__="gelu" , A__=False , A__=True , A__=0.0_2 , A__=1e-5 , A__=True , A__=None , A__=True , A__=10 , A__=8 , A__=["stage1", "stage2"] , A__=[1, 2] , ) -> Any: snake_case = parent snake_case = batch_size snake_case = image_size snake_case = patch_size snake_case = num_channels snake_case = embed_dim snake_case = hidden_sizes snake_case = depths snake_case = num_heads snake_case = window_size snake_case = mlp_ratio snake_case = qkv_bias snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = drop_path_rate snake_case = hidden_act snake_case = use_absolute_embeddings snake_case = patch_norm snake_case = layer_norm_eps snake_case = initializer_range snake_case = is_training snake_case = scope snake_case = use_labels snake_case = type_sequence_label_size snake_case = encoder_stride snake_case = out_features snake_case = out_indices def UpperCamelCase ( self ) -> List[str]: snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case = None if self.use_labels: snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ) -> Any: return FocalNetConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , hidden_sizes=self.hidden_sizes , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , out_features=self.out_features , out_indices=self.out_indices , ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Any: snake_case = FocalNetModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() snake_case = model(lowerCamelCase__ ) snake_case = ((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) snake_case = int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Dict: snake_case = FocalNetBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() snake_case = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size, 8, 8] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[:-1] ) # verify backbone works with out_features=None snake_case = None snake_case = FocalNetBackbone(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() snake_case = model(lowerCamelCase__ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.image_size * 2, 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> int: snake_case = FocalNetForMaskedImageModeling(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() snake_case = model(lowerCamelCase__ ) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images snake_case = 1 snake_case = FocalNetForMaskedImageModeling(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() snake_case = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case = model(lowerCamelCase__ ) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Optional[Any]: snake_case = self.type_sequence_label_size snake_case = FocalNetForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() snake_case = model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images snake_case = 1 snake_case = FocalNetForImageClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() snake_case = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case = model(lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase ( self ) -> Optional[Any]: snake_case = self.prepare_config_and_inputs() snake_case = config_and_inputs snake_case = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class _lowercase ( UpperCAmelCase__ , UpperCAmelCase__ , unittest.TestCase ): _UpperCAmelCase = ( ( FocalNetModel, FocalNetForImageClassification, FocalNetForMaskedImageModeling, FocalNetBackbone, ) if is_torch_available() else () ) _UpperCAmelCase = ( {"feature-extraction": FocalNetModel, "image-classification": FocalNetForImageClassification} if is_torch_available() else {} ) _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False def UpperCamelCase ( self ) -> str: snake_case = FocalNetModelTester(self ) snake_case = ConfigTester(self , config_class=lowerCamelCase__ , embed_dim=37 , has_text_modality=lowerCamelCase__ ) def UpperCamelCase ( self ) -> int: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase ( self ) -> Any: return def UpperCamelCase ( self ) -> Optional[int]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase__ ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowerCamelCase__ ) def UpperCamelCase ( self ) -> List[str]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCamelCase__ ) def UpperCamelCase ( self ) -> List[str]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase__ ) @unittest.skip(reason='''FocalNet does not use inputs_embeds''' ) def UpperCamelCase ( self ) -> Tuple: pass @unittest.skip(reason='''FocalNet does not use feedforward chunking''' ) def UpperCamelCase ( self ) -> int: pass def UpperCamelCase ( self ) -> str: snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: snake_case = model_class(lowerCamelCase__ ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCamelCase__ , nn.Linear ) ) def UpperCamelCase ( self ) -> str: snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes[:-1]: snake_case = model_class(lowerCamelCase__ ) snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case = [*signature.parameters.keys()] snake_case = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCamelCase__ ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ ) -> str: snake_case = model_class(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() with torch.no_grad(): snake_case = model(**self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) ) snake_case = outputs.hidden_states snake_case = getattr( self.model_tester , '''expected_num_hidden_layers''' , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) # FocalNet has a different seq_length snake_case = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) snake_case = outputs.reshaped_hidden_states self.assertEqual(len(lowerCamelCase__ ) , lowerCamelCase__ ) snake_case = reshaped_hidden_states[0].shape snake_case = ( reshaped_hidden_states[0].view(lowerCamelCase__ , lowerCamelCase__ , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def UpperCamelCase ( self ) -> Optional[Any]: snake_case = self.model_tester.prepare_config_and_inputs_for_common() snake_case = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes[:-1]: snake_case = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.model_tester.prepare_config_and_inputs_for_common() snake_case = 3 snake_case = ( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) snake_case = ( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) snake_case = image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) snake_case = image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes[:-1]: snake_case = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case = True self.check_hidden_states_output(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , (padded_height, padded_width) ) @slow def UpperCamelCase ( self ) -> List[str]: for model_name in FOCALNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case = FocalNetModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) def UpperCamelCase ( self ) -> Dict: snake_case = self.model_tester.prepare_config_and_inputs_for_common() snake_case = _config_zero_init(lowerCamelCase__ ) for model_class in self.all_model_classes: snake_case = model_class(config=lowerCamelCase__ ) for name, param in model.named_parameters(): if "embeddings" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @require_vision @require_torch class _lowercase ( unittest.TestCase ): @cached_property def UpperCamelCase ( self ) -> Tuple: return AutoImageProcessor.from_pretrained('''microsoft/focalnet-tiny''' ) if is_vision_available() else None @slow def UpperCamelCase ( self ) -> Tuple: snake_case = FocalNetForImageClassification.from_pretrained('''microsoft/focalnet-tiny''' ).to(lowerCamelCase__ ) snake_case = self.default_image_processor snake_case = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) snake_case = image_processor(images=lowerCamelCase__ , return_tensors='''pt''' ).to(lowerCamelCase__ ) # forward pass with torch.no_grad(): snake_case = model(**lowerCamelCase__ ) # verify the logits snake_case = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , lowerCamelCase__ ) snake_case = torch.tensor([0.2_1_6_6, -0.4_3_6_8, 0.2_1_9_1] ).to(lowerCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCamelCase__ , atol=1e-4 ) ) self.assertTrue(outputs.logits.argmax(dim=-1 ).item() , 2_81 ) @require_torch class _lowercase ( UpperCAmelCase__ , unittest.TestCase ): _UpperCAmelCase = (FocalNetBackbone,) if is_torch_available() else () _UpperCAmelCase = FocalNetConfig _UpperCAmelCase = False def UpperCamelCase ( self ) -> int: snake_case = FocalNetModelTester(self )
718
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator class _lowercase : def __init__( self , A__ ) -> None: snake_case = value snake_case = None snake_case = None class _lowercase : def __init__( self , A__ ) -> None: snake_case = tree def UpperCamelCase ( self , A__ ) -> int: if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self ) -> Iterator[int]: yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
44
0
'''simple docstring''' import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class _lowercase ( __a ): _UpperCAmelCase = ["image_processor", "tokenizer"] _UpperCAmelCase = "CLIPImageProcessor" _UpperCAmelCase = ("XLMRobertaTokenizer", "XLMRobertaTokenizerFast") def __init__( self , A__=None , A__=None , **A__ ) -> List[str]: snake_case = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , lowerCAmelCase_ , ) snake_case = kwargs.pop('''feature_extractor''' ) snake_case = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(lowerCAmelCase_ , lowerCAmelCase_ ) def __call__( self , A__=None , A__=None , A__=None , **A__ ) -> Optional[int]: if text is None and images is None: raise ValueError('''You have to specify either text or images. Both cannot be none.''' ) if text is not None: snake_case = self.tokenizer(lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , **lowerCAmelCase_ ) if images is not None: snake_case = self.image_processor(lowerCAmelCase_ , return_tensors=lowerCAmelCase_ , **lowerCAmelCase_ ) if text is not None and images is not None: snake_case = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**lowerCAmelCase_ ) , tensor_type=lowerCAmelCase_ ) def UpperCamelCase ( self , *A__ , **A__ ) -> Tuple: return self.tokenizer.batch_decode(*lowerCAmelCase_ , **lowerCAmelCase_ ) def UpperCamelCase ( self , *A__ , **A__ ) -> Tuple: return self.tokenizer.decode(*lowerCAmelCase_ , **lowerCAmelCase_ ) @property def UpperCamelCase ( self ) -> Optional[Any]: snake_case = self.tokenizer.model_input_names snake_case = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
719
'''simple docstring''' import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() _lowercase = logging.get_logger(__name__) _lowercase = [ ('bert.bert', 'visual_bert'), ('bert.cls', 'cls'), ('bert.classifier', 'cls'), ('token_type_embeddings_visual', 'visual_token_type_embeddings'), ('position_embeddings_visual', 'visual_position_embeddings'), ('projection', 'visual_projection'), ] _lowercase = [ 'nlvr2_coco_pre_trained.th', 'nlvr2_fine_tuned.th', 'nlvr2_pre_trained.th', 'vcr_coco_pre_train.th', 'vcr_fine_tune.th', 'vcr_pre_train.th', 'vqa_coco_pre_trained.th', 'vqa_fine_tuned.th', 'vqa_pre_trained.th', ] def __UpperCamelCase ( a : List[str] ) ->Optional[int]: snake_case = torch.load(a , map_location='''cpu''' ) return sd def __UpperCamelCase ( a : Optional[int] , a : Union[str, Any] , a : int=rename_keys_prefix ) ->Tuple: snake_case = OrderedDict() snake_case = torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue snake_case = key for name_pair in rename_keys_prefix: snake_case = new_key.replace(name_pair[0] , name_pair[1] ) snake_case = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately snake_case = new_d['''cls.predictions.bias'''] return new_d @torch.no_grad() def __UpperCamelCase ( a : Optional[int] , a : int ) ->Union[str, Any]: assert ( checkpoint_path.split('''/''' )[-1] in ACCEPTABLE_CHECKPOINTS ), f"""The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.""" # Get Config if "pre" in checkpoint_path: snake_case = '''pretraining''' if "vcr" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 512} elif "vqa_advanced" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048} elif "vqa" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048} elif "nlvr" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 1024} else: raise NotImplementedError(f"""No implementation found for `{checkpoint_path}`.""" ) else: if "vcr" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 512} snake_case = '''multichoice''' elif "vqa_advanced" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048} snake_case = '''vqa_advanced''' elif "vqa" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048, '''num_labels''': 3129} snake_case = '''vqa''' elif "nlvr" in checkpoint_path: snake_case = { '''visual_embedding_dim''': 1024, '''num_labels''': 2, } snake_case = '''nlvr''' snake_case = VisualBertConfig(**a ) # Load State Dict snake_case = load_state_dict(a ) snake_case = get_new_dict(a , a ) if model_type == "pretraining": snake_case = VisualBertForPreTraining(a ) elif model_type == "vqa": snake_case = VisualBertForQuestionAnswering(a ) elif model_type == "nlvr": snake_case = VisualBertForVisualReasoning(a ) elif model_type == "multichoice": snake_case = VisualBertForMultipleChoice(a ) model.load_state_dict(a ) # Save Checkpoints Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument('orig_checkpoint_path', type=str, help='A path to .th on local filesystem.') parser.add_argument('pytorch_dump_folder_path', type=str, help='Path to the output PyTorch model.') _lowercase = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
44
0
'''simple docstring''' import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class _lowercase : @staticmethod def UpperCamelCase ( *A__ , **A__ ) -> Dict: pass def __UpperCamelCase ( a : List[Any] ) ->Union[str, Any]: return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. _lowercase = ( 'https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png' ) @is_pipeline_test @require_torch @require_vision class _lowercase ( unittest.TestCase ): _UpperCAmelCase = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def UpperCamelCase ( self , A__ , A__ , A__ ) -> int: snake_case = pipeline( '''document-question-answering''' , model=__a , tokenizer=__a , image_processor=__a ) snake_case = INVOICE_URL snake_case = list(zip(*apply_tesseract(load_image(__a ) , __a , '''''' ) ) ) snake_case = """What is the placebo?""" snake_case = [ { """image""": load_image(__a ), """question""": question, }, { """image""": image, """question""": question, }, { """image""": image, """question""": question, """word_boxes""": word_boxes, }, ] return dqa_pipeline, examples def UpperCamelCase ( self , A__ , A__ ) -> str: snake_case = dqa_pipeline(__a , top_k=2 ) self.assertEqual( __a , [ [ {'''score''': ANY(__a ), '''answer''': ANY(__a ), '''start''': ANY(__a ), '''end''': ANY(__a )}, {'''score''': ANY(__a ), '''answer''': ANY(__a ), '''start''': ANY(__a ), '''end''': ANY(__a )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def UpperCamelCase ( self ) -> Any: snake_case = pipeline('''document-question-answering''' , model='''hf-internal-testing/tiny-random-layoutlmv2''' ) snake_case = INVOICE_URL snake_case = """How many cats are there?""" snake_case = [ {"""score""": 0.0_0_0_1, """answer""": """oy 2312/2019""", """start""": 38, """end""": 39}, {"""score""": 0.0_0_0_1, """answer""": """oy 2312/2019 DUE""", """start""": 38, """end""": 40}, ] snake_case = dqa_pipeline(image=__a , question=__a , top_k=2 ) self.assertEqual(nested_simplify(__a , decimals=4 ) , __a ) snake_case = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual(nested_simplify(__a , decimals=4 ) , __a ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably snake_case = """./tests/fixtures/tests_samples/COCO/000000039769.png""" snake_case = dqa_pipeline(image=__a , question=__a , top_k=2 ) self.assertEqual(__a , [] ) # We can optionnally pass directly the words and bounding boxes snake_case = """./tests/fixtures/tests_samples/COCO/000000039769.png""" snake_case = [] snake_case = [] snake_case = dqa_pipeline(image=__a , question=__a , words=__a , boxes=__a , top_k=2 ) self.assertEqual(__a , [] ) @slow @require_torch @require_detectrona @require_pytesseract def UpperCamelCase ( self ) -> Optional[int]: snake_case = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , ) snake_case = INVOICE_URL snake_case = """What is the invoice number?""" snake_case = dqa_pipeline(image=__a , question=__a , top_k=2 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {'''score''': 0.9_9_4_4, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_0_0_9, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) snake_case = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {'''score''': 0.9_9_4_4, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_0_0_9, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) snake_case = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ [ {'''score''': 0.9_9_4_4, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_0_0_9, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def UpperCamelCase ( self ) -> str: snake_case = pipeline( '''document-question-answering''' , model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' , revision='''9977165''' , max_seq_len=50 , ) snake_case = INVOICE_URL snake_case = """What is the invoice number?""" snake_case = dqa_pipeline(image=__a , question=__a , top_k=2 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {'''score''': 0.9_9_7_4, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_9_4_8, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) snake_case = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {'''score''': 0.9_9_7_4, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_9_4_8, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) snake_case = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ [ {'''score''': 0.9_9_7_4, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, {'''score''': 0.9_9_4_8, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def UpperCamelCase ( self ) -> List[Any]: snake_case = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__a ) snake_case = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__a , revision='''3dc6de3''' , ) snake_case = INVOICE_URL snake_case = """What is the invoice number?""" snake_case = dqa_pipeline(image=__a , question=__a , top_k=2 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {'''score''': 0.4_2_5_1, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_8_1_9, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) snake_case = dqa_pipeline({'''image''': image, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {'''score''': 0.4_2_5_1, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_8_1_9, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) snake_case = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ [ {'''score''': 0.4_2_5_1, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_8_1_9, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] ] * 2 , ) snake_case = list(zip(*apply_tesseract(load_image(__a ) , __a , '''''' ) ) ) # This model should also work if `image` is set to None snake_case = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {'''score''': 0.4_2_5_1, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.0_8_1_9, '''answer''': '''1110212019''', '''start''': 23, '''end''': 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def UpperCamelCase ( self ) -> Dict: snake_case = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' , revision='''3dc6de3''' , add_prefix_space=__a ) snake_case = pipeline( '''document-question-answering''' , model='''impira/layoutlm-document-qa''' , tokenizer=__a , revision='''3dc6de3''' , max_seq_len=50 , ) snake_case = INVOICE_URL snake_case = """What is the invoice number?""" snake_case = dqa_pipeline(image=__a , question=__a , top_k=2 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {'''score''': 0.9_9_9_9, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_9_9_8, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) snake_case = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ [ {'''score''': 0.9_9_9_9, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_9_9_8, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] ] * 2 , ) snake_case = list(zip(*apply_tesseract(load_image(__a ) , __a , '''''' ) ) ) # This model should also work if `image` is set to None snake_case = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} , top_k=2 ) self.assertEqual( nested_simplify(__a , decimals=4 ) , [ {'''score''': 0.9_9_9_9, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, {'''score''': 0.9_9_9_8, '''answer''': '''us-001''', '''start''': 16, '''end''': 16}, ] , ) @slow @require_torch def UpperCamelCase ( self ) -> Dict: snake_case = pipeline( '''document-question-answering''' , model='''naver-clova-ix/donut-base-finetuned-docvqa''' , tokenizer=AutoTokenizer.from_pretrained('''naver-clova-ix/donut-base-finetuned-docvqa''' ) , feature_extractor='''naver-clova-ix/donut-base-finetuned-docvqa''' , ) snake_case = INVOICE_URL snake_case = """What is the invoice number?""" snake_case = dqa_pipeline(image=__a , question=__a , top_k=2 ) self.assertEqual(nested_simplify(__a , decimals=4 ) , [{'''answer''': '''us-001'''}] ) @require_tf @unittest.skip('''Document question answering not implemented in TF''' ) def UpperCamelCase ( self ) -> Tuple: pass
720
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() _lowercase = logging.get_logger(__name__) def __UpperCamelCase ( a : Dict , a : Optional[int] , a : Dict , a : Dict ) ->Union[str, Any]: snake_case = original_name.split('''.''' )[0] snake_case = key.split('''.''' ) snake_case = int(key_list[key_list.index(a ) - 2] ) snake_case = int(key_list[key_list.index(a ) - 1] ) snake_case = orig_block_num - offset snake_case = key.replace(f"""{orig_block_num}.{layer_num}.{original_name}""" , f"""block.{new_block_num}.{layer_num}.{new_name}""" ) return key def __UpperCamelCase ( a : Tuple ) ->Dict: snake_case = OrderedDict() snake_case , snake_case = 0, 0 for key, value in state_dict.items(): if key.startswith('''network''' ): snake_case = key.replace('''network''' , '''poolformer.encoder''' ) if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith('''bias''' ) and "patch_embed" not in key: patch_emb_offset += 1 snake_case = key[: key.find('''proj''' )] snake_case = key.replace(a , f"""patch_embeddings.{total_embed_found}.""" ) snake_case = key.replace('''proj''' , '''projection''' ) if key.endswith('''bias''' ): total_embed_found += 1 if "patch_embeddings" in key: snake_case = '''poolformer.encoder.''' + key if "mlp.fc1" in key: snake_case = replace_key_with_offset(a , a , '''mlp.fc1''' , '''output.conv1''' ) if "mlp.fc2" in key: snake_case = replace_key_with_offset(a , a , '''mlp.fc2''' , '''output.conv2''' ) if "norm1" in key: snake_case = replace_key_with_offset(a , a , '''norm1''' , '''before_norm''' ) if "norm2" in key: snake_case = replace_key_with_offset(a , a , '''norm2''' , '''after_norm''' ) if "layer_scale_1" in key: snake_case = replace_key_with_offset(a , a , '''layer_scale_1''' , '''layer_scale_1''' ) if "layer_scale_2" in key: snake_case = replace_key_with_offset(a , a , '''layer_scale_2''' , '''layer_scale_2''' ) if "head" in key: snake_case = key.replace('''head''' , '''classifier''' ) snake_case = value return new_state_dict def __UpperCamelCase ( ) ->Optional[int]: snake_case = '''http://images.cocodataset.org/val2017/000000039769.jpg''' snake_case = Image.open(requests.get(a , stream=a ).raw ) return image @torch.no_grad() def __UpperCamelCase ( a : Dict , a : Optional[Any] , a : Tuple ) ->List[str]: snake_case = PoolFormerConfig() # set attributes based on model_name snake_case = '''huggingface/label-files''' snake_case = model_name[-3:] snake_case = 1000 snake_case = '''imagenet-1k-id2label.json''' snake_case = (1, 1000) # set config attributes snake_case = json.load(open(hf_hub_download(a , a , repo_type='''dataset''' ) , '''r''' ) ) snake_case = {int(a ): v for k, v in idalabel.items()} snake_case = idalabel snake_case = {v: k for k, v in idalabel.items()} if size == "s12": snake_case = [2, 2, 6, 2] snake_case = [64, 128, 320, 512] snake_case = 4.0 snake_case = 0.9 elif size == "s24": snake_case = [4, 4, 12, 4] snake_case = [64, 128, 320, 512] snake_case = 4.0 snake_case = 0.9 elif size == "s36": snake_case = [6, 6, 18, 6] snake_case = [64, 128, 320, 512] snake_case = 4.0 snake_case = 1e-6 snake_case = 0.9 elif size == "m36": snake_case = [6, 6, 18, 6] snake_case = [96, 192, 384, 768] snake_case = 4.0 snake_case = 1e-6 snake_case = 0.95 elif size == "m48": snake_case = [8, 8, 24, 8] snake_case = [96, 192, 384, 768] snake_case = 4.0 snake_case = 1e-6 snake_case = 0.95 else: raise ValueError(f"""Size {size} not supported""" ) # load image processor snake_case = PoolFormerImageProcessor(crop_pct=a ) # Prepare image snake_case = prepare_img() snake_case = image_processor(images=a , return_tensors='''pt''' ).pixel_values logger.info(f"""Converting model {model_name}...""" ) # load original state dict snake_case = torch.load(a , map_location=torch.device('''cpu''' ) ) # rename keys snake_case = rename_keys(a ) # create HuggingFace model and load state dict snake_case = PoolFormerForImageClassification(a ) model.load_state_dict(a ) model.eval() # Define image processor snake_case = PoolFormerImageProcessor(crop_pct=a ) snake_case = image_processor(images=prepare_img() , return_tensors='''pt''' ).pixel_values # forward pass snake_case = model(a ) snake_case = outputs.logits # define expected logit slices for different models if size == "s12": snake_case = torch.tensor([-0.3045, -0.6758, -0.4869] ) elif size == "s24": snake_case = torch.tensor([0.4402, -0.1374, -0.8045] ) elif size == "s36": snake_case = torch.tensor([-0.6080, -0.5133, -0.5898] ) elif size == "m36": snake_case = torch.tensor([0.3952, 0.2263, -1.2668] ) elif size == "m48": snake_case = torch.tensor([0.1167, -0.0656, -0.3423] ) else: raise ValueError(f"""Size {size} not supported""" ) # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] , a , atol=1e-2 ) # finally, save model and image processor logger.info(f"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(a ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument( '--model_name', default='poolformer_s12', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) _lowercase = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
44
0
from __future__ import annotations import math def __UpperCamelCase ( a : int , a : int , a : bool , a : list[int] , a : float ) ->str: if depth < 0: raise ValueError('''Depth cannot be less than 0''' ) if not scores: raise ValueError('''Scores cannot be empty''' ) if depth == height: return scores[node_index] return ( max( minimax(depth + 1 , node_index * 2 , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) , ) if is_max else min( minimax(depth + 1 , node_index * 2 , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) , minimax(depth + 1 , node_index * 2 + 1 , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) , ) ) def __UpperCamelCase ( ) ->Optional[int]: snake_case = [90, 23, 6, 33, 21, 65, 123, 3_4423] snake_case = math.log(len(_UpperCamelCase ) , 2 ) print(f"""Optimal value : {minimax(0 , 0 , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase )}""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
721
'''simple docstring''' import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow _lowercase = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ 'text-classification', 'language-modeling', 'summarization', 'token-classification', 'question-answering', ] ] sys.path.extend(SRC_DIRS) if SRC_DIRS is not None: import run_clm_flax import run_flax_glue import run_flax_ner import run_mlm_flax import run_qa import run_summarization_flax import run_ta_mlm_flax logging.basicConfig(level=logging.DEBUG) _lowercase = logging.getLogger() def __UpperCamelCase ( ) ->Tuple: snake_case = argparse.ArgumentParser() parser.add_argument('''-f''' ) snake_case = parser.parse_args() return args.f def __UpperCamelCase ( a : Dict , a : Tuple="eval" ) ->List[Any]: snake_case = os.path.join(a , f"""{split}_results.json""" ) if os.path.exists(a ): with open(a , '''r''' ) as f: return json.load(a ) raise ValueError(f"""can't find {path}""" ) _lowercase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class _lowercase ( __a ): def UpperCamelCase ( self ) -> List[str]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --eval_steps=2 --warmup_steps=2 --seed=42 --max_seq_length=128 """.split() with patch.object(A__ , '''argv''' , A__ ): run_flax_glue.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.7_5 ) @slow def UpperCamelCase ( self ) -> List[Any]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_clm_flax.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --block_size 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(A__ , '''argv''' , A__ ): run_clm_flax.main() snake_case = get_results(A__ ) self.assertLess(result['''eval_perplexity'''] , 1_00 ) @slow def UpperCamelCase ( self ) -> int: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_summarization.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --test_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=8 --do_train --do_eval --do_predict --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --predict_with_generate """.split() with patch.object(A__ , '''argv''' , A__ ): run_summarization_flax.main() snake_case = get_results(A__ , split='''test''' ) self.assertGreaterEqual(result['''test_rouge1'''] , 10 ) self.assertGreaterEqual(result['''test_rouge2'''] , 2 ) self.assertGreaterEqual(result['''test_rougeL'''] , 7 ) self.assertGreaterEqual(result['''test_rougeLsum'''] , 7 ) @slow def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_mlm.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --overwrite_output_dir --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --logging_steps 2 --eval_steps 2 --do_train --do_eval --num_train_epochs=1 """.split() with patch.object(A__ , '''argv''' , A__ ): run_mlm_flax.main() snake_case = get_results(A__ ) self.assertLess(result['''eval_perplexity'''] , 42 ) @slow def UpperCamelCase ( self ) -> Dict: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_t5_mlm_flax.py --model_name_or_path t5-small --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(A__ , '''argv''' , A__ ): run_ta_mlm_flax.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.4_2 ) @slow def UpperCamelCase ( self ) -> int: # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu snake_case = 7 if get_gpu_count() > 1 else 2 snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_flax_ner.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --overwrite_output_dir --do_train --do_eval --warmup_steps=2 --learning_rate=2e-4 --logging_steps 2 --eval_steps 2 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 """.split() with patch.object(A__ , '''argv''' , A__ ): run_flax_ner.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.7_5 ) self.assertGreaterEqual(result['''eval_f1'''] , 0.3 ) @slow def UpperCamelCase ( self ) -> Any: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_qa.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=2 --do_train --do_eval --logging_steps 2 --eval_steps 2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 """.split() with patch.object(A__ , '''argv''' , A__ ): run_qa.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_f1'''] , 30 ) self.assertGreaterEqual(result['''eval_exact'''] , 30 )
44
0
'''simple docstring''' import copy from typing import Dict, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING from ..detr import DetrConfig from ..swin import SwinConfig _lowercase = { 'facebook/maskformer-swin-base-ade': ( 'https://huggingface.co/facebook/maskformer-swin-base-ade/blob/main/config.json' ) # See all MaskFormer models at https://huggingface.co/models?filter=maskformer } _lowercase = logging.get_logger(__name__) class _lowercase ( __a ): _UpperCAmelCase = 'maskformer' _UpperCAmelCase = {'hidden_size': 'mask_feature_size'} _UpperCAmelCase = ['resnet', 'swin'] _UpperCAmelCase = ['detr'] def __init__( self , A__ = 2_56 , A__ = 2_56 , A__ = 0.1 , A__ = False , A__ = None , A__ = None , A__ = 0.0_2 , A__ = 1.0 , A__ = 1.0 , A__ = 1.0 , A__ = 2_0.0 , A__ = None , **A__ , ) -> str: if backbone_config is None: # fall back to https://huggingface.co/microsoft/swin-base-patch4-window12-384-in22k snake_case = SwinConfig( image_size=3_84 , in_channels=3 , patch_size=4 , embed_dim=1_28 , depths=[2, 2, 18, 2] , num_heads=[4, 8, 16, 32] , window_size=12 , drop_path_rate=0.3 , out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] , ) if isinstance(A__ , A__ ): snake_case = backbone_config.pop('''model_type''' ) snake_case = CONFIG_MAPPING[backbone_model_type] snake_case = config_class.from_dict(A__ ) # verify that the backbone is supported if backbone_config.model_type not in self.backbones_supported: logger.warning_once( F"""Backbone {backbone_config.model_type} is not a supported model and may not be compatible with MaskFormer. """ F"""Supported model types: {','.join(self.backbones_supported )}""" ) if decoder_config is None: # fall back to https://huggingface.co/facebook/detr-resnet-50 snake_case = DetrConfig() else: # verify that the decoder is supported snake_case = ( decoder_config.pop('''model_type''' ) if isinstance(A__ , A__ ) else decoder_config.model_type ) if decoder_type not in self.decoders_supported: raise ValueError( F"""Transformer Decoder {decoder_type} not supported, please use one of""" F""" {','.join(self.decoders_supported )}""" ) if isinstance(A__ , A__ ): snake_case = CONFIG_MAPPING[decoder_type] snake_case = config_class.from_dict(A__ ) snake_case = backbone_config snake_case = decoder_config # main feature dimension for the model snake_case = fpn_feature_size snake_case = mask_feature_size # initializer snake_case = init_std snake_case = init_xavier_std # Hungarian matcher && loss snake_case = cross_entropy_weight snake_case = dice_weight snake_case = mask_weight snake_case = use_auxiliary_loss snake_case = no_object_weight snake_case = output_auxiliary_logits snake_case = self.decoder_config.encoder_attention_heads snake_case = self.decoder_config.num_hidden_layers super().__init__(**A__ ) @classmethod def UpperCamelCase ( cls , A__ , A__ , **A__ ) -> List[str]: return cls( backbone_config=A__ , decoder_config=A__ , **A__ , ) def UpperCamelCase ( self ) -> Dict[str, any]: snake_case = copy.deepcopy(self.__dict__ ) snake_case = self.backbone_config.to_dict() snake_case = self.decoder_config.to_dict() snake_case = self.__class__.model_type return output
700
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS _lowercase = logging.get_logger(__name__) _lowercase = { 'linear': get_linear_schedule_with_warmup, 'cosine': get_cosine_schedule_with_warmup, 'cosine_w_restarts': get_cosine_with_hard_restarts_schedule_with_warmup, 'polynomial': get_polynomial_decay_schedule_with_warmup, 'constant': get_constant_schedule, 'constant_w_warmup': get_constant_schedule_with_warmup, } class _lowercase ( __a ): def __init__( self , A__=None , A__=None , *A__ , **A__ ) -> Union[str, Any]: super().__init__(*A__ , **A__ ) if config is None: assert isinstance(self.model , A__ ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) snake_case = self.model.config else: snake_case = config snake_case = data_args snake_case = self.config.tgt_vocab_size if isinstance(self.config , A__ ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ''' padding..''' ) if self.args.label_smoothing == 0: snake_case = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss snake_case = label_smoothed_nll_loss def UpperCamelCase ( self , A__ ) -> Tuple: if self.optimizer is None: snake_case = ['''bias''', '''LayerNorm.weight'''] snake_case = [ { '''params''': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], '''weight_decay''': self.args.weight_decay, }, { '''params''': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0, }, ] snake_case = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: snake_case = Adafactor snake_case = {'''scale_parameter''': False, '''relative_step''': False} else: snake_case = AdamW snake_case = { '''betas''': (self.args.adam_betaa, self.args.adam_betaa), '''eps''': self.args.adam_epsilon, } snake_case = self.args.learning_rate if self.sharded_ddp: snake_case = OSS( params=A__ , optim=A__ , **A__ , ) else: snake_case = optimizer_cls(A__ , **A__ ) if self.lr_scheduler is None: snake_case = self._get_lr_scheduler(A__ ) else: # ignoring --lr_scheduler logger.warning('''scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.''' ) def UpperCamelCase ( self , A__ ) -> Tuple: snake_case = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": snake_case = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": snake_case = schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: snake_case = schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=A__ ) return scheduler def UpperCamelCase ( self ) -> Optional[torch.utils.data.Sampler]: if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> List[Any]: if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token snake_case = model(**A__ , use_cache=A__ )[0] snake_case = self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models snake_case , snake_case = model(**A__ , labels=A__ , use_cache=A__ )[:2] else: # compute label smoothed loss snake_case = model(**A__ , use_cache=A__ )[0] snake_case = torch.nn.functional.log_softmax(A__ , dim=-1 ) snake_case , snake_case = self.loss_fn(A__ , A__ , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def UpperCamelCase ( self , A__ , A__ ) -> Any: snake_case = inputs.pop('''labels''' ) snake_case , snake_case = self._compute_loss(A__ , A__ , A__ ) return loss def UpperCamelCase ( self , A__ , A__ , A__ , A__ = None , ) -> Tuple[Optional[float], Optional[torch.Tensor], Optional[torch.Tensor]]: snake_case = self._prepare_inputs(A__ ) snake_case = { '''max_length''': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, '''num_beams''': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: snake_case = self.model.generate( inputs['''input_ids'''] , attention_mask=inputs['''attention_mask'''] , **A__ , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: snake_case = self._pad_tensors_to_max_len(A__ , gen_kwargs['''max_length'''] ) snake_case = inputs.pop('''labels''' ) with torch.no_grad(): # compute loss on predict data snake_case , snake_case = self._compute_loss(A__ , A__ , A__ ) snake_case = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) snake_case = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: snake_case = self._pad_tensors_to_max_len(A__ , gen_kwargs['''max_length'''] ) return (loss, logits, labels) def UpperCamelCase ( self , A__ , A__ ) -> List[str]: # If PAD token is not defined at least EOS token has to be defined snake_case = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( '''Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be''' F""" padded to `max_length`={max_length}""" ) snake_case = pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) snake_case = tensor return padded_tensor
44
0
'''simple docstring''' def __UpperCamelCase ( a : List[str]=2_8123 ) ->int: snake_case = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i snake_case = set() snake_case = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(lowercase__ ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
701
'''simple docstring''' import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def __UpperCamelCase ( a : List[str] ) ->str: snake_case = [] for line in lines: snake_case = re.sub(R'''#.*''' , '''''' , a ) # remove comments if line: filtered_lines.append(a ) snake_case = '''\n'''.join(a ) # Make a hash from all this code snake_case = full_str.encode('''utf-8''' ) return shaaaa(a ).hexdigest() # get importable module names and hash for caching _lowercase = { 'csv': (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), 'json': (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), 'pandas': (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), 'parquet': (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), 'arrow': (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), 'text': (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), 'imagefolder': (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), 'audiofolder': (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions _lowercase = { '.csv': ('csv', {}), '.tsv': ('csv', {'sep': '\t'}), '.json': ('json', {}), '.jsonl': ('json', {}), '.parquet': ('parquet', {}), '.arrow': ('arrow', {}), '.txt': ('text', {}), } _EXTENSION_TO_MODULE.update({ext: ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _lowercase = {'imagefolder', 'audiofolder'} # Used to filter data files based on extensions given a module name _lowercase = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append('.zip') _MODULE_TO_EXTENSIONS["audiofolder"].append('.zip')
44
0
'''simple docstring''' from ..utils import DummyObject, requires_backends class _lowercase ( metaclass=a__ ): _UpperCAmelCase = ["""flax"""] def __init__( self , *A__ , **A__ ): requires_backends(self , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) class _lowercase ( metaclass=a__ ): _UpperCAmelCase = ["""flax"""] def __init__( self , *A__ , **A__ ): requires_backends(self , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) class _lowercase ( metaclass=a__ ): _UpperCAmelCase = ["""flax"""] def __init__( self , *A__ , **A__ ): requires_backends(self , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) class _lowercase ( metaclass=a__ ): _UpperCAmelCase = ["""flax"""] def __init__( self , *A__ , **A__ ): requires_backends(self , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) class _lowercase ( metaclass=a__ ): _UpperCAmelCase = ["""flax"""] def __init__( self , *A__ , **A__ ): requires_backends(self , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) class _lowercase ( metaclass=a__ ): _UpperCAmelCase = ["""flax"""] def __init__( self , *A__ , **A__ ): requires_backends(self , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) class _lowercase ( metaclass=a__ ): _UpperCAmelCase = ["""flax"""] def __init__( self , *A__ , **A__ ): requires_backends(self , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) class _lowercase ( metaclass=a__ ): _UpperCAmelCase = ["""flax"""] def __init__( self , *A__ , **A__ ): requires_backends(self , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) class _lowercase ( metaclass=a__ ): _UpperCAmelCase = ["""flax"""] def __init__( self , *A__ , **A__ ): requires_backends(self , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) class _lowercase ( metaclass=a__ ): _UpperCAmelCase = ["""flax"""] def __init__( self , *A__ , **A__ ): requires_backends(self , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) class _lowercase ( metaclass=a__ ): _UpperCAmelCase = ["""flax"""] def __init__( self , *A__ , **A__ ): requires_backends(self , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) class _lowercase ( metaclass=a__ ): _UpperCAmelCase = ["""flax"""] def __init__( self , *A__ , **A__ ): requires_backends(self , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) class _lowercase ( metaclass=a__ ): _UpperCAmelCase = ["""flax"""] def __init__( self , *A__ , **A__ ): requires_backends(self , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ): requires_backends(cls , ['''flax'''] )
702
'''simple docstring''' _lowercase = { 'Pillow': 'Pillow', 'accelerate': 'accelerate>=0.11.0', 'compel': 'compel==0.1.8', 'black': 'black~=23.1', 'datasets': 'datasets', 'filelock': 'filelock', 'flax': 'flax>=0.4.1', 'hf-doc-builder': 'hf-doc-builder>=0.3.0', 'huggingface-hub': 'huggingface-hub>=0.13.2', 'requests-mock': 'requests-mock==1.10.0', 'importlib_metadata': 'importlib_metadata', 'invisible-watermark': 'invisible-watermark', 'isort': 'isort>=5.5.4', 'jax': 'jax>=0.2.8,!=0.3.2', 'jaxlib': 'jaxlib>=0.1.65', 'Jinja2': 'Jinja2', 'k-diffusion': 'k-diffusion>=0.0.12', 'torchsde': 'torchsde', 'note_seq': 'note_seq', 'librosa': 'librosa', 'numpy': 'numpy', 'omegaconf': 'omegaconf', 'parameterized': 'parameterized', 'protobuf': 'protobuf>=3.20.3,<4', 'pytest': 'pytest', 'pytest-timeout': 'pytest-timeout', 'pytest-xdist': 'pytest-xdist', 'ruff': 'ruff>=0.0.241', 'safetensors': 'safetensors', 'sentencepiece': 'sentencepiece>=0.1.91,!=0.1.92', 'scipy': 'scipy', 'onnx': 'onnx', 'regex': 'regex!=2019.12.17', 'requests': 'requests', 'tensorboard': 'tensorboard', 'torch': 'torch>=1.4', 'torchvision': 'torchvision', 'transformers': 'transformers>=4.25.1', 'urllib3': 'urllib3<=2.0.0', }
44
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import BlenderbotSmallConfig, BlenderbotSmallTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel @require_tf class _lowercase : _UpperCAmelCase = BlenderbotSmallConfig _UpperCAmelCase = {} _UpperCAmelCase = """gelu""" def __init__( self , A__ , A__=13 , A__=7 , A__=True , A__=False , A__=99 , A__=32 , A__=2 , A__=4 , A__=37 , A__=0.1 , A__=0.1 , A__=20 , A__=2 , A__=1 , A__=0 , ) -> List[str]: snake_case = parent snake_case = batch_size snake_case = seq_length snake_case = is_training snake_case = use_labels snake_case = vocab_size snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_size snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = max_position_embeddings snake_case = eos_token_id snake_case = pad_token_id snake_case = bos_token_id def UpperCamelCase ( self ) -> Dict: snake_case = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) snake_case = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) snake_case = tf.concat([input_ids, eos_tensor] , axis=1 ) snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) snake_case = prepare_blenderbot_small_inputs_dict(lowercase__ , lowercase__ , lowercase__ ) return config, inputs_dict def UpperCamelCase ( self , A__ , A__ ) -> int: snake_case = TFBlenderbotSmallModel(config=lowercase__ ).get_decoder() snake_case = inputs_dict["""input_ids"""] snake_case = input_ids[:1, :] snake_case = inputs_dict["""attention_mask"""][:1, :] snake_case = inputs_dict["""head_mask"""] snake_case = 1 # first forward pass snake_case = model(lowercase__ , attention_mask=lowercase__ , head_mask=lowercase__ , use_cache=lowercase__ ) snake_case = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and snake_case = tf.concat([input_ids, next_tokens] , axis=-1 ) snake_case = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) snake_case = model(lowercase__ , attention_mask=lowercase__ )[0] snake_case = model(lowercase__ , attention_mask=lowercase__ , past_key_values=lowercase__ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice snake_case = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) snake_case = output_from_no_past[:, -3:, random_slice_idx] snake_case = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowercase__ , lowercase__ , rtol=1e-3 ) def __UpperCamelCase ( a : Any , a : int , a : str , a : Tuple=None , a : Optional[int]=None , a : Union[str, Any]=None , a : Union[str, Any]=None , a : int=None , ) ->Optional[Any]: if attention_mask is None: snake_case = tf.cast(tf.math.not_equal(SCREAMING_SNAKE_CASE__ , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: snake_case = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: snake_case = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: snake_case = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: snake_case = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class _lowercase ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): _UpperCAmelCase = ( (TFBlenderbotSmallForConditionalGeneration, TFBlenderbotSmallModel) if is_tf_available() else () ) _UpperCAmelCase = (TFBlenderbotSmallForConditionalGeneration,) if is_tf_available() else () _UpperCAmelCase = ( { """conversational""": TFBlenderbotSmallForConditionalGeneration, """feature-extraction""": TFBlenderbotSmallModel, """summarization""": TFBlenderbotSmallForConditionalGeneration, """text2text-generation""": TFBlenderbotSmallForConditionalGeneration, """translation""": TFBlenderbotSmallForConditionalGeneration, } if is_tf_available() else {} ) _UpperCAmelCase = True _UpperCAmelCase = False _UpperCAmelCase = False def UpperCamelCase ( self ) -> Dict: snake_case = TFBlenderbotSmallModelTester(self ) snake_case = ConfigTester(self , config_class=lowercase__ ) def UpperCamelCase ( self ) -> Optional[int]: self.config_tester.run_common_tests() def UpperCamelCase ( self ) -> Any: snake_case = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowercase__ ) @require_tokenizers @require_tf class _lowercase ( unittest.TestCase ): _UpperCAmelCase = [ """Social anxiety\nWow, I am never shy. Do you have anxiety?\nYes. I end up sweating and blushing and feel like """ """ i'm going to throw up.\nand why is that?""" ] _UpperCAmelCase = """facebook/blenderbot_small-90M""" @cached_property def UpperCamelCase ( self ) -> Optional[Any]: # use "old" tokenizer here because of bug when downloading new tokenizer return BlenderbotSmallTokenizer.from_pretrained('''facebook/blenderbot-90M''' ) @cached_property def UpperCamelCase ( self ) -> Optional[int]: snake_case = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def UpperCamelCase ( self ) -> List[Any]: snake_case = self.tokenizer(self.src_text , return_tensors='''tf''' ) snake_case = self.model.generate( model_inputs.input_ids , attention_mask=model_inputs.attention_mask , num_beams=2 , use_cache=lowercase__ , ) snake_case = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=lowercase__ )[0] assert generated_words in ( "i don't know. i just feel like i'm going to throw up. it's not fun.", "i'm not sure. i just feel like i've been feeling like i have to be in a certain place", "i'm not sure. i just feel like i've been in a bad situation.", )
703
'''simple docstring''' import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _lowercase ( __a , __a , unittest.TestCase ): _UpperCAmelCase = IFInpaintingSuperResolutionPipeline _UpperCAmelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} _UpperCAmelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({'''original_image'''} ) _UpperCAmelCase = PipelineTesterMixin.required_optional_params - {'''latents'''} def UpperCamelCase ( self ) -> int: return self._get_superresolution_dummy_components() def UpperCamelCase ( self , A__ , A__=0 ) -> Union[str, Any]: if str(A__ ).startswith('''mps''' ): snake_case = torch.manual_seed(A__ ) else: snake_case = torch.Generator(device=A__ ).manual_seed(A__ ) snake_case = floats_tensor((1, 3, 16, 16) , rng=random.Random(A__ ) ).to(A__ ) snake_case = floats_tensor((1, 3, 32, 32) , rng=random.Random(A__ ) ).to(A__ ) snake_case = floats_tensor((1, 3, 32, 32) , rng=random.Random(A__ ) ).to(A__ ) snake_case = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase ( self ) -> List[Any]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def UpperCamelCase ( self ) -> Optional[Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def UpperCamelCase ( self ) -> List[str]: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1 ) def UpperCamelCase ( self ) -> int: self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def UpperCamelCase ( self ) -> Optional[Any]: self._test_save_load_local() def UpperCamelCase ( self ) -> Dict: self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
44
0
'''simple docstring''' import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, logging, ) logging.set_verbosity_info() _lowercase = logging.get_logger(__name__) _lowercase = { 'post_extract_proj': 'feature_projection.projection', 'encoder.pos_conv.0': 'encoder.pos_conv_embed.conv', 'self_attn.k_proj': 'encoder.layers.*.attention.k_proj', 'self_attn.v_proj': 'encoder.layers.*.attention.v_proj', 'self_attn.q_proj': 'encoder.layers.*.attention.q_proj', 'self_attn.out_proj': 'encoder.layers.*.attention.out_proj', 'self_attn_layer_norm': 'encoder.layers.*.layer_norm', 'fc1': 'encoder.layers.*.feed_forward.intermediate_dense', 'fc2': 'encoder.layers.*.feed_forward.output_dense', 'final_layer_norm': 'encoder.layers.*.final_layer_norm', 'encoder.layer_norm': 'encoder.layer_norm', 'w2v_model.layer_norm': 'feature_projection.layer_norm', 'quantizer.weight_proj': 'quantizer.weight_proj', 'quantizer.vars': 'quantizer.codevectors', 'project_q': 'project_q', 'final_proj': 'project_hid', 'w2v_encoder.proj': 'lm_head', 'mask_emb': 'masked_spec_embed', } _lowercase = [ 'lm_head', 'quantizer.weight_proj', 'quantizer.codevectors', 'project_q', 'project_hid', ] def __UpperCamelCase ( a : List[str] , a : List[str] , a : str , a : Union[str, Any] , a : Optional[Any] ) ->Tuple: for attribute in key.split('''.''' ): snake_case = getattr(a_ , a_ ) if weight_type is not None: snake_case = getattr(a_ , a_ ).shape else: snake_case = hf_pointer.shape assert hf_shape == value.shape, ( f"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": snake_case = value elif weight_type == "weight_g": snake_case = value elif weight_type == "weight_v": snake_case = value elif weight_type == "bias": snake_case = value else: snake_case = value logger.info(f"""{key + '.' + weight_type if weight_type is not None else ''} was initialized from {full_name}.""" ) def __UpperCamelCase ( a : Optional[Any] , a : int ) ->Any: snake_case = [] snake_case = fairseq_model.state_dict() snake_case = hf_model.feature_extractor # if encoder has different dim to decoder -> use proj_weight snake_case = None for name, value in fairseq_dict.items(): snake_case = False if "conv_layers" in name: load_conv_layer( a_ , a_ , a_ , a_ , hf_model.config.feat_extract_norm == '''group''' , ) snake_case = True elif name.split('''.''' )[0] == "proj": snake_case = fairseq_model.proj snake_case = True else: for key, mapped_key in MAPPING.items(): if key in name or key.split('''w2v_model.''' )[-1] == name.split('''.''' )[0]: snake_case = True if "*" in mapped_key: snake_case = name.split(a_ )[0].split('''.''' )[-2] snake_case = mapped_key.replace('''*''' , a_ ) if "weight_g" in name: snake_case = '''weight_g''' elif "weight_v" in name: snake_case = '''weight_v''' elif "bias" in name: snake_case = '''bias''' elif "weight" in name: snake_case = '''weight''' else: snake_case = None set_recursively(a_ , a_ , a_ , a_ , a_ ) continue if not is_used: unused_weights.append(a_ ) logger.warning(f"""Unused weights: {unused_weights}""" ) return proj_weight def __UpperCamelCase ( a : List[Any] , a : int , a : Optional[int] , a : int , a : Tuple ) ->str: snake_case = full_name.split('''conv_layers.''' )[-1] snake_case = name.split('''.''' ) snake_case = int(items[0] ) snake_case = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.""" ) snake_case = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.""" ) snake_case = value logger.info(f"""Feat extract conv layer {layer_id} was initialized from {full_name}.""" ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f"""{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was""" " found." ) snake_case = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f"""{full_name} has size {value.shape}, but""" f""" {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.""" ) snake_case = value logger.info(f"""Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.""" ) else: unused_weights.append(a_ ) def __UpperCamelCase ( a : Tuple ) ->Union[str, Any]: snake_case = emb.weight.shape snake_case = nn.Linear(a_ , a_ , bias=a_ ) snake_case = emb.weight.data return lin_layer def __UpperCamelCase ( a : Union[str, Any] ) ->Optional[Any]: with open(a_ , '''r''' , encoding='''utf-8''' ) as f: snake_case = f.readlines() snake_case = [line.split(''' ''' )[0] for line in lines] snake_case = len(a_ ) snake_case = { '''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3, } vocab_dict.update(dict(zip(a_ , range(4 , num_words + 4 ) ) ) ) return vocab_dict @torch.no_grad() def __UpperCamelCase ( a : List[Any] , a : Dict , a : Optional[int] , a : str , a : Tuple , a : Any , a : List[str] , ) ->Union[str, Any]: snake_case = WavaVecaConfig.from_pretrained(a_ ) snake_case = SpeechaTextaConfig.from_pretrained( a_ , vocab_size=a_ , decoder_layers=a_ , do_stable_layer_norm=a_ ) snake_case = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0 , do_normalize=a_ , return_attention_mask=a_ , ) snake_case = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={'''data''': '''/'''.join(dict_path.split('''/''' )[:-1] )} ) snake_case = model[0].eval() # set weights for wav2vec2 encoder snake_case = WavaVecaModel(a_ ) snake_case = recursively_load_weights_wavaveca(model.encoder , a_ ) snake_case = SpeechaTextaForCausalLM(a_ ) snake_case = hf_decoder.model.decoder.load_state_dict(model.decoder.state_dict() , strict=a_ ) # set output linear layer unexpected_keys.remove('''embed_out''' ) snake_case = nn.Parameter(model.decoder.embed_out.detach() ) # layer norm is init to identity matrix so leaving it is fine logger.warning(f"""The following keys are missing when loading the decoder weights: {missing_keys}""" ) logger.warning(f"""The following keys are unexpected when loading the decoder weights: {unexpected_keys}""" ) snake_case = SpeechEncoderDecoderModel(encoder=a_ , decoder=a_ ) snake_case = False # add projection layer snake_case = nn.Parameter(projection_layer.weight ) snake_case = nn.Parameter(projection_layer.bias ) snake_case = create_vocab_dict(a_ ) with open(os.path.join(a_ , '''vocab.json''' ) , '''w''' ) as fp: json.dump(a_ , a_ ) snake_case = SpeechaTextaTokenizer(os.path.join(a_ , '''vocab.json''' ) ) tokenizer.save_pretrained(a_ ) snake_case = hf_wavavec.config.to_dict() snake_case = tokenizer.pad_token_id snake_case = tokenizer.bos_token_id snake_case = tokenizer.eos_token_id snake_case = '''speech_to_text_2''' snake_case = '''wav2vec2''' snake_case = SpeechEncoderDecoderConfig.from_dict(a_ ) hf_wavavec.save_pretrained(a_ ) feature_extractor.save_pretrained(a_ ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--dict_path', default=None, type=str, help='Path to dict of fine-tuned model') parser.add_argument( '--encoder_config_path', default='facebook/wav2vec2-large-lv60', type=str, help='Path to hf encoder wav2vec2 checkpoint config', ) parser.add_argument( '--decoder_config_path', default='facebook/s2t-small-mustc-en-fr-st', type=str, help='Path to hf decoder s2t checkpoint config', ) parser.add_argument('--vocab_size', default=10_224, type=int, help='Vocab size of decoder') parser.add_argument('--num_decoder_layers', default=7, type=int, help='Number of decoder layers') _lowercase = parser.parse_args() convert_wavaveca_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.dict_path, encoder_config_path=args.encoder_config_path, decoder_config_path=args.decoder_config_path, vocab_size=args.vocab_size, num_decoder_layers=args.num_decoder_layers, )
704
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy _lowercase = logging.get_logger(__name__) class _lowercase ( __a ): def __init__( self , A__ , A__ , A__ , **A__ ) -> Union[str, Any]: snake_case = feature_size snake_case = sampling_rate snake_case = padding_value snake_case = kwargs.pop('''padding_side''' , '''right''' ) snake_case = kwargs.pop('''return_attention_mask''' , A__ ) super().__init__(**A__ ) def UpperCamelCase ( self , A__ , A__ = True , A__ = None , A__ = False , A__ = None , A__ = None , A__ = None , ) -> BatchFeature: # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(A__ , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): snake_case = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( '''You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`''' F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) snake_case = processed_features[self.model_input_names[0]] snake_case = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(A__ ) == 0: if return_attention_mask: snake_case = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch snake_case = required_input[0] if isinstance(A__ , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. snake_case = 0 while len(required_input[index] ) == 0: index += 1 if index < len(A__ ): snake_case = required_input[index][0] if return_tensors is None: if is_tf_tensor(A__ ): snake_case = '''tf''' elif is_torch_tensor(A__ ): snake_case = '''pt''' elif isinstance(A__ , (int, float, list, tuple, np.ndarray) ): snake_case = '''np''' else: raise ValueError( F"""type of {first_element} unknown: {type(A__ )}. """ '''Should be one of a python, numpy, pytorch or tensorflow object.''' ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): snake_case = to_numpy(A__ ) else: snake_case = [to_numpy(A__ ) for v in value] # Convert padding_strategy in PaddingStrategy snake_case = self._get_padding_strategies(padding=A__ , max_length=A__ ) snake_case = processed_features[self.model_input_names[0]] snake_case = len(A__ ) if not all(len(A__ ) == batch_size for v in processed_features.values() ): raise ValueError('''Some items in the output dictionary have a different batch size than others.''' ) snake_case = [] for i in range(A__ ): snake_case = {k: v[i] for k, v in processed_features.items()} # truncation snake_case = self._truncate( A__ , max_length=A__ , pad_to_multiple_of=A__ , truncation=A__ , ) truncated_inputs.append(A__ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length snake_case = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) snake_case = PaddingStrategy.MAX_LENGTH snake_case = {} for i in range(A__ ): # padding snake_case = self._pad( truncated_inputs[i] , max_length=A__ , padding_strategy=A__ , pad_to_multiple_of=A__ , return_attention_mask=A__ , ) for key, value in outputs.items(): if key not in batch_outputs: snake_case = [] if value.dtype is np.dtype(np.floataa ): snake_case = value.astype(np.floataa ) batch_outputs[key].append(A__ ) return BatchFeature(A__ , tensor_type=A__ ) def UpperCamelCase ( self , A__ , A__ = None , A__ = PaddingStrategy.DO_NOT_PAD , A__ = None , A__ = None , ) -> dict: snake_case = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: snake_case = len(A__ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): snake_case = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of snake_case = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(A__ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: snake_case = np.ones(len(A__ ) , dtype=np.intaa ) if needs_to_be_padded: snake_case = max_length - len(A__ ) if self.padding_side == "right": if return_attention_mask: snake_case = np.pad( processed_features['''attention_mask'''] , (0, difference) ) snake_case = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) snake_case = np.pad( A__ , A__ , '''constant''' , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: snake_case = np.pad( processed_features['''attention_mask'''] , (difference, 0) ) snake_case = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) snake_case = np.pad( A__ , A__ , '''constant''' , constant_values=self.padding_value ) else: raise ValueError('''Invalid padding strategy:''' + str(self.padding_side ) ) return processed_features def UpperCamelCase ( self , A__ , A__ = None , A__ = None , A__ = None , ) -> Union[str, Any]: if not truncation: return processed_features elif truncation and max_length is None: raise ValueError('''When setting ``truncation=True``, make sure that ``max_length`` is defined.''' ) snake_case = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): snake_case = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of snake_case = len(A__ ) > max_length if needs_to_be_truncated: snake_case = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: snake_case = processed_features['''attention_mask'''][:max_length] return processed_features def UpperCamelCase ( self , A__=False , A__=None ) -> Union[str, Any]: # Get padding strategy if padding is not False: if padding is True: snake_case = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(A__ , A__ ): snake_case = PaddingStrategy(A__ ) elif isinstance(A__ , A__ ): snake_case = padding else: snake_case = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( '''Asking to pad but the feature_extractor does not have a padding value. Please select a value to use''' ''' as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.''' ) return padding_strategy
44
0
'''simple docstring''' import torch from diffusers import DDIMParallelScheduler from .test_schedulers import SchedulerCommonTest class _lowercase ( __a ): _UpperCAmelCase = (DDIMParallelScheduler,) _UpperCAmelCase = (("eta", 0.0), ("num_inference_steps", 50)) def UpperCamelCase ( self , **A__ ) -> Dict: snake_case = { '''num_train_timesteps''': 10_00, '''beta_start''': 0.0_0_0_1, '''beta_end''': 0.0_2, '''beta_schedule''': '''linear''', '''clip_sample''': True, } config.update(**A__ ) return config def UpperCamelCase ( self , **A__ ) -> int: snake_case = self.scheduler_classes[0] snake_case = self.get_scheduler_config(**A__ ) snake_case = scheduler_class(**A__ ) snake_case , snake_case = 10, 0.0 snake_case = self.dummy_model() snake_case = self.dummy_sample_deter scheduler.set_timesteps(A__ ) for t in scheduler.timesteps: snake_case = model(A__ , A__ ) snake_case = scheduler.step(A__ , A__ , A__ , A__ ).prev_sample return sample def UpperCamelCase ( self ) -> Tuple: for timesteps in [1_00, 5_00, 10_00]: self.check_over_configs(num_train_timesteps=A__ ) def UpperCamelCase ( self ) -> List[Any]: for steps_offset in [0, 1]: self.check_over_configs(steps_offset=A__ ) snake_case = self.scheduler_classes[0] snake_case = self.get_scheduler_config(steps_offset=1 ) snake_case = scheduler_class(**A__ ) scheduler.set_timesteps(5 ) assert torch.equal(scheduler.timesteps , torch.LongTensor([8_01, 6_01, 4_01, 2_01, 1] ) ) def UpperCamelCase ( self ) -> Tuple: for beta_start, beta_end in zip([0.0_0_0_1, 0.0_0_1, 0.0_1, 0.1] , [0.0_0_2, 0.0_2, 0.2, 2] ): self.check_over_configs(beta_start=A__ , beta_end=A__ ) def UpperCamelCase ( self ) -> List[str]: for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=A__ ) def UpperCamelCase ( self ) -> str: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=A__ ) def UpperCamelCase ( self ) -> Union[str, Any]: for clip_sample in [True, False]: self.check_over_configs(clip_sample=A__ ) def UpperCamelCase ( self ) -> Any: for timestep_spacing in ["trailing", "leading"]: self.check_over_configs(timestep_spacing=A__ ) def UpperCamelCase ( self ) -> str: for rescale_betas_zero_snr in [True, False]: self.check_over_configs(rescale_betas_zero_snr=A__ ) def UpperCamelCase ( self ) -> int: self.check_over_configs(thresholding=A__ ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs( thresholding=A__ , prediction_type=A__ , sample_max_value=A__ , ) def UpperCamelCase ( self ) -> List[Any]: for t in [1, 10, 49]: self.check_over_forward(time_step=A__ ) def UpperCamelCase ( self ) -> List[Any]: for t, num_inference_steps in zip([1, 10, 50] , [10, 50, 5_00] ): self.check_over_forward(time_step=A__ , num_inference_steps=A__ ) def UpperCamelCase ( self ) -> List[Any]: for t, eta in zip([1, 10, 49] , [0.0, 0.5, 1.0] ): self.check_over_forward(time_step=A__ , eta=A__ ) def UpperCamelCase ( self ) -> Any: snake_case = self.scheduler_classes[0] snake_case = self.get_scheduler_config() snake_case = scheduler_class(**A__ ) assert torch.sum(torch.abs(scheduler._get_variance(0 , 0 ) - 0.0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(4_20 , 4_00 ) - 0.1_4_7_7_1 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(9_80 , 9_60 ) - 0.3_2_4_6_0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(0 , 0 ) - 0.0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(4_87 , 4_86 ) - 0.0_0_9_7_9 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(9_99 , 9_98 ) - 0.0_2 ) ) < 1e-5 def UpperCamelCase ( self ) -> Optional[Any]: snake_case = self.scheduler_classes[0] snake_case = self.get_scheduler_config() snake_case = scheduler_class(**A__ ) snake_case , snake_case = 10, 0.0 scheduler.set_timesteps(A__ ) snake_case = self.dummy_model() snake_case = self.dummy_sample_deter snake_case = self.dummy_sample_deter + 0.1 snake_case = self.dummy_sample_deter - 0.1 snake_case = samplea.shape[0] snake_case = torch.stack([samplea, samplea, samplea] , dim=0 ) snake_case = torch.arange(A__ )[0:3, None].repeat(1 , A__ ) snake_case = model(samples.flatten(0 , 1 ) , timesteps.flatten(0 , 1 ) ) snake_case = scheduler.batch_step_no_noise(A__ , timesteps.flatten(0 , 1 ) , samples.flatten(0 , 1 ) , A__ ) snake_case = torch.sum(torch.abs(A__ ) ) snake_case = torch.mean(torch.abs(A__ ) ) assert abs(result_sum.item() - 1_1_4_7.7_9_0_4 ) < 1e-2 assert abs(result_mean.item() - 0.4_9_8_2 ) < 1e-3 def UpperCamelCase ( self ) -> Tuple: snake_case = self.full_loop() snake_case = torch.sum(torch.abs(A__ ) ) snake_case = torch.mean(torch.abs(A__ ) ) assert abs(result_sum.item() - 1_7_2.0_0_6_7 ) < 1e-2 assert abs(result_mean.item() - 0.2_2_3_9_6_7 ) < 1e-3 def UpperCamelCase ( self ) -> Any: snake_case = self.full_loop(prediction_type='''v_prediction''' ) snake_case = torch.sum(torch.abs(A__ ) ) snake_case = torch.mean(torch.abs(A__ ) ) assert abs(result_sum.item() - 5_2.5_3_0_2 ) < 1e-2 assert abs(result_mean.item() - 0.0_6_8_4 ) < 1e-3 def UpperCamelCase ( self ) -> List[str]: snake_case = self.full_loop(set_alpha_to_one=A__ , beta_start=0.0_1 ) snake_case = torch.sum(torch.abs(A__ ) ) snake_case = torch.mean(torch.abs(A__ ) ) assert abs(result_sum.item() - 1_4_9.8_2_9_5 ) < 1e-2 assert abs(result_mean.item() - 0.1_9_5_1 ) < 1e-3 def UpperCamelCase ( self ) -> Optional[Any]: snake_case = self.full_loop(set_alpha_to_one=A__ , beta_start=0.0_1 ) snake_case = torch.sum(torch.abs(A__ ) ) snake_case = torch.mean(torch.abs(A__ ) ) assert abs(result_sum.item() - 1_4_9.0_7_8_4 ) < 1e-2 assert abs(result_mean.item() - 0.1_9_4_1 ) < 1e-3
705
'''simple docstring''' from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class _lowercase ( yaml.SafeLoader ): def UpperCamelCase ( self , A__ ) -> List[str]: snake_case = [self.constructed_objects[key_node] for key_node, _ in node.value] snake_case = [tuple(A__ ) if isinstance(A__ , A__ ) else key for key in keys] snake_case = Counter(A__ ) snake_case = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F"""Got duplicate yaml keys: {duplicate_keys}""" ) def UpperCamelCase ( self , A__ , A__=False ) -> List[Any]: snake_case = super().construct_mapping(A__ , deep=A__ ) self._check_no_duplicates_on_constructed_node(A__ ) return mapping def __UpperCamelCase ( a : str ) ->Tuple[Optional[str], str]: snake_case = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: snake_case = full_content[1:].index('''---''' ) + 1 snake_case = '''\n'''.join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(a ) class _lowercase ( __a ): # class attributes _UpperCAmelCase = {'''train_eval_index'''} # train-eval-index in the YAML metadata @classmethod def UpperCamelCase ( cls , A__ ) -> "DatasetMetadata": with open(A__ , encoding='''utf-8''' ) as readme_file: snake_case , snake_case = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(A__ ) else: return cls() def UpperCamelCase ( self , A__ ) -> str: if path.exists(): with open(A__ , encoding='''utf-8''' ) as readme_file: snake_case = readme_file.read() else: snake_case = None snake_case = self._to_readme(A__ ) with open(A__ , '''w''' , encoding='''utf-8''' ) as readme_file: readme_file.write(A__ ) def UpperCamelCase ( self , A__ = None ) -> str: if readme_content is not None: snake_case , snake_case = _split_yaml_from_readme(A__ ) snake_case = '''---\n''' + self.to_yaml_string() + '''---\n''' + content else: snake_case = '''---\n''' + self.to_yaml_string() + '''---\n''' return full_content @classmethod def UpperCamelCase ( cls , A__ ) -> "DatasetMetadata": snake_case = yaml.load(A__ , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields snake_case = { (key.replace('''-''' , '''_''' ) if key.replace('''-''' , '''_''' ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**A__ ) def UpperCamelCase ( self ) -> str: return yaml.safe_dump( { (key.replace('''_''' , '''-''' ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=A__ , allow_unicode=A__ , encoding='''utf-8''' , ).decode('''utf-8''' ) _lowercase = { 'image-classification': [], 'translation': [], 'image-segmentation': [], 'fill-mask': [], 'automatic-speech-recognition': [], 'token-classification': [], 'sentence-similarity': [], 'audio-classification': [], 'question-answering': [], 'summarization': [], 'zero-shot-classification': [], 'table-to-text': [], 'feature-extraction': [], 'other': [], 'multiple-choice': [], 'text-classification': [], 'text-to-image': [], 'text2text-generation': [], 'zero-shot-image-classification': [], 'tabular-classification': [], 'tabular-regression': [], 'image-to-image': [], 'tabular-to-text': [], 'unconditional-image-generation': [], 'text-retrieval': [], 'text-to-speech': [], 'object-detection': [], 'audio-to-audio': [], 'text-generation': [], 'conversational': [], 'table-question-answering': [], 'visual-question-answering': [], 'image-to-text': [], 'reinforcement-learning': [], 'voice-activity-detection': [], 'time-series-forecasting': [], 'document-question-answering': [], } if __name__ == "__main__": from argparse import ArgumentParser _lowercase = ArgumentParser(usage='Validate the yaml metadata block of a README.md file.') ap.add_argument('readme_filepath') _lowercase = ap.parse_args() _lowercase = Path(args.readme_filepath) _lowercase = DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
44
0
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class _lowercase ( unittest.TestCase ): def UpperCamelCase ( self ) -> str: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def UpperCamelCase ( self ) -> Tuple: torch.manual_seed(0 ) snake_case = UNetaDModel( sample_size=(32, 64) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(1_28, 1_28) , down_block_types=('''AttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''AttnUpBlock2D''') , ) return model @property def UpperCamelCase ( self ) -> List[str]: torch.manual_seed(0 ) snake_case = UNetaDConditionModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(1_28, 1_28) , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , cross_attention_dim=10 , ) return model @property def UpperCamelCase ( self ) -> Union[str, Any]: torch.manual_seed(0 ) snake_case = AutoencoderKL( sample_size=(1_28, 64) , in_channels=1 , out_channels=1 , latent_channels=1 , layers_per_block=2 , block_out_channels=(1_28, 1_28) , down_block_types=('''DownEncoderBlock2D''', '''DownEncoderBlock2D''') , up_block_types=('''UpDecoderBlock2D''', '''UpDecoderBlock2D''') , ) snake_case = UNetaDModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(1_28, 1_28) , down_block_types=('''AttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''AttnUpBlock2D''') , ) return vqvae, unet @slow def UpperCamelCase ( self ) -> Tuple: snake_case = "cpu" # ensure determinism for the device-dependent torch.Generator snake_case = Mel( x_res=self.dummy_unet.config.sample_size[1] , y_res=self.dummy_unet.config.sample_size[0] , ) snake_case = DDPMScheduler() snake_case = AudioDiffusionPipeline(vqvae=__a , unet=self.dummy_unet , mel=__a , scheduler=__a ) snake_case = pipe.to(__a ) pipe.set_progress_bar_config(disable=__a ) snake_case = torch.Generator(device=__a ).manual_seed(42 ) snake_case = pipe(generator=__a , steps=4 ) snake_case = output.audios[0] snake_case = output.images[0] snake_case = torch.Generator(device=__a ).manual_seed(42 ) snake_case = pipe(generator=__a , steps=4 , return_dict=__a ) snake_case = output[0][0] assert audio.shape == (1, (self.dummy_unet.config.sample_size[1] - 1) * mel.hop_length) assert ( image.height == self.dummy_unet.config.sample_size[0] and image.width == self.dummy_unet.config.sample_size[1] ) snake_case = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] snake_case = np.frombuffer(image_from_tuple.tobytes() , dtype='''uint8''' )[:10] snake_case = np.array([69, 2_55, 2_55, 2_55, 0, 0, 77, 1_81, 12, 1_27] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() == 0 snake_case = Mel( x_res=self.dummy_vqvae_and_unet[0].config.sample_size[1] , y_res=self.dummy_vqvae_and_unet[0].config.sample_size[0] , ) snake_case = DDIMScheduler() snake_case = self.dummy_vqvae_and_unet snake_case = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=dummy_vqvae_and_unet[1] , mel=__a , scheduler=__a ) snake_case = pipe.to(__a ) pipe.set_progress_bar_config(disable=__a ) np.random.seed(0 ) snake_case = np.random.uniform(-1 , 1 , ((dummy_vqvae_and_unet[0].config.sample_size[1] - 1) * mel.hop_length,) ) snake_case = torch.Generator(device=__a ).manual_seed(42 ) snake_case = pipe(raw_audio=__a , generator=__a , start_step=5 , steps=10 ) snake_case = output.images[0] assert ( image.height == self.dummy_vqvae_and_unet[0].config.sample_size[0] and image.width == self.dummy_vqvae_and_unet[0].config.sample_size[1] ) snake_case = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] snake_case = np.array([1_20, 1_17, 1_10, 1_09, 1_38, 1_67, 1_38, 1_48, 1_32, 1_21] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 snake_case = self.dummy_unet_condition snake_case = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=__a , mel=__a , scheduler=__a ) snake_case = pipe.to(__a ) pipe.set_progress_bar_config(disable=__a ) np.random.seed(0 ) snake_case = torch.rand((1, 1, 10) ) snake_case = pipe(generator=__a , encoding=__a ) snake_case = output.images[0] snake_case = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] snake_case = np.array([1_07, 1_03, 1_20, 1_27, 1_42, 1_22, 1_13, 1_22, 97, 1_11] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 @slow @require_torch_gpu class _lowercase ( unittest.TestCase ): def UpperCamelCase ( self ) -> Optional[int]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self ) -> str: snake_case = torch_device snake_case = DiffusionPipeline.from_pretrained('''teticio/audio-diffusion-ddim-256''' ) snake_case = pipe.to(__a ) pipe.set_progress_bar_config(disable=__a ) snake_case = torch.Generator(device=__a ).manual_seed(42 ) snake_case = pipe(generator=__a ) snake_case = output.audios[0] snake_case = output.images[0] assert audio.shape == (1, (pipe.unet.config.sample_size[1] - 1) * pipe.mel.hop_length) assert image.height == pipe.unet.config.sample_size[0] and image.width == pipe.unet.config.sample_size[1] snake_case = np.frombuffer(image.tobytes() , dtype='''uint8''' )[:10] snake_case = np.array([1_51, 1_67, 1_54, 1_44, 1_22, 1_34, 1_21, 1_05, 70, 26] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0
706
'''simple docstring''' import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowercase ( __a , unittest.TestCase ): _UpperCAmelCase = CodeGenTokenizer _UpperCAmelCase = CodeGenTokenizerFast _UpperCAmelCase = True _UpperCAmelCase = {'''add_prefix_space''': True} _UpperCAmelCase = False def UpperCamelCase ( self ) -> Tuple: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt snake_case = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', '''<|endoftext|>''', ] snake_case = dict(zip(A__ , range(len(A__ ) ) ) ) snake_case = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] snake_case = {'''unk_token''': '''<unk>'''} snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(A__ ) ) def UpperCamelCase ( self , **A__ ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **A__ ) def UpperCamelCase ( self , **A__ ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **A__ ) def UpperCamelCase ( self , A__ ) -> Tuple: snake_case = '''lower newer''' snake_case = '''lower newer''' return input_text, output_text def UpperCamelCase ( self ) -> List[Any]: snake_case = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) snake_case = '''lower newer''' snake_case = ['''\u0120low''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] snake_case = tokenizer.tokenize(A__ , add_prefix_space=A__ ) self.assertListEqual(A__ , A__ ) snake_case = tokens + [tokenizer.unk_token] snake_case = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(A__ ) , A__ ) def UpperCamelCase ( self ) -> Optional[int]: if not self.test_rust_tokenizer: return snake_case = self.get_tokenizer() snake_case = self.get_rust_tokenizer(add_prefix_space=A__ ) snake_case = '''lower newer''' # Testing tokenization snake_case = tokenizer.tokenize(A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.tokenize(A__ ) self.assertListEqual(A__ , A__ ) # Testing conversion to ids without special tokens snake_case = tokenizer.encode(A__ , add_special_tokens=A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.encode(A__ , add_special_tokens=A__ ) self.assertListEqual(A__ , A__ ) # Testing conversion to ids with special tokens snake_case = self.get_rust_tokenizer(add_prefix_space=A__ ) snake_case = tokenizer.encode(A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.encode(A__ ) self.assertListEqual(A__ , A__ ) # Testing the unknown token snake_case = tokens + [rust_tokenizer.unk_token] snake_case = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(A__ ) , A__ ) def UpperCamelCase ( self , *A__ , **A__ ) -> List[str]: # It's very difficult to mix/test pretokenization with byte-level # And get both CodeGen and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def UpperCamelCase ( self , A__=15 ) -> Tuple: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): snake_case = self.rust_tokenizer_class.from_pretrained(A__ , **A__ ) # Simple input snake_case = '''This is a simple input''' snake_case = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case = ('''This is a simple input''', '''This is a pair''') snake_case = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(A__ , tokenizer_r.encode , A__ , max_length=A__ , padding='''max_length''' ) # Simple input self.assertRaises(A__ , tokenizer_r.encode_plus , A__ , max_length=A__ , padding='''max_length''' ) # Simple input self.assertRaises( A__ , tokenizer_r.batch_encode_plus , A__ , max_length=A__ , padding='''max_length''' , ) # Pair input self.assertRaises(A__ , tokenizer_r.encode , A__ , max_length=A__ , padding='''max_length''' ) # Pair input self.assertRaises(A__ , tokenizer_r.encode_plus , A__ , max_length=A__ , padding='''max_length''' ) # Pair input self.assertRaises( A__ , tokenizer_r.batch_encode_plus , A__ , max_length=A__ , padding='''max_length''' , ) def UpperCamelCase ( self ) -> Tuple: snake_case = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token='''<pad>''' ) # Simple input snake_case = '''This is a simple input''' snake_case = ['''This is a simple input looooooooong''', '''This is a simple input'''] snake_case = ('''This is a simple input''', '''This is a pair''') snake_case = [ ('''This is a simple input loooooong''', '''This is a simple input'''), ('''This is a simple pair loooooong''', '''This is a simple pair'''), ] snake_case = tokenizer.pad_token_id snake_case = tokenizer(A__ , padding='''max_length''' , max_length=30 , return_tensors='''np''' ) snake_case = tokenizer(A__ , padding=A__ , truncate=A__ , return_tensors='''np''' ) snake_case = tokenizer(*A__ , padding='''max_length''' , max_length=60 , return_tensors='''np''' ) snake_case = tokenizer(A__ , padding=A__ , truncate=A__ , return_tensors='''np''' ) # s # test single string max_length padding self.assertEqual(out_s['''input_ids'''].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s['''input_ids'''] ) self.assertTrue(0 in out_s['''attention_mask'''] ) # s2 # test automatic padding self.assertEqual(out_sa['''input_ids'''].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['''input_ids'''][0] ) self.assertFalse(0 in out_sa['''attention_mask'''][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['''input_ids'''][1] ) self.assertTrue(0 in out_sa['''attention_mask'''][1] ) # p # test single pair max_length padding self.assertEqual(out_p['''input_ids'''].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p['''input_ids'''] ) self.assertTrue(0 in out_p['''attention_mask'''] ) # p2 # test automatic padding pair self.assertEqual(out_pa['''input_ids'''].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['''input_ids'''][0] ) self.assertFalse(0 in out_pa['''attention_mask'''][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['''input_ids'''][1] ) self.assertTrue(0 in out_pa['''attention_mask'''][1] ) def UpperCamelCase ( self ) -> str: snake_case = '''$$$''' snake_case = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=A__ , add_bos_token=A__ ) snake_case = '''This is a simple input''' snake_case = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case = tokenizer.bos_token_id snake_case = tokenizer(A__ ) snake_case = tokenizer(A__ ) self.assertEqual(out_s.input_ids[0] , A__ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) snake_case = tokenizer.decode(out_s.input_ids ) snake_case = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , A__ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def UpperCamelCase ( self ) -> Any: snake_case = CodeGenTokenizer.from_pretrained('''Salesforce/codegen-350M-mono''' ) snake_case = '''\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#''' snake_case = '''\nif len_a > len_b: result = a\nelse: result = b''' snake_case = tokenizer.encode(A__ ) snake_case = ['''^#''', re.escape('''<|endoftext|>''' ), '''^\'\'\'''', '''^"""''', '''\n\n\n'''] snake_case = tokenizer.decode(A__ , truncate_before_pattern=A__ ) self.assertEqual(A__ , A__ ) def UpperCamelCase ( self ) -> Union[str, Any]: pass
44
0
'''simple docstring''' import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() _lowercase = logging.get_logger('transformers.models.encodec') _lowercase = { 'quantizer.vq.layers.*._codebook.inited': 'quantizer.layers.*.codebook.inited', 'quantizer.vq.layers.*._codebook.cluster_size': 'quantizer.layers.*.codebook.cluster_size', 'quantizer.vq.layers.*._codebook.embed': 'quantizer.layers.*.codebook.embed', 'quantizer.vq.layers.*._codebook.embed_avg': 'quantizer.layers.*.codebook.embed_avg', } _lowercase = { 'encoder.model.0.conv.conv': 'encoder.layers.0.conv', 'encoder.model.1.block.1.conv.conv': 'encoder.layers.1.block.1.conv', 'encoder.model.1.block.3.conv.conv': 'encoder.layers.1.block.3.conv', 'encoder.model.1.shortcut.conv.conv': 'encoder.layers.1.shortcut.conv', 'encoder.model.3.conv.conv': 'encoder.layers.3.conv', 'encoder.model.4.block.1.conv.conv': 'encoder.layers.4.block.1.conv', 'encoder.model.4.block.3.conv.conv': 'encoder.layers.4.block.3.conv', 'encoder.model.4.shortcut.conv.conv': 'encoder.layers.4.shortcut.conv', 'encoder.model.6.conv.conv': 'encoder.layers.6.conv', 'encoder.model.7.block.1.conv.conv': 'encoder.layers.7.block.1.conv', 'encoder.model.7.block.3.conv.conv': 'encoder.layers.7.block.3.conv', 'encoder.model.7.shortcut.conv.conv': 'encoder.layers.7.shortcut.conv', 'encoder.model.9.conv.conv': 'encoder.layers.9.conv', 'encoder.model.10.block.1.conv.conv': 'encoder.layers.10.block.1.conv', 'encoder.model.10.block.3.conv.conv': 'encoder.layers.10.block.3.conv', 'encoder.model.10.shortcut.conv.conv': 'encoder.layers.10.shortcut.conv', 'encoder.model.12.conv.conv': 'encoder.layers.12.conv', 'encoder.model.13.lstm': 'encoder.layers.13.lstm', 'encoder.model.15.conv.conv': 'encoder.layers.15.conv', } _lowercase = { 'encoder.model.0.conv.norm': 'encoder.layers.0.norm', 'encoder.model.1.block.1.conv.norm': 'encoder.layers.1.block.1.norm', 'encoder.model.1.block.3.conv.norm': 'encoder.layers.1.block.3.norm', 'encoder.model.1.shortcut.conv.norm': 'encoder.layers.1.shortcut.norm', 'encoder.model.3.conv.norm': 'encoder.layers.3.norm', 'encoder.model.4.block.1.conv.norm': 'encoder.layers.4.block.1.norm', 'encoder.model.4.block.3.conv.norm': 'encoder.layers.4.block.3.norm', 'encoder.model.4.shortcut.conv.norm': 'encoder.layers.4.shortcut.norm', 'encoder.model.6.conv.norm': 'encoder.layers.6.norm', 'encoder.model.7.block.1.conv.norm': 'encoder.layers.7.block.1.norm', 'encoder.model.7.block.3.conv.norm': 'encoder.layers.7.block.3.norm', 'encoder.model.7.shortcut.conv.norm': 'encoder.layers.7.shortcut.norm', 'encoder.model.9.conv.norm': 'encoder.layers.9.norm', 'encoder.model.10.block.1.conv.norm': 'encoder.layers.10.block.1.norm', 'encoder.model.10.block.3.conv.norm': 'encoder.layers.10.block.3.norm', 'encoder.model.10.shortcut.conv.norm': 'encoder.layers.10.shortcut.norm', 'encoder.model.12.conv.norm': 'encoder.layers.12.norm', 'encoder.model.15.conv.norm': 'encoder.layers.15.norm', } _lowercase = { 'decoder.model.0.conv.conv': 'decoder.layers.0.conv', 'decoder.model.1.lstm': 'decoder.layers.1.lstm', 'decoder.model.3.convtr.convtr': 'decoder.layers.3.conv', 'decoder.model.4.block.1.conv.conv': 'decoder.layers.4.block.1.conv', 'decoder.model.4.block.3.conv.conv': 'decoder.layers.4.block.3.conv', 'decoder.model.4.shortcut.conv.conv': 'decoder.layers.4.shortcut.conv', 'decoder.model.6.convtr.convtr': 'decoder.layers.6.conv', 'decoder.model.7.block.1.conv.conv': 'decoder.layers.7.block.1.conv', 'decoder.model.7.block.3.conv.conv': 'decoder.layers.7.block.3.conv', 'decoder.model.7.shortcut.conv.conv': 'decoder.layers.7.shortcut.conv', 'decoder.model.9.convtr.convtr': 'decoder.layers.9.conv', 'decoder.model.10.block.1.conv.conv': 'decoder.layers.10.block.1.conv', 'decoder.model.10.block.3.conv.conv': 'decoder.layers.10.block.3.conv', 'decoder.model.10.shortcut.conv.conv': 'decoder.layers.10.shortcut.conv', 'decoder.model.12.convtr.convtr': 'decoder.layers.12.conv', 'decoder.model.13.block.1.conv.conv': 'decoder.layers.13.block.1.conv', 'decoder.model.13.block.3.conv.conv': 'decoder.layers.13.block.3.conv', 'decoder.model.13.shortcut.conv.conv': 'decoder.layers.13.shortcut.conv', 'decoder.model.15.conv.conv': 'decoder.layers.15.conv', } _lowercase = { 'decoder.model.0.conv.norm': 'decoder.layers.0.norm', 'decoder.model.3.convtr.norm': 'decoder.layers.3.norm', 'decoder.model.4.block.1.conv.norm': 'decoder.layers.4.block.1.norm', 'decoder.model.4.block.3.conv.norm': 'decoder.layers.4.block.3.norm', 'decoder.model.4.shortcut.conv.norm': 'decoder.layers.4.shortcut.norm', 'decoder.model.6.convtr.norm': 'decoder.layers.6.norm', 'decoder.model.7.block.1.conv.norm': 'decoder.layers.7.block.1.norm', 'decoder.model.7.block.3.conv.norm': 'decoder.layers.7.block.3.norm', 'decoder.model.7.shortcut.conv.norm': 'decoder.layers.7.shortcut.norm', 'decoder.model.9.convtr.norm': 'decoder.layers.9.norm', 'decoder.model.10.block.1.conv.norm': 'decoder.layers.10.block.1.norm', 'decoder.model.10.block.3.conv.norm': 'decoder.layers.10.block.3.norm', 'decoder.model.10.shortcut.conv.norm': 'decoder.layers.10.shortcut.norm', 'decoder.model.12.convtr.norm': 'decoder.layers.12.norm', 'decoder.model.13.block.1.conv.norm': 'decoder.layers.13.block.1.norm', 'decoder.model.13.block.3.conv.norm': 'decoder.layers.13.block.3.norm', 'decoder.model.13.shortcut.conv.norm': 'decoder.layers.13.shortcut.norm', 'decoder.model.15.conv.norm': 'decoder.layers.15.norm', } _lowercase = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } _lowercase = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } _lowercase = [] _lowercase = [] def __UpperCamelCase ( a : Any , a : List[Any] , a : List[Any] , a : List[str] , a : str ) ->Any: for attribute in key.split('''.''' ): snake_case = getattr(__UpperCamelCase , __UpperCamelCase ) if weight_type is not None: snake_case = getattr(__UpperCamelCase , __UpperCamelCase ).shape else: snake_case = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f"""Shape of hf {key + '.' + weight_type if weight_type is not None else ''} is {hf_shape}, but should be""" f""" {value.shape} for {full_name}""" ) if weight_type == "weight": snake_case = value elif weight_type == "weight_g": snake_case = value elif weight_type == "weight_v": snake_case = value elif weight_type == "bias": snake_case = value elif weight_type == "running_mean": snake_case = value elif weight_type == "running_var": snake_case = value elif weight_type == "num_batches_tracked": snake_case = value elif weight_type == "weight_ih_l0": snake_case = value elif weight_type == "weight_hh_l0": snake_case = value elif weight_type == "bias_ih_l0": snake_case = value elif weight_type == "bias_hh_l0": snake_case = value elif weight_type == "weight_ih_l1": snake_case = value elif weight_type == "weight_hh_l1": snake_case = value elif weight_type == "bias_ih_l1": snake_case = value elif weight_type == "bias_hh_l1": snake_case = value else: snake_case = value logger.info(f"""{key + ('.' + weight_type if weight_type is not None else '')} was initialized from {full_name}.""" ) def __UpperCamelCase ( a : Dict , a : str ) ->Union[str, Any]: for key in ignore_keys: if key.endswith('''.*''' ): if name.startswith(key[:-1] ): return True elif ".*." in key: snake_case , snake_case = key.split('''.*.''' ) if prefix in name and suffix in name: return True elif key in name: return True return False def __UpperCamelCase ( a : Tuple , a : Tuple , a : Union[str, Any] ) ->Any: snake_case = [] if model_name == "encodec_24khz" or "encodec_32khz": snake_case = MAPPING_24K elif model_name == "encodec_48khz": snake_case = MAPPING_48K else: raise ValueError(f"""Unsupported model: {model_name}""" ) for name, value in orig_dict.items(): if should_ignore(__UpperCamelCase , __UpperCamelCase ): logger.info(f"""{name} was ignored""" ) continue snake_case = False for key, mapped_key in MAPPING.items(): if "*" in key: snake_case , snake_case = key.split('''.*.''' ) if prefix in name and suffix in name: snake_case = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith('''embed''' ) and name.endswith('''embed_avg''' ): continue snake_case = True if "*" in mapped_key: snake_case = name.split(__UpperCamelCase )[0].split('''.''' )[-2] snake_case = mapped_key.replace('''*''' , __UpperCamelCase ) if "weight_g" in name: snake_case = '''weight_g''' elif "weight_v" in name: snake_case = '''weight_v''' elif "weight_ih_l0" in name: snake_case = '''weight_ih_l0''' elif "weight_hh_l0" in name: snake_case = '''weight_hh_l0''' elif "bias_ih_l0" in name: snake_case = '''bias_ih_l0''' elif "bias_hh_l0" in name: snake_case = '''bias_hh_l0''' elif "weight_ih_l1" in name: snake_case = '''weight_ih_l1''' elif "weight_hh_l1" in name: snake_case = '''weight_hh_l1''' elif "bias_ih_l1" in name: snake_case = '''bias_ih_l1''' elif "bias_hh_l1" in name: snake_case = '''bias_hh_l1''' elif "bias" in name: snake_case = '''bias''' elif "weight" in name: snake_case = '''weight''' elif "running_mean" in name: snake_case = '''running_mean''' elif "running_var" in name: snake_case = '''running_var''' elif "num_batches_tracked" in name: snake_case = '''num_batches_tracked''' else: snake_case = None set_recursively(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) continue if not is_used: unused_weights.append(__UpperCamelCase ) logger.warning(f"""Unused weights: {unused_weights}""" ) @torch.no_grad() def __UpperCamelCase ( a : List[str] , a : int , a : Optional[int] , a : Dict=None , a : List[Any]=None , ) ->int: if config_path is not None: snake_case = EncodecConfig.from_pretrained(__UpperCamelCase ) else: snake_case = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": snake_case = [8, 5, 4, 4] snake_case = [2.2] snake_case = 64 snake_case = 3_2000 snake_case = 2048 snake_case = False snake_case = False snake_case = False elif model_name == "encodec_48khz": snake_case = [8, 5, 4, 2] snake_case = [3.0, 6.0, 12.0, 24.0] snake_case = 4_8000 snake_case = 2 snake_case = False snake_case = '''time_group_norm''' snake_case = True snake_case = 1.0 snake_case = 0.01 else: raise ValueError(f"""Unknown model name: {model_name}""" ) snake_case = EncodecModel(__UpperCamelCase ) snake_case = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(__UpperCamelCase ) snake_case = torch.load(__UpperCamelCase ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights snake_case = original_checkpoint['''best_state'''] recursively_load_weights(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) model.save_pretrained(__UpperCamelCase ) if repo_id: print('''Pushing to the hub...''' ) feature_extractor.push_to_hub(__UpperCamelCase ) model.push_to_hub(__UpperCamelCase ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument( '--model', default='encodec_24khz', type=str, help='The model to convert. Should be one of \'encodec_24khz\', \'encodec_32khz\', \'encodec_48khz\'.', ) parser.add_argument('--checkpoint_path', required=True, default=None, type=str, help='Path to original checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument( '--pytorch_dump_folder_path', required=True, default=None, type=str, help='Path to the output PyTorch model.' ) parser.add_argument( '--push_to_hub', default=None, type=str, help='Where to upload the converted model on the 🤗 hub.' ) _lowercase = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
707
'''simple docstring''' from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowercase : def __init__( self , A__ , A__=13 , A__=30 , A__=2 , A__=3 , A__=True , A__=True , A__=32 , A__=2 , A__=4 , A__=37 , A__="gelu" , A__=0.1 , A__=0.1 , A__=10 , A__=0.0_2 , A__=3 , A__=None , ) -> List[Any]: snake_case = parent snake_case = batch_size snake_case = image_size snake_case = patch_size snake_case = num_channels snake_case = is_training snake_case = use_labels snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_size snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = type_sequence_label_size snake_case = initializer_range snake_case = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) snake_case = (image_size // patch_size) ** 2 snake_case = num_patches + 1 def UpperCamelCase ( self ) -> int: snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case = None if self.use_labels: snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ) -> int: return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A__ , initializer_range=self.initializer_range , ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Union[str, Any]: snake_case = TFViTModel(config=A__ ) snake_case = model(A__ , training=A__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. snake_case = self.image_size // 2 snake_case = pixel_values[:, :, :image_size, :image_size] snake_case = model(A__ , interpolate_pos_encoding=A__ , training=A__ ) snake_case = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Optional[int]: snake_case = self.type_sequence_label_size snake_case = TFViTForImageClassification(A__ ) snake_case = model(A__ , labels=A__ , training=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. snake_case = self.image_size // 2 snake_case = pixel_values[:, :, :image_size, :image_size] snake_case = model(A__ , interpolate_pos_encoding=A__ , training=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images snake_case = 1 snake_case = TFViTForImageClassification(A__ ) snake_case = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case = model(A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.prepare_config_and_inputs() snake_case , snake_case , snake_case = config_and_inputs snake_case = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class _lowercase ( __a , __a , unittest.TestCase ): _UpperCAmelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () _UpperCAmelCase = ( {'''feature-extraction''': TFViTModel, '''image-classification''': TFViTForImageClassification} if is_tf_available() else {} ) _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False def UpperCamelCase ( self ) -> List[Any]: snake_case = TFViTModelTester(self ) snake_case = ConfigTester(self , config_class=A__ , has_text_modality=A__ , hidden_size=37 ) def UpperCamelCase ( self ) -> int: self.config_tester.run_common_tests() @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def UpperCamelCase ( self ) -> int: pass @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def UpperCamelCase ( self ) -> str: pass def UpperCamelCase ( self ) -> Union[str, Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(A__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) snake_case = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A__ , tf.keras.layers.Layer ) ) def UpperCamelCase ( self ) -> List[Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(A__ ) snake_case = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case = [*signature.parameters.keys()] snake_case = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A__ ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A__ ) def UpperCamelCase ( self ) -> Optional[Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A__ ) @slow def UpperCamelCase ( self ) -> Any: snake_case = TFViTModel.from_pretrained('''google/vit-base-patch16-224''' ) self.assertIsNotNone(A__ ) def __UpperCamelCase ( ) ->Any: snake_case = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class _lowercase ( unittest.TestCase ): @cached_property def UpperCamelCase ( self ) -> Optional[int]: return ViTImageProcessor.from_pretrained('''google/vit-base-patch16-224''' ) if is_vision_available() else None @slow def UpperCamelCase ( self ) -> Dict: snake_case = TFViTForImageClassification.from_pretrained('''google/vit-base-patch16-224''' ) snake_case = self.default_image_processor snake_case = prepare_img() snake_case = image_processor(images=A__ , return_tensors='''tf''' ) # forward pass snake_case = model(**A__ ) # verify the logits snake_case = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , A__ ) snake_case = tf.constant([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ) tf.debugging.assert_near(outputs.logits[0, :3] , A__ , atol=1e-4 )
44
0
'''simple docstring''' from __future__ import annotations def __UpperCamelCase ( a : List[Any] , a : Optional[Any] , a : Dict , a : Tuple ) ->Optional[int]: if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and array[indexa] < array[indexa] ): snake_case , snake_case = array[indexa], array[indexa] def __UpperCamelCase ( a : Optional[int] , a : Optional[Any] , a : Union[str, Any] , a : Tuple ) ->List[str]: if length > 1: snake_case = int(length / 2 ) for i in range(a , low + middle ): comp_and_swap(a , a , i + middle , a ) bitonic_merge(a , a , a , a ) bitonic_merge(a , low + middle , a , a ) def __UpperCamelCase ( a : Optional[Any] , a : Any , a : Optional[int] , a : Dict ) ->int: if length > 1: snake_case = int(length / 2 ) bitonic_sort(a , a , a , 1 ) bitonic_sort(a , low + middle , a , 0 ) bitonic_merge(a , a , a , a ) if __name__ == "__main__": _lowercase = input('Enter numbers separated by a comma:\n').strip() _lowercase = [int(item.strip()) for item in user_input.split(',')] bitonic_sort(unsorted, 0, len(unsorted), 1) print('\nSorted array in ascending order is: ', end='') print(*unsorted, sep=', ') bitonic_merge(unsorted, 0, len(unsorted), 0) print('Sorted array in descending order is: ', end='') print(*unsorted, sep=', ')
708
'''simple docstring''' import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from datasets.iterable_dataset import IterableDataset from datasets.load import dataset_module_factory, import_main_class from datasets.utils.file_utils import cached_path _lowercase = [ {'dataset': 'wikipedia', 'config_name': '20220301.de'}, {'dataset': 'wikipedia', 'config_name': '20220301.en'}, {'dataset': 'wikipedia', 'config_name': '20220301.fr'}, {'dataset': 'wikipedia', 'config_name': '20220301.frr'}, {'dataset': 'wikipedia', 'config_name': '20220301.it'}, {'dataset': 'wikipedia', 'config_name': '20220301.simple'}, {'dataset': 'snli', 'config_name': 'plain_text'}, {'dataset': 'eli5', 'config_name': 'LFQA_reddit'}, {'dataset': 'wiki40b', 'config_name': 'en'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.compressed'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.no_index'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.multiset.no_index'}, {'dataset': 'natural_questions', 'config_name': 'default'}, ] def __UpperCamelCase ( a : Dict=True ) ->str: if with_config: return [ { "testcase_name": d["dataset"] + "/" + d["config_name"], "dataset": d["dataset"], "config_name": d["config_name"], } for d in DATASETS_ON_HF_GCP ] else: return [ {"testcase_name": dataset, "dataset": dataset} for dataset in {d["dataset"] for d in DATASETS_ON_HF_GCP} ] @parameterized.named_parameters(list_datasets_on_hf_gcp_parameters(with_config=__a ) ) class _lowercase ( __a ): _UpperCAmelCase = None _UpperCAmelCase = None def UpperCamelCase ( self , A__ , A__ ) -> str: with TemporaryDirectory() as tmp_dir: snake_case = dataset_module_factory(A__ , cache_dir=A__ ) snake_case = import_main_class(dataset_module.module_path , dataset=A__ ) snake_case = builder_cls( cache_dir=A__ , config_name=A__ , hash=dataset_module.hash , ) snake_case = '''/'''.join( [ HF_GCP_BASE_URL, builder_instance._relative_data_dir(with_hash=A__ ).replace(os.sep , '''/''' ), config.DATASET_INFO_FILENAME, ] ) snake_case = cached_path(A__ , cache_dir=A__ ) self.assertTrue(os.path.exists(A__ ) ) @pytest.mark.integration def __UpperCamelCase ( a : List[str] ) ->Any: snake_case = tmp_path_factory.mktemp('''test_hf_gcp''' ) / '''test_wikipedia_simple''' snake_case = dataset_module_factory('''wikipedia''' , cache_dir=a ) snake_case = import_main_class(dataset_module.module_path ) snake_case = builder_cls( cache_dir=a , config_name='''20220301.frr''' , hash=dataset_module.hash , ) # use the HF cloud storage, not the original download_and_prepare that uses apache-beam snake_case = None builder_instance.download_and_prepare() snake_case = builder_instance.as_dataset() assert ds @pytest.mark.integration def __UpperCamelCase ( a : Any ) ->Union[str, Any]: snake_case = dataset_module_factory('''wikipedia''' , cache_dir=a ) snake_case = import_main_class(dataset_module.module_path , dataset=a ) snake_case = builder_cls( cache_dir=a , config_name='''20220301.frr''' , hash=dataset_module.hash , ) snake_case = builder_instance.as_streaming_dataset() assert ds assert isinstance(a , a ) assert "train" in ds assert isinstance(ds['''train'''] , a ) assert next(iter(ds['''train'''] ) )
44
0
'''simple docstring''' from __future__ import annotations def __UpperCamelCase ( a : Tuple , a : Union[str, Any] ) ->Tuple: if b == 0: return (1, 0) (snake_case) = extended_euclid(snake_case_ , a % b ) snake_case = a // b return (y, x - k * y) def __UpperCamelCase ( a : Dict , a : Any , a : List[Any] , a : Tuple ) ->Tuple: (snake_case) = extended_euclid(snake_case_ , snake_case_ ) snake_case = na * na snake_case = ra * x * na + ra * y * na return (n % m + m) % m def __UpperCamelCase ( a : Tuple , a : int ) ->int: (snake_case) = extended_euclid(snake_case_ , snake_case_ ) if b < 0: snake_case = (b % n + n) % n return b def __UpperCamelCase ( a : Union[str, Any] , a : Optional[int] , a : Any , a : int ) ->List[str]: snake_case = invert_modulo(snake_case_ , snake_case_ ), invert_modulo(snake_case_ , snake_case_ ) snake_case = na * na snake_case = ra * x * na + ra * y * na return (n % m + m) % m if __name__ == "__main__": from doctest import testmod testmod(name='chinese_remainder_theorem', verbose=True) testmod(name='chinese_remainder_theorem2', verbose=True) testmod(name='invert_modulo', verbose=True) testmod(name='extended_euclid', verbose=True)
709
'''simple docstring''' def __UpperCamelCase ( a : int , a : int ) ->int: while b: snake_case , snake_case = b, a % b return a def __UpperCamelCase ( a : int , a : int ) ->int: return a if b == 0 else euclidean_gcd_recursive(a , a % b ) def __UpperCamelCase ( ) ->Optional[Any]: print(f"""euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}""" ) print(f"""euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}""" ) print(f"""euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}""" ) print(f"""euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}""" ) print(f"""euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}""" ) print(f"""euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}""" ) print(f"""euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}""" ) print(f"""euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}""" ) if __name__ == "__main__": main()
44
0
'''simple docstring''' import argparse import json import logging import os import shutil import sys import tempfile import unittest from unittest import mock import torch from accelerate.utils import write_basic_config from transformers.testing_utils import TestCasePlus, get_gpu_count, run_command, slow, torch_device from transformers.utils import is_apex_available logging.basicConfig(level=logging.DEBUG) _lowercase = logging.getLogger() def __UpperCamelCase ( ) ->List[Any]: snake_case = argparse.ArgumentParser() parser.add_argument('''-f''' ) snake_case = parser.parse_args() return args.f def __UpperCamelCase ( a : Any ) ->Optional[Any]: snake_case = {} snake_case = os.path.join(__UpperCamelCase , '''all_results.json''' ) if os.path.exists(__UpperCamelCase ): with open(__UpperCamelCase , '''r''' ) as f: snake_case = json.load(__UpperCamelCase ) else: raise ValueError(f"""can't find {path}""" ) return results def __UpperCamelCase ( ) ->Dict: snake_case = torch.cuda.is_available() and torch_device == '''cuda''' return is_using_cuda and is_apex_available() _lowercase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class _lowercase ( SCREAMING_SNAKE_CASE_ ): @classmethod def UpperCamelCase ( cls ) -> List[str]: # Write Accelerate config, will pick up on CPU, GPU, and multi-GPU snake_case = tempfile.mkdtemp() snake_case = os.path.join(cls.tmpdir , '''default_config.yml''' ) write_basic_config(save_location=cls.configPath ) snake_case = ['''accelerate''', '''launch''', '''--config_file''', cls.configPath] @classmethod def UpperCamelCase ( cls ) -> List[Any]: shutil.rmtree(cls.tmpdir ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def UpperCamelCase ( self ) -> List[str]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F"""\n {self.examples_dir}/pytorch/text-classification/run_glue_no_trainer.py\n --model_name_or_path distilbert-base-uncased\n --output_dir {tmp_dir}\n --train_file ./tests/fixtures/tests_samples/MRPC/train.csv\n --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --learning_rate=1e-4\n --seed=42\n --checkpointing_steps epoch\n --with_tracking\n """.split() if is_cuda_and_apex_available(): testargs.append('''--fp16''' ) run_command(self._launch_args + testargs ) snake_case = get_results(UpperCamelCase__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.7_5 ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''glue_no_trainer''' ) ) ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def UpperCamelCase ( self ) -> Dict: snake_case = self.get_auto_remove_tmp_dir() snake_case = F"""\n {self.examples_dir}/pytorch/language-modeling/run_clm_no_trainer.py\n --model_name_or_path distilgpt2\n --train_file ./tests/fixtures/sample_text.txt\n --validation_file ./tests/fixtures/sample_text.txt\n --block_size 128\n --per_device_train_batch_size 5\n --per_device_eval_batch_size 5\n --num_train_epochs 2\n --output_dir {tmp_dir}\n --checkpointing_steps epoch\n --with_tracking\n """.split() if torch.cuda.device_count() > 1: # Skipping because there are not enough batches to train the model + would need a drop_last to work. return run_command(self._launch_args + testargs ) snake_case = get_results(UpperCamelCase__ ) self.assertLess(result['''perplexity'''] , 1_00 ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''clm_no_trainer''' ) ) ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def UpperCamelCase ( self ) -> int: snake_case = self.get_auto_remove_tmp_dir() snake_case = F"""\n {self.examples_dir}/pytorch/language-modeling/run_mlm_no_trainer.py\n --model_name_or_path distilroberta-base\n --train_file ./tests/fixtures/sample_text.txt\n --validation_file ./tests/fixtures/sample_text.txt\n --output_dir {tmp_dir}\n --num_train_epochs=1\n --checkpointing_steps epoch\n --with_tracking\n """.split() run_command(self._launch_args + testargs ) snake_case = get_results(UpperCamelCase__ ) self.assertLess(result['''perplexity'''] , 42 ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''mlm_no_trainer''' ) ) ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def UpperCamelCase ( self ) -> Optional[int]: # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu snake_case = 7 if get_gpu_count() > 1 else 2 snake_case = self.get_auto_remove_tmp_dir() snake_case = F"""\n {self.examples_dir}/pytorch/token-classification/run_ner_no_trainer.py\n --model_name_or_path bert-base-uncased\n --train_file tests/fixtures/tests_samples/conll/sample.json\n --validation_file tests/fixtures/tests_samples/conll/sample.json\n --output_dir {tmp_dir}\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=2\n --num_train_epochs={epochs}\n --seed 7\n --checkpointing_steps epoch\n --with_tracking\n """.split() run_command(self._launch_args + testargs ) snake_case = get_results(UpperCamelCase__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.7_5 ) self.assertLess(result['''train_loss'''] , 0.5 ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''ner_no_trainer''' ) ) ) @unittest.skip(reason='''Fix me @muellerzr''' ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def UpperCamelCase ( self ) -> List[Any]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F"""\n {self.examples_dir}/pytorch/question-answering/run_qa_no_trainer.py\n --model_name_or_path bert-base-uncased\n --version_2_with_negative\n --train_file tests/fixtures/tests_samples/SQUAD/sample.json\n --validation_file tests/fixtures/tests_samples/SQUAD/sample.json\n --output_dir {tmp_dir}\n --seed=42\n --max_train_steps=10\n --num_warmup_steps=2\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --checkpointing_steps epoch\n --with_tracking\n """.split() run_command(self._launch_args + testargs ) snake_case = get_results(UpperCamelCase__ ) # Because we use --version_2_with_negative the testing script uses SQuAD v2 metrics. self.assertGreaterEqual(result['''eval_f1'''] , 28 ) self.assertGreaterEqual(result['''eval_exact'''] , 28 ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''qa_no_trainer''' ) ) ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def UpperCamelCase ( self ) -> Optional[int]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F"""\n {self.examples_dir}/pytorch/multiple-choice/run_swag_no_trainer.py\n --model_name_or_path bert-base-uncased\n --train_file tests/fixtures/tests_samples/swag/sample.json\n --validation_file tests/fixtures/tests_samples/swag/sample.json\n --output_dir {tmp_dir}\n --max_train_steps=20\n --num_warmup_steps=2\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --with_tracking\n """.split() run_command(self._launch_args + testargs ) snake_case = get_results(UpperCamelCase__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.8 ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''swag_no_trainer''' ) ) ) @slow @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def UpperCamelCase ( self ) -> Any: snake_case = self.get_auto_remove_tmp_dir() snake_case = F"""\n {self.examples_dir}/pytorch/summarization/run_summarization_no_trainer.py\n --model_name_or_path t5-small\n --train_file tests/fixtures/tests_samples/xsum/sample.json\n --validation_file tests/fixtures/tests_samples/xsum/sample.json\n --output_dir {tmp_dir}\n --max_train_steps=50\n --num_warmup_steps=8\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --checkpointing_steps epoch\n --with_tracking\n """.split() run_command(self._launch_args + testargs ) snake_case = get_results(UpperCamelCase__ ) self.assertGreaterEqual(result['''eval_rouge1'''] , 10 ) self.assertGreaterEqual(result['''eval_rouge2'''] , 2 ) self.assertGreaterEqual(result['''eval_rougeL'''] , 7 ) self.assertGreaterEqual(result['''eval_rougeLsum'''] , 7 ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''summarization_no_trainer''' ) ) ) @slow @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def UpperCamelCase ( self ) -> Tuple: snake_case = self.get_auto_remove_tmp_dir() snake_case = F"""\n {self.examples_dir}/pytorch/translation/run_translation_no_trainer.py\n --model_name_or_path sshleifer/student_marian_en_ro_6_1\n --source_lang en\n --target_lang ro\n --train_file tests/fixtures/tests_samples/wmt16/sample.json\n --validation_file tests/fixtures/tests_samples/wmt16/sample.json\n --output_dir {tmp_dir}\n --max_train_steps=50\n --num_warmup_steps=8\n --num_beams=6\n --learning_rate=3e-3\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --source_lang en_XX\n --target_lang ro_RO\n --checkpointing_steps epoch\n --with_tracking\n """.split() run_command(self._launch_args + testargs ) snake_case = get_results(UpperCamelCase__ ) self.assertGreaterEqual(result['''eval_bleu'''] , 30 ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''epoch_0''' ) ) ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''translation_no_trainer''' ) ) ) @slow def UpperCamelCase ( self ) -> List[str]: snake_case = logging.StreamHandler(sys.stdout ) logger.addHandler(UpperCamelCase__ ) snake_case = self.get_auto_remove_tmp_dir() snake_case = F"""\n {self.examples_dir}/pytorch/semantic-segmentation/run_semantic_segmentation_no_trainer.py\n --dataset_name huggingface/semantic-segmentation-test-sample\n --output_dir {tmp_dir}\n --max_train_steps=10\n --num_warmup_steps=2\n --learning_rate=2e-4\n --per_device_train_batch_size=2\n --per_device_eval_batch_size=1\n --checkpointing_steps epoch\n """.split() run_command(self._launch_args + testargs ) snake_case = get_results(UpperCamelCase__ ) self.assertGreaterEqual(result['''eval_overall_accuracy'''] , 0.1_0 ) @mock.patch.dict(os.environ , {'''WANDB_MODE''': '''offline'''} ) def UpperCamelCase ( self ) -> List[Any]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F"""\n {self.examples_dir}/pytorch/image-classification/run_image_classification_no_trainer.py\n --model_name_or_path google/vit-base-patch16-224-in21k\n --dataset_name hf-internal-testing/cats_vs_dogs_sample\n --learning_rate 1e-4\n --per_device_train_batch_size 2\n --per_device_eval_batch_size 1\n --max_train_steps 2\n --train_val_split 0.1\n --seed 42\n --output_dir {tmp_dir}\n --with_tracking\n --checkpointing_steps 1\n """.split() if is_cuda_and_apex_available(): testargs.append('''--fp16''' ) run_command(self._launch_args + testargs ) snake_case = get_results(UpperCamelCase__ ) # The base model scores a 25% self.assertGreaterEqual(result['''eval_accuracy'''] , 0.6 ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''step_1''' ) ) ) self.assertTrue(os.path.exists(os.path.join(UpperCamelCase__ , '''image_classification_no_trainer''' ) ) )
710
'''simple docstring''' import argparse import copy def __UpperCamelCase ( a : Union[str, Any] ) ->Tuple: snake_case = {} with open(a ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: snake_case = [] _list.append([line.split()[1], line.split()[2]] ) snake_case = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: snake_case = [] _list.append([line.split()[0], line.split()[2]] ) snake_case = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def __UpperCamelCase ( a : Dict , a : Tuple ) ->int: with open(a ) as f: snake_case = f.read(1 ) snake_case = start_node snake_case = [] snake_case = start_node snake_case = 0 while visiting not in first_solution: snake_case = 1_0000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(a ) and k[0] not in first_solution: snake_case = k[1] snake_case = k[0] first_solution.append(a ) snake_case = distance_of_first_solution + int(a ) snake_case = best_node first_solution.append(a ) snake_case = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 snake_case = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_0000 ) return first_solution, distance_of_first_solution def __UpperCamelCase ( a : Optional[int] , a : str ) ->str: snake_case = [] for n in solution[1:-1]: snake_case = solution.index(a ) for kn in solution[1:-1]: snake_case = solution.index(a ) if n == kn: continue snake_case = copy.deepcopy(a ) snake_case = kn snake_case = n snake_case = 0 for k in _tmp[:-1]: snake_case = _tmp[_tmp.index(a ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: snake_case = distance + int(i[1] ) _tmp.append(a ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) snake_case = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda a : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def __UpperCamelCase ( a : Any , a : Optional[Any] , a : int , a : Optional[int] , a : Union[str, Any] ) ->List[Any]: snake_case = 1 snake_case = first_solution snake_case = [] snake_case = distance_of_first_solution snake_case = solution while count <= iters: snake_case = find_neighborhood(a , a ) snake_case = 0 snake_case = neighborhood[index_of_best_solution] snake_case = len(a ) - 1 snake_case = False while not found: snake_case = 0 while i < len(a ): if best_solution[i] != solution[i]: snake_case = best_solution[i] snake_case = solution[i] break snake_case = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) snake_case = True snake_case = best_solution[:-1] snake_case = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: snake_case = cost snake_case = solution else: snake_case = index_of_best_solution + 1 snake_case = neighborhood[index_of_best_solution] if len(a ) >= size: tabu_list.pop(0 ) snake_case = count + 1 return best_solution_ever, best_cost def __UpperCamelCase ( a : Union[str, Any]=None ) ->Optional[Any]: snake_case = generate_neighbours(args.File ) snake_case , snake_case = generate_first_solution( args.File , a ) snake_case , snake_case = tabu_search( a , a , a , args.Iterations , args.Size , ) print(f"""Best solution: {best_sol}, with total distance: {best_cost}.""" ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser(description='Tabu Search') parser.add_argument( '-f', '--File', type=str, help='Path to the file containing the data', required=True, ) parser.add_argument( '-i', '--Iterations', type=int, help='How many iterations the algorithm should perform', required=True, ) parser.add_argument( '-s', '--Size', type=int, help='Size of the tabu list', required=True ) # Pass the arguments to main method main(parser.parse_args())
44
0
'''simple docstring''' from ...processing_utils import ProcessorMixin class _lowercase ( lowerCAmelCase__ ): _UpperCAmelCase = "SpeechT5FeatureExtractor" _UpperCAmelCase = "SpeechT5Tokenizer" def __init__( self , A__ , A__ ) -> Tuple: super().__init__(_lowerCamelCase , _lowerCamelCase ) def __call__( self , *A__ , **A__ ) -> str: snake_case = kwargs.pop('''audio''' , _lowerCamelCase ) snake_case = kwargs.pop('''text''' , _lowerCamelCase ) snake_case = kwargs.pop('''text_target''' , _lowerCamelCase ) snake_case = kwargs.pop('''audio_target''' , _lowerCamelCase ) snake_case = kwargs.pop('''sampling_rate''' , _lowerCamelCase ) if audio is not None and text is not None: raise ValueError( '''Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?''' ) if audio_target is not None and text_target is not None: raise ValueError( '''Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?''' ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( '''You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process.''' ) if audio is not None: snake_case = self.feature_extractor(_lowerCamelCase , *_lowerCamelCase , sampling_rate=_lowerCamelCase , **_lowerCamelCase ) elif text is not None: snake_case = self.tokenizer(_lowerCamelCase , **_lowerCamelCase ) else: snake_case = None if audio_target is not None: snake_case = self.feature_extractor(audio_target=_lowerCamelCase , *_lowerCamelCase , sampling_rate=_lowerCamelCase , **_lowerCamelCase ) snake_case = targets['''input_values'''] elif text_target is not None: snake_case = self.tokenizer(_lowerCamelCase , **_lowerCamelCase ) snake_case = targets['''input_ids'''] else: snake_case = None if inputs is None: return targets if targets is not None: snake_case = labels snake_case = targets.get('''attention_mask''' ) if decoder_attention_mask is not None: snake_case = decoder_attention_mask return inputs def UpperCamelCase ( self , *A__ , **A__ ) -> List[Any]: snake_case = kwargs.pop('''input_values''' , _lowerCamelCase ) snake_case = kwargs.pop('''input_ids''' , _lowerCamelCase ) snake_case = kwargs.pop('''labels''' , _lowerCamelCase ) if input_values is not None and input_ids is not None: raise ValueError('''Cannot process both `input_values` and `input_ids` inputs.''' ) if input_values is None and input_ids is None and labels is None: raise ValueError( '''You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded.''' ) if input_values is not None: snake_case = self.feature_extractor.pad(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) elif input_ids is not None: snake_case = self.tokenizer.pad(_lowerCamelCase , **_lowerCamelCase ) else: snake_case = None if labels is not None: if "input_ids" in labels or (isinstance(_lowerCamelCase , _lowerCamelCase ) and "input_ids" in labels[0]): snake_case = self.tokenizer.pad(_lowerCamelCase , **_lowerCamelCase ) snake_case = targets['''input_ids'''] else: snake_case = self.feature_extractor.feature_size snake_case = self.feature_extractor.num_mel_bins snake_case = self.feature_extractor.pad(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) snake_case = feature_size_hack snake_case = targets['''input_values'''] else: snake_case = None if inputs is None: return targets if targets is not None: snake_case = labels snake_case = targets.get('''attention_mask''' ) if decoder_attention_mask is not None: snake_case = decoder_attention_mask return inputs def UpperCamelCase ( self , *A__ , **A__ ) -> int: return self.tokenizer.batch_decode(*_lowerCamelCase , **_lowerCamelCase ) def UpperCamelCase ( self , *A__ , **A__ ) -> Tuple: return self.tokenizer.decode(*_lowerCamelCase , **_lowerCamelCase )
711
'''simple docstring''' from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .notes_encoder import SpectrogramNotesEncoder from .continous_encoder import SpectrogramContEncoder from .pipeline_spectrogram_diffusion import ( SpectrogramContEncoder, SpectrogramDiffusionPipeline, TaFilmDecoder, ) try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .midi_utils import MidiProcessor
44
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { "facebook/vit-mae-base": "https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json", # See all ViT MAE models at https://huggingface.co/models?filter=vit-mae } class _lowercase ( _A ): _UpperCAmelCase = '''vit_mae''' def __init__( self , A__=7_68 , A__=12 , A__=12 , A__=30_72 , A__="gelu" , A__=0.0 , A__=0.0 , A__=0.0_2 , A__=1e-12 , A__=2_24 , A__=16 , A__=3 , A__=True , A__=16 , A__=5_12 , A__=8 , A__=20_48 , A__=0.7_5 , A__=False , **A__ , ) -> int: super().__init__(**A__ ) snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_size snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = initializer_range snake_case = layer_norm_eps snake_case = image_size snake_case = patch_size snake_case = num_channels snake_case = qkv_bias snake_case = decoder_num_attention_heads snake_case = decoder_hidden_size snake_case = decoder_num_hidden_layers snake_case = decoder_intermediate_size snake_case = mask_ratio snake_case = norm_pix_loss
712
'''simple docstring''' from ...processing_utils import ProcessorMixin class _lowercase ( __a ): _UpperCAmelCase = '''WhisperFeatureExtractor''' _UpperCAmelCase = '''WhisperTokenizer''' def __init__( self , A__ , A__ ) -> Optional[Any]: super().__init__(A__ , A__ ) snake_case = self.feature_extractor snake_case = False def UpperCamelCase ( self , A__=None , A__=None , A__=True ) -> Union[str, Any]: return self.tokenizer.get_decoder_prompt_ids(task=A__ , language=A__ , no_timestamps=A__ ) def __call__( self , *A__ , **A__ ) -> Dict: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*A__ , **A__ ) snake_case = kwargs.pop('''audio''' , A__ ) snake_case = kwargs.pop('''sampling_rate''' , A__ ) snake_case = kwargs.pop('''text''' , A__ ) if len(A__ ) > 0: snake_case = args[0] snake_case = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if audio is not None: snake_case = self.feature_extractor(A__ , *A__ , sampling_rate=A__ , **A__ ) if text is not None: snake_case = self.tokenizer(A__ , **A__ ) if text is None: return inputs elif audio is None: return encodings else: snake_case = encodings['''input_ids'''] return inputs def UpperCamelCase ( self , *A__ , **A__ ) -> Optional[Any]: return self.tokenizer.batch_decode(*A__ , **A__ ) def UpperCamelCase ( self , *A__ , **A__ ) -> str: return self.tokenizer.decode(*A__ , **A__ ) def UpperCamelCase ( self , A__ , A__="np" ) -> Optional[Any]: return self.tokenizer.get_prompt_ids(A__ , return_tensors=A__ )
44
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowercase = { 'configuration_time_series_transformer': [ 'TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TimeSeriesTransformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ 'TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TimeSeriesTransformerForPrediction', 'TimeSeriesTransformerModel', 'TimeSeriesTransformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
713
'''simple docstring''' import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class _lowercase ( __a ): _UpperCAmelCase = '''char''' _UpperCAmelCase = '''bpe''' _UpperCAmelCase = '''wp''' _lowercase = (DecodeType.CHARACTER, DecodeType.BPE, DecodeType.WORDPIECE) class _lowercase ( __a ): _UpperCAmelCase = ['''image_processor''', '''char_tokenizer'''] _UpperCAmelCase = '''ViTImageProcessor''' _UpperCAmelCase = '''MgpstrTokenizer''' def __init__( self , A__=None , A__=None , **A__ ) -> List[Any]: snake_case = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , A__ , ) snake_case = kwargs.pop('''feature_extractor''' ) snake_case = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) snake_case = tokenizer snake_case = AutoTokenizer.from_pretrained('''gpt2''' ) snake_case = AutoTokenizer.from_pretrained('''bert-base-uncased''' ) super().__init__(A__ , A__ ) def __call__( self , A__=None , A__=None , A__=None , **A__ ) -> List[str]: if images is None and text is None: raise ValueError('''You need to specify either an `images` or `text` input to process.''' ) if images is not None: snake_case = self.image_processor(A__ , return_tensors=A__ , **A__ ) if text is not None: snake_case = self.char_tokenizer(A__ , return_tensors=A__ , **A__ ) if text is None: return inputs elif images is None: return encodings else: snake_case = encodings['''input_ids'''] return inputs def UpperCamelCase ( self , A__ ) -> Dict: snake_case , snake_case , snake_case = sequences snake_case = char_preds.size(0 ) snake_case , snake_case = self._decode_helper(A__ , '''char''' ) snake_case , snake_case = self._decode_helper(A__ , '''bpe''' ) snake_case , snake_case = self._decode_helper(A__ , '''wp''' ) snake_case = [] snake_case = [] for i in range(A__ ): snake_case = [char_scores[i], bpe_scores[i], wp_scores[i]] snake_case = [char_strs[i], bpe_strs[i], wp_strs[i]] snake_case = scores.index(max(A__ ) ) final_strs.append(strs[max_score_index] ) final_scores.append(scores[max_score_index] ) snake_case = {} snake_case = final_strs snake_case = final_scores snake_case = char_strs snake_case = bpe_strs snake_case = wp_strs return out def UpperCamelCase ( self , A__ , A__ ) -> Optional[Any]: if format == DecodeType.CHARACTER: snake_case = self.char_decode snake_case = 1 snake_case = '''[s]''' elif format == DecodeType.BPE: snake_case = self.bpe_decode snake_case = 2 snake_case = '''#''' elif format == DecodeType.WORDPIECE: snake_case = self.wp_decode snake_case = 1_02 snake_case = '''[SEP]''' else: raise ValueError(F"""Format {format} is not supported.""" ) snake_case , snake_case = [], [] snake_case = pred_logits.size(0 ) snake_case = pred_logits.size(1 ) snake_case , snake_case = pred_logits.topk(1 , dim=-1 , largest=A__ , sorted=A__ ) snake_case = preds_index.view(-1 , A__ )[:, 1:] snake_case = decoder(A__ ) snake_case , snake_case = torch.nn.functional.softmax(A__ , dim=2 ).max(dim=2 ) snake_case = preds_max_prob[:, 1:] for index in range(A__ ): snake_case = preds_str[index].find(A__ ) snake_case = preds_str[index][:pred_eos] snake_case = preds_index[index].cpu().tolist() snake_case = pred_index.index(A__ ) if eos_token in pred_index else -1 snake_case = preds_max_prob[index][: pred_eos_index + 1] snake_case = pred_max_prob.cumprod(dim=0 )[-1] if pred_max_prob.nelement() != 0 else 0.0 dec_strs.append(A__ ) conf_scores.append(A__ ) return dec_strs, conf_scores def UpperCamelCase ( self , A__ ) -> int: snake_case = [seq.replace(''' ''' , '''''' ) for seq in self.char_tokenizer.batch_decode(A__ )] return decode_strs def UpperCamelCase ( self , A__ ) -> List[str]: return self.bpe_tokenizer.batch_decode(A__ ) def UpperCamelCase ( self , A__ ) -> Union[str, Any]: snake_case = [seq.replace(''' ''' , '''''' ) for seq in self.wp_tokenizer.batch_decode(A__ )] return decode_strs
44
0
'''simple docstring''' from __future__ import annotations from collections.abc import Callable from typing import Any, Generic, TypeVar _lowercase = TypeVar('T') class _lowercase ( Generic[T] ): def __init__( self , A__ , A__ ) -> None: snake_case = None snake_case = len(__A ) snake_case = [any_type for _ in range(self.N )] + arr snake_case = fnc self.build() def UpperCamelCase ( self ) -> None: for p in range(self.N - 1 , 0 , -1 ): snake_case = self.fn(self.st[p * 2] , self.st[p * 2 + 1] ) def UpperCamelCase ( self , A__ , A__ ) -> None: p += self.N snake_case = v while p > 1: snake_case = p // 2 snake_case = self.fn(self.st[p * 2] , self.st[p * 2 + 1] ) def UpperCamelCase ( self , A__ , A__ ) -> T | None: # noqa: E741 snake_case , snake_case = l + self.N, r + self.N snake_case = None while l <= r: if l % 2 == 1: snake_case = self.st[l] if res is None else self.fn(__A , self.st[l] ) if r % 2 == 0: snake_case = self.st[r] if res is None else self.fn(__A , self.st[r] ) snake_case , snake_case = (l + 1) // 2, (r - 1) // 2 return res if __name__ == "__main__": from functools import reduce _lowercase = [1, 10, -2, 9, -3, 8, 4, -7, 5, 6, 11, -12] _lowercase = { 0: 7, 1: 2, 2: 6, 3: -14, 4: 5, 5: 4, 6: 7, 7: -10, 8: 9, 9: 10, 10: 12, 11: 1, } _lowercase = SegmentTree(test_array, min) _lowercase = SegmentTree(test_array, max) _lowercase = SegmentTree(test_array, lambda a, b: a + b) def __UpperCamelCase ( ) ->str: for i in range(len(UpperCamelCase__ ) ): for j in range(UpperCamelCase__ , len(UpperCamelCase__ ) ): snake_case = reduce(UpperCamelCase__ , test_array[i : j + 1] ) snake_case = reduce(UpperCamelCase__ , test_array[i : j + 1] ) snake_case = reduce(lambda a , a : a + b , test_array[i : j + 1] ) assert min_range == min_segment_tree.query(UpperCamelCase__ , UpperCamelCase__ ) assert max_range == max_segment_tree.query(UpperCamelCase__ , UpperCamelCase__ ) assert sum_range == sum_segment_tree.query(UpperCamelCase__ , UpperCamelCase__ ) test_all_segments() for index, value in test_updates.items(): _lowercase = value min_segment_tree.update(index, value) max_segment_tree.update(index, value) sum_segment_tree.update(index, value) test_all_segments()
714
'''simple docstring''' import os from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen, xsplitext from ..table import array_cast from ..utils.py_utils import no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: from .features import FeatureType _lowercase , _lowercase , _lowercase = False, False, False @dataclass class _lowercase : _UpperCAmelCase = None _UpperCAmelCase = True _UpperCAmelCase = True _UpperCAmelCase = None # Automatically constructed _UpperCAmelCase = "dict" _UpperCAmelCase = pa.struct({'''bytes''': pa.binary(), '''path''': pa.string()} ) _UpperCAmelCase = field(default='''Audio''' , init=__a , repr=__a ) def __call__( self ) -> Optional[Any]: return self.pa_type def UpperCamelCase ( self , A__ ) -> dict: try: import soundfile as sf # soundfile is a dependency of librosa, needed to decode audio files. except ImportError as err: raise ImportError('''To support encoding audio data, please install \'soundfile\'.''' ) from err if isinstance(A__ , A__ ): return {"bytes": None, "path": value} elif isinstance(A__ , A__ ): return {"bytes": value, "path": None} elif "array" in value: # convert the audio array to wav bytes snake_case = BytesIO() sf.write(A__ , value['''array'''] , value['''sampling_rate'''] , format='''wav''' ) return {"bytes": buffer.getvalue(), "path": None} elif value.get('''path''' ) is not None and os.path.isfile(value['''path'''] ): # we set "bytes": None to not duplicate the data if they're already available locally if value["path"].endswith('''pcm''' ): # "PCM" only has raw audio bytes if value.get('''sampling_rate''' ) is None: # At least, If you want to convert "PCM-byte" to "WAV-byte", you have to know sampling rate raise KeyError('''To use PCM files, please specify a \'sampling_rate\' in Audio object''' ) if value.get('''bytes''' ): # If we already had PCM-byte, we don`t have to make "read file, make bytes" (just use it!) snake_case = np.frombuffer(value['''bytes'''] , dtype=np.intaa ).astype(np.floataa ) / 3_27_67 else: snake_case = np.memmap(value['''path'''] , dtype='''h''' , mode='''r''' ).astype(np.floataa ) / 3_27_67 snake_case = BytesIO(bytes() ) sf.write(A__ , A__ , value['''sampling_rate'''] , format='''wav''' ) return {"bytes": buffer.getvalue(), "path": None} else: return {"bytes": None, "path": value.get('''path''' )} elif value.get('''bytes''' ) is not None or value.get('''path''' ) is not None: # store the audio bytes, and path is used to infer the audio format using the file extension return {"bytes": value.get('''bytes''' ), "path": value.get('''path''' )} else: raise ValueError( F"""An audio sample should have one of 'path' or 'bytes' but they are missing or None in {value}.""" ) def UpperCamelCase ( self , A__ , A__ = None ) -> dict: if not self.decode: raise RuntimeError('''Decoding is disabled for this feature. Please use Audio(decode=True) instead.''' ) snake_case , snake_case = (value['''path'''], BytesIO(value['''bytes'''] )) if value['''bytes'''] is not None else (value['''path'''], None) if path is None and file is None: raise ValueError(F"""An audio sample should have one of 'path' or 'bytes' but both are None in {value}.""" ) try: import librosa import soundfile as sf except ImportError as err: raise ImportError('''To support decoding audio files, please install \'librosa\' and \'soundfile\'.''' ) from err snake_case = xsplitext(A__ )[1][1:].lower() if path is not None else None if not config.IS_OPUS_SUPPORTED and audio_format == "opus": raise RuntimeError( '''Decoding \'opus\' files requires system library \'libsndfile\'>=1.0.31, ''' '''You can try to update `soundfile` python library: `pip install "soundfile>=0.12.1"`. ''' ) elif not config.IS_MP3_SUPPORTED and audio_format == "mp3": raise RuntimeError( '''Decoding \'mp3\' files requires system library \'libsndfile\'>=1.1.0, ''' '''You can try to update `soundfile` python library: `pip install "soundfile>=0.12.1"`. ''' ) if file is None: snake_case = token_per_repo_id or {} snake_case = path.split('''::''' )[-1] try: snake_case = string_to_dict(A__ , config.HUB_DATASETS_URL )['''repo_id'''] snake_case = token_per_repo_id[repo_id] except (ValueError, KeyError): snake_case = None with xopen(A__ , '''rb''' , use_auth_token=A__ ) as f: snake_case , snake_case = sf.read(A__ ) else: snake_case , snake_case = sf.read(A__ ) snake_case = array.T if self.mono: snake_case = librosa.to_mono(A__ ) if self.sampling_rate and self.sampling_rate != sampling_rate: snake_case = librosa.resample(A__ , orig_sr=A__ , target_sr=self.sampling_rate ) snake_case = self.sampling_rate return {"path": path, "array": array, "sampling_rate": sampling_rate} def UpperCamelCase ( self ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value if self.decode: raise ValueError('''Cannot flatten a decoded Audio feature.''' ) return { "bytes": Value('''binary''' ), "path": Value('''string''' ), } def UpperCamelCase ( self , A__ ) -> pa.StructArray: if pa.types.is_string(storage.type ): snake_case = pa.array([None] * len(A__ ) , type=pa.binary() ) snake_case = pa.StructArray.from_arrays([bytes_array, storage] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): snake_case = pa.array([None] * len(A__ ) , type=pa.string() ) snake_case = pa.StructArray.from_arrays([storage, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ) and storage.type.get_all_field_indices('''array''' ): snake_case = pa.array([Audio().encode_example(A__ ) if x is not None else None for x in storage.to_pylist()] ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index('''bytes''' ) >= 0: snake_case = storage.field('''bytes''' ) else: snake_case = pa.array([None] * len(A__ ) , type=pa.binary() ) if storage.type.get_field_index('''path''' ) >= 0: snake_case = storage.field('''path''' ) else: snake_case = pa.array([None] * len(A__ ) , type=pa.string() ) snake_case = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null() ) return array_cast(A__ , self.pa_type ) def UpperCamelCase ( self , A__ ) -> pa.StructArray: @no_op_if_value_is_null def path_to_bytes(A__ ): with xopen(A__ , '''rb''' ) as f: snake_case = f.read() return bytes_ snake_case = pa.array( [ (path_to_bytes(x['''path'''] ) if x['''bytes'''] is None else x['''bytes''']) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) snake_case = pa.array( [os.path.basename(A__ ) if path is not None else None for path in storage.field('''path''' ).to_pylist()] , type=pa.string() , ) snake_case = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=bytes_array.is_null() ) return array_cast(A__ , self.pa_type )
44
0
'''simple docstring''' import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint _lowercase = { '''169M''': 12, '''430M''': 24, '''1B5''': 24, '''3B''': 32, '''7B''': 32, '''14B''': 40, } _lowercase = { '''169M''': 768, '''430M''': 1_024, '''1B5''': 2_048, '''3B''': 2_560, '''7B''': 4_096, '''14B''': 5_120, } def __UpperCamelCase ( a : Tuple ) ->Any: snake_case = list(state_dict.keys() ) for name in state_dict_keys: snake_case = state_dict.pop(_UpperCAmelCase ) # emb -> embedding if name.startswith('''emb.''' ): snake_case = name.replace('''emb.''' , '''embeddings.''' ) # ln_0 -> pre_ln (only present at block 0) if name.startswith('''blocks.0.ln0''' ): snake_case = name.replace('''blocks.0.ln0''' , '''blocks.0.pre_ln''' ) # att -> attention snake_case = re.sub(R'''blocks\.(\d+)\.att''' , R'''blocks.\1.attention''' , _UpperCAmelCase ) # ffn -> feed_forward snake_case = re.sub(R'''blocks\.(\d+)\.ffn''' , R'''blocks.\1.feed_forward''' , _UpperCAmelCase ) # time_mix_k -> time_mix_key and reshape if name.endswith('''.time_mix_k''' ): snake_case = name.replace('''.time_mix_k''' , '''.time_mix_key''' ) # time_mix_v -> time_mix_value and reshape if name.endswith('''.time_mix_v''' ): snake_case = name.replace('''.time_mix_v''' , '''.time_mix_value''' ) # time_mix_r -> time_mix_key and reshape if name.endswith('''.time_mix_r''' ): snake_case = name.replace('''.time_mix_r''' , '''.time_mix_receptance''' ) if name != "head.weight": snake_case = '''rwkv.''' + name snake_case = weight return state_dict def __UpperCamelCase ( a : Optional[Any] , a : Optional[int] , a : Optional[Any] , a : Union[str, Any]=None , a : Optional[Any]=None , a : List[str]=False , a : int=None ) ->Union[str, Any]: # 1. If possible, build the tokenizer. if tokenizer_file is None: print('''No `--tokenizer_file` provided, we will use the default tokenizer.''' ) snake_case = 5_0277 snake_case = AutoTokenizer.from_pretrained('''EleutherAI/gpt-neox-20b''' ) else: snake_case = PreTrainedTokenizerFast(tokenizer_file=_UpperCAmelCase ) snake_case = len(_UpperCAmelCase ) tokenizer.save_pretrained(_UpperCAmelCase ) # 2. Build the config snake_case = list(NUM_HIDDEN_LAYERS_MAPPING.keys() ) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: snake_case = candidate break if size is None: raise ValueError('''Could not infer the size, please provide it with the `--size` argument.''' ) if size not in possible_sizes: raise ValueError(f"""`size` should be one of {possible_sizes}, got {size}.""" ) snake_case = RwkvConfig( vocab_size=_UpperCAmelCase , num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size] , hidden_size=HIDEN_SIZE_MAPPING[size] , ) config.save_pretrained(_UpperCAmelCase ) # 3. Download model file then convert state_dict snake_case = hf_hub_download(_UpperCAmelCase , _UpperCAmelCase ) snake_case = torch.load(_UpperCAmelCase , map_location='''cpu''' ) snake_case = convert_state_dict(_UpperCAmelCase ) # 4. Split in shards and save snake_case , snake_case = shard_checkpoint(_UpperCAmelCase ) for shard_file, shard in shards.items(): torch.save(_UpperCAmelCase , os.path.join(_UpperCAmelCase , _UpperCAmelCase ) ) if index is not None: snake_case = os.path.join(_UpperCAmelCase , _UpperCAmelCase ) # Save the index as well with open(_UpperCAmelCase , '''w''' , encoding='''utf-8''' ) as f: snake_case = json.dumps(_UpperCAmelCase , indent=2 , sort_keys=_UpperCAmelCase ) + '''\n''' f.write(_UpperCAmelCase ) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( '''Cleaning up shards. This may error with an OOM error, it this is the case don\'t worry you still have converted the model.''' ) snake_case = list(shards.keys() ) del state_dict del shards gc.collect() for shard_file in shard_files: snake_case = torch.load(os.path.join(_UpperCAmelCase , _UpperCAmelCase ) ) torch.save({k: v.cpu().clone() for k, v in state_dict.items()} , os.path.join(_UpperCAmelCase , _UpperCAmelCase ) ) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError('''Please provide a `model_name` to push the model to the Hub.''' ) snake_case = AutoModelForCausalLM.from_pretrained(_UpperCAmelCase ) model.push_to_hub(_UpperCAmelCase , max_shard_size='''2GB''' ) tokenizer.push_to_hub(_UpperCAmelCase ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--repo_id', default=None, type=str, required=True, help='Repo ID from which to pull the checkpoint.' ) parser.add_argument( '--checkpoint_file', default=None, type=str, required=True, help='Name of the checkpoint file in the repo.' ) parser.add_argument( '--output_dir', default=None, type=str, required=True, help='Where to save the converted model.' ) parser.add_argument( '--tokenizer_file', default=None, type=str, help='Path to the tokenizer file to use (if not provided, only the model is converted).', ) parser.add_argument( '--size', default=None, type=str, help='Size of the model. Will be inferred from the `checkpoint_file` if not passed.', ) parser.add_argument( '--push_to_hub', action='store_true', help='Push to the Hub the converted model.', ) parser.add_argument( '--model_name', default=None, type=str, help='Name of the pushed model on the Hub, including the username / organization.', ) _lowercase = parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
715
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class _lowercase : @staticmethod def UpperCamelCase ( *A__ , **A__ ) -> List[Any]: pass def __UpperCamelCase ( a : Image ) ->str: snake_case = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class _lowercase ( unittest.TestCase ): _UpperCAmelCase = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def UpperCamelCase ( self , A__ , A__ , A__ ) -> Union[str, Any]: snake_case = DepthEstimationPipeline(model=A__ , image_processor=A__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def UpperCamelCase ( self , A__ , A__ ) -> List[Any]: snake_case = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} , A__ ) import datasets snake_case = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' , '''image''' , split='''test''' ) snake_case = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] , A__ , ) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def UpperCamelCase ( self ) -> Optional[Any]: pass @slow @require_torch def UpperCamelCase ( self ) -> Dict: snake_case = '''Intel/dpt-large''' snake_case = pipeline('''depth-estimation''' , model=A__ ) snake_case = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) snake_case = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) , 2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) , 2.6_6_2 ) @require_torch def UpperCamelCase ( self ) -> Any: # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
44
0
'''simple docstring''' import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, get_image_size, is_torch_available, is_torch_tensor, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_torch_available(): import torch if is_vision_available(): import PIL _lowercase = logging.get_logger(__name__) def __UpperCamelCase ( a : np.ndarray , a : Union[int, Iterable[int]] , a : bool , a : int ) ->Tuple[int, int]: def constraint_to_multiple_of(a : Any , a : Optional[Any] , a : List[Any]=0 , a : Union[str, Any]=None ): snake_case = round(val / multiple ) * multiple if max_val is not None and x > max_val: snake_case = math.floor(val / multiple ) * multiple if x < min_val: snake_case = math.ceil(val / multiple ) * multiple return x snake_case = (output_size, output_size) if isinstance(a , a ) else output_size snake_case = get_image_size(a ) snake_case = output_size # determine new height and width snake_case = output_height / input_height snake_case = output_width / input_width if keep_aspect_ratio: # scale as little as possible if abs(1 - scale_width ) < abs(1 - scale_height ): # fit width snake_case = scale_width else: # fit height snake_case = scale_height snake_case = constraint_to_multiple_of(scale_height * input_height , multiple=a ) snake_case = constraint_to_multiple_of(scale_width * input_width , multiple=a ) return (new_height, new_width) class _lowercase ( UpperCamelCase_ ): _UpperCAmelCase = ['''pixel_values'''] def __init__( self , A__ = True , A__ = None , A__ = PILImageResampling.BILINEAR , A__ = False , A__ = 1 , A__ = True , A__ = 1 / 2_55 , A__ = True , A__ = None , A__ = None , **A__ , ) -> None: super().__init__(**__a ) snake_case = size if size is not None else {'height': 3_84, 'width': 3_84} snake_case = get_size_dict(__a ) snake_case = do_resize snake_case = size snake_case = keep_aspect_ratio snake_case = ensure_multiple_of snake_case = resample snake_case = do_rescale snake_case = rescale_factor snake_case = do_normalize snake_case = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN snake_case = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCamelCase ( self , A__ , A__ , A__ = False , A__ = 1 , A__ = PILImageResampling.BICUBIC , A__ = None , **A__ , ) -> np.ndarray: snake_case = get_size_dict(__a ) if "height" not in size or "width" not in size: raise ValueError(F"""The size dictionary must contain the keys 'height' and 'width'. Got {size.keys()}""" ) snake_case = get_resize_output_image_size( __a , output_size=(size['''height'''], size['''width''']) , keep_aspect_ratio=__a , multiple=__a , ) return resize(__a , size=__a , resample=__a , data_format=__a , **__a ) def UpperCamelCase ( self , A__ , A__ , A__ = None , **A__ , ) -> Union[str, Any]: return rescale(__a , scale=__a , data_format=__a , **__a ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ = None , **A__ , ) -> np.ndarray: return normalize(__a , mean=__a , std=__a , data_format=__a , **__a ) def UpperCamelCase ( self , A__ , A__ = None , A__ = None , A__ = None , A__ = None , A__ = None , A__ = None , A__ = None , A__ = None , A__ = None , A__ = None , A__ = None , A__ = ChannelDimension.FIRST , **A__ , ) -> PIL.Image.Image: snake_case = do_resize if do_resize is not None else self.do_resize snake_case = size if size is not None else self.size snake_case = get_size_dict(__a ) snake_case = keep_aspect_ratio if keep_aspect_ratio is not None else self.keep_aspect_ratio snake_case = ensure_multiple_of if ensure_multiple_of is not None else self.ensure_multiple_of snake_case = resample if resample is not None else self.resample snake_case = do_rescale if do_rescale is not None else self.do_rescale snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor snake_case = do_normalize if do_normalize is not None else self.do_normalize snake_case = image_mean if image_mean is not None else self.image_mean snake_case = image_std if image_std is not None else self.image_std snake_case = make_list_of_images(__a ) if not valid_images(__a ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. snake_case = [to_numpy_array(__a ) for image in images] if do_resize: snake_case = [self.resize(image=__a , size=__a , resample=__a ) for image in images] if do_rescale: snake_case = [self.rescale(image=__a , scale=__a ) for image in images] if do_normalize: snake_case = [self.normalize(image=__a , mean=__a , std=__a ) for image in images] snake_case = [to_channel_dimension_format(__a , __a ) for image in images] snake_case = {'pixel_values': images} return BatchFeature(data=__a , tensor_type=__a ) def UpperCamelCase ( self , A__ , A__ = None ) -> str: snake_case = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(__a ) != len(__a ): raise ValueError( '''Make sure that you pass in as many target sizes as the batch dimension of the logits''' ) if is_torch_tensor(__a ): snake_case = target_sizes.numpy() snake_case = [] for idx in range(len(__a ) ): snake_case = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) , size=target_sizes[idx] , mode='''bilinear''' , align_corners=__a ) snake_case = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(__a ) else: snake_case = logits.argmax(dim=1 ) snake_case = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
716
'''simple docstring''' import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def __UpperCamelCase ( a : Optional[int] ) ->Dict: snake_case = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(a , a ) def __UpperCamelCase ( a : Optional[Any] ) ->int: snake_case = list(s_dict.keys() ) for key in keys: if "transformer_layers" in key: snake_case = s_dict.pop(a ) elif "subsample" in key: snake_case = s_dict.pop(a ) def __UpperCamelCase ( a : Optional[int] ) ->Optional[int]: snake_case , snake_case = emb.weight.shape snake_case = nn.Linear(a , a , bias=a ) snake_case = emb.weight.data return lin_layer def __UpperCamelCase ( a : Any , a : Tuple ) ->Tuple: snake_case = torch.load(a , map_location='''cpu''' ) snake_case = mam_aaa['''args'''] snake_case = mam_aaa['''model'''] snake_case = state_dict['''decoder.output_projection.weight'''] remove_ignore_keys_(a ) rename_keys(a ) snake_case = state_dict['''decoder.embed_tokens.weight'''].shape[0] snake_case = args.share_decoder_input_output_embed snake_case = [int(a ) for i in args.conv_kernel_sizes.split(''',''' )] snake_case = SpeechaTextConfig( vocab_size=a , max_source_positions=args.max_source_positions , max_target_positions=args.max_target_positions , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='''relu''' , num_conv_layers=len(a ) , conv_channels=args.conv_channels , conv_kernel_sizes=a , input_feat_per_channel=args.input_feat_per_channel , input_channels=args.input_channels , tie_word_embeddings=a , num_beams=5 , max_length=200 , use_cache=a , decoder_start_token_id=2 , early_stopping=a , ) snake_case = SpeechaTextForConditionalGeneration(a ) snake_case , snake_case = model.model.load_state_dict(a , strict=a ) if len(a ) > 0 and not set(a ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( '''Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,''' f""" but all the following weights are missing {missing}""" ) if tie_embeds: snake_case = make_linear_from_emb(model.model.decoder.embed_tokens ) else: snake_case = lm_head_weights model.save_pretrained(a ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument('--fairseq_path', type=str, help='Path to the fairseq model (.pt) file.') parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') _lowercase = parser.parse_args() convert_fairseq_sat_checkpoint_to_tfms(args.fairseq_path, args.pytorch_dump_folder_path)
44
0
'''simple docstring''' import baseaa def __UpperCamelCase ( a : Dict ) ->List[Any]: return baseaa.aaaencode(string.encode('''utf-8''' ) ) def __UpperCamelCase ( a : Dict ) ->List[str]: return baseaa.aaadecode(_lowercase ).decode('''utf-8''' ) if __name__ == "__main__": import doctest doctest.testmod()
717
'''simple docstring''' from ..utils import DummyObject, requires_backends class _lowercase ( metaclass=__a ): _UpperCAmelCase = ['''transformers''', '''torch''', '''note_seq'''] def __init__( self , *A__ , **A__ ) -> Union[str, Any]: requires_backends(self , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ) -> Optional[Any]: requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ) -> Any: requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] )
44
0
'''simple docstring''' import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class _lowercase ( __snake_case ): _UpperCAmelCase = ['''image_processor''', '''tokenizer'''] _UpperCAmelCase = '''OwlViTImageProcessor''' _UpperCAmelCase = ('''CLIPTokenizer''', '''CLIPTokenizerFast''') def __init__( self , A__=None , A__=None , **A__ ) -> Tuple: snake_case = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _lowercase , ) snake_case = kwargs.pop('''feature_extractor''' ) snake_case = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_lowercase , _lowercase ) def __call__( self , A__=None , A__=None , A__=None , A__="max_length" , A__="np" , **A__ ) -> int: if text is None and query_images is None and images is None: raise ValueError( '''You have to specify at least one text or query image or image. All three cannot be none.''' ) if text is not None: if isinstance(_lowercase , _lowercase ) or (isinstance(_lowercase , _lowercase ) and not isinstance(text[0] , _lowercase )): snake_case = [self.tokenizer(_lowercase , padding=_lowercase , return_tensors=_lowercase , **_lowercase )] elif isinstance(_lowercase , _lowercase ) and isinstance(text[0] , _lowercase ): snake_case = [] # Maximum number of queries across batch snake_case = max([len(_lowercase ) for t in text] ) # Pad all batch samples to max number of text queries for t in text: if len(_lowercase ) != max_num_queries: snake_case = t + [""" """] * (max_num_queries - len(_lowercase )) snake_case = self.tokenizer(_lowercase , padding=_lowercase , return_tensors=_lowercase , **_lowercase ) encodings.append(_lowercase ) else: raise TypeError('''Input text should be a string, a list of strings or a nested list of strings''' ) if return_tensors == "np": snake_case = np.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0 ) snake_case = np.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0 ) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp snake_case = jnp.concatenate([encoding['''input_ids'''] for encoding in encodings] , axis=0 ) snake_case = jnp.concatenate([encoding['''attention_mask'''] for encoding in encodings] , axis=0 ) elif return_tensors == "pt" and is_torch_available(): import torch snake_case = torch.cat([encoding['''input_ids'''] for encoding in encodings] , dim=0 ) snake_case = torch.cat([encoding['''attention_mask'''] for encoding in encodings] , dim=0 ) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf snake_case = tf.stack([encoding['''input_ids'''] for encoding in encodings] , axis=0 ) snake_case = tf.stack([encoding['''attention_mask'''] for encoding in encodings] , axis=0 ) else: raise ValueError('''Target return tensor type could not be returned''' ) snake_case = BatchEncoding() snake_case = input_ids snake_case = attention_mask if query_images is not None: snake_case = BatchEncoding() snake_case = self.image_processor( _lowercase , return_tensors=_lowercase , **_lowercase ).pixel_values snake_case = query_pixel_values if images is not None: snake_case = self.image_processor(_lowercase , return_tensors=_lowercase , **_lowercase ) if text is not None and images is not None: snake_case = image_features.pixel_values return encoding elif query_images is not None and images is not None: snake_case = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**_lowercase ) , tensor_type=_lowercase ) def UpperCamelCase ( self , *A__ , **A__ ) -> int: return self.image_processor.post_process(*_lowercase , **_lowercase ) def UpperCamelCase ( self , *A__ , **A__ ) -> int: return self.image_processor.post_process_object_detection(*_lowercase , **_lowercase ) def UpperCamelCase ( self , *A__ , **A__ ) -> Any: return self.image_processor.post_process_image_guided_detection(*_lowercase , **_lowercase ) def UpperCamelCase ( self , *A__ , **A__ ) -> Any: return self.tokenizer.batch_decode(*_lowercase , **_lowercase ) def UpperCamelCase ( self , *A__ , **A__ ) -> Union[str, Any]: return self.tokenizer.decode(*_lowercase , **_lowercase ) @property def UpperCamelCase ( self ) -> str: warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , _lowercase , ) return self.image_processor_class @property def UpperCamelCase ( self ) -> Any: warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , _lowercase , ) return self.image_processor
718
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator class _lowercase : def __init__( self , A__ ) -> None: snake_case = value snake_case = None snake_case = None class _lowercase : def __init__( self , A__ ) -> None: snake_case = tree def UpperCamelCase ( self , A__ ) -> int: if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self ) -> Iterator[int]: yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
44
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowercase = { 'configuration_poolformer': [ 'POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PoolFormerConfig', 'PoolFormerOnnxConfig', ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ['PoolFormerFeatureExtractor'] _lowercase = ['PoolFormerImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ 'POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'PoolFormerForImageClassification', 'PoolFormerModel', 'PoolFormerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_poolformer import ( POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, PoolFormerConfig, PoolFormerOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_poolformer import PoolFormerFeatureExtractor from .image_processing_poolformer import PoolFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_poolformer import ( POOLFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, PoolFormerForImageClassification, PoolFormerModel, PoolFormerPreTrainedModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['__file__'], _import_structure)
719
'''simple docstring''' import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() _lowercase = logging.get_logger(__name__) _lowercase = [ ('bert.bert', 'visual_bert'), ('bert.cls', 'cls'), ('bert.classifier', 'cls'), ('token_type_embeddings_visual', 'visual_token_type_embeddings'), ('position_embeddings_visual', 'visual_position_embeddings'), ('projection', 'visual_projection'), ] _lowercase = [ 'nlvr2_coco_pre_trained.th', 'nlvr2_fine_tuned.th', 'nlvr2_pre_trained.th', 'vcr_coco_pre_train.th', 'vcr_fine_tune.th', 'vcr_pre_train.th', 'vqa_coco_pre_trained.th', 'vqa_fine_tuned.th', 'vqa_pre_trained.th', ] def __UpperCamelCase ( a : List[str] ) ->Optional[int]: snake_case = torch.load(a , map_location='''cpu''' ) return sd def __UpperCamelCase ( a : Optional[int] , a : Union[str, Any] , a : int=rename_keys_prefix ) ->Tuple: snake_case = OrderedDict() snake_case = torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue snake_case = key for name_pair in rename_keys_prefix: snake_case = new_key.replace(name_pair[0] , name_pair[1] ) snake_case = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately snake_case = new_d['''cls.predictions.bias'''] return new_d @torch.no_grad() def __UpperCamelCase ( a : Optional[int] , a : int ) ->Union[str, Any]: assert ( checkpoint_path.split('''/''' )[-1] in ACCEPTABLE_CHECKPOINTS ), f"""The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.""" # Get Config if "pre" in checkpoint_path: snake_case = '''pretraining''' if "vcr" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 512} elif "vqa_advanced" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048} elif "vqa" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048} elif "nlvr" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 1024} else: raise NotImplementedError(f"""No implementation found for `{checkpoint_path}`.""" ) else: if "vcr" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 512} snake_case = '''multichoice''' elif "vqa_advanced" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048} snake_case = '''vqa_advanced''' elif "vqa" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048, '''num_labels''': 3129} snake_case = '''vqa''' elif "nlvr" in checkpoint_path: snake_case = { '''visual_embedding_dim''': 1024, '''num_labels''': 2, } snake_case = '''nlvr''' snake_case = VisualBertConfig(**a ) # Load State Dict snake_case = load_state_dict(a ) snake_case = get_new_dict(a , a ) if model_type == "pretraining": snake_case = VisualBertForPreTraining(a ) elif model_type == "vqa": snake_case = VisualBertForQuestionAnswering(a ) elif model_type == "nlvr": snake_case = VisualBertForVisualReasoning(a ) elif model_type == "multichoice": snake_case = VisualBertForMultipleChoice(a ) model.load_state_dict(a ) # Save Checkpoints Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument('orig_checkpoint_path', type=str, help='A path to .th on local filesystem.') parser.add_argument('pytorch_dump_folder_path', type=str, help='Path to the output PyTorch model.') _lowercase = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
44
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import _LazyModule _lowercase = {'tokenization_wav2vec2_phoneme': ['Wav2Vec2PhonemeCTCTokenizer']} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys _lowercase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
720
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() _lowercase = logging.get_logger(__name__) def __UpperCamelCase ( a : Dict , a : Optional[int] , a : Dict , a : Dict ) ->Union[str, Any]: snake_case = original_name.split('''.''' )[0] snake_case = key.split('''.''' ) snake_case = int(key_list[key_list.index(a ) - 2] ) snake_case = int(key_list[key_list.index(a ) - 1] ) snake_case = orig_block_num - offset snake_case = key.replace(f"""{orig_block_num}.{layer_num}.{original_name}""" , f"""block.{new_block_num}.{layer_num}.{new_name}""" ) return key def __UpperCamelCase ( a : Tuple ) ->Dict: snake_case = OrderedDict() snake_case , snake_case = 0, 0 for key, value in state_dict.items(): if key.startswith('''network''' ): snake_case = key.replace('''network''' , '''poolformer.encoder''' ) if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith('''bias''' ) and "patch_embed" not in key: patch_emb_offset += 1 snake_case = key[: key.find('''proj''' )] snake_case = key.replace(a , f"""patch_embeddings.{total_embed_found}.""" ) snake_case = key.replace('''proj''' , '''projection''' ) if key.endswith('''bias''' ): total_embed_found += 1 if "patch_embeddings" in key: snake_case = '''poolformer.encoder.''' + key if "mlp.fc1" in key: snake_case = replace_key_with_offset(a , a , '''mlp.fc1''' , '''output.conv1''' ) if "mlp.fc2" in key: snake_case = replace_key_with_offset(a , a , '''mlp.fc2''' , '''output.conv2''' ) if "norm1" in key: snake_case = replace_key_with_offset(a , a , '''norm1''' , '''before_norm''' ) if "norm2" in key: snake_case = replace_key_with_offset(a , a , '''norm2''' , '''after_norm''' ) if "layer_scale_1" in key: snake_case = replace_key_with_offset(a , a , '''layer_scale_1''' , '''layer_scale_1''' ) if "layer_scale_2" in key: snake_case = replace_key_with_offset(a , a , '''layer_scale_2''' , '''layer_scale_2''' ) if "head" in key: snake_case = key.replace('''head''' , '''classifier''' ) snake_case = value return new_state_dict def __UpperCamelCase ( ) ->Optional[int]: snake_case = '''http://images.cocodataset.org/val2017/000000039769.jpg''' snake_case = Image.open(requests.get(a , stream=a ).raw ) return image @torch.no_grad() def __UpperCamelCase ( a : Dict , a : Optional[Any] , a : Tuple ) ->List[str]: snake_case = PoolFormerConfig() # set attributes based on model_name snake_case = '''huggingface/label-files''' snake_case = model_name[-3:] snake_case = 1000 snake_case = '''imagenet-1k-id2label.json''' snake_case = (1, 1000) # set config attributes snake_case = json.load(open(hf_hub_download(a , a , repo_type='''dataset''' ) , '''r''' ) ) snake_case = {int(a ): v for k, v in idalabel.items()} snake_case = idalabel snake_case = {v: k for k, v in idalabel.items()} if size == "s12": snake_case = [2, 2, 6, 2] snake_case = [64, 128, 320, 512] snake_case = 4.0 snake_case = 0.9 elif size == "s24": snake_case = [4, 4, 12, 4] snake_case = [64, 128, 320, 512] snake_case = 4.0 snake_case = 0.9 elif size == "s36": snake_case = [6, 6, 18, 6] snake_case = [64, 128, 320, 512] snake_case = 4.0 snake_case = 1e-6 snake_case = 0.9 elif size == "m36": snake_case = [6, 6, 18, 6] snake_case = [96, 192, 384, 768] snake_case = 4.0 snake_case = 1e-6 snake_case = 0.95 elif size == "m48": snake_case = [8, 8, 24, 8] snake_case = [96, 192, 384, 768] snake_case = 4.0 snake_case = 1e-6 snake_case = 0.95 else: raise ValueError(f"""Size {size} not supported""" ) # load image processor snake_case = PoolFormerImageProcessor(crop_pct=a ) # Prepare image snake_case = prepare_img() snake_case = image_processor(images=a , return_tensors='''pt''' ).pixel_values logger.info(f"""Converting model {model_name}...""" ) # load original state dict snake_case = torch.load(a , map_location=torch.device('''cpu''' ) ) # rename keys snake_case = rename_keys(a ) # create HuggingFace model and load state dict snake_case = PoolFormerForImageClassification(a ) model.load_state_dict(a ) model.eval() # Define image processor snake_case = PoolFormerImageProcessor(crop_pct=a ) snake_case = image_processor(images=prepare_img() , return_tensors='''pt''' ).pixel_values # forward pass snake_case = model(a ) snake_case = outputs.logits # define expected logit slices for different models if size == "s12": snake_case = torch.tensor([-0.3045, -0.6758, -0.4869] ) elif size == "s24": snake_case = torch.tensor([0.4402, -0.1374, -0.8045] ) elif size == "s36": snake_case = torch.tensor([-0.6080, -0.5133, -0.5898] ) elif size == "m36": snake_case = torch.tensor([0.3952, 0.2263, -1.2668] ) elif size == "m48": snake_case = torch.tensor([0.1167, -0.0656, -0.3423] ) else: raise ValueError(f"""Size {size} not supported""" ) # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] , a , atol=1e-2 ) # finally, save model and image processor logger.info(f"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(a ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument( '--model_name', default='poolformer_s12', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) _lowercase = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
44
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING _lowercase = logging.get_logger(__name__) class _lowercase ( __a ): _UpperCAmelCase = "upernet" def __init__( self , A__=None , A__=5_12 , A__=0.0_2 , A__=[1, 2, 3, 6] , A__=True , A__=0.4 , A__=3_84 , A__=2_56 , A__=1 , A__=False , A__=2_55 , **A__ , ) -> str: super().__init__(**UpperCamelCase__ ) if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) snake_case = CONFIG_MAPPING['''resnet'''](out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] ) elif isinstance(UpperCamelCase__ , UpperCamelCase__ ): snake_case = backbone_config.get('''model_type''' ) snake_case = CONFIG_MAPPING[backbone_model_type] snake_case = config_class.from_dict(UpperCamelCase__ ) snake_case = backbone_config snake_case = hidden_size snake_case = initializer_range snake_case = pool_scales snake_case = use_auxiliary_head snake_case = auxiliary_loss_weight snake_case = auxiliary_in_channels snake_case = auxiliary_channels snake_case = auxiliary_num_convs snake_case = auxiliary_concat_input snake_case = loss_ignore_index def UpperCamelCase ( self ) -> List[Any]: snake_case = copy.deepcopy(self.__dict__ ) snake_case = self.backbone_config.to_dict() snake_case = self.__class__.model_type return output
721
'''simple docstring''' import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow _lowercase = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ 'text-classification', 'language-modeling', 'summarization', 'token-classification', 'question-answering', ] ] sys.path.extend(SRC_DIRS) if SRC_DIRS is not None: import run_clm_flax import run_flax_glue import run_flax_ner import run_mlm_flax import run_qa import run_summarization_flax import run_ta_mlm_flax logging.basicConfig(level=logging.DEBUG) _lowercase = logging.getLogger() def __UpperCamelCase ( ) ->Tuple: snake_case = argparse.ArgumentParser() parser.add_argument('''-f''' ) snake_case = parser.parse_args() return args.f def __UpperCamelCase ( a : Dict , a : Tuple="eval" ) ->List[Any]: snake_case = os.path.join(a , f"""{split}_results.json""" ) if os.path.exists(a ): with open(a , '''r''' ) as f: return json.load(a ) raise ValueError(f"""can't find {path}""" ) _lowercase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class _lowercase ( __a ): def UpperCamelCase ( self ) -> List[str]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --eval_steps=2 --warmup_steps=2 --seed=42 --max_seq_length=128 """.split() with patch.object(A__ , '''argv''' , A__ ): run_flax_glue.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.7_5 ) @slow def UpperCamelCase ( self ) -> List[Any]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_clm_flax.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --block_size 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(A__ , '''argv''' , A__ ): run_clm_flax.main() snake_case = get_results(A__ ) self.assertLess(result['''eval_perplexity'''] , 1_00 ) @slow def UpperCamelCase ( self ) -> int: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_summarization.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --test_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=8 --do_train --do_eval --do_predict --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --predict_with_generate """.split() with patch.object(A__ , '''argv''' , A__ ): run_summarization_flax.main() snake_case = get_results(A__ , split='''test''' ) self.assertGreaterEqual(result['''test_rouge1'''] , 10 ) self.assertGreaterEqual(result['''test_rouge2'''] , 2 ) self.assertGreaterEqual(result['''test_rougeL'''] , 7 ) self.assertGreaterEqual(result['''test_rougeLsum'''] , 7 ) @slow def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_mlm.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --overwrite_output_dir --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --logging_steps 2 --eval_steps 2 --do_train --do_eval --num_train_epochs=1 """.split() with patch.object(A__ , '''argv''' , A__ ): run_mlm_flax.main() snake_case = get_results(A__ ) self.assertLess(result['''eval_perplexity'''] , 42 ) @slow def UpperCamelCase ( self ) -> Dict: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_t5_mlm_flax.py --model_name_or_path t5-small --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(A__ , '''argv''' , A__ ): run_ta_mlm_flax.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.4_2 ) @slow def UpperCamelCase ( self ) -> int: # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu snake_case = 7 if get_gpu_count() > 1 else 2 snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_flax_ner.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --overwrite_output_dir --do_train --do_eval --warmup_steps=2 --learning_rate=2e-4 --logging_steps 2 --eval_steps 2 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 """.split() with patch.object(A__ , '''argv''' , A__ ): run_flax_ner.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.7_5 ) self.assertGreaterEqual(result['''eval_f1'''] , 0.3 ) @slow def UpperCamelCase ( self ) -> Any: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_qa.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=2 --do_train --do_eval --logging_steps 2 --eval_steps 2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 """.split() with patch.object(A__ , '''argv''' , A__ ): run_qa.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_f1'''] , 30 ) self.assertGreaterEqual(result['''eval_exact'''] , 30 )
44
0
'''simple docstring''' def __UpperCamelCase ( a : Dict , a : Dict ) ->Tuple: snake_case = len(a ) print('''The following activities are selected:''' ) # The first activity is always selected snake_case = 0 print(a , end=''',''' ) # Consider rest of the activities for j in range(a ): # If this activity has start time greater than # or equal to the finish time of previously # selected activity, then select it if start[j] >= finish[i]: print(a , end=''',''' ) snake_case = j if __name__ == "__main__": import doctest doctest.testmod() _lowercase = [1, 3, 0, 5, 8, 5] _lowercase = [2, 4, 6, 7, 9, 9] print_max_activities(start, finish)
700
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS _lowercase = logging.get_logger(__name__) _lowercase = { 'linear': get_linear_schedule_with_warmup, 'cosine': get_cosine_schedule_with_warmup, 'cosine_w_restarts': get_cosine_with_hard_restarts_schedule_with_warmup, 'polynomial': get_polynomial_decay_schedule_with_warmup, 'constant': get_constant_schedule, 'constant_w_warmup': get_constant_schedule_with_warmup, } class _lowercase ( __a ): def __init__( self , A__=None , A__=None , *A__ , **A__ ) -> Union[str, Any]: super().__init__(*A__ , **A__ ) if config is None: assert isinstance(self.model , A__ ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) snake_case = self.model.config else: snake_case = config snake_case = data_args snake_case = self.config.tgt_vocab_size if isinstance(self.config , A__ ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ''' padding..''' ) if self.args.label_smoothing == 0: snake_case = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss snake_case = label_smoothed_nll_loss def UpperCamelCase ( self , A__ ) -> Tuple: if self.optimizer is None: snake_case = ['''bias''', '''LayerNorm.weight'''] snake_case = [ { '''params''': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], '''weight_decay''': self.args.weight_decay, }, { '''params''': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0, }, ] snake_case = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: snake_case = Adafactor snake_case = {'''scale_parameter''': False, '''relative_step''': False} else: snake_case = AdamW snake_case = { '''betas''': (self.args.adam_betaa, self.args.adam_betaa), '''eps''': self.args.adam_epsilon, } snake_case = self.args.learning_rate if self.sharded_ddp: snake_case = OSS( params=A__ , optim=A__ , **A__ , ) else: snake_case = optimizer_cls(A__ , **A__ ) if self.lr_scheduler is None: snake_case = self._get_lr_scheduler(A__ ) else: # ignoring --lr_scheduler logger.warning('''scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.''' ) def UpperCamelCase ( self , A__ ) -> Tuple: snake_case = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": snake_case = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": snake_case = schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: snake_case = schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=A__ ) return scheduler def UpperCamelCase ( self ) -> Optional[torch.utils.data.Sampler]: if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> List[Any]: if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token snake_case = model(**A__ , use_cache=A__ )[0] snake_case = self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models snake_case , snake_case = model(**A__ , labels=A__ , use_cache=A__ )[:2] else: # compute label smoothed loss snake_case = model(**A__ , use_cache=A__ )[0] snake_case = torch.nn.functional.log_softmax(A__ , dim=-1 ) snake_case , snake_case = self.loss_fn(A__ , A__ , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def UpperCamelCase ( self , A__ , A__ ) -> Any: snake_case = inputs.pop('''labels''' ) snake_case , snake_case = self._compute_loss(A__ , A__ , A__ ) return loss def UpperCamelCase ( self , A__ , A__ , A__ , A__ = None , ) -> Tuple[Optional[float], Optional[torch.Tensor], Optional[torch.Tensor]]: snake_case = self._prepare_inputs(A__ ) snake_case = { '''max_length''': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, '''num_beams''': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: snake_case = self.model.generate( inputs['''input_ids'''] , attention_mask=inputs['''attention_mask'''] , **A__ , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: snake_case = self._pad_tensors_to_max_len(A__ , gen_kwargs['''max_length'''] ) snake_case = inputs.pop('''labels''' ) with torch.no_grad(): # compute loss on predict data snake_case , snake_case = self._compute_loss(A__ , A__ , A__ ) snake_case = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) snake_case = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: snake_case = self._pad_tensors_to_max_len(A__ , gen_kwargs['''max_length'''] ) return (loss, logits, labels) def UpperCamelCase ( self , A__ , A__ ) -> List[str]: # If PAD token is not defined at least EOS token has to be defined snake_case = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( '''Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be''' F""" padded to `max_length`={max_length}""" ) snake_case = pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) snake_case = tensor return padded_tensor
44
0
'''simple docstring''' import unittest from queue import Empty from threading import Thread from transformers import AutoTokenizer, TextIteratorStreamer, TextStreamer, is_torch_available from transformers.testing_utils import CaptureStdout, require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers import AutoModelForCausalLM @require_torch class _lowercase ( unittest.TestCase ): def UpperCamelCase ( self ) -> int: snake_case = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) snake_case = AutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ).to(lowercase__ ) snake_case = -1 snake_case = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(lowercase__ ) snake_case = model.generate(lowercase__ , max_new_tokens=10 , do_sample=lowercase__ ) snake_case = tokenizer.decode(greedy_ids[0] ) with CaptureStdout() as cs: snake_case = TextStreamer(lowercase__ ) model.generate(lowercase__ , max_new_tokens=10 , do_sample=lowercase__ , streamer=lowercase__ ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer snake_case = cs.out[:-1] self.assertEqual(lowercase__ , lowercase__ ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) snake_case = AutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ).to(lowercase__ ) snake_case = -1 snake_case = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(lowercase__ ) snake_case = model.generate(lowercase__ , max_new_tokens=10 , do_sample=lowercase__ ) snake_case = tokenizer.decode(greedy_ids[0] ) snake_case = TextIteratorStreamer(lowercase__ ) snake_case = {"input_ids": input_ids, "max_new_tokens": 10, "do_sample": False, "streamer": streamer} snake_case = Thread(target=model.generate , kwargs=lowercase__ ) thread.start() snake_case = "" for new_text in streamer: streamer_text += new_text self.assertEqual(lowercase__ , lowercase__ ) def UpperCamelCase ( self ) -> Any: snake_case = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) snake_case = AutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ).to(lowercase__ ) snake_case = -1 snake_case = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(lowercase__ ) snake_case = model.generate(lowercase__ , max_new_tokens=10 , do_sample=lowercase__ ) snake_case = greedy_ids[:, input_ids.shape[1] :] snake_case = tokenizer.decode(new_greedy_ids[0] ) with CaptureStdout() as cs: snake_case = TextStreamer(lowercase__ , skip_prompt=lowercase__ ) model.generate(lowercase__ , max_new_tokens=10 , do_sample=lowercase__ , streamer=lowercase__ ) # The greedy text should be printed to stdout, except for the final "\n" in the streamer snake_case = cs.out[:-1] self.assertEqual(lowercase__ , lowercase__ ) def UpperCamelCase ( self ) -> str: snake_case = AutoTokenizer.from_pretrained('''distilgpt2''' ) snake_case = AutoModelForCausalLM.from_pretrained('''distilgpt2''' ).to(lowercase__ ) snake_case = -1 snake_case = torch.ones((1, 5) , device=lowercase__ ).long() * model.config.bos_token_id with CaptureStdout() as cs: snake_case = TextStreamer(lowercase__ , skip_special_tokens=lowercase__ ) model.generate(lowercase__ , max_new_tokens=1 , do_sample=lowercase__ , streamer=lowercase__ ) # The prompt contains a special token, so the streamer should not print it. As such, the output text, when # re-tokenized, must only contain one token snake_case = cs.out[:-1] # Remove the final "\n" snake_case = tokenizer(lowercase__ , return_tensors='''pt''' ) self.assertEqual(streamer_text_tokenized.input_ids.shape , (1, 1) ) def UpperCamelCase ( self ) -> str: snake_case = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ) snake_case = AutoModelForCausalLM.from_pretrained('''hf-internal-testing/tiny-random-gpt2''' ).to(lowercase__ ) snake_case = -1 snake_case = ids_tensor((1, 5) , vocab_size=model.config.vocab_size ).to(lowercase__ ) snake_case = TextIteratorStreamer(lowercase__ , timeout=0.0_0_1 ) snake_case = {"input_ids": input_ids, "max_new_tokens": 10, "do_sample": False, "streamer": streamer} snake_case = Thread(target=model.generate , kwargs=lowercase__ ) thread.start() # The streamer will timeout after 0.001 seconds, so an exception will be raised with self.assertRaises(lowercase__ ): snake_case = "" for new_text in streamer: streamer_text += new_text
701
'''simple docstring''' import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def __UpperCamelCase ( a : List[str] ) ->str: snake_case = [] for line in lines: snake_case = re.sub(R'''#.*''' , '''''' , a ) # remove comments if line: filtered_lines.append(a ) snake_case = '''\n'''.join(a ) # Make a hash from all this code snake_case = full_str.encode('''utf-8''' ) return shaaaa(a ).hexdigest() # get importable module names and hash for caching _lowercase = { 'csv': (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), 'json': (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), 'pandas': (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), 'parquet': (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), 'arrow': (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), 'text': (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), 'imagefolder': (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), 'audiofolder': (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions _lowercase = { '.csv': ('csv', {}), '.tsv': ('csv', {'sep': '\t'}), '.json': ('json', {}), '.jsonl': ('json', {}), '.parquet': ('parquet', {}), '.arrow': ('arrow', {}), '.txt': ('text', {}), } _EXTENSION_TO_MODULE.update({ext: ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _lowercase = {'imagefolder', 'audiofolder'} # Used to filter data files based on extensions given a module name _lowercase = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append('.zip') _MODULE_TO_EXTENSIONS["audiofolder"].append('.zip')
44
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import MvpTokenizer _lowercase = logging.get_logger(__name__) _lowercase = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} # See all MVP models at https://huggingface.co/models?filter=mvp _lowercase = { 'vocab_file': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/vocab.json', }, 'added_tokens.json': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/added_tokens.json', }, 'merges_file': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/merges.txt', }, 'tokenizer_file': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/tokenizer.json', }, } _lowercase = { 'RUCAIBox/mvp': 1_024, } class _lowercase ( __a ): _UpperCAmelCase = VOCAB_FILES_NAMES _UpperCAmelCase = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase = ['''input_ids''', '''attention_mask'''] _UpperCAmelCase = MvpTokenizer def __init__( self , A__=None , A__=None , A__=None , A__="replace" , A__="<s>" , A__="</s>" , A__="</s>" , A__="<s>" , A__="<unk>" , A__="<pad>" , A__="<mask>" , A__=False , A__=True , **A__ , ): super().__init__( A__ , A__ , tokenizer_file=A__ , errors=A__ , bos_token=A__ , eos_token=A__ , sep_token=A__ , cls_token=A__ , unk_token=A__ , pad_token=A__ , mask_token=A__ , add_prefix_space=A__ , trim_offsets=A__ , **A__ , ) snake_case = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , A__ ) != add_prefix_space: snake_case = getattr(A__ , pre_tok_state.pop('''type''' ) ) snake_case = add_prefix_space snake_case = pre_tok_class(**A__ ) snake_case = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` snake_case = 'post_processor' snake_case = getattr(self.backend_tokenizer , A__ , A__ ) if tokenizer_component_instance: snake_case = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: snake_case = tuple(state['''sep'''] ) if "cls" in state: snake_case = tuple(state['''cls'''] ) snake_case = False if state.get('''add_prefix_space''' , A__ ) != add_prefix_space: snake_case = add_prefix_space snake_case = True if state.get('''trim_offsets''' , A__ ) != trim_offsets: snake_case = trim_offsets snake_case = True if changes_to_apply: snake_case = getattr(A__ , state.pop('''type''' ) ) snake_case = component_class(**A__ ) setattr(self.backend_tokenizer , A__ , A__ ) @property def UpperCamelCase ( self ): if self._mask_token is None: if self.verbose: logger.error('''Using mask_token, but it is not set yet.''' ) return None return str(self._mask_token ) @mask_token.setter def UpperCamelCase ( self , A__ ): snake_case = AddedToken(A__ , lstrip=A__ , rstrip=A__ ) if isinstance(A__ , A__ ) else value snake_case = value def UpperCamelCase ( self , *A__ , **A__ ): snake_case = kwargs.get('''is_split_into_words''' , A__ ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ '''to use it with pretokenized inputs.''' ) return super()._batch_encode_plus(*A__ , **A__ ) def UpperCamelCase ( self , *A__ , **A__ ): snake_case = kwargs.get('''is_split_into_words''' , A__ ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"""You need to instantiate {self.__class__.__name__} with add_prefix_space=True """ '''to use it with pretokenized inputs.''' ) return super()._encode_plus(*A__ , **A__ ) def UpperCamelCase ( self , A__ , A__ = None ): snake_case = self._tokenizer.model.save(A__ , name=A__ ) return tuple(A__ ) def UpperCamelCase ( self , A__ , A__=None ): snake_case = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def UpperCamelCase ( self , A__ , A__ = None ): snake_case = [self.sep_token_id] snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
702
'''simple docstring''' _lowercase = { 'Pillow': 'Pillow', 'accelerate': 'accelerate>=0.11.0', 'compel': 'compel==0.1.8', 'black': 'black~=23.1', 'datasets': 'datasets', 'filelock': 'filelock', 'flax': 'flax>=0.4.1', 'hf-doc-builder': 'hf-doc-builder>=0.3.0', 'huggingface-hub': 'huggingface-hub>=0.13.2', 'requests-mock': 'requests-mock==1.10.0', 'importlib_metadata': 'importlib_metadata', 'invisible-watermark': 'invisible-watermark', 'isort': 'isort>=5.5.4', 'jax': 'jax>=0.2.8,!=0.3.2', 'jaxlib': 'jaxlib>=0.1.65', 'Jinja2': 'Jinja2', 'k-diffusion': 'k-diffusion>=0.0.12', 'torchsde': 'torchsde', 'note_seq': 'note_seq', 'librosa': 'librosa', 'numpy': 'numpy', 'omegaconf': 'omegaconf', 'parameterized': 'parameterized', 'protobuf': 'protobuf>=3.20.3,<4', 'pytest': 'pytest', 'pytest-timeout': 'pytest-timeout', 'pytest-xdist': 'pytest-xdist', 'ruff': 'ruff>=0.0.241', 'safetensors': 'safetensors', 'sentencepiece': 'sentencepiece>=0.1.91,!=0.1.92', 'scipy': 'scipy', 'onnx': 'onnx', 'regex': 'regex!=2019.12.17', 'requests': 'requests', 'tensorboard': 'tensorboard', 'torch': 'torch>=1.4', 'torchvision': 'torchvision', 'transformers': 'transformers>=4.25.1', 'urllib3': 'urllib3<=2.0.0', }
44
0
'''simple docstring''' import unittest from transformers import XLMConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMWithLMHeadModel, ) from transformers.models.xlm.modeling_xlm import XLM_PRETRAINED_MODEL_ARCHIVE_LIST class _lowercase : def __init__( self , A__ , A__=13 , A__=7 , A__=True , A__=True , A__=True , A__=True , A__=True , A__=False , A__=False , A__=False , A__=2 , A__=99 , A__=0 , A__=32 , A__=5 , A__=4 , A__=0.1 , A__=0.1 , A__=5_12 , A__=2 , A__=0.0_2 , A__=2 , A__=4 , A__="last" , A__=True , A__=None , A__=0 , ) -> List[Any]: snake_case = parent snake_case = batch_size snake_case = seq_length snake_case = is_training snake_case = use_input_lengths snake_case = use_token_type_ids snake_case = use_labels snake_case = gelu_activation snake_case = sinusoidal_embeddings snake_case = causal snake_case = asm snake_case = n_langs snake_case = vocab_size snake_case = n_special snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = max_position_embeddings snake_case = type_sequence_label_size snake_case = initializer_range snake_case = num_labels snake_case = num_choices snake_case = summary_type snake_case = use_proj snake_case = scope snake_case = bos_token_id def UpperCamelCase ( self ) -> Any: snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case = random_attention_mask([self.batch_size, self.seq_length] ) snake_case = None if self.use_input_lengths: snake_case = ( ids_tensor([self.batch_size] , vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length snake_case = None if self.use_token_type_ids: snake_case = ids_tensor([self.batch_size, self.seq_length] , self.n_langs ) snake_case = None snake_case = None snake_case = None if self.use_labels: snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case = ids_tensor([self.batch_size] , 2 ).float() snake_case = ids_tensor([self.batch_size] , self.num_choices ) snake_case = self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def UpperCamelCase ( self ) -> Optional[int]: return XLMConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , num_labels=self.num_labels , bos_token_id=self.bos_token_id , ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ) -> List[str]: snake_case = XLMModel(config=A__ ) model.to(A__ ) model.eval() snake_case = model(A__ , lengths=A__ , langs=A__ ) snake_case = model(A__ , langs=A__ ) snake_case = model(A__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ) -> Any: snake_case = XLMWithLMHeadModel(A__ ) model.to(A__ ) model.eval() snake_case = model(A__ , token_type_ids=A__ , labels=A__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ) -> int: snake_case = XLMForQuestionAnsweringSimple(A__ ) model.to(A__ ) model.eval() snake_case = model(A__ ) snake_case = model(A__ , start_positions=A__ , end_positions=A__ ) snake_case = outputs self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ) -> str: snake_case = XLMForQuestionAnswering(A__ ) model.to(A__ ) model.eval() snake_case = model(A__ ) snake_case = model( A__ , start_positions=A__ , end_positions=A__ , cls_index=A__ , is_impossible=A__ , p_mask=A__ , ) snake_case = model( A__ , start_positions=A__ , end_positions=A__ , cls_index=A__ , is_impossible=A__ , ) (snake_case ) = result_with_labels.to_tuple() snake_case = model(A__ , start_positions=A__ , end_positions=A__ ) (snake_case ) = result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape , () ) self.parent.assertEqual(result.start_top_log_probs.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual(result.start_top_index.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual( result.end_top_log_probs.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual( result.end_top_index.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual(result.cls_logits.shape , (self.batch_size,) ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ) -> List[Any]: snake_case = XLMForSequenceClassification(A__ ) model.to(A__ ) model.eval() snake_case = model(A__ ) snake_case = model(A__ , labels=A__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ) -> Optional[Any]: snake_case = self.num_labels snake_case = XLMForTokenClassification(A__ ) model.to(A__ ) model.eval() snake_case = model(A__ , attention_mask=A__ , labels=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ) -> Optional[Any]: snake_case = self.num_choices snake_case = XLMForMultipleChoice(config=A__ ) model.to(A__ ) model.eval() snake_case = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case = model( A__ , attention_mask=A__ , token_type_ids=A__ , labels=A__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCamelCase ( self ) -> Optional[int]: snake_case = self.prepare_config_and_inputs() ( snake_case ) = config_and_inputs snake_case = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'lengths': input_lengths} return config, inputs_dict @require_torch class _lowercase ( __A , __A , __A , unittest.TestCase ): _UpperCAmelCase = ( ( XLMModel, XLMWithLMHeadModel, XLMForQuestionAnswering, XLMForSequenceClassification, XLMForQuestionAnsweringSimple, XLMForTokenClassification, XLMForMultipleChoice, ) if is_torch_available() else () ) _UpperCAmelCase = ( (XLMWithLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Check other models whether language generation is also applicable _UpperCAmelCase = ( { '''feature-extraction''': XLMModel, '''fill-mask''': XLMWithLMHeadModel, '''question-answering''': XLMForQuestionAnsweringSimple, '''text-classification''': XLMForSequenceClassification, '''text-generation''': XLMWithLMHeadModel, '''token-classification''': XLMForTokenClassification, '''zero-shot''': XLMForSequenceClassification, } if is_torch_available() else {} ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ ) -> Optional[Any]: if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith('''Fast''' ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def UpperCamelCase ( self , A__ , A__ , A__=False ) -> Optional[int]: snake_case = super()._prepare_for_class(A__ , A__ , return_labels=A__ ) if return_labels: if model_class.__name__ == "XLMForQuestionAnswering": snake_case = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=A__ ) snake_case = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=A__ ) return inputs_dict def UpperCamelCase ( self ) -> Dict: snake_case = XLMModelTester(self ) snake_case = ConfigTester(self , config_class=A__ , emb_dim=37 ) def UpperCamelCase ( self ) -> Any: self.config_tester.run_common_tests() def UpperCamelCase ( self ) -> Optional[int]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_model(*A__ ) def UpperCamelCase ( self ) -> int: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_lm_head(*A__ ) def UpperCamelCase ( self ) -> List[str]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_simple_qa(*A__ ) def UpperCamelCase ( self ) -> List[str]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_qa(*A__ ) def UpperCamelCase ( self ) -> Optional[Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_sequence_classif(*A__ ) def UpperCamelCase ( self ) -> int: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_token_classif(*A__ ) def UpperCamelCase ( self ) -> str: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_xlm_for_multiple_choice(*A__ ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__=False , A__=1 ) -> Dict: self.assertIsInstance(A__ , A__ ) self.assertListEqual( [isinstance(A__ , A__ ) for iter_attentions in attentions] , [True] * len(A__ ) ) self.assertEqual(len(A__ ) , (max_length - min_length) * num_beam_groups ) for idx, iter_attentions in enumerate(A__ ): # adds PAD dummy token snake_case = min_length + idx + 1 snake_case = min_length + idx + 1 snake_case = ( batch_size * num_beam_groups, config.num_attention_heads, tgt_len, src_len, ) # check attn size self.assertListEqual( [layer_attention.shape for layer_attention in iter_attentions] , [expected_shape] * len(A__ ) ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__=False , A__=1 ) -> Dict: self.assertIsInstance(A__ , A__ ) self.assertListEqual( [isinstance(A__ , A__ ) for iter_hidden_states in hidden_states] , [True] * len(A__ ) , ) self.assertEqual(len(A__ ) , (max_length - min_length) * num_beam_groups ) for idx, iter_hidden_states in enumerate(A__ ): # adds PAD dummy token snake_case = min_length + idx + 1 snake_case = (batch_size * num_beam_groups, seq_len, config.hidden_size) # check hidden size self.assertListEqual( [layer_hidden_states.shape for layer_hidden_states in iter_hidden_states] , [expected_shape] * len(A__ ) , ) pass @slow def UpperCamelCase ( self ) -> Dict: for model_name in XLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case = XLMModel.from_pretrained(A__ ) self.assertIsNotNone(A__ ) @require_torch class _lowercase ( unittest.TestCase ): @slow def UpperCamelCase ( self ) -> int: snake_case = XLMWithLMHeadModel.from_pretrained('''xlm-mlm-en-2048''' ) model.to(A__ ) snake_case = torch.tensor([[14, 4_47]] , dtype=torch.long , device=A__ ) # the president snake_case = [ 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, 14, 4_47, ] # the president the president the president the president the president the president the president the president the president the president # TODO(PVP): this and other input_ids I tried for generation give pretty bad results. Not sure why. Model might just not be made for auto-regressive inference snake_case = model.generate(A__ , do_sample=A__ ) self.assertListEqual(output_ids[0].cpu().numpy().tolist() , A__ )
703
'''simple docstring''' import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _lowercase ( __a , __a , unittest.TestCase ): _UpperCAmelCase = IFInpaintingSuperResolutionPipeline _UpperCAmelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} _UpperCAmelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({'''original_image'''} ) _UpperCAmelCase = PipelineTesterMixin.required_optional_params - {'''latents'''} def UpperCamelCase ( self ) -> int: return self._get_superresolution_dummy_components() def UpperCamelCase ( self , A__ , A__=0 ) -> Union[str, Any]: if str(A__ ).startswith('''mps''' ): snake_case = torch.manual_seed(A__ ) else: snake_case = torch.Generator(device=A__ ).manual_seed(A__ ) snake_case = floats_tensor((1, 3, 16, 16) , rng=random.Random(A__ ) ).to(A__ ) snake_case = floats_tensor((1, 3, 32, 32) , rng=random.Random(A__ ) ).to(A__ ) snake_case = floats_tensor((1, 3, 32, 32) , rng=random.Random(A__ ) ).to(A__ ) snake_case = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase ( self ) -> List[Any]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def UpperCamelCase ( self ) -> Optional[Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def UpperCamelCase ( self ) -> List[str]: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1 ) def UpperCamelCase ( self ) -> int: self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def UpperCamelCase ( self ) -> Optional[Any]: self._test_save_load_local() def UpperCamelCase ( self ) -> Dict: self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
44
0
'''simple docstring''' # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowercase = { 'configuration_xmod': [ 'XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XmodConfig', 'XmodOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ 'XMOD_PRETRAINED_MODEL_ARCHIVE_LIST', 'XmodForCausalLM', 'XmodForMaskedLM', 'XmodForMultipleChoice', 'XmodForQuestionAnswering', 'XmodForSequenceClassification', 'XmodForTokenClassification', 'XmodModel', 'XmodPreTrainedModel', ] if TYPE_CHECKING: from .configuration_xmod import XMOD_PRETRAINED_CONFIG_ARCHIVE_MAP, XmodConfig, XmodOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xmod import ( XMOD_PRETRAINED_MODEL_ARCHIVE_LIST, XmodForCausalLM, XmodForMaskedLM, XmodForMultipleChoice, XmodForQuestionAnswering, XmodForSequenceClassification, XmodForTokenClassification, XmodModel, XmodPreTrainedModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
704
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy _lowercase = logging.get_logger(__name__) class _lowercase ( __a ): def __init__( self , A__ , A__ , A__ , **A__ ) -> Union[str, Any]: snake_case = feature_size snake_case = sampling_rate snake_case = padding_value snake_case = kwargs.pop('''padding_side''' , '''right''' ) snake_case = kwargs.pop('''return_attention_mask''' , A__ ) super().__init__(**A__ ) def UpperCamelCase ( self , A__ , A__ = True , A__ = None , A__ = False , A__ = None , A__ = None , A__ = None , ) -> BatchFeature: # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(A__ , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): snake_case = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( '''You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`''' F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) snake_case = processed_features[self.model_input_names[0]] snake_case = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(A__ ) == 0: if return_attention_mask: snake_case = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch snake_case = required_input[0] if isinstance(A__ , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. snake_case = 0 while len(required_input[index] ) == 0: index += 1 if index < len(A__ ): snake_case = required_input[index][0] if return_tensors is None: if is_tf_tensor(A__ ): snake_case = '''tf''' elif is_torch_tensor(A__ ): snake_case = '''pt''' elif isinstance(A__ , (int, float, list, tuple, np.ndarray) ): snake_case = '''np''' else: raise ValueError( F"""type of {first_element} unknown: {type(A__ )}. """ '''Should be one of a python, numpy, pytorch or tensorflow object.''' ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): snake_case = to_numpy(A__ ) else: snake_case = [to_numpy(A__ ) for v in value] # Convert padding_strategy in PaddingStrategy snake_case = self._get_padding_strategies(padding=A__ , max_length=A__ ) snake_case = processed_features[self.model_input_names[0]] snake_case = len(A__ ) if not all(len(A__ ) == batch_size for v in processed_features.values() ): raise ValueError('''Some items in the output dictionary have a different batch size than others.''' ) snake_case = [] for i in range(A__ ): snake_case = {k: v[i] for k, v in processed_features.items()} # truncation snake_case = self._truncate( A__ , max_length=A__ , pad_to_multiple_of=A__ , truncation=A__ , ) truncated_inputs.append(A__ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length snake_case = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) snake_case = PaddingStrategy.MAX_LENGTH snake_case = {} for i in range(A__ ): # padding snake_case = self._pad( truncated_inputs[i] , max_length=A__ , padding_strategy=A__ , pad_to_multiple_of=A__ , return_attention_mask=A__ , ) for key, value in outputs.items(): if key not in batch_outputs: snake_case = [] if value.dtype is np.dtype(np.floataa ): snake_case = value.astype(np.floataa ) batch_outputs[key].append(A__ ) return BatchFeature(A__ , tensor_type=A__ ) def UpperCamelCase ( self , A__ , A__ = None , A__ = PaddingStrategy.DO_NOT_PAD , A__ = None , A__ = None , ) -> dict: snake_case = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: snake_case = len(A__ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): snake_case = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of snake_case = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(A__ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: snake_case = np.ones(len(A__ ) , dtype=np.intaa ) if needs_to_be_padded: snake_case = max_length - len(A__ ) if self.padding_side == "right": if return_attention_mask: snake_case = np.pad( processed_features['''attention_mask'''] , (0, difference) ) snake_case = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) snake_case = np.pad( A__ , A__ , '''constant''' , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: snake_case = np.pad( processed_features['''attention_mask'''] , (difference, 0) ) snake_case = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) snake_case = np.pad( A__ , A__ , '''constant''' , constant_values=self.padding_value ) else: raise ValueError('''Invalid padding strategy:''' + str(self.padding_side ) ) return processed_features def UpperCamelCase ( self , A__ , A__ = None , A__ = None , A__ = None , ) -> Union[str, Any]: if not truncation: return processed_features elif truncation and max_length is None: raise ValueError('''When setting ``truncation=True``, make sure that ``max_length`` is defined.''' ) snake_case = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): snake_case = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of snake_case = len(A__ ) > max_length if needs_to_be_truncated: snake_case = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: snake_case = processed_features['''attention_mask'''][:max_length] return processed_features def UpperCamelCase ( self , A__=False , A__=None ) -> Union[str, Any]: # Get padding strategy if padding is not False: if padding is True: snake_case = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(A__ , A__ ): snake_case = PaddingStrategy(A__ ) elif isinstance(A__ , A__ ): snake_case = padding else: snake_case = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( '''Asking to pad but the feature_extractor does not have a padding value. Please select a value to use''' ''' as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.''' ) return padding_strategy
44
0
'''simple docstring''' import json import logging import os import re import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import datasets import numpy as np import torch import torchaudio from packaging import version from torch import nn import transformers from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaProcessor, is_apex_available, set_seed, ) from transformers.trainer_utils import get_last_checkpoint, is_main_process if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('1.6'): _lowercase = True from torch.cuda.amp import autocast _lowercase = logging.getLogger(__name__) def __UpperCamelCase ( a : Optional[Any]=None , a : List[str]=None ) ->Optional[Any]: return field(default_factory=lambda: default , metadata=lowerCAmelCase__ ) @dataclass class _lowercase : _UpperCAmelCase = field( metadata={'''help''': '''Path to pretrained model or model identifier from huggingface.co/models'''} ) _UpperCAmelCase = field( default=__a , metadata={'''help''': '''Where do you want to store the pretrained models downloaded from huggingface.co'''} , ) _UpperCAmelCase = field( default=__a , metadata={'''help''': '''Whether to freeze the feature extractor layers of the model.'''} ) _UpperCAmelCase = field( default=0.1 , metadata={'''help''': '''The dropout ratio for the attention probabilities.'''} ) _UpperCAmelCase = field( default=0.1 , metadata={'''help''': '''The dropout ratio for activations inside the fully connected layer.'''} ) _UpperCAmelCase = field( default=0.1 , metadata={ '''help''': '''The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler.''' } , ) _UpperCAmelCase = field( default=0.1 , metadata={'''help''': '''The dropout probabilitiy for all 1D convolutional layers in feature extractor.'''} , ) _UpperCAmelCase = field( default=0.05 , metadata={ '''help''': ( '''Propability of each feature vector along the time axis to be chosen as the start of the vector''' '''span to be masked. Approximately ``mask_time_prob * sequence_length // mask_time_length`` feature''' '''vectors will be masked along the time axis. This is only relevant if ``apply_spec_augment is True``.''' ) } , ) _UpperCAmelCase = field(default=0.0 , metadata={'''help''': '''The LayerDrop probability.'''} ) @dataclass class _lowercase : _UpperCAmelCase = field( default=__a , metadata={'''help''': '''The configuration name of the dataset to use (via the datasets library).'''} ) _UpperCAmelCase = field( default='''train+validation''' , metadata={ '''help''': '''The name of the training data set split to use (via the datasets library). Defaults to \'train\'''' } , ) _UpperCAmelCase = field( default=__a , metadata={'''help''': '''Overwrite the cached preprocessed datasets or not.'''} ) _UpperCAmelCase = field( default=__a , metadata={'''help''': '''The number of processes to use for the preprocessing.'''} , ) _UpperCAmelCase = field( default=__a , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of training examples to this ''' '''value if set.''' ) } , ) _UpperCAmelCase = field( default=__a , metadata={ '''help''': ( '''For debugging purposes or quicker training, truncate the number of validation examples to this ''' '''value if set.''' ) } , ) _UpperCAmelCase = list_field( default=[''',''', '''?''', '''.''', '''!''', '''-''', ''';''', ''':''', '''""''', '''%''', '''\'''', '''"''', '''�'''] , metadata={'''help''': '''A list of characters to remove from the transcripts.'''} , ) @dataclass class _lowercase : _UpperCAmelCase = 42 _UpperCAmelCase = True _UpperCAmelCase = None _UpperCAmelCase = None _UpperCAmelCase = None _UpperCAmelCase = None def __call__( self , A__ ) -> Dict: # split inputs and labels since they have to be of different lenghts and need # different padding methods snake_case = [{'''input_values''': feature['''input_values''']} for feature in features] snake_case = [{'''input_ids''': feature['''labels''']} for feature in features] snake_case = self.processor.pad( snake_case__ , padding=self.padding , max_length=self.max_length , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' , ) snake_case = self.processor.pad( labels=snake_case__ , padding=self.padding , max_length=self.max_length_labels , pad_to_multiple_of=self.pad_to_multiple_of_labels , return_tensors='''pt''' , ) # replace padding with -100 to ignore loss correctly snake_case = labels_batch['''input_ids'''].masked_fill(labels_batch.attention_mask.ne(1 ) , -1_00 ) snake_case = labels return batch class _lowercase ( __a ): def UpperCamelCase ( self , A__ , A__ ) -> int: model.train() snake_case = self._prepare_inputs(snake_case__ ) if self.use_amp: with autocast(): snake_case = self.compute_loss(snake_case__ , snake_case__ ) else: snake_case = self.compute_loss(snake_case__ , snake_case__ ) if self.args.n_gpu > 1: if model.module.config.ctc_loss_reduction == "mean": snake_case = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": snake_case = loss.sum() / (inputs['''labels'''] >= 0).sum() else: raise ValueError(F"""{model.config.ctc_loss_reduction} is not valid. Choose one of ['mean', 'sum']""" ) if self.args.gradient_accumulation_steps > 1: snake_case = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(snake_case__ ).backward() elif self.use_apex: with amp.scale_loss(snake_case__ , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(snake_case__ ) else: loss.backward() return loss.detach() def __UpperCamelCase ( ) ->Optional[int]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. snake_case = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('''.json''' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. snake_case , snake_case , snake_case = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: snake_case , snake_case , snake_case = parser.parse_args_into_dataclasses() # Detecting last checkpoint. snake_case = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: snake_case = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. """ '''Use --overwrite_output_dir to overcome.''' ) elif last_checkpoint is not None: logger.info( f"""Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change """ '''the `--output_dir` or add `--overwrite_output_dir` to train from scratch.''' ) # Setup logging logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) logger.setLevel(logging.INFO if is_main_process(training_args.local_rank ) else logging.WARN ) # Log on each process the small summary: logger.warning( f"""Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}""" + f"""distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}""" ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() logger.info('''Training/evaluation parameters %s''' , lowerCAmelCase__ ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: snake_case = datasets.load_dataset( '''common_voice''' , data_args.dataset_config_name , split=data_args.train_split_name ) snake_case = datasets.load_dataset('''common_voice''' , data_args.dataset_config_name , split='''test''' ) # Create and save tokenizer snake_case = f"""[{''.join(data_args.chars_to_ignore )}]""" def remove_special_characters(a : Optional[int] ): snake_case = re.sub(lowerCAmelCase__ , '''''' , batch['''sentence'''] ).lower() + ''' ''' return batch snake_case = train_dataset.map(lowerCAmelCase__ , remove_columns=['''sentence'''] ) snake_case = eval_dataset.map(lowerCAmelCase__ , remove_columns=['''sentence'''] ) def extract_all_chars(a : str ): snake_case = ''' '''.join(batch['''text'''] ) snake_case = list(set(lowerCAmelCase__ ) ) return {"vocab": [vocab], "all_text": [all_text]} snake_case = train_dataset.map( lowerCAmelCase__ , batched=lowerCAmelCase__ , batch_size=-1 , keep_in_memory=lowerCAmelCase__ , remove_columns=train_dataset.column_names , ) snake_case = train_dataset.map( lowerCAmelCase__ , batched=lowerCAmelCase__ , batch_size=-1 , keep_in_memory=lowerCAmelCase__ , remove_columns=eval_dataset.column_names , ) snake_case = list(set(vocab_train['''vocab'''][0] ) | set(vocab_test['''vocab'''][0] ) ) snake_case = {v: k for k, v in enumerate(lowerCAmelCase__ )} snake_case = vocab_dict[''' '''] del vocab_dict[" "] snake_case = len(lowerCAmelCase__ ) snake_case = len(lowerCAmelCase__ ) with open('''vocab.json''' , '''w''' ) as vocab_file: json.dump(lowerCAmelCase__ , lowerCAmelCase__ ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. snake_case = WavaVecaCTCTokenizer( '''vocab.json''' , unk_token='''[UNK]''' , pad_token='''[PAD]''' , word_delimiter_token='''|''' , ) snake_case = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_6000 , padding_value=0.0 , do_normalize=lowerCAmelCase__ , return_attention_mask=lowerCAmelCase__ ) snake_case = WavaVecaProcessor(feature_extractor=lowerCAmelCase__ , tokenizer=lowerCAmelCase__ ) snake_case = WavaVecaForCTC.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , activation_dropout=model_args.activation_dropout , attention_dropout=model_args.attention_dropout , hidden_dropout=model_args.hidden_dropout , feat_proj_dropout=model_args.feat_proj_dropout , mask_time_prob=model_args.mask_time_prob , gradient_checkpointing=training_args.gradient_checkpointing , layerdrop=model_args.layerdrop , ctc_loss_reduction='''mean''' , pad_token_id=processor.tokenizer.pad_token_id , vocab_size=len(processor.tokenizer ) , ) if data_args.max_train_samples is not None: snake_case = min(len(lowerCAmelCase__ ) , data_args.max_train_samples ) snake_case = train_dataset.select(range(lowerCAmelCase__ ) ) if data_args.max_val_samples is not None: snake_case = eval_dataset.select(range(data_args.max_val_samples ) ) snake_case = torchaudio.transforms.Resample(4_8000 , 1_6000 ) # Preprocessing the datasets. # We need to read the aduio files as arrays and tokenize the targets. def speech_file_to_array_fn(a : Tuple ): snake_case , snake_case = torchaudio.load(batch['''path'''] ) snake_case = resampler(lowerCAmelCase__ ).squeeze().numpy() snake_case = 1_6000 snake_case = batch['''text'''] return batch snake_case = train_dataset.map( lowerCAmelCase__ , remove_columns=train_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) snake_case = eval_dataset.map( lowerCAmelCase__ , remove_columns=eval_dataset.column_names , num_proc=data_args.preprocessing_num_workers , ) def prepare_dataset(a : Optional[Any] ): # check that all files have the correct sampling rate assert ( len(set(batch['''sampling_rate'''] ) ) == 1 ), f"""Make sure all inputs have the same sampling rate of {processor.feature_extractor.sampling_rate}.""" snake_case = processor( audio=batch['''speech'''] , text=batch['''target_text'''] , sampling_rate=batch['''sampling_rate'''][0] ) batch.update(lowerCAmelCase__ ) return batch snake_case = train_dataset.map( lowerCAmelCase__ , remove_columns=train_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=lowerCAmelCase__ , num_proc=data_args.preprocessing_num_workers , ) snake_case = eval_dataset.map( lowerCAmelCase__ , remove_columns=eval_dataset.column_names , batch_size=training_args.per_device_train_batch_size , batched=lowerCAmelCase__ , num_proc=data_args.preprocessing_num_workers , ) # Metric snake_case = datasets.load_metric('''wer''' ) def compute_metrics(a : int ): snake_case = pred.predictions snake_case = np.argmax(lowerCAmelCase__ , axis=-1 ) snake_case = processor.tokenizer.pad_token_id snake_case = processor.batch_decode(lowerCAmelCase__ ) # we do not want to group tokens when computing the metrics snake_case = processor.batch_decode(pred.label_ids , group_tokens=lowerCAmelCase__ ) snake_case = wer_metric.compute(predictions=lowerCAmelCase__ , references=lowerCAmelCase__ ) return {"wer": wer} if model_args.freeze_feature_extractor: model.freeze_feature_extractor() # Data collator snake_case = DataCollatorCTCWithPadding(processor=lowerCAmelCase__ , padding=lowerCAmelCase__ ) # Initialize our Trainer snake_case = CTCTrainer( model=lowerCAmelCase__ , data_collator=lowerCAmelCase__ , args=lowerCAmelCase__ , compute_metrics=lowerCAmelCase__ , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , tokenizer=processor.feature_extractor , ) # Training if training_args.do_train: if last_checkpoint is not None: snake_case = last_checkpoint elif os.path.isdir(model_args.model_name_or_path ): snake_case = model_args.model_name_or_path else: snake_case = None # Save the feature_extractor and the tokenizer if is_main_process(training_args.local_rank ): processor.save_pretrained(training_args.output_dir ) snake_case = trainer.train(resume_from_checkpoint=lowerCAmelCase__ ) trainer.save_model() snake_case = train_result.metrics snake_case = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowerCAmelCase__ ) ) snake_case = min(lowerCAmelCase__ , len(lowerCAmelCase__ ) ) trainer.log_metrics('''train''' , lowerCAmelCase__ ) trainer.save_metrics('''train''' , lowerCAmelCase__ ) trainer.save_state() # Evaluation snake_case = {} if training_args.do_eval: logger.info('''*** Evaluate ***''' ) snake_case = trainer.evaluate() snake_case = data_args.max_val_samples if data_args.max_val_samples is not None else len(lowerCAmelCase__ ) snake_case = min(lowerCAmelCase__ , len(lowerCAmelCase__ ) ) trainer.log_metrics('''eval''' , lowerCAmelCase__ ) trainer.save_metrics('''eval''' , lowerCAmelCase__ ) return results if __name__ == "__main__": main()
705
'''simple docstring''' from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class _lowercase ( yaml.SafeLoader ): def UpperCamelCase ( self , A__ ) -> List[str]: snake_case = [self.constructed_objects[key_node] for key_node, _ in node.value] snake_case = [tuple(A__ ) if isinstance(A__ , A__ ) else key for key in keys] snake_case = Counter(A__ ) snake_case = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F"""Got duplicate yaml keys: {duplicate_keys}""" ) def UpperCamelCase ( self , A__ , A__=False ) -> List[Any]: snake_case = super().construct_mapping(A__ , deep=A__ ) self._check_no_duplicates_on_constructed_node(A__ ) return mapping def __UpperCamelCase ( a : str ) ->Tuple[Optional[str], str]: snake_case = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: snake_case = full_content[1:].index('''---''' ) + 1 snake_case = '''\n'''.join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(a ) class _lowercase ( __a ): # class attributes _UpperCAmelCase = {'''train_eval_index'''} # train-eval-index in the YAML metadata @classmethod def UpperCamelCase ( cls , A__ ) -> "DatasetMetadata": with open(A__ , encoding='''utf-8''' ) as readme_file: snake_case , snake_case = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(A__ ) else: return cls() def UpperCamelCase ( self , A__ ) -> str: if path.exists(): with open(A__ , encoding='''utf-8''' ) as readme_file: snake_case = readme_file.read() else: snake_case = None snake_case = self._to_readme(A__ ) with open(A__ , '''w''' , encoding='''utf-8''' ) as readme_file: readme_file.write(A__ ) def UpperCamelCase ( self , A__ = None ) -> str: if readme_content is not None: snake_case , snake_case = _split_yaml_from_readme(A__ ) snake_case = '''---\n''' + self.to_yaml_string() + '''---\n''' + content else: snake_case = '''---\n''' + self.to_yaml_string() + '''---\n''' return full_content @classmethod def UpperCamelCase ( cls , A__ ) -> "DatasetMetadata": snake_case = yaml.load(A__ , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields snake_case = { (key.replace('''-''' , '''_''' ) if key.replace('''-''' , '''_''' ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**A__ ) def UpperCamelCase ( self ) -> str: return yaml.safe_dump( { (key.replace('''_''' , '''-''' ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=A__ , allow_unicode=A__ , encoding='''utf-8''' , ).decode('''utf-8''' ) _lowercase = { 'image-classification': [], 'translation': [], 'image-segmentation': [], 'fill-mask': [], 'automatic-speech-recognition': [], 'token-classification': [], 'sentence-similarity': [], 'audio-classification': [], 'question-answering': [], 'summarization': [], 'zero-shot-classification': [], 'table-to-text': [], 'feature-extraction': [], 'other': [], 'multiple-choice': [], 'text-classification': [], 'text-to-image': [], 'text2text-generation': [], 'zero-shot-image-classification': [], 'tabular-classification': [], 'tabular-regression': [], 'image-to-image': [], 'tabular-to-text': [], 'unconditional-image-generation': [], 'text-retrieval': [], 'text-to-speech': [], 'object-detection': [], 'audio-to-audio': [], 'text-generation': [], 'conversational': [], 'table-question-answering': [], 'visual-question-answering': [], 'image-to-text': [], 'reinforcement-learning': [], 'voice-activity-detection': [], 'time-series-forecasting': [], 'document-question-answering': [], } if __name__ == "__main__": from argparse import ArgumentParser _lowercase = ArgumentParser(usage='Validate the yaml metadata block of a README.md file.') ap.add_argument('readme_filepath') _lowercase = ap.parse_args() _lowercase = Path(args.readme_filepath) _lowercase = DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
44
0
'''simple docstring''' _lowercase = 'ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/' def __UpperCamelCase ( a : Union[str, Any] ) ->List[Any]: if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): snake_case = f"""a bytes-like object is required, not '{data.__class__.__name__}'""" raise TypeError(lowerCAmelCase_ ) snake_case = "".join(bin(lowerCAmelCase_ )[2:].zfill(8 ) for byte in data ) snake_case = len(lowerCAmelCase_ ) % 6 != 0 if padding_needed: # The padding that will be added later snake_case = b"=" * ((6 - len(lowerCAmelCase_ ) % 6) // 2) # Append binary_stream with arbitrary binary digits (0's by default) to make its # length a multiple of 6. binary_stream += "0" * (6 - len(lowerCAmelCase_ ) % 6) else: snake_case = b"" # Encode every 6 binary digits to their corresponding Base64 character return ( "".join( B64_CHARSET[int(binary_stream[index : index + 6] , 2 )] for index in range(0 , len(lowerCAmelCase_ ) , 6 ) ).encode() + padding ) def __UpperCamelCase ( a : List[Any] ) ->int: if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): snake_case = ( "argument should be a bytes-like object or ASCII string, " f"""not '{encoded_data.__class__.__name__}'""" ) raise TypeError(lowerCAmelCase_ ) # In case encoded_data is a bytes-like object, make sure it contains only # ASCII characters so we convert it to a string object if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): try: snake_case = encoded_data.decode('''utf-8''' ) except UnicodeDecodeError: raise ValueError('''base64 encoded data should only contain ASCII characters''' ) snake_case = encoded_data.count('''=''' ) # Check if the encoded string contains non base64 characters if padding: assert all( char in B64_CHARSET for char in encoded_data[:-padding] ), "Invalid base64 character(s) found." else: assert all( char in B64_CHARSET for char in encoded_data ), "Invalid base64 character(s) found." # Check the padding assert len(lowerCAmelCase_ ) % 4 == 0 and padding < 3, "Incorrect padding" if padding: # Remove padding if there is one snake_case = encoded_data[:-padding] snake_case = "".join( bin(B64_CHARSET.index(lowerCAmelCase_ ) )[2:].zfill(6 ) for char in encoded_data )[: -padding * 2] else: snake_case = "".join( bin(B64_CHARSET.index(lowerCAmelCase_ ) )[2:].zfill(6 ) for char in encoded_data ) snake_case = [ int(binary_stream[index : index + 8] , 2 ) for index in range(0 , len(lowerCAmelCase_ ) , 8 ) ] return bytes(lowerCAmelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
706
'''simple docstring''' import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowercase ( __a , unittest.TestCase ): _UpperCAmelCase = CodeGenTokenizer _UpperCAmelCase = CodeGenTokenizerFast _UpperCAmelCase = True _UpperCAmelCase = {'''add_prefix_space''': True} _UpperCAmelCase = False def UpperCamelCase ( self ) -> Tuple: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt snake_case = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', '''<|endoftext|>''', ] snake_case = dict(zip(A__ , range(len(A__ ) ) ) ) snake_case = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] snake_case = {'''unk_token''': '''<unk>'''} snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(A__ ) ) def UpperCamelCase ( self , **A__ ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **A__ ) def UpperCamelCase ( self , **A__ ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **A__ ) def UpperCamelCase ( self , A__ ) -> Tuple: snake_case = '''lower newer''' snake_case = '''lower newer''' return input_text, output_text def UpperCamelCase ( self ) -> List[Any]: snake_case = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) snake_case = '''lower newer''' snake_case = ['''\u0120low''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] snake_case = tokenizer.tokenize(A__ , add_prefix_space=A__ ) self.assertListEqual(A__ , A__ ) snake_case = tokens + [tokenizer.unk_token] snake_case = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(A__ ) , A__ ) def UpperCamelCase ( self ) -> Optional[int]: if not self.test_rust_tokenizer: return snake_case = self.get_tokenizer() snake_case = self.get_rust_tokenizer(add_prefix_space=A__ ) snake_case = '''lower newer''' # Testing tokenization snake_case = tokenizer.tokenize(A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.tokenize(A__ ) self.assertListEqual(A__ , A__ ) # Testing conversion to ids without special tokens snake_case = tokenizer.encode(A__ , add_special_tokens=A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.encode(A__ , add_special_tokens=A__ ) self.assertListEqual(A__ , A__ ) # Testing conversion to ids with special tokens snake_case = self.get_rust_tokenizer(add_prefix_space=A__ ) snake_case = tokenizer.encode(A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.encode(A__ ) self.assertListEqual(A__ , A__ ) # Testing the unknown token snake_case = tokens + [rust_tokenizer.unk_token] snake_case = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(A__ ) , A__ ) def UpperCamelCase ( self , *A__ , **A__ ) -> List[str]: # It's very difficult to mix/test pretokenization with byte-level # And get both CodeGen and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def UpperCamelCase ( self , A__=15 ) -> Tuple: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): snake_case = self.rust_tokenizer_class.from_pretrained(A__ , **A__ ) # Simple input snake_case = '''This is a simple input''' snake_case = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case = ('''This is a simple input''', '''This is a pair''') snake_case = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(A__ , tokenizer_r.encode , A__ , max_length=A__ , padding='''max_length''' ) # Simple input self.assertRaises(A__ , tokenizer_r.encode_plus , A__ , max_length=A__ , padding='''max_length''' ) # Simple input self.assertRaises( A__ , tokenizer_r.batch_encode_plus , A__ , max_length=A__ , padding='''max_length''' , ) # Pair input self.assertRaises(A__ , tokenizer_r.encode , A__ , max_length=A__ , padding='''max_length''' ) # Pair input self.assertRaises(A__ , tokenizer_r.encode_plus , A__ , max_length=A__ , padding='''max_length''' ) # Pair input self.assertRaises( A__ , tokenizer_r.batch_encode_plus , A__ , max_length=A__ , padding='''max_length''' , ) def UpperCamelCase ( self ) -> Tuple: snake_case = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token='''<pad>''' ) # Simple input snake_case = '''This is a simple input''' snake_case = ['''This is a simple input looooooooong''', '''This is a simple input'''] snake_case = ('''This is a simple input''', '''This is a pair''') snake_case = [ ('''This is a simple input loooooong''', '''This is a simple input'''), ('''This is a simple pair loooooong''', '''This is a simple pair'''), ] snake_case = tokenizer.pad_token_id snake_case = tokenizer(A__ , padding='''max_length''' , max_length=30 , return_tensors='''np''' ) snake_case = tokenizer(A__ , padding=A__ , truncate=A__ , return_tensors='''np''' ) snake_case = tokenizer(*A__ , padding='''max_length''' , max_length=60 , return_tensors='''np''' ) snake_case = tokenizer(A__ , padding=A__ , truncate=A__ , return_tensors='''np''' ) # s # test single string max_length padding self.assertEqual(out_s['''input_ids'''].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s['''input_ids'''] ) self.assertTrue(0 in out_s['''attention_mask'''] ) # s2 # test automatic padding self.assertEqual(out_sa['''input_ids'''].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['''input_ids'''][0] ) self.assertFalse(0 in out_sa['''attention_mask'''][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['''input_ids'''][1] ) self.assertTrue(0 in out_sa['''attention_mask'''][1] ) # p # test single pair max_length padding self.assertEqual(out_p['''input_ids'''].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p['''input_ids'''] ) self.assertTrue(0 in out_p['''attention_mask'''] ) # p2 # test automatic padding pair self.assertEqual(out_pa['''input_ids'''].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['''input_ids'''][0] ) self.assertFalse(0 in out_pa['''attention_mask'''][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['''input_ids'''][1] ) self.assertTrue(0 in out_pa['''attention_mask'''][1] ) def UpperCamelCase ( self ) -> str: snake_case = '''$$$''' snake_case = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=A__ , add_bos_token=A__ ) snake_case = '''This is a simple input''' snake_case = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case = tokenizer.bos_token_id snake_case = tokenizer(A__ ) snake_case = tokenizer(A__ ) self.assertEqual(out_s.input_ids[0] , A__ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) snake_case = tokenizer.decode(out_s.input_ids ) snake_case = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , A__ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def UpperCamelCase ( self ) -> Any: snake_case = CodeGenTokenizer.from_pretrained('''Salesforce/codegen-350M-mono''' ) snake_case = '''\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#''' snake_case = '''\nif len_a > len_b: result = a\nelse: result = b''' snake_case = tokenizer.encode(A__ ) snake_case = ['''^#''', re.escape('''<|endoftext|>''' ), '''^\'\'\'''', '''^"""''', '''\n\n\n'''] snake_case = tokenizer.decode(A__ , truncate_before_pattern=A__ ) self.assertEqual(A__ , A__ ) def UpperCamelCase ( self ) -> Union[str, Any]: pass
44
0
'''simple docstring''' from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available from ...utils import OptionalDependencyNotAvailable _lowercase = {'configuration_dpt': ['DPT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'DPTConfig']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ['DPTFeatureExtractor'] _lowercase = ['DPTImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ 'DPT_PRETRAINED_MODEL_ARCHIVE_LIST', 'DPTForDepthEstimation', 'DPTForSemanticSegmentation', 'DPTModel', 'DPTPreTrainedModel', ] if TYPE_CHECKING: from .configuration_dpt import DPT_PRETRAINED_CONFIG_ARCHIVE_MAP, DPTConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_dpt import DPTFeatureExtractor from .image_processing_dpt import DPTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dpt import ( DPT_PRETRAINED_MODEL_ARCHIVE_LIST, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel, DPTPreTrainedModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
707
'''simple docstring''' from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowercase : def __init__( self , A__ , A__=13 , A__=30 , A__=2 , A__=3 , A__=True , A__=True , A__=32 , A__=2 , A__=4 , A__=37 , A__="gelu" , A__=0.1 , A__=0.1 , A__=10 , A__=0.0_2 , A__=3 , A__=None , ) -> List[Any]: snake_case = parent snake_case = batch_size snake_case = image_size snake_case = patch_size snake_case = num_channels snake_case = is_training snake_case = use_labels snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_size snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = type_sequence_label_size snake_case = initializer_range snake_case = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) snake_case = (image_size // patch_size) ** 2 snake_case = num_patches + 1 def UpperCamelCase ( self ) -> int: snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case = None if self.use_labels: snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ) -> int: return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A__ , initializer_range=self.initializer_range , ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Union[str, Any]: snake_case = TFViTModel(config=A__ ) snake_case = model(A__ , training=A__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. snake_case = self.image_size // 2 snake_case = pixel_values[:, :, :image_size, :image_size] snake_case = model(A__ , interpolate_pos_encoding=A__ , training=A__ ) snake_case = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Optional[int]: snake_case = self.type_sequence_label_size snake_case = TFViTForImageClassification(A__ ) snake_case = model(A__ , labels=A__ , training=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. snake_case = self.image_size // 2 snake_case = pixel_values[:, :, :image_size, :image_size] snake_case = model(A__ , interpolate_pos_encoding=A__ , training=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images snake_case = 1 snake_case = TFViTForImageClassification(A__ ) snake_case = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case = model(A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.prepare_config_and_inputs() snake_case , snake_case , snake_case = config_and_inputs snake_case = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class _lowercase ( __a , __a , unittest.TestCase ): _UpperCAmelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () _UpperCAmelCase = ( {'''feature-extraction''': TFViTModel, '''image-classification''': TFViTForImageClassification} if is_tf_available() else {} ) _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False def UpperCamelCase ( self ) -> List[Any]: snake_case = TFViTModelTester(self ) snake_case = ConfigTester(self , config_class=A__ , has_text_modality=A__ , hidden_size=37 ) def UpperCamelCase ( self ) -> int: self.config_tester.run_common_tests() @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def UpperCamelCase ( self ) -> int: pass @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def UpperCamelCase ( self ) -> str: pass def UpperCamelCase ( self ) -> Union[str, Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(A__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) snake_case = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A__ , tf.keras.layers.Layer ) ) def UpperCamelCase ( self ) -> List[Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(A__ ) snake_case = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case = [*signature.parameters.keys()] snake_case = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A__ ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A__ ) def UpperCamelCase ( self ) -> Optional[Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A__ ) @slow def UpperCamelCase ( self ) -> Any: snake_case = TFViTModel.from_pretrained('''google/vit-base-patch16-224''' ) self.assertIsNotNone(A__ ) def __UpperCamelCase ( ) ->Any: snake_case = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class _lowercase ( unittest.TestCase ): @cached_property def UpperCamelCase ( self ) -> Optional[int]: return ViTImageProcessor.from_pretrained('''google/vit-base-patch16-224''' ) if is_vision_available() else None @slow def UpperCamelCase ( self ) -> Dict: snake_case = TFViTForImageClassification.from_pretrained('''google/vit-base-patch16-224''' ) snake_case = self.default_image_processor snake_case = prepare_img() snake_case = image_processor(images=A__ , return_tensors='''tf''' ) # forward pass snake_case = model(**A__ ) # verify the logits snake_case = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , A__ ) snake_case = tf.constant([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ) tf.debugging.assert_near(outputs.logits[0, :3] , A__ , atol=1e-4 )
44
0
'''simple docstring''' import inspect import unittest from transformers import YolosConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import YolosForObjectDetection, YolosModel from transformers.models.yolos.modeling_yolos import YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowercase : def __init__( self , A__ , A__=13 , A__=[30, 30] , A__=2 , A__=3 , A__=True , A__=True , A__=32 , A__=5 , A__=4 , A__=37 , A__="gelu" , A__=0.1 , A__=0.1 , A__=10 , A__=0.0_2 , A__=3 , A__=None , A__=8 , A__=10 , ) -> List[Any]: snake_case = parent snake_case = batch_size snake_case = image_size snake_case = patch_size snake_case = num_channels snake_case = is_training snake_case = use_labels snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_size snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = type_sequence_label_size snake_case = initializer_range snake_case = num_labels snake_case = scope snake_case = n_targets snake_case = num_detection_tokens # we set the expected sequence length (which is used in several tests) # expected sequence length = num_patches + 1 (we add 1 for the [CLS] token) + num_detection_tokens snake_case = (image_size[1] // patch_size) * (image_size[0] // patch_size) snake_case = num_patches + 1 + self.num_detection_tokens def UpperCamelCase ( self ) -> Any: snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size[0], self.image_size[1]] ) snake_case = None if self.use_labels: # labels is a list of Dict (each Dict being the labels for a given example in the batch) snake_case = [] for i in range(self.batch_size ): snake_case = {} snake_case = torch.randint( high=self.num_labels , size=(self.n_targets,) , device=_SCREAMING_SNAKE_CASE ) snake_case = torch.rand(self.n_targets , 4 , device=_SCREAMING_SNAKE_CASE ) labels.append(_SCREAMING_SNAKE_CASE ) snake_case = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ) -> List[Any]: return YolosConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , num_detection_tokens=self.num_detection_tokens , num_labels=self.num_labels , ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Tuple: snake_case = YolosModel(config=_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() snake_case = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.expected_seq_len, self.hidden_size) ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> int: snake_case = YolosForObjectDetection(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() snake_case = model(pixel_values=_SCREAMING_SNAKE_CASE ) snake_case = model(_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) snake_case = model(pixel_values=_SCREAMING_SNAKE_CASE , labels=_SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_detection_tokens, self.num_labels + 1) ) self.parent.assertEqual(result.pred_boxes.shape , (self.batch_size, self.num_detection_tokens, 4) ) def UpperCamelCase ( self ) -> List[Any]: snake_case = self.prepare_config_and_inputs() snake_case , snake_case , snake_case = config_and_inputs snake_case = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class _lowercase ( lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase ): _UpperCAmelCase = (YolosModel, YolosForObjectDetection) if is_torch_available() else () _UpperCAmelCase = ( {"feature-extraction": YolosModel, "object-detection": YolosForObjectDetection} if is_torch_available() else {} ) _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False def UpperCamelCase ( self , A__ , A__ , A__=False ) -> Any: snake_case = super()._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , return_labels=_SCREAMING_SNAKE_CASE ) if return_labels: if model_class.__name__ == "YolosForObjectDetection": snake_case = [] for i in range(self.model_tester.batch_size ): snake_case = {} snake_case = torch.ones( size=(self.model_tester.n_targets,) , device=_SCREAMING_SNAKE_CASE , dtype=torch.long ) snake_case = torch.ones( self.model_tester.n_targets , 4 , device=_SCREAMING_SNAKE_CASE , dtype=torch.float ) labels.append(_SCREAMING_SNAKE_CASE ) snake_case = labels return inputs_dict def UpperCamelCase ( self ) -> List[str]: snake_case = YolosModelTester(self ) snake_case = ConfigTester(self , config_class=_SCREAMING_SNAKE_CASE , has_text_modality=_SCREAMING_SNAKE_CASE , hidden_size=37 ) def UpperCamelCase ( self ) -> int: self.config_tester.run_common_tests() def UpperCamelCase ( self ) -> Any: # YOLOS does not use inputs_embeds pass def UpperCamelCase ( self ) -> Optional[Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(_SCREAMING_SNAKE_CASE ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) snake_case = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_SCREAMING_SNAKE_CASE , nn.Linear ) ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(_SCREAMING_SNAKE_CASE ) snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case = [*signature.parameters.keys()] snake_case = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _SCREAMING_SNAKE_CASE ) def UpperCamelCase ( self ) -> List[str]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_SCREAMING_SNAKE_CASE ) def UpperCamelCase ( self ) -> Optional[int]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() snake_case = True # in YOLOS, the seq_len is different snake_case = self.model_tester.expected_seq_len for model_class in self.all_model_classes: snake_case = True snake_case = False snake_case = True snake_case = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): snake_case = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) snake_case = outputs.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , self.model_tester.num_hidden_layers ) # check that output_attentions also work using config del inputs_dict["output_attentions"] snake_case = True snake_case = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): snake_case = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) snake_case = outputs.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) snake_case = len(_SCREAMING_SNAKE_CASE ) # Check attention is always last and order is fine snake_case = True snake_case = True snake_case = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): snake_case = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) snake_case = 1 self.assertEqual(out_len + added_hidden_states , len(_SCREAMING_SNAKE_CASE ) ) snake_case = outputs.attentions self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, seq_len, seq_len] , ) def UpperCamelCase ( self ) -> str: def check_hidden_states_output(A__ , A__ , A__ ): snake_case = model_class(_SCREAMING_SNAKE_CASE ) model.to(_SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): snake_case = model(**self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ) snake_case = outputs.hidden_states snake_case = getattr( self.model_tester , '''expected_num_hidden_layers''' , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , _SCREAMING_SNAKE_CASE ) # YOLOS has a different seq_length snake_case = self.model_tester.expected_seq_len self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case = True check_hidden_states_output(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def UpperCamelCase ( self ) -> str: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_object_detection(*_SCREAMING_SNAKE_CASE ) @slow def UpperCamelCase ( self ) -> List[str]: for model_name in YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case = YolosModel.from_pretrained(_SCREAMING_SNAKE_CASE ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE ) def __UpperCamelCase ( ) ->List[Any]: snake_case = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class _lowercase ( unittest.TestCase ): @cached_property def UpperCamelCase ( self ) -> int: return AutoImageProcessor.from_pretrained('''hustvl/yolos-small''' ) if is_vision_available() else None @slow def UpperCamelCase ( self ) -> Any: snake_case = YolosForObjectDetection.from_pretrained('''hustvl/yolos-small''' ).to(_SCREAMING_SNAKE_CASE ) snake_case = self.default_image_processor snake_case = prepare_img() snake_case = image_processor(images=_SCREAMING_SNAKE_CASE , return_tensors='''pt''' ).to(_SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): snake_case = model(inputs.pixel_values ) # verify outputs snake_case = torch.Size((1, 1_00, 92) ) self.assertEqual(outputs.logits.shape , _SCREAMING_SNAKE_CASE ) snake_case = torch.tensor( [[-24.02_48, -10.30_24, -14.82_90], [-42.03_92, -16.82_00, -27.43_34], [-27.27_43, -11.81_54, -18.71_48]] , device=_SCREAMING_SNAKE_CASE , ) snake_case = torch.tensor( [[0.2_5_5_9, 0.5_4_5_5, 0.4_7_0_6], [0.2_9_8_9, 0.7_2_7_9, 0.1_8_7_5], [0.7_7_3_2, 0.4_0_1_7, 0.4_4_6_2]] , device=_SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ) ) self.assertTrue(torch.allclose(outputs.pred_boxes[0, :3, :3] , _SCREAMING_SNAKE_CASE , atol=1e-4 ) ) # verify postprocessing snake_case = image_processor.post_process_object_detection( _SCREAMING_SNAKE_CASE , threshold=0.3 , target_sizes=[image.size[::-1]] )[0] snake_case = torch.tensor([0.9_9_9_4, 0.9_7_9_0, 0.9_9_6_4, 0.9_9_7_2, 0.9_8_6_1] ).to(_SCREAMING_SNAKE_CASE ) snake_case = [75, 75, 17, 63, 17] snake_case = torch.tensor([3_35.06_09, 79.38_48, 3_75.42_16, 1_87.24_95] ).to(_SCREAMING_SNAKE_CASE ) self.assertEqual(len(results['''scores'''] ) , 5 ) self.assertTrue(torch.allclose(results['''scores'''] , _SCREAMING_SNAKE_CASE , atol=1e-4 ) ) self.assertSequenceEqual(results['''labels'''].tolist() , _SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(results['''boxes'''][0, :] , _SCREAMING_SNAKE_CASE ) )
708
'''simple docstring''' import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from datasets.iterable_dataset import IterableDataset from datasets.load import dataset_module_factory, import_main_class from datasets.utils.file_utils import cached_path _lowercase = [ {'dataset': 'wikipedia', 'config_name': '20220301.de'}, {'dataset': 'wikipedia', 'config_name': '20220301.en'}, {'dataset': 'wikipedia', 'config_name': '20220301.fr'}, {'dataset': 'wikipedia', 'config_name': '20220301.frr'}, {'dataset': 'wikipedia', 'config_name': '20220301.it'}, {'dataset': 'wikipedia', 'config_name': '20220301.simple'}, {'dataset': 'snli', 'config_name': 'plain_text'}, {'dataset': 'eli5', 'config_name': 'LFQA_reddit'}, {'dataset': 'wiki40b', 'config_name': 'en'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.compressed'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.no_index'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.multiset.no_index'}, {'dataset': 'natural_questions', 'config_name': 'default'}, ] def __UpperCamelCase ( a : Dict=True ) ->str: if with_config: return [ { "testcase_name": d["dataset"] + "/" + d["config_name"], "dataset": d["dataset"], "config_name": d["config_name"], } for d in DATASETS_ON_HF_GCP ] else: return [ {"testcase_name": dataset, "dataset": dataset} for dataset in {d["dataset"] for d in DATASETS_ON_HF_GCP} ] @parameterized.named_parameters(list_datasets_on_hf_gcp_parameters(with_config=__a ) ) class _lowercase ( __a ): _UpperCAmelCase = None _UpperCAmelCase = None def UpperCamelCase ( self , A__ , A__ ) -> str: with TemporaryDirectory() as tmp_dir: snake_case = dataset_module_factory(A__ , cache_dir=A__ ) snake_case = import_main_class(dataset_module.module_path , dataset=A__ ) snake_case = builder_cls( cache_dir=A__ , config_name=A__ , hash=dataset_module.hash , ) snake_case = '''/'''.join( [ HF_GCP_BASE_URL, builder_instance._relative_data_dir(with_hash=A__ ).replace(os.sep , '''/''' ), config.DATASET_INFO_FILENAME, ] ) snake_case = cached_path(A__ , cache_dir=A__ ) self.assertTrue(os.path.exists(A__ ) ) @pytest.mark.integration def __UpperCamelCase ( a : List[str] ) ->Any: snake_case = tmp_path_factory.mktemp('''test_hf_gcp''' ) / '''test_wikipedia_simple''' snake_case = dataset_module_factory('''wikipedia''' , cache_dir=a ) snake_case = import_main_class(dataset_module.module_path ) snake_case = builder_cls( cache_dir=a , config_name='''20220301.frr''' , hash=dataset_module.hash , ) # use the HF cloud storage, not the original download_and_prepare that uses apache-beam snake_case = None builder_instance.download_and_prepare() snake_case = builder_instance.as_dataset() assert ds @pytest.mark.integration def __UpperCamelCase ( a : Any ) ->Union[str, Any]: snake_case = dataset_module_factory('''wikipedia''' , cache_dir=a ) snake_case = import_main_class(dataset_module.module_path , dataset=a ) snake_case = builder_cls( cache_dir=a , config_name='''20220301.frr''' , hash=dataset_module.hash , ) snake_case = builder_instance.as_streaming_dataset() assert ds assert isinstance(a , a ) assert "train" in ds assert isinstance(ds['''train'''] , a ) assert next(iter(ds['''train'''] ) )
44
0
'''simple docstring''' import copy import random from transformers import CLIPTokenizer class _lowercase ( _A ): def __init__( self , *A__ , **A__ ) -> Any: super().__init__(*UpperCamelCase__ , **UpperCamelCase__ ) snake_case = {} def UpperCamelCase ( self , A__ , *A__ , **A__ ) -> int: snake_case = super().add_tokens(UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ) if num_added_tokens == 0: raise ValueError( F"""The tokenizer already contains the token {placeholder_token}. Please pass a different""" ''' `placeholder_token` that is not already in the tokenizer.''' ) def UpperCamelCase ( self , A__ , *A__ , A__=1 , **A__ ) -> str: snake_case = [] if num_vec_per_token == 1: self.try_adding_tokens(UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ) output.append(UpperCamelCase__ ) else: snake_case = [] for i in range(UpperCamelCase__ ): snake_case = placeholder_token + F"""_{i}""" self.try_adding_tokens(UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ) output.append(UpperCamelCase__ ) # handle cases where there is a new placeholder token that contains the current placeholder token but is larger for token in self.token_map: if token in placeholder_token: raise ValueError( F"""The tokenizer already has placeholder token {token} that can get confused with""" F""" {placeholder_token}keep placeholder tokens independent""" ) snake_case = output def UpperCamelCase ( self , A__ , A__=False , A__=1.0 ) -> Optional[int]: if isinstance(UpperCamelCase__ , UpperCamelCase__ ): snake_case = [] for i in range(len(UpperCamelCase__ ) ): output.append(self.replace_placeholder_tokens_in_text(text[i] , vector_shuffle=UpperCamelCase__ ) ) return output for placeholder_token in self.token_map: if placeholder_token in text: snake_case = self.token_map[placeholder_token] snake_case = tokens[: 1 + int(len(UpperCamelCase__ ) * prop_tokens_to_load )] if vector_shuffle: snake_case = copy.copy(UpperCamelCase__ ) random.shuffle(UpperCamelCase__ ) snake_case = text.replace(UpperCamelCase__ , ''' '''.join(UpperCamelCase__ ) ) return text def __call__( self , A__ , *A__ , A__=False , A__=1.0 , **A__ ) -> List[str]: return super().__call__( self.replace_placeholder_tokens_in_text( UpperCamelCase__ , vector_shuffle=UpperCamelCase__ , prop_tokens_to_load=UpperCamelCase__ ) , *UpperCamelCase__ , **UpperCamelCase__ , ) def UpperCamelCase ( self , A__ , *A__ , A__=False , A__=1.0 , **A__ ) -> Optional[Any]: return super().encode( self.replace_placeholder_tokens_in_text( UpperCamelCase__ , vector_shuffle=UpperCamelCase__ , prop_tokens_to_load=UpperCamelCase__ ) , *UpperCamelCase__ , **UpperCamelCase__ , )
709
'''simple docstring''' def __UpperCamelCase ( a : int , a : int ) ->int: while b: snake_case , snake_case = b, a % b return a def __UpperCamelCase ( a : int , a : int ) ->int: return a if b == 0 else euclidean_gcd_recursive(a , a % b ) def __UpperCamelCase ( ) ->Optional[Any]: print(f"""euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}""" ) print(f"""euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}""" ) print(f"""euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}""" ) print(f"""euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}""" ) print(f"""euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}""" ) print(f"""euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}""" ) print(f"""euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}""" ) print(f"""euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}""" ) if __name__ == "__main__": main()
44
0
'''simple docstring''' import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors _lowercase = logging.getLogger(__name__) class _lowercase ( snake_case__ ): _UpperCAmelCase = '''sequence-classification''' def __init__( self , A__ ) -> Any: if type(_A ) == dict: snake_case = Namespace(**_A ) snake_case = glue_output_modes[hparams.task] snake_case = glue_tasks_num_labels[hparams.task] super().__init__(_A , _A , self.mode ) def UpperCamelCase ( self , **A__ ) -> Optional[Any]: return self.model(**_A ) def UpperCamelCase ( self , A__ , A__ ) -> Union[str, Any]: snake_case = {'input_ids': batch[0], 'attention_mask': batch[1], 'labels': batch[3]} if self.config.model_type not in ["distilbert", "bart"]: snake_case = batch[2] if self.config.model_type in ['bert', 'xlnet', 'albert'] else None snake_case = self(**_A ) snake_case = outputs[0] snake_case = self.trainer.lr_schedulers[0]['scheduler'] snake_case = {'loss': loss, 'rate': lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def UpperCamelCase ( self ) -> Optional[Any]: snake_case = self.hparams snake_case = processors[args.task]() snake_case = processor.get_labels() for mode in ["train", "dev"]: snake_case = self._feature_file(_A ) if os.path.exists(_A ) and not args.overwrite_cache: logger.info('''Loading features from cached file %s''' , _A ) else: logger.info('''Creating features from dataset file at %s''' , args.data_dir ) snake_case = ( processor.get_dev_examples(args.data_dir ) if mode == 'dev' else processor.get_train_examples(args.data_dir ) ) snake_case = convert_examples_to_features( _A , self.tokenizer , max_length=args.max_seq_length , label_list=self.labels , output_mode=args.glue_output_mode , ) logger.info('''Saving features into cached file %s''' , _A ) torch.save(_A , _A ) def UpperCamelCase ( self , A__ , A__ , A__ = False ) -> Tuple: snake_case = 'dev' if mode == 'test' else mode snake_case = self._feature_file(_A ) logger.info('''Loading features from cached file %s''' , _A ) snake_case = torch.load(_A ) snake_case = torch.tensor([f.input_ids for f in features] , dtype=torch.long ) snake_case = torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) snake_case = torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) if self.hparams.glue_output_mode == "classification": snake_case = torch.tensor([f.label for f in features] , dtype=torch.long ) elif self.hparams.glue_output_mode == "regression": snake_case = torch.tensor([f.label for f in features] , dtype=torch.float ) return DataLoader( TensorDataset(_A , _A , _A , _A ) , batch_size=_A , shuffle=_A , ) def UpperCamelCase ( self , A__ , A__ ) -> List[str]: snake_case = {'input_ids': batch[0], 'attention_mask': batch[1], 'labels': batch[3]} if self.config.model_type not in ["distilbert", "bart"]: snake_case = batch[2] if self.config.model_type in ['bert', 'xlnet', 'albert'] else None snake_case = self(**_A ) snake_case = outputs[:2] snake_case = logits.detach().cpu().numpy() snake_case = inputs['labels'].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def UpperCamelCase ( self , A__ ) -> Any: snake_case = torch.stack([x['''val_loss'''] for x in outputs] ).mean().detach().cpu().item() snake_case = np.concatenate([x['''pred'''] for x in outputs] , axis=0 ) if self.hparams.glue_output_mode == "classification": snake_case = np.argmax(_A , axis=1 ) elif self.hparams.glue_output_mode == "regression": snake_case = np.squeeze(_A ) snake_case = np.concatenate([x['''target'''] for x in outputs] , axis=0 ) snake_case = [[] for _ in range(out_label_ids.shape[0] )] snake_case = [[] for _ in range(out_label_ids.shape[0] )] snake_case = {**{'val_loss': val_loss_mean}, **compute_metrics(self.hparams.task , _A , _A )} snake_case = dict(results.items() ) snake_case = results return ret, preds_list, out_label_list def UpperCamelCase ( self , A__ ) -> Optional[Any]: snake_case = self._eval_end(_A ) snake_case = ret['log'] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def UpperCamelCase ( self , A__ ) -> Tuple: snake_case = self._eval_end(_A ) snake_case = ret['log'] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def UpperCamelCase ( A__ , A__ ) -> Any: BaseTransformer.add_model_specific_args(_A , _A ) parser.add_argument( '''--max_seq_length''' , default=1_28 , type=_A , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--task''' , default='''''' , type=_A , required=_A , help='''The GLUE task to run''' , ) parser.add_argument( '''--gpus''' , default=0 , type=_A , help='''The number of GPUs allocated for this, it is by default 0 meaning none''' , ) parser.add_argument( '''--overwrite_cache''' , action='''store_true''' , help='''Overwrite the cached training and evaluation sets''' ) return parser def __UpperCamelCase ( ) ->Tuple: snake_case = argparse.ArgumentParser() add_generic_args(a , os.getcwd() ) snake_case = GLUETransformer.add_model_specific_args(a , os.getcwd() ) snake_case = parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: snake_case = os.path.join( '''./results''' , f"""{args.task}_{time.strftime('%Y%m%d_%H%M%S' )}""" , ) os.makedirs(args.output_dir ) snake_case = GLUETransformer(a ) snake_case = generic_train(a , a ) # Optionally, predict on dev set and write to output_dir if args.do_predict: snake_case = sorted(glob.glob(os.path.join(args.output_dir , '''checkpoint-epoch=*.ckpt''' ) , recursive=a ) ) snake_case = model.load_from_checkpoint(checkpoints[-1] ) return trainer.test(a ) if __name__ == "__main__": main()
710
'''simple docstring''' import argparse import copy def __UpperCamelCase ( a : Union[str, Any] ) ->Tuple: snake_case = {} with open(a ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: snake_case = [] _list.append([line.split()[1], line.split()[2]] ) snake_case = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: snake_case = [] _list.append([line.split()[0], line.split()[2]] ) snake_case = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def __UpperCamelCase ( a : Dict , a : Tuple ) ->int: with open(a ) as f: snake_case = f.read(1 ) snake_case = start_node snake_case = [] snake_case = start_node snake_case = 0 while visiting not in first_solution: snake_case = 1_0000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(a ) and k[0] not in first_solution: snake_case = k[1] snake_case = k[0] first_solution.append(a ) snake_case = distance_of_first_solution + int(a ) snake_case = best_node first_solution.append(a ) snake_case = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 snake_case = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_0000 ) return first_solution, distance_of_first_solution def __UpperCamelCase ( a : Optional[int] , a : str ) ->str: snake_case = [] for n in solution[1:-1]: snake_case = solution.index(a ) for kn in solution[1:-1]: snake_case = solution.index(a ) if n == kn: continue snake_case = copy.deepcopy(a ) snake_case = kn snake_case = n snake_case = 0 for k in _tmp[:-1]: snake_case = _tmp[_tmp.index(a ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: snake_case = distance + int(i[1] ) _tmp.append(a ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) snake_case = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda a : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def __UpperCamelCase ( a : Any , a : Optional[Any] , a : int , a : Optional[int] , a : Union[str, Any] ) ->List[Any]: snake_case = 1 snake_case = first_solution snake_case = [] snake_case = distance_of_first_solution snake_case = solution while count <= iters: snake_case = find_neighborhood(a , a ) snake_case = 0 snake_case = neighborhood[index_of_best_solution] snake_case = len(a ) - 1 snake_case = False while not found: snake_case = 0 while i < len(a ): if best_solution[i] != solution[i]: snake_case = best_solution[i] snake_case = solution[i] break snake_case = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) snake_case = True snake_case = best_solution[:-1] snake_case = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: snake_case = cost snake_case = solution else: snake_case = index_of_best_solution + 1 snake_case = neighborhood[index_of_best_solution] if len(a ) >= size: tabu_list.pop(0 ) snake_case = count + 1 return best_solution_ever, best_cost def __UpperCamelCase ( a : Union[str, Any]=None ) ->Optional[Any]: snake_case = generate_neighbours(args.File ) snake_case , snake_case = generate_first_solution( args.File , a ) snake_case , snake_case = tabu_search( a , a , a , args.Iterations , args.Size , ) print(f"""Best solution: {best_sol}, with total distance: {best_cost}.""" ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser(description='Tabu Search') parser.add_argument( '-f', '--File', type=str, help='Path to the file containing the data', required=True, ) parser.add_argument( '-i', '--Iterations', type=int, help='How many iterations the algorithm should perform', required=True, ) parser.add_argument( '-s', '--Size', type=int, help='Size of the tabu list', required=True ) # Pass the arguments to main method main(parser.parse_args())
44
0
'''simple docstring''' def __UpperCamelCase ( a : Optional[Any] ) ->None: snake_case = generate_pascal_triangle(__snake_case ) for row_idx in range(__snake_case ): # Print left spaces for _ in range(num_rows - row_idx - 1 ): print(end=''' ''' ) # Print row values for col_idx in range(row_idx + 1 ): if col_idx != row_idx: print(triangle[row_idx][col_idx] , end=''' ''' ) else: print(triangle[row_idx][col_idx] , end='''''' ) print() def __UpperCamelCase ( a : int ) ->list[list[int]]: if not isinstance(__snake_case , __snake_case ): raise TypeError('''The input value of \'num_rows\' should be \'int\'''' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( '''The input value of \'num_rows\' should be greater than or equal to 0''' ) snake_case = [] for current_row_idx in range(__snake_case ): snake_case = populate_current_row(__snake_case , __snake_case ) triangle.append(__snake_case ) return triangle def __UpperCamelCase ( a : Any , a : str ) ->list[int]: snake_case = [-1] * (current_row_idx + 1) # first and last elements of current row are equal to 1 snake_case , snake_case = 1, 1 for current_col_idx in range(1 , __snake_case ): calculate_current_element( __snake_case , __snake_case , __snake_case , __snake_case ) return current_row def __UpperCamelCase ( a : Dict , a : Union[str, Any] , a : Tuple , a : List[str] , ) ->None: snake_case = triangle[current_row_idx - 1][current_col_idx - 1] snake_case = triangle[current_row_idx - 1][current_col_idx] snake_case = above_to_left_elt + above_to_right_elt def __UpperCamelCase ( a : Tuple ) ->list[list[int]]: if not isinstance(__snake_case , __snake_case ): raise TypeError('''The input value of \'num_rows\' should be \'int\'''' ) if num_rows == 0: return [] elif num_rows < 0: raise ValueError( '''The input value of \'num_rows\' should be greater than or equal to 0''' ) snake_case = [[1]] for row_index in range(1 , __snake_case ): snake_case = [0] + result[-1] + [0] snake_case = row_index + 1 # Calculate the number of distinct elements in a row snake_case = sum(divmod(__snake_case , 2 ) ) snake_case = [ temp_row[i - 1] + temp_row[i] for i in range(1 , distinct_elements + 1 ) ] snake_case = row_first_half[: (row_index + 1) // 2] row_second_half.reverse() snake_case = row_first_half + row_second_half result.append(__snake_case ) return result def __UpperCamelCase ( ) ->None: from collections.abc import Callable from timeit import timeit def benchmark_a_function(a : str , a : int ) -> None: snake_case = f"""{func.__name__}({value})""" snake_case = timeit(f"""__main__.{call}""" , setup='''import __main__''' ) # print(f"{call:38} = {func(value)} -- {timing:.4f} seconds") print(f"""{call:38} -- {timing:.4f} seconds""" ) for value in range(15 ): # (1, 7, 14): for func in (generate_pascal_triangle, generate_pascal_triangle_optimized): benchmark_a_function(__snake_case , __snake_case ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
711
'''simple docstring''' from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .notes_encoder import SpectrogramNotesEncoder from .continous_encoder import SpectrogramContEncoder from .pipeline_spectrogram_diffusion import ( SpectrogramContEncoder, SpectrogramDiffusionPipeline, TaFilmDecoder, ) try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .midi_utils import MidiProcessor
44
0
import json import sys def __UpperCamelCase ( a : Optional[int] , a : Optional[Any] ) ->Optional[Any]: with open(lowerCamelCase_ , encoding='''utf-8''' ) as f: snake_case = json.load(lowerCamelCase_ ) snake_case = ["""<details>""", """<summary>Show updated benchmarks!</summary>""", """ """] for benchmark_name in sorted(lowerCamelCase_ ): snake_case = results[benchmark_name] snake_case = benchmark_name.split('''/''' )[-1] output_md.append(f"""### Benchmark: {benchmark_file_name}""" ) snake_case = """| metric |""" snake_case = """|--------|""" snake_case = """| new / old (diff) |""" for metric_name in sorted(lowerCamelCase_ ): snake_case = benchmark_res[metric_name] snake_case = metric_vals["""new"""] snake_case = metric_vals.get('''old''' , lowerCamelCase_ ) snake_case = metric_vals.get('''diff''' , lowerCamelCase_ ) snake_case = f""" {new_val:f}""" if isinstance(lowerCamelCase_ , (int, float) ) else """None""" if old_val is not None: val_str += f""" / {old_val:f}""" if isinstance(lowerCamelCase_ , (int, float) ) else "None" if dif_val is not None: val_str += f""" ({dif_val:f})""" if isinstance(lowerCamelCase_ , (int, float) ) else "None" title += " " + metric_name + " |" lines += "---|" value += val_str + " |" output_md += [title, lines, value, " "] output_md.append('''</details>''' ) with open(lowerCamelCase_ , '''w''' , encoding='''utf-8''' ) as f: f.writelines('''\n'''.join(lowerCamelCase_ ) ) if __name__ == "__main__": _lowercase = sys.argv[1] _lowercase = sys.argv[2] format_json_to_md(input_json_file, output_md_file)
712
'''simple docstring''' from ...processing_utils import ProcessorMixin class _lowercase ( __a ): _UpperCAmelCase = '''WhisperFeatureExtractor''' _UpperCAmelCase = '''WhisperTokenizer''' def __init__( self , A__ , A__ ) -> Optional[Any]: super().__init__(A__ , A__ ) snake_case = self.feature_extractor snake_case = False def UpperCamelCase ( self , A__=None , A__=None , A__=True ) -> Union[str, Any]: return self.tokenizer.get_decoder_prompt_ids(task=A__ , language=A__ , no_timestamps=A__ ) def __call__( self , *A__ , **A__ ) -> Dict: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*A__ , **A__ ) snake_case = kwargs.pop('''audio''' , A__ ) snake_case = kwargs.pop('''sampling_rate''' , A__ ) snake_case = kwargs.pop('''text''' , A__ ) if len(A__ ) > 0: snake_case = args[0] snake_case = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if audio is not None: snake_case = self.feature_extractor(A__ , *A__ , sampling_rate=A__ , **A__ ) if text is not None: snake_case = self.tokenizer(A__ , **A__ ) if text is None: return inputs elif audio is None: return encodings else: snake_case = encodings['''input_ids'''] return inputs def UpperCamelCase ( self , *A__ , **A__ ) -> Optional[Any]: return self.tokenizer.batch_decode(*A__ , **A__ ) def UpperCamelCase ( self , *A__ , **A__ ) -> str: return self.tokenizer.decode(*A__ , **A__ ) def UpperCamelCase ( self , A__ , A__="np" ) -> Optional[Any]: return self.tokenizer.get_prompt_ids(A__ , return_tensors=A__ )
44
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _lowercase = { 'configuration_rembert': ['REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RemBertConfig', 'RemBertOnnxConfig'] } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ['RemBertTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ['RemBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ 'REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'RemBertForCausalLM', 'RemBertForMaskedLM', 'RemBertForMultipleChoice', 'RemBertForQuestionAnswering', 'RemBertForSequenceClassification', 'RemBertForTokenClassification', 'RemBertLayer', 'RemBertModel', 'RemBertPreTrainedModel', 'load_tf_weights_in_rembert', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ 'TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFRemBertForCausalLM', 'TFRemBertForMaskedLM', 'TFRemBertForMultipleChoice', 'TFRemBertForQuestionAnswering', 'TFRemBertForSequenceClassification', 'TFRemBertForTokenClassification', 'TFRemBertLayer', 'TFRemBertModel', 'TFRemBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_rembert import REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RemBertConfig, RemBertOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert import RemBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert_fast import RemBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rembert import ( REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, RemBertForCausalLM, RemBertForMaskedLM, RemBertForMultipleChoice, RemBertForQuestionAnswering, RemBertForSequenceClassification, RemBertForTokenClassification, RemBertLayer, RemBertModel, RemBertPreTrainedModel, load_tf_weights_in_rembert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rembert import ( TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFRemBertForCausalLM, TFRemBertForMaskedLM, TFRemBertForMultipleChoice, TFRemBertForQuestionAnswering, TFRemBertForSequenceClassification, TFRemBertForTokenClassification, TFRemBertLayer, TFRemBertModel, TFRemBertPreTrainedModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
713
'''simple docstring''' import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class _lowercase ( __a ): _UpperCAmelCase = '''char''' _UpperCAmelCase = '''bpe''' _UpperCAmelCase = '''wp''' _lowercase = (DecodeType.CHARACTER, DecodeType.BPE, DecodeType.WORDPIECE) class _lowercase ( __a ): _UpperCAmelCase = ['''image_processor''', '''char_tokenizer'''] _UpperCAmelCase = '''ViTImageProcessor''' _UpperCAmelCase = '''MgpstrTokenizer''' def __init__( self , A__=None , A__=None , **A__ ) -> List[Any]: snake_case = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , A__ , ) snake_case = kwargs.pop('''feature_extractor''' ) snake_case = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) snake_case = tokenizer snake_case = AutoTokenizer.from_pretrained('''gpt2''' ) snake_case = AutoTokenizer.from_pretrained('''bert-base-uncased''' ) super().__init__(A__ , A__ ) def __call__( self , A__=None , A__=None , A__=None , **A__ ) -> List[str]: if images is None and text is None: raise ValueError('''You need to specify either an `images` or `text` input to process.''' ) if images is not None: snake_case = self.image_processor(A__ , return_tensors=A__ , **A__ ) if text is not None: snake_case = self.char_tokenizer(A__ , return_tensors=A__ , **A__ ) if text is None: return inputs elif images is None: return encodings else: snake_case = encodings['''input_ids'''] return inputs def UpperCamelCase ( self , A__ ) -> Dict: snake_case , snake_case , snake_case = sequences snake_case = char_preds.size(0 ) snake_case , snake_case = self._decode_helper(A__ , '''char''' ) snake_case , snake_case = self._decode_helper(A__ , '''bpe''' ) snake_case , snake_case = self._decode_helper(A__ , '''wp''' ) snake_case = [] snake_case = [] for i in range(A__ ): snake_case = [char_scores[i], bpe_scores[i], wp_scores[i]] snake_case = [char_strs[i], bpe_strs[i], wp_strs[i]] snake_case = scores.index(max(A__ ) ) final_strs.append(strs[max_score_index] ) final_scores.append(scores[max_score_index] ) snake_case = {} snake_case = final_strs snake_case = final_scores snake_case = char_strs snake_case = bpe_strs snake_case = wp_strs return out def UpperCamelCase ( self , A__ , A__ ) -> Optional[Any]: if format == DecodeType.CHARACTER: snake_case = self.char_decode snake_case = 1 snake_case = '''[s]''' elif format == DecodeType.BPE: snake_case = self.bpe_decode snake_case = 2 snake_case = '''#''' elif format == DecodeType.WORDPIECE: snake_case = self.wp_decode snake_case = 1_02 snake_case = '''[SEP]''' else: raise ValueError(F"""Format {format} is not supported.""" ) snake_case , snake_case = [], [] snake_case = pred_logits.size(0 ) snake_case = pred_logits.size(1 ) snake_case , snake_case = pred_logits.topk(1 , dim=-1 , largest=A__ , sorted=A__ ) snake_case = preds_index.view(-1 , A__ )[:, 1:] snake_case = decoder(A__ ) snake_case , snake_case = torch.nn.functional.softmax(A__ , dim=2 ).max(dim=2 ) snake_case = preds_max_prob[:, 1:] for index in range(A__ ): snake_case = preds_str[index].find(A__ ) snake_case = preds_str[index][:pred_eos] snake_case = preds_index[index].cpu().tolist() snake_case = pred_index.index(A__ ) if eos_token in pred_index else -1 snake_case = preds_max_prob[index][: pred_eos_index + 1] snake_case = pred_max_prob.cumprod(dim=0 )[-1] if pred_max_prob.nelement() != 0 else 0.0 dec_strs.append(A__ ) conf_scores.append(A__ ) return dec_strs, conf_scores def UpperCamelCase ( self , A__ ) -> int: snake_case = [seq.replace(''' ''' , '''''' ) for seq in self.char_tokenizer.batch_decode(A__ )] return decode_strs def UpperCamelCase ( self , A__ ) -> List[str]: return self.bpe_tokenizer.batch_decode(A__ ) def UpperCamelCase ( self , A__ ) -> Union[str, Any]: snake_case = [seq.replace(''' ''' , '''''' ) for seq in self.wp_tokenizer.batch_decode(A__ )] return decode_strs
44
0
'''simple docstring''' from __future__ import annotations def __UpperCamelCase ( a : Any ) ->Tuple: snake_case = str(lowerCamelCase__ ) return n == n[::-1] def __UpperCamelCase ( a : Tuple = 100_0000 ) ->Tuple: snake_case = 0 for i in range(1 , lowerCamelCase__ ): if is_palindrome(lowerCamelCase__ ) and is_palindrome(bin(lowerCamelCase__ ).split('''b''' )[1] ): total += i return total if __name__ == "__main__": print(solution(int(str(input().strip()))))
714
'''simple docstring''' import os from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen, xsplitext from ..table import array_cast from ..utils.py_utils import no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: from .features import FeatureType _lowercase , _lowercase , _lowercase = False, False, False @dataclass class _lowercase : _UpperCAmelCase = None _UpperCAmelCase = True _UpperCAmelCase = True _UpperCAmelCase = None # Automatically constructed _UpperCAmelCase = "dict" _UpperCAmelCase = pa.struct({'''bytes''': pa.binary(), '''path''': pa.string()} ) _UpperCAmelCase = field(default='''Audio''' , init=__a , repr=__a ) def __call__( self ) -> Optional[Any]: return self.pa_type def UpperCamelCase ( self , A__ ) -> dict: try: import soundfile as sf # soundfile is a dependency of librosa, needed to decode audio files. except ImportError as err: raise ImportError('''To support encoding audio data, please install \'soundfile\'.''' ) from err if isinstance(A__ , A__ ): return {"bytes": None, "path": value} elif isinstance(A__ , A__ ): return {"bytes": value, "path": None} elif "array" in value: # convert the audio array to wav bytes snake_case = BytesIO() sf.write(A__ , value['''array'''] , value['''sampling_rate'''] , format='''wav''' ) return {"bytes": buffer.getvalue(), "path": None} elif value.get('''path''' ) is not None and os.path.isfile(value['''path'''] ): # we set "bytes": None to not duplicate the data if they're already available locally if value["path"].endswith('''pcm''' ): # "PCM" only has raw audio bytes if value.get('''sampling_rate''' ) is None: # At least, If you want to convert "PCM-byte" to "WAV-byte", you have to know sampling rate raise KeyError('''To use PCM files, please specify a \'sampling_rate\' in Audio object''' ) if value.get('''bytes''' ): # If we already had PCM-byte, we don`t have to make "read file, make bytes" (just use it!) snake_case = np.frombuffer(value['''bytes'''] , dtype=np.intaa ).astype(np.floataa ) / 3_27_67 else: snake_case = np.memmap(value['''path'''] , dtype='''h''' , mode='''r''' ).astype(np.floataa ) / 3_27_67 snake_case = BytesIO(bytes() ) sf.write(A__ , A__ , value['''sampling_rate'''] , format='''wav''' ) return {"bytes": buffer.getvalue(), "path": None} else: return {"bytes": None, "path": value.get('''path''' )} elif value.get('''bytes''' ) is not None or value.get('''path''' ) is not None: # store the audio bytes, and path is used to infer the audio format using the file extension return {"bytes": value.get('''bytes''' ), "path": value.get('''path''' )} else: raise ValueError( F"""An audio sample should have one of 'path' or 'bytes' but they are missing or None in {value}.""" ) def UpperCamelCase ( self , A__ , A__ = None ) -> dict: if not self.decode: raise RuntimeError('''Decoding is disabled for this feature. Please use Audio(decode=True) instead.''' ) snake_case , snake_case = (value['''path'''], BytesIO(value['''bytes'''] )) if value['''bytes'''] is not None else (value['''path'''], None) if path is None and file is None: raise ValueError(F"""An audio sample should have one of 'path' or 'bytes' but both are None in {value}.""" ) try: import librosa import soundfile as sf except ImportError as err: raise ImportError('''To support decoding audio files, please install \'librosa\' and \'soundfile\'.''' ) from err snake_case = xsplitext(A__ )[1][1:].lower() if path is not None else None if not config.IS_OPUS_SUPPORTED and audio_format == "opus": raise RuntimeError( '''Decoding \'opus\' files requires system library \'libsndfile\'>=1.0.31, ''' '''You can try to update `soundfile` python library: `pip install "soundfile>=0.12.1"`. ''' ) elif not config.IS_MP3_SUPPORTED and audio_format == "mp3": raise RuntimeError( '''Decoding \'mp3\' files requires system library \'libsndfile\'>=1.1.0, ''' '''You can try to update `soundfile` python library: `pip install "soundfile>=0.12.1"`. ''' ) if file is None: snake_case = token_per_repo_id or {} snake_case = path.split('''::''' )[-1] try: snake_case = string_to_dict(A__ , config.HUB_DATASETS_URL )['''repo_id'''] snake_case = token_per_repo_id[repo_id] except (ValueError, KeyError): snake_case = None with xopen(A__ , '''rb''' , use_auth_token=A__ ) as f: snake_case , snake_case = sf.read(A__ ) else: snake_case , snake_case = sf.read(A__ ) snake_case = array.T if self.mono: snake_case = librosa.to_mono(A__ ) if self.sampling_rate and self.sampling_rate != sampling_rate: snake_case = librosa.resample(A__ , orig_sr=A__ , target_sr=self.sampling_rate ) snake_case = self.sampling_rate return {"path": path, "array": array, "sampling_rate": sampling_rate} def UpperCamelCase ( self ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value if self.decode: raise ValueError('''Cannot flatten a decoded Audio feature.''' ) return { "bytes": Value('''binary''' ), "path": Value('''string''' ), } def UpperCamelCase ( self , A__ ) -> pa.StructArray: if pa.types.is_string(storage.type ): snake_case = pa.array([None] * len(A__ ) , type=pa.binary() ) snake_case = pa.StructArray.from_arrays([bytes_array, storage] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): snake_case = pa.array([None] * len(A__ ) , type=pa.string() ) snake_case = pa.StructArray.from_arrays([storage, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ) and storage.type.get_all_field_indices('''array''' ): snake_case = pa.array([Audio().encode_example(A__ ) if x is not None else None for x in storage.to_pylist()] ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index('''bytes''' ) >= 0: snake_case = storage.field('''bytes''' ) else: snake_case = pa.array([None] * len(A__ ) , type=pa.binary() ) if storage.type.get_field_index('''path''' ) >= 0: snake_case = storage.field('''path''' ) else: snake_case = pa.array([None] * len(A__ ) , type=pa.string() ) snake_case = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null() ) return array_cast(A__ , self.pa_type ) def UpperCamelCase ( self , A__ ) -> pa.StructArray: @no_op_if_value_is_null def path_to_bytes(A__ ): with xopen(A__ , '''rb''' ) as f: snake_case = f.read() return bytes_ snake_case = pa.array( [ (path_to_bytes(x['''path'''] ) if x['''bytes'''] is None else x['''bytes''']) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) snake_case = pa.array( [os.path.basename(A__ ) if path is not None else None for path in storage.field('''path''' ).to_pylist()] , type=pa.string() , ) snake_case = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=bytes_array.is_null() ) return array_cast(A__ , self.pa_type )
44
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _lowercase = { 'configuration_resnet': ['RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ResNetConfig', 'ResNetOnnxConfig'] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ 'RESNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'ResNetForImageClassification', 'ResNetModel', 'ResNetPreTrainedModel', 'ResNetBackbone', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ 'TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFResNetForImageClassification', 'TFResNetModel', 'TFResNetPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ 'FlaxResNetForImageClassification', 'FlaxResNetModel', 'FlaxResNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys _lowercase = _LazyModule(__name__, globals()['__file__'], _import_structure)
715
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class _lowercase : @staticmethod def UpperCamelCase ( *A__ , **A__ ) -> List[Any]: pass def __UpperCamelCase ( a : Image ) ->str: snake_case = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class _lowercase ( unittest.TestCase ): _UpperCAmelCase = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def UpperCamelCase ( self , A__ , A__ , A__ ) -> Union[str, Any]: snake_case = DepthEstimationPipeline(model=A__ , image_processor=A__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def UpperCamelCase ( self , A__ , A__ ) -> List[Any]: snake_case = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} , A__ ) import datasets snake_case = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' , '''image''' , split='''test''' ) snake_case = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] , A__ , ) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def UpperCamelCase ( self ) -> Optional[Any]: pass @slow @require_torch def UpperCamelCase ( self ) -> Dict: snake_case = '''Intel/dpt-large''' snake_case = pipeline('''depth-estimation''' , model=A__ ) snake_case = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) snake_case = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) , 2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) , 2.6_6_2 ) @require_torch def UpperCamelCase ( self ) -> Any: # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
44
0
'''simple docstring''' def __UpperCamelCase ( a : int , a : list ) ->Any: _enforce_args(snake_case__ , snake_case__ ) if n == 0: return 0 snake_case = float('''-inf''' ) for i in range(1 , n + 1 ): snake_case = max( snake_case__ , prices[i - 1] + naive_cut_rod_recursive(n - i , snake_case__ ) ) return max_revue def __UpperCamelCase ( a : int , a : list ) ->Optional[Any]: _enforce_args(snake_case__ , snake_case__ ) snake_case = [float('''-inf''' ) for _ in range(n + 1 )] return _top_down_cut_rod_recursive(snake_case__ , snake_case__ , snake_case__ ) def __UpperCamelCase ( a : int , a : list , a : list ) ->Optional[Any]: if max_rev[n] >= 0: return max_rev[n] elif n == 0: return 0 else: snake_case = float('''-inf''' ) for i in range(1 , n + 1 ): snake_case = max( snake_case__ , prices[i - 1] + _top_down_cut_rod_recursive(n - i , snake_case__ , snake_case__ ) , ) snake_case = max_revenue return max_rev[n] def __UpperCamelCase ( a : int , a : list ) ->Optional[Any]: _enforce_args(snake_case__ , snake_case__ ) # length(max_rev) = n + 1, to accommodate for the revenue obtainable from a rod of # length 0. snake_case = [float('''-inf''' ) for _ in range(n + 1 )] snake_case = 0 for i in range(1 , n + 1 ): snake_case = max_rev[i] for j in range(1 , i + 1 ): snake_case = max(snake_case__ , prices[j - 1] + max_rev[i - j] ) snake_case = max_revenue_i return max_rev[n] def __UpperCamelCase ( a : int , a : list ) ->Union[str, Any]: if n < 0: snake_case = f"""n must be greater than or equal to 0. Got n = {n}""" raise ValueError(snake_case__ ) if n > len(snake_case__ ): snake_case = ( '''Each integral piece of rod must have a corresponding price. ''' f"""Got n = {n} but length of prices = {len(snake_case__ )}""" ) raise ValueError(snake_case__ ) def __UpperCamelCase ( ) ->List[str]: snake_case = [6, 10, 12, 15, 20, 23] snake_case = len(snake_case__ ) # the best revenue comes from cutting the rod into 6 pieces, each # of length 1 resulting in a revenue of 6 * 6 = 36. snake_case = 36 snake_case = top_down_cut_rod(snake_case__ , snake_case__ ) snake_case = bottom_up_cut_rod(snake_case__ , snake_case__ ) snake_case = naive_cut_rod_recursive(snake_case__ , snake_case__ ) assert expected_max_revenue == max_rev_top_down assert max_rev_top_down == max_rev_bottom_up assert max_rev_bottom_up == max_rev_naive if __name__ == "__main__": main()
716
'''simple docstring''' import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def __UpperCamelCase ( a : Optional[int] ) ->Dict: snake_case = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(a , a ) def __UpperCamelCase ( a : Optional[Any] ) ->int: snake_case = list(s_dict.keys() ) for key in keys: if "transformer_layers" in key: snake_case = s_dict.pop(a ) elif "subsample" in key: snake_case = s_dict.pop(a ) def __UpperCamelCase ( a : Optional[int] ) ->Optional[int]: snake_case , snake_case = emb.weight.shape snake_case = nn.Linear(a , a , bias=a ) snake_case = emb.weight.data return lin_layer def __UpperCamelCase ( a : Any , a : Tuple ) ->Tuple: snake_case = torch.load(a , map_location='''cpu''' ) snake_case = mam_aaa['''args'''] snake_case = mam_aaa['''model'''] snake_case = state_dict['''decoder.output_projection.weight'''] remove_ignore_keys_(a ) rename_keys(a ) snake_case = state_dict['''decoder.embed_tokens.weight'''].shape[0] snake_case = args.share_decoder_input_output_embed snake_case = [int(a ) for i in args.conv_kernel_sizes.split(''',''' )] snake_case = SpeechaTextConfig( vocab_size=a , max_source_positions=args.max_source_positions , max_target_positions=args.max_target_positions , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='''relu''' , num_conv_layers=len(a ) , conv_channels=args.conv_channels , conv_kernel_sizes=a , input_feat_per_channel=args.input_feat_per_channel , input_channels=args.input_channels , tie_word_embeddings=a , num_beams=5 , max_length=200 , use_cache=a , decoder_start_token_id=2 , early_stopping=a , ) snake_case = SpeechaTextForConditionalGeneration(a ) snake_case , snake_case = model.model.load_state_dict(a , strict=a ) if len(a ) > 0 and not set(a ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( '''Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,''' f""" but all the following weights are missing {missing}""" ) if tie_embeds: snake_case = make_linear_from_emb(model.model.decoder.embed_tokens ) else: snake_case = lm_head_weights model.save_pretrained(a ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument('--fairseq_path', type=str, help='Path to the fairseq model (.pt) file.') parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') _lowercase = parser.parse_args() convert_fairseq_sat_checkpoint_to_tfms(args.fairseq_path, args.pytorch_dump_folder_path)
44
0
'''simple docstring''' import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger _lowercase = get_logger(__name__) class _lowercase ( enum.Enum ): _UpperCAmelCase = 'all_checks' _UpperCAmelCase = 'basic_checks' _UpperCAmelCase = 'no_checks' class _lowercase ( __a ): pass class _lowercase ( __a ): pass class _lowercase ( __a ): pass class _lowercase ( __a ): pass def __UpperCamelCase ( a : List[str] , a : Optional[Any] , a : Union[str, Any]=None ) ->List[Any]: if expected_checksums is None: logger.info('''Unable to verify checksums.''' ) return if len(set(a ) - set(a ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(a ) - set(a ) ) ) if len(set(a ) - set(a ) ) > 0: raise UnexpectedDownloadedFile(str(set(a ) - set(a ) ) ) snake_case = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] snake_case = ''' for ''' + verification_name if verification_name is not None else '''''' if len(a ) > 0: raise NonMatchingChecksumError( f"""Checksums didn't match{for_verification_name}:\n""" f"""{bad_urls}\n""" '''Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error''' ) logger.info('''All the checksums matched successfully''' + for_verification_name ) class _lowercase ( __a ): pass class _lowercase ( __a ): pass class _lowercase ( __a ): pass class _lowercase ( __a ): pass def __UpperCamelCase ( a : int , a : Union[str, Any] ) ->Optional[int]: if expected_splits is None: logger.info('''Unable to verify splits sizes.''' ) return if len(set(a ) - set(a ) ) > 0: raise ExpectedMoreSplits(str(set(a ) - set(a ) ) ) if len(set(a ) - set(a ) ) > 0: raise UnexpectedSplits(str(set(a ) - set(a ) ) ) snake_case = [ {'''expected''': expected_splits[name], '''recorded''': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(a ) > 0: raise NonMatchingSplitsSizesError(str(a ) ) logger.info('''All the splits matched successfully.''' ) def __UpperCamelCase ( a : Optional[int] , a : Union[str, Any] = True ) ->dict: if record_checksum: snake_case = shaaaa() with open(a , '''rb''' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , b'''''' ): m.update(a ) snake_case = m.hexdigest() else: snake_case = None return {"num_bytes": os.path.getsize(a ), "checksum": checksum} def __UpperCamelCase ( a : Optional[Any] ) ->Union[str, Any]: if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
717
'''simple docstring''' from ..utils import DummyObject, requires_backends class _lowercase ( metaclass=__a ): _UpperCAmelCase = ['''transformers''', '''torch''', '''note_seq'''] def __init__( self , *A__ , **A__ ) -> Union[str, Any]: requires_backends(self , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ) -> Optional[Any]: requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ) -> Any: requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] )
44
0
'''simple docstring''' import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class _lowercase : def __init__( self , A__ , A__=13 , A__=7 , A__=True , A__=True , A__=False , A__=True , A__=99 , A__=64 , A__=5 , A__=4 , A__=64 , A__="gelu" , A__=0.1 , A__=0.1 , A__=5_12 , A__=16 , A__=2 , A__=0.0_2 , A__=3 , A__=4 , A__=None , ) -> Tuple: snake_case = parent snake_case = batch_size snake_case = seq_length snake_case = is_training snake_case = use_input_mask snake_case = use_token_type_ids snake_case = use_labels snake_case = vocab_size snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_size snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = max_position_embeddings snake_case = type_vocab_size snake_case = type_sequence_label_size snake_case = initializer_range snake_case = num_labels snake_case = num_choices snake_case = scope def UpperCamelCase ( self ) -> Tuple: return MPNetConfig.from_pretrained('''microsoft/mpnet-base''' ) def UpperCamelCase ( self ) -> Dict: snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case = None if self.use_input_mask: snake_case = random_attention_mask([self.batch_size, self.seq_length] ) snake_case = None snake_case = None snake_case = None if self.use_labels: snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case = ids_tensor([self.batch_size] , self.num_choices ) snake_case = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def UpperCamelCase ( self ) -> List[Any]: return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__ ) -> Any: snake_case = MPNetModel(config=snake_case__ ) model.to(snake_case__ ) model.eval() snake_case = model(snake_case__ , snake_case__ ) snake_case = model(snake_case__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__ ) -> Dict: snake_case = MPNetForQuestionAnswering(config=snake_case__ ) model.to(snake_case__ ) model.eval() snake_case = model( snake_case__ , attention_mask=snake_case__ , start_positions=snake_case__ , end_positions=snake_case__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__ ) -> Optional[Any]: snake_case = self.num_labels snake_case = MPNetForSequenceClassification(snake_case__ ) model.to(snake_case__ ) model.eval() snake_case = model(snake_case__ , attention_mask=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__ ) -> Any: snake_case = self.num_choices snake_case = MPNetForMultipleChoice(config=snake_case__ ) model.to(snake_case__ ) model.eval() snake_case = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() snake_case = model( snake_case__ , attention_mask=snake_case__ , labels=snake_case__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def UpperCamelCase ( self , A__ , A__ , A__ , A__ , A__ , A__ ) -> Tuple: snake_case = self.num_labels snake_case = MPNetForTokenClassification(config=snake_case__ ) model.to(snake_case__ ) model.eval() snake_case = model(snake_case__ , attention_mask=snake_case__ , labels=snake_case__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def UpperCamelCase ( self ) -> int: snake_case = self.prepare_config_and_inputs() (snake_case) = config_and_inputs snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class _lowercase ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , unittest.TestCase ): _UpperCAmelCase = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) _UpperCAmelCase = ( { '''feature-extraction''': MPNetModel, '''fill-mask''': MPNetForMaskedLM, '''question-answering''': MPNetForQuestionAnswering, '''text-classification''': MPNetForSequenceClassification, '''token-classification''': MPNetForTokenClassification, '''zero-shot''': MPNetForSequenceClassification, } if is_torch_available() else {} ) _UpperCAmelCase = False _UpperCAmelCase = True def UpperCamelCase ( self ) -> List[str]: snake_case = MPNetModelTester(self ) snake_case = ConfigTester(self , config_class=snake_case__ , hidden_size=37 ) def UpperCamelCase ( self ) -> Optional[int]: self.config_tester.run_common_tests() def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*snake_case__ ) def UpperCamelCase ( self ) -> List[str]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*snake_case__ ) def UpperCamelCase ( self ) -> Tuple: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*snake_case__ ) def UpperCamelCase ( self ) -> Optional[int]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*snake_case__ ) def UpperCamelCase ( self ) -> Dict: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*snake_case__ ) @require_torch class _lowercase ( unittest.TestCase ): @slow def UpperCamelCase ( self ) -> List[Any]: snake_case = MPNetModel.from_pretrained('''microsoft/mpnet-base''' ) snake_case = torch.tensor([[0, 3_45, 2_32, 3_28, 7_40, 1_40, 16_95, 69, 60_78, 15_88, 2]] ) snake_case = model(snake_case__ )[0] snake_case = torch.Size((1, 11, 7_68) ) self.assertEqual(output.shape , snake_case__ ) snake_case = torch.tensor( [[[-0.0_5_5_0, 0.1_9_4_3, -0.0_7_4_0], [-0.0_5_6_2, 0.2_2_1_1, -0.0_5_7_9], [-0.0_4_3_7, 0.3_3_3_7, -0.0_6_4_1]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case__ , atol=1e-4 ) )
718
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator class _lowercase : def __init__( self , A__ ) -> None: snake_case = value snake_case = None snake_case = None class _lowercase : def __init__( self , A__ ) -> None: snake_case = tree def UpperCamelCase ( self , A__ ) -> int: if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self ) -> Iterator[int]: yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
44
0
'''simple docstring''' import sys _lowercase = ( '73167176531330624919225119674426574742355349194934' '96983520312774506326239578318016984801869478851843' '85861560789112949495459501737958331952853208805511' '12540698747158523863050715693290963295227443043557' '66896648950445244523161731856403098711121722383113' '62229893423380308135336276614282806444486645238749' '30358907296290491560440772390713810515859307960866' '70172427121883998797908792274921901699720888093776' '65727333001053367881220235421809751254540594752243' '52584907711670556013604839586446706324415722155397' '53697817977846174064955149290862569321978468622482' '83972241375657056057490261407972968652414535100474' '82166370484403199890008895243450658541227588666881' '16427171479924442928230863465674813919123162824586' '17866458359124566529476545682848912883142607690042' '24219022671055626321111109370544217506941658960408' '07198403850962455444362981230987879927244284909188' '84580156166097919133875499200524063689912560717606' '05886116467109405077541002256983155200055935729725' '71636269561882670428252483600823257530420752963450' ) def __UpperCamelCase ( a : str ) ->Dict: snake_case = 1 for digit in s: product *= int(__a ) return product def __UpperCamelCase ( a : str = N ) ->Optional[int]: snake_case = -sys.maxsize - 1 snake_case = n[:13] snake_case = 13 while cur_index < len(__a ) - 13: if int(n[cur_index] ) >= int(substr[0] ): snake_case = substr[1:] + n[cur_index] cur_index += 1 else: snake_case = max(__a , str_eval(__a ) ) snake_case = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f'{solution() = }')
719
'''simple docstring''' import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() _lowercase = logging.get_logger(__name__) _lowercase = [ ('bert.bert', 'visual_bert'), ('bert.cls', 'cls'), ('bert.classifier', 'cls'), ('token_type_embeddings_visual', 'visual_token_type_embeddings'), ('position_embeddings_visual', 'visual_position_embeddings'), ('projection', 'visual_projection'), ] _lowercase = [ 'nlvr2_coco_pre_trained.th', 'nlvr2_fine_tuned.th', 'nlvr2_pre_trained.th', 'vcr_coco_pre_train.th', 'vcr_fine_tune.th', 'vcr_pre_train.th', 'vqa_coco_pre_trained.th', 'vqa_fine_tuned.th', 'vqa_pre_trained.th', ] def __UpperCamelCase ( a : List[str] ) ->Optional[int]: snake_case = torch.load(a , map_location='''cpu''' ) return sd def __UpperCamelCase ( a : Optional[int] , a : Union[str, Any] , a : int=rename_keys_prefix ) ->Tuple: snake_case = OrderedDict() snake_case = torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue snake_case = key for name_pair in rename_keys_prefix: snake_case = new_key.replace(name_pair[0] , name_pair[1] ) snake_case = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately snake_case = new_d['''cls.predictions.bias'''] return new_d @torch.no_grad() def __UpperCamelCase ( a : Optional[int] , a : int ) ->Union[str, Any]: assert ( checkpoint_path.split('''/''' )[-1] in ACCEPTABLE_CHECKPOINTS ), f"""The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.""" # Get Config if "pre" in checkpoint_path: snake_case = '''pretraining''' if "vcr" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 512} elif "vqa_advanced" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048} elif "vqa" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048} elif "nlvr" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 1024} else: raise NotImplementedError(f"""No implementation found for `{checkpoint_path}`.""" ) else: if "vcr" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 512} snake_case = '''multichoice''' elif "vqa_advanced" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048} snake_case = '''vqa_advanced''' elif "vqa" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048, '''num_labels''': 3129} snake_case = '''vqa''' elif "nlvr" in checkpoint_path: snake_case = { '''visual_embedding_dim''': 1024, '''num_labels''': 2, } snake_case = '''nlvr''' snake_case = VisualBertConfig(**a ) # Load State Dict snake_case = load_state_dict(a ) snake_case = get_new_dict(a , a ) if model_type == "pretraining": snake_case = VisualBertForPreTraining(a ) elif model_type == "vqa": snake_case = VisualBertForQuestionAnswering(a ) elif model_type == "nlvr": snake_case = VisualBertForVisualReasoning(a ) elif model_type == "multichoice": snake_case = VisualBertForMultipleChoice(a ) model.load_state_dict(a ) # Save Checkpoints Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument('orig_checkpoint_path', type=str, help='A path to .th on local filesystem.') parser.add_argument('pytorch_dump_folder_path', type=str, help='Path to the output PyTorch model.') _lowercase = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
44
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { """MIT/ast-finetuned-audioset-10-10-0.4593""": ( """https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json""" ), } class _lowercase ( _snake_case ): _UpperCAmelCase = """audio-spectrogram-transformer""" def __init__( self , A__=7_68 , A__=12 , A__=12 , A__=30_72 , A__="gelu" , A__=0.0 , A__=0.0 , A__=0.0_2 , A__=1e-12 , A__=16 , A__=True , A__=10 , A__=10 , A__=10_24 , A__=1_28 , **A__ , ) -> Optional[int]: super().__init__(**snake_case_ ) snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_size snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = initializer_range snake_case = layer_norm_eps snake_case = patch_size snake_case = qkv_bias snake_case = frequency_stride snake_case = time_stride snake_case = max_length snake_case = num_mel_bins
720
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() _lowercase = logging.get_logger(__name__) def __UpperCamelCase ( a : Dict , a : Optional[int] , a : Dict , a : Dict ) ->Union[str, Any]: snake_case = original_name.split('''.''' )[0] snake_case = key.split('''.''' ) snake_case = int(key_list[key_list.index(a ) - 2] ) snake_case = int(key_list[key_list.index(a ) - 1] ) snake_case = orig_block_num - offset snake_case = key.replace(f"""{orig_block_num}.{layer_num}.{original_name}""" , f"""block.{new_block_num}.{layer_num}.{new_name}""" ) return key def __UpperCamelCase ( a : Tuple ) ->Dict: snake_case = OrderedDict() snake_case , snake_case = 0, 0 for key, value in state_dict.items(): if key.startswith('''network''' ): snake_case = key.replace('''network''' , '''poolformer.encoder''' ) if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith('''bias''' ) and "patch_embed" not in key: patch_emb_offset += 1 snake_case = key[: key.find('''proj''' )] snake_case = key.replace(a , f"""patch_embeddings.{total_embed_found}.""" ) snake_case = key.replace('''proj''' , '''projection''' ) if key.endswith('''bias''' ): total_embed_found += 1 if "patch_embeddings" in key: snake_case = '''poolformer.encoder.''' + key if "mlp.fc1" in key: snake_case = replace_key_with_offset(a , a , '''mlp.fc1''' , '''output.conv1''' ) if "mlp.fc2" in key: snake_case = replace_key_with_offset(a , a , '''mlp.fc2''' , '''output.conv2''' ) if "norm1" in key: snake_case = replace_key_with_offset(a , a , '''norm1''' , '''before_norm''' ) if "norm2" in key: snake_case = replace_key_with_offset(a , a , '''norm2''' , '''after_norm''' ) if "layer_scale_1" in key: snake_case = replace_key_with_offset(a , a , '''layer_scale_1''' , '''layer_scale_1''' ) if "layer_scale_2" in key: snake_case = replace_key_with_offset(a , a , '''layer_scale_2''' , '''layer_scale_2''' ) if "head" in key: snake_case = key.replace('''head''' , '''classifier''' ) snake_case = value return new_state_dict def __UpperCamelCase ( ) ->Optional[int]: snake_case = '''http://images.cocodataset.org/val2017/000000039769.jpg''' snake_case = Image.open(requests.get(a , stream=a ).raw ) return image @torch.no_grad() def __UpperCamelCase ( a : Dict , a : Optional[Any] , a : Tuple ) ->List[str]: snake_case = PoolFormerConfig() # set attributes based on model_name snake_case = '''huggingface/label-files''' snake_case = model_name[-3:] snake_case = 1000 snake_case = '''imagenet-1k-id2label.json''' snake_case = (1, 1000) # set config attributes snake_case = json.load(open(hf_hub_download(a , a , repo_type='''dataset''' ) , '''r''' ) ) snake_case = {int(a ): v for k, v in idalabel.items()} snake_case = idalabel snake_case = {v: k for k, v in idalabel.items()} if size == "s12": snake_case = [2, 2, 6, 2] snake_case = [64, 128, 320, 512] snake_case = 4.0 snake_case = 0.9 elif size == "s24": snake_case = [4, 4, 12, 4] snake_case = [64, 128, 320, 512] snake_case = 4.0 snake_case = 0.9 elif size == "s36": snake_case = [6, 6, 18, 6] snake_case = [64, 128, 320, 512] snake_case = 4.0 snake_case = 1e-6 snake_case = 0.9 elif size == "m36": snake_case = [6, 6, 18, 6] snake_case = [96, 192, 384, 768] snake_case = 4.0 snake_case = 1e-6 snake_case = 0.95 elif size == "m48": snake_case = [8, 8, 24, 8] snake_case = [96, 192, 384, 768] snake_case = 4.0 snake_case = 1e-6 snake_case = 0.95 else: raise ValueError(f"""Size {size} not supported""" ) # load image processor snake_case = PoolFormerImageProcessor(crop_pct=a ) # Prepare image snake_case = prepare_img() snake_case = image_processor(images=a , return_tensors='''pt''' ).pixel_values logger.info(f"""Converting model {model_name}...""" ) # load original state dict snake_case = torch.load(a , map_location=torch.device('''cpu''' ) ) # rename keys snake_case = rename_keys(a ) # create HuggingFace model and load state dict snake_case = PoolFormerForImageClassification(a ) model.load_state_dict(a ) model.eval() # Define image processor snake_case = PoolFormerImageProcessor(crop_pct=a ) snake_case = image_processor(images=prepare_img() , return_tensors='''pt''' ).pixel_values # forward pass snake_case = model(a ) snake_case = outputs.logits # define expected logit slices for different models if size == "s12": snake_case = torch.tensor([-0.3045, -0.6758, -0.4869] ) elif size == "s24": snake_case = torch.tensor([0.4402, -0.1374, -0.8045] ) elif size == "s36": snake_case = torch.tensor([-0.6080, -0.5133, -0.5898] ) elif size == "m36": snake_case = torch.tensor([0.3952, 0.2263, -1.2668] ) elif size == "m48": snake_case = torch.tensor([0.1167, -0.0656, -0.3423] ) else: raise ValueError(f"""Size {size} not supported""" ) # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] , a , atol=1e-2 ) # finally, save model and image processor logger.info(f"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(a ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument( '--model_name', default='poolformer_s12', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) _lowercase = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
44
0
from __future__ import annotations from collections.abc import Callable def __UpperCamelCase ( a : Callable[[int | float], int | float] , a : int | float , a : int | float , a : int = 100 , ) ->float: snake_case = x_start snake_case = fnc(a ) snake_case = 0.0 for _ in range(a ): # Approximates small segments of curve as linear and solve # for trapezoidal area snake_case = (x_end - x_start) / steps + xa snake_case = fnc(a ) area += abs(fxa + fxa ) * (xa - xa) / 2 # Increment step snake_case = xa snake_case = fxa return area if __name__ == "__main__": def __UpperCamelCase ( a : Any ) ->int: return x**3 + x**2 print('f(x) = x^3 + x^2') print('The area between the curve, x = -5, x = 5 and the x axis is:') _lowercase = 10 while i <= 100_000: print(f'with {i} steps: {trapezoidal_area(f, -5, 5, i)}') i *= 10
721
'''simple docstring''' import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow _lowercase = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ 'text-classification', 'language-modeling', 'summarization', 'token-classification', 'question-answering', ] ] sys.path.extend(SRC_DIRS) if SRC_DIRS is not None: import run_clm_flax import run_flax_glue import run_flax_ner import run_mlm_flax import run_qa import run_summarization_flax import run_ta_mlm_flax logging.basicConfig(level=logging.DEBUG) _lowercase = logging.getLogger() def __UpperCamelCase ( ) ->Tuple: snake_case = argparse.ArgumentParser() parser.add_argument('''-f''' ) snake_case = parser.parse_args() return args.f def __UpperCamelCase ( a : Dict , a : Tuple="eval" ) ->List[Any]: snake_case = os.path.join(a , f"""{split}_results.json""" ) if os.path.exists(a ): with open(a , '''r''' ) as f: return json.load(a ) raise ValueError(f"""can't find {path}""" ) _lowercase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class _lowercase ( __a ): def UpperCamelCase ( self ) -> List[str]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --eval_steps=2 --warmup_steps=2 --seed=42 --max_seq_length=128 """.split() with patch.object(A__ , '''argv''' , A__ ): run_flax_glue.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.7_5 ) @slow def UpperCamelCase ( self ) -> List[Any]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_clm_flax.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --block_size 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(A__ , '''argv''' , A__ ): run_clm_flax.main() snake_case = get_results(A__ ) self.assertLess(result['''eval_perplexity'''] , 1_00 ) @slow def UpperCamelCase ( self ) -> int: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_summarization.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --test_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=8 --do_train --do_eval --do_predict --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --predict_with_generate """.split() with patch.object(A__ , '''argv''' , A__ ): run_summarization_flax.main() snake_case = get_results(A__ , split='''test''' ) self.assertGreaterEqual(result['''test_rouge1'''] , 10 ) self.assertGreaterEqual(result['''test_rouge2'''] , 2 ) self.assertGreaterEqual(result['''test_rougeL'''] , 7 ) self.assertGreaterEqual(result['''test_rougeLsum'''] , 7 ) @slow def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_mlm.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --overwrite_output_dir --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --logging_steps 2 --eval_steps 2 --do_train --do_eval --num_train_epochs=1 """.split() with patch.object(A__ , '''argv''' , A__ ): run_mlm_flax.main() snake_case = get_results(A__ ) self.assertLess(result['''eval_perplexity'''] , 42 ) @slow def UpperCamelCase ( self ) -> Dict: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_t5_mlm_flax.py --model_name_or_path t5-small --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(A__ , '''argv''' , A__ ): run_ta_mlm_flax.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.4_2 ) @slow def UpperCamelCase ( self ) -> int: # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu snake_case = 7 if get_gpu_count() > 1 else 2 snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_flax_ner.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --overwrite_output_dir --do_train --do_eval --warmup_steps=2 --learning_rate=2e-4 --logging_steps 2 --eval_steps 2 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 """.split() with patch.object(A__ , '''argv''' , A__ ): run_flax_ner.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.7_5 ) self.assertGreaterEqual(result['''eval_f1'''] , 0.3 ) @slow def UpperCamelCase ( self ) -> Any: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_qa.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=2 --do_train --do_eval --logging_steps 2 --eval_steps 2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 """.split() with patch.object(A__ , '''argv''' , A__ ): run_qa.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_f1'''] , 30 ) self.assertGreaterEqual(result['''eval_exact'''] , 30 )
44
0
'''simple docstring''' import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class _lowercase ( __UpperCAmelCase ): _UpperCAmelCase = "M-CLIP" def __init__( self , A__=10_24 , A__=7_68 , **A__ ) -> Any: snake_case = transformerDimSize snake_case = imageDimSize super().__init__(**_lowerCamelCase ) class _lowercase ( __UpperCAmelCase ): _UpperCAmelCase = MCLIPConfig def __init__( self , A__ , *A__ , **A__ ) -> Dict: super().__init__(_lowerCamelCase , *_lowerCamelCase , **_lowerCamelCase ) snake_case = XLMRobertaModel(_lowerCamelCase ) snake_case = torch.nn.Linear( in_features=config.transformerDimensions , out_features=config.numDims ) def UpperCamelCase ( self , A__ , A__ ) -> int: snake_case = self.transformer(input_ids=_lowerCamelCase , attention_mask=_lowerCamelCase )[0] snake_case = (embs * attention_mask.unsqueeze(2 )).sum(dim=1 ) / attention_mask.sum(dim=1 )[:, None] return self.LinearTransformation(_lowerCamelCase ), embs
700
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS _lowercase = logging.get_logger(__name__) _lowercase = { 'linear': get_linear_schedule_with_warmup, 'cosine': get_cosine_schedule_with_warmup, 'cosine_w_restarts': get_cosine_with_hard_restarts_schedule_with_warmup, 'polynomial': get_polynomial_decay_schedule_with_warmup, 'constant': get_constant_schedule, 'constant_w_warmup': get_constant_schedule_with_warmup, } class _lowercase ( __a ): def __init__( self , A__=None , A__=None , *A__ , **A__ ) -> Union[str, Any]: super().__init__(*A__ , **A__ ) if config is None: assert isinstance(self.model , A__ ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) snake_case = self.model.config else: snake_case = config snake_case = data_args snake_case = self.config.tgt_vocab_size if isinstance(self.config , A__ ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ''' padding..''' ) if self.args.label_smoothing == 0: snake_case = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss snake_case = label_smoothed_nll_loss def UpperCamelCase ( self , A__ ) -> Tuple: if self.optimizer is None: snake_case = ['''bias''', '''LayerNorm.weight'''] snake_case = [ { '''params''': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], '''weight_decay''': self.args.weight_decay, }, { '''params''': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0, }, ] snake_case = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: snake_case = Adafactor snake_case = {'''scale_parameter''': False, '''relative_step''': False} else: snake_case = AdamW snake_case = { '''betas''': (self.args.adam_betaa, self.args.adam_betaa), '''eps''': self.args.adam_epsilon, } snake_case = self.args.learning_rate if self.sharded_ddp: snake_case = OSS( params=A__ , optim=A__ , **A__ , ) else: snake_case = optimizer_cls(A__ , **A__ ) if self.lr_scheduler is None: snake_case = self._get_lr_scheduler(A__ ) else: # ignoring --lr_scheduler logger.warning('''scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.''' ) def UpperCamelCase ( self , A__ ) -> Tuple: snake_case = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": snake_case = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": snake_case = schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: snake_case = schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=A__ ) return scheduler def UpperCamelCase ( self ) -> Optional[torch.utils.data.Sampler]: if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> List[Any]: if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token snake_case = model(**A__ , use_cache=A__ )[0] snake_case = self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models snake_case , snake_case = model(**A__ , labels=A__ , use_cache=A__ )[:2] else: # compute label smoothed loss snake_case = model(**A__ , use_cache=A__ )[0] snake_case = torch.nn.functional.log_softmax(A__ , dim=-1 ) snake_case , snake_case = self.loss_fn(A__ , A__ , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def UpperCamelCase ( self , A__ , A__ ) -> Any: snake_case = inputs.pop('''labels''' ) snake_case , snake_case = self._compute_loss(A__ , A__ , A__ ) return loss def UpperCamelCase ( self , A__ , A__ , A__ , A__ = None , ) -> Tuple[Optional[float], Optional[torch.Tensor], Optional[torch.Tensor]]: snake_case = self._prepare_inputs(A__ ) snake_case = { '''max_length''': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, '''num_beams''': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: snake_case = self.model.generate( inputs['''input_ids'''] , attention_mask=inputs['''attention_mask'''] , **A__ , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: snake_case = self._pad_tensors_to_max_len(A__ , gen_kwargs['''max_length'''] ) snake_case = inputs.pop('''labels''' ) with torch.no_grad(): # compute loss on predict data snake_case , snake_case = self._compute_loss(A__ , A__ , A__ ) snake_case = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) snake_case = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: snake_case = self._pad_tensors_to_max_len(A__ , gen_kwargs['''max_length'''] ) return (loss, logits, labels) def UpperCamelCase ( self , A__ , A__ ) -> List[str]: # If PAD token is not defined at least EOS token has to be defined snake_case = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( '''Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be''' F""" padded to `max_length`={max_length}""" ) snake_case = pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) snake_case = tensor return padded_tensor
44
0
'''simple docstring''' import unittest from transformers import SPIECE_UNDERLINE, XLNetTokenizer, XLNetTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _lowercase = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class _lowercase ( lowerCamelCase__ , unittest.TestCase ): _UpperCAmelCase = XLNetTokenizer _UpperCAmelCase = XLNetTokenizerFast _UpperCAmelCase = True _UpperCAmelCase = True def UpperCamelCase ( self ) -> str: super().setUp() # We have a SentencePiece fixture for testing snake_case = XLNetTokenizer(__lowerCamelCase , keep_accents=__lowerCamelCase ) tokenizer.sanitize_special_tokens() tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase ( self ) -> int: snake_case = '''<s>''' snake_case = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowerCamelCase ) , __lowerCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowerCamelCase ) , __lowerCamelCase ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '''<unk>''' ) self.assertEqual(vocab_keys[1] , '''<s>''' ) self.assertEqual(vocab_keys[-1] , '''<eod>''' ) self.assertEqual(len(__lowerCamelCase ) , 10_06 ) def UpperCamelCase ( self ) -> Union[str, Any]: self.assertEqual(self.get_tokenizer().vocab_size , 10_00 ) def UpperCamelCase ( self ) -> List[Any]: snake_case = XLNetTokenizer(__lowerCamelCase , keep_accents=__lowerCamelCase ) snake_case = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(__lowerCamelCase , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCamelCase ) , [2_85, 46, 10, 1_70, 3_82] ) snake_case = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) snake_case = tokenizer.convert_tokens_to_ids(__lowerCamelCase ) self.assertListEqual(__lowerCamelCase , [8, 21, 84, 55, 24, 19, 7, 0, 6_02, 3_47, 3_47, 3_47, 3, 12, 66, 46, 72, 80, 6, 0, 4] ) snake_case = tokenizer.convert_ids_to_tokens(__lowerCamelCase ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def UpperCamelCase ( self ) -> Dict: snake_case = XLNetTokenizer(__lowerCamelCase , do_lower_case=__lowerCamelCase ) snake_case = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + '''''', '''i''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''se''', '''.''', ] , ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) , ['''▁he''', '''ll''', '''o'''] ) def UpperCamelCase ( self ) -> Tuple: snake_case = XLNetTokenizer(__lowerCamelCase , do_lower_case=__lowerCamelCase ) snake_case = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( __lowerCamelCase , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''se''', '''.''', ] , ) @slow def UpperCamelCase ( self ) -> List[Any]: snake_case = XLNetTokenizer.from_pretrained('''xlnet-base-cased''' ) snake_case = tokenizer.encode('''sequence builders''' , add_special_tokens=__lowerCamelCase ) snake_case = tokenizer.encode('''multi-sequence build''' , add_special_tokens=__lowerCamelCase ) snake_case = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase ) snake_case = tokenizer.build_inputs_with_special_tokens(__lowerCamelCase , __lowerCamelCase ) assert encoded_sentence == text + [4, 3] assert encoded_pair == text + [4] + text_a + [4, 3] @slow def UpperCamelCase ( self ) -> int: snake_case = {'''input_ids''': [[17, 2_14_42, 2_70, 17, 10, 1_46_45, 3_18, 34, 17, 45_46, 31_45, 7_87, 13, 77_52, 2_20_18, 23, 21, 17, 45_46, 31_45, 7_87, 13, 33_52, 1_44_31, 13, 55_00, 11, 11_76, 5_80, 13, 1_68_19, 47_97, 23, 17, 10, 1_71_35, 6_58, 19, 4_57, 79_32, 13, 1_84, 19, 31_54, 1_71_35, 64_68, 19, 14_04, 1_22_69, 19, 42_29, 53_56, 1_62_64, 46, 19, 17, 2_05_45, 1_03_95, 9, 9, 9, 11, 28, 64_21, 95_31, 2_07_29, 17, 10, 3_53, 1_70_22, 11, 21, 64_21, 95_31, 1_69_49, 17, 10, 1_15_09, 7_53, 11, 33, 95, 24_21, 73_85, 9_56, 1_44_31, 26_26, 25, 8_42, 73_85, 48_36, 21, 14_29, 22_72, 98_55, 31_20, 1_61, 2_47_38, 19, 1_32_03, 6_58, 2_18, 7_87, 21, 4_30, 1_84_82, 8_47, 26_37, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 3_22, 2_21_78, 27, 10_64, 22, 9_56, 13, 1_11_01, 14_29, 58_54, 2_43_13, 1_89_53, 40, 4_22, 2_43_66, 68, 17_58, 37, 1_04_83, 1_42_57, 31, 2_07, 2_63, 21, 2_03, 37_73, 25, 71, 97_35, 9, 4, 3], [5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 32, 20_49, 34_42, 17, 1_38_94, 33_80, 23, 95, 18, 1_76_34, 22_88, 9, 4, 3]], '''token_type_ids''': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2], [3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowerCamelCase , model_name='''xlnet-base-cased''' , revision='''c841166438c31ec7ca9a106dee7bb312b73ae511''' , )
701
'''simple docstring''' import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def __UpperCamelCase ( a : List[str] ) ->str: snake_case = [] for line in lines: snake_case = re.sub(R'''#.*''' , '''''' , a ) # remove comments if line: filtered_lines.append(a ) snake_case = '''\n'''.join(a ) # Make a hash from all this code snake_case = full_str.encode('''utf-8''' ) return shaaaa(a ).hexdigest() # get importable module names and hash for caching _lowercase = { 'csv': (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), 'json': (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), 'pandas': (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), 'parquet': (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), 'arrow': (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), 'text': (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), 'imagefolder': (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), 'audiofolder': (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions _lowercase = { '.csv': ('csv', {}), '.tsv': ('csv', {'sep': '\t'}), '.json': ('json', {}), '.jsonl': ('json', {}), '.parquet': ('parquet', {}), '.arrow': ('arrow', {}), '.txt': ('text', {}), } _EXTENSION_TO_MODULE.update({ext: ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _lowercase = {'imagefolder', 'audiofolder'} # Used to filter data files based on extensions given a module name _lowercase = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append('.zip') _MODULE_TO_EXTENSIONS["audiofolder"].append('.zip')
44
0
'''simple docstring''' import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sentencepiece @require_tokenizers @require_flax class _lowercase ( unittest.TestCase ): @slow def UpperCamelCase ( self ): snake_case = FlaxXLMRobertaModel.from_pretrained('''xlm-roberta-base''' ) snake_case = AutoTokenizer.from_pretrained('''xlm-roberta-base''' ) snake_case = '''The dog is cute and lives in the garden house''' snake_case = jnp.array([tokenizer.encode(_lowercase )] ) snake_case = (1, 12, 7_68) # batch_size, sequence_length, embedding_vector_dim snake_case = jnp.array( [[-0.0_1_0_1, 0.1_2_1_8, -0.0_8_0_3, 0.0_8_0_1, 0.1_3_2_7, 0.0_7_7_6, -0.1_2_1_5, 0.2_3_8_3, 0.3_3_3_8, 0.3_1_0_6, 0.0_3_0_0, 0.0_2_5_2]] ) snake_case = model(_lowercase )['''last_hidden_state'''] self.assertEqual(output.shape , _lowercase ) # compare the actual values for a slice of last dim self.assertTrue(jnp.allclose(output[:, :, -1] , _lowercase , atol=1e-3 ) )
702
'''simple docstring''' _lowercase = { 'Pillow': 'Pillow', 'accelerate': 'accelerate>=0.11.0', 'compel': 'compel==0.1.8', 'black': 'black~=23.1', 'datasets': 'datasets', 'filelock': 'filelock', 'flax': 'flax>=0.4.1', 'hf-doc-builder': 'hf-doc-builder>=0.3.0', 'huggingface-hub': 'huggingface-hub>=0.13.2', 'requests-mock': 'requests-mock==1.10.0', 'importlib_metadata': 'importlib_metadata', 'invisible-watermark': 'invisible-watermark', 'isort': 'isort>=5.5.4', 'jax': 'jax>=0.2.8,!=0.3.2', 'jaxlib': 'jaxlib>=0.1.65', 'Jinja2': 'Jinja2', 'k-diffusion': 'k-diffusion>=0.0.12', 'torchsde': 'torchsde', 'note_seq': 'note_seq', 'librosa': 'librosa', 'numpy': 'numpy', 'omegaconf': 'omegaconf', 'parameterized': 'parameterized', 'protobuf': 'protobuf>=3.20.3,<4', 'pytest': 'pytest', 'pytest-timeout': 'pytest-timeout', 'pytest-xdist': 'pytest-xdist', 'ruff': 'ruff>=0.0.241', 'safetensors': 'safetensors', 'sentencepiece': 'sentencepiece>=0.1.91,!=0.1.92', 'scipy': 'scipy', 'onnx': 'onnx', 'regex': 'regex!=2019.12.17', 'requests': 'requests', 'tensorboard': 'tensorboard', 'torch': 'torch>=1.4', 'torchvision': 'torchvision', 'transformers': 'transformers>=4.25.1', 'urllib3': 'urllib3<=2.0.0', }
44
0
'''simple docstring''' from __future__ import annotations from collections.abc import Callable from typing import Generic, TypeVar _lowercase = TypeVar('T') _lowercase = TypeVar('U') class _lowercase ( Generic[T, U] ): def __init__( self , A__ , A__ ) -> str: snake_case = key snake_case = val snake_case = None snake_case = None def __repr__( self ) -> List[str]: return ( F"""Node: key: {self.key}, val: {self.val}, """ F"""has next: {bool(self.next )}, has prev: {bool(self.prev )}""" ) class _lowercase ( Generic[T, U] ): def __init__( self ) -> List[Any]: snake_case = DoubleLinkedListNode(lowercase_ , lowercase_ ) snake_case = DoubleLinkedListNode(lowercase_ , lowercase_ ) snake_case = self.rear, self.head def __repr__( self ) -> Tuple: snake_case = ["DoubleLinkedList"] snake_case = self.head while node.next is not None: rep.append(str(lowercase_ ) ) snake_case = node.next rep.append(str(self.rear ) ) return ",\n ".join(lowercase_ ) def UpperCamelCase ( self , A__ ) -> Union[str, Any]: snake_case = self.rear.prev # All nodes other than self.head are guaranteed to have non-None previous assert previous is not None snake_case = node snake_case = previous snake_case = node snake_case = self.rear def UpperCamelCase ( self , A__ ) -> Tuple: if node.prev is None or node.next is None: return None snake_case = node.next snake_case = node.prev snake_case = None snake_case = None return node class _lowercase ( Generic[T, U] ): _UpperCAmelCase = {} def __init__( self , A__ ) -> List[str]: snake_case = DoubleLinkedList() snake_case = capacity snake_case = 0 snake_case = 0 snake_case = 0 snake_case = {} def __repr__( self ) -> str: return ( F"""CacheInfo(hits={self.hits}, misses={self.miss}, """ F"""capacity={self.capacity}, current size={self.num_keys})""" ) def __contains__( self , A__ ) -> Dict: return key in self.cache def UpperCamelCase ( self , A__ ) -> List[Any]: # Note: pythonic interface would throw KeyError rather than return None if key in self.cache: self.hits += 1 snake_case = self.cache[key] snake_case = self.list.remove(self.cache[key] ) assert node == value_node # node is guaranteed not None because it is in self.cache assert node is not None self.list.add(lowercase_ ) return node.val self.miss += 1 return None def UpperCamelCase ( self , A__ , A__ ) -> Tuple: if key not in self.cache: if self.num_keys >= self.capacity: # delete first node (oldest) when over capacity snake_case = self.list.head.next # guaranteed to have a non-None first node when num_keys > 0 # explain to type checker via assertions assert first_node is not None assert first_node.key is not None assert ( self.list.remove(lowercase_ ) is not None ) # node guaranteed to be in list assert node.key is not None del self.cache[first_node.key] self.num_keys -= 1 snake_case = DoubleLinkedListNode(lowercase_ , lowercase_ ) self.list.add(self.cache[key] ) self.num_keys += 1 else: # bump node to the end of the list, update value snake_case = self.list.remove(self.cache[key] ) assert node is not None # node guaranteed to be in list snake_case = value self.list.add(lowercase_ ) @classmethod def UpperCamelCase ( cls , A__ = 1_28 ) -> List[str]: def cache_decorator_inner(A__ ) -> Callable[..., U]: def cache_decorator_wrapper(*A__ ) -> U: if func not in cls.decorator_function_to_instance_map: snake_case = LRUCache(lowercase_ ) snake_case = cls.decorator_function_to_instance_map[func].get(args[0] ) if result is None: snake_case = func(*lowercase_ ) cls.decorator_function_to_instance_map[func].put(args[0] , lowercase_ ) return result def cache_info() -> LRUCache[T, U]: return cls.decorator_function_to_instance_map[func] setattr(lowercase_ , '''cache_info''' , lowercase_ ) # noqa: B010 return cache_decorator_wrapper return cache_decorator_inner if __name__ == "__main__": import doctest doctest.testmod()
703
'''simple docstring''' import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _lowercase ( __a , __a , unittest.TestCase ): _UpperCAmelCase = IFInpaintingSuperResolutionPipeline _UpperCAmelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} _UpperCAmelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({'''original_image'''} ) _UpperCAmelCase = PipelineTesterMixin.required_optional_params - {'''latents'''} def UpperCamelCase ( self ) -> int: return self._get_superresolution_dummy_components() def UpperCamelCase ( self , A__ , A__=0 ) -> Union[str, Any]: if str(A__ ).startswith('''mps''' ): snake_case = torch.manual_seed(A__ ) else: snake_case = torch.Generator(device=A__ ).manual_seed(A__ ) snake_case = floats_tensor((1, 3, 16, 16) , rng=random.Random(A__ ) ).to(A__ ) snake_case = floats_tensor((1, 3, 32, 32) , rng=random.Random(A__ ) ).to(A__ ) snake_case = floats_tensor((1, 3, 32, 32) , rng=random.Random(A__ ) ).to(A__ ) snake_case = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase ( self ) -> List[Any]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def UpperCamelCase ( self ) -> Optional[Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def UpperCamelCase ( self ) -> List[str]: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1 ) def UpperCamelCase ( self ) -> int: self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def UpperCamelCase ( self ) -> Optional[Any]: self._test_save_load_local() def UpperCamelCase ( self ) -> Dict: self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
44
0
'''simple docstring''' import random def __UpperCamelCase ( a : str , a : Optional[Any] , a : Union[str, Any] ) ->Any: snake_case = a[left_index] snake_case = left_index + 1 for j in range(left_index + 1 , a ): if a[j] < pivot: snake_case , snake_case = a[i], a[j] i += 1 snake_case , snake_case = a[i - 1], a[left_index] return i - 1 def __UpperCamelCase ( a : Optional[int] , a : List[str] , a : List[str] ) ->str: if left < right: snake_case = random.randint(a , right - 1 ) snake_case , snake_case = ( a[left], a[pivot], ) # switches the pivot with the left most bound snake_case = partition(a , a , a ) quick_sort_random( a , a , a ) # recursive quicksort to the left of the pivot point quick_sort_random( a , pivot_index + 1 , a ) # recursive quicksort to the right of the pivot point def __UpperCamelCase ( ) ->List[str]: snake_case = input('''Enter numbers separated by a comma:\n''' ).strip() snake_case = [int(a ) for item in user_input.split(''',''' )] quick_sort_random(a , 0 , len(a ) ) print(a ) if __name__ == "__main__": main()
704
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy _lowercase = logging.get_logger(__name__) class _lowercase ( __a ): def __init__( self , A__ , A__ , A__ , **A__ ) -> Union[str, Any]: snake_case = feature_size snake_case = sampling_rate snake_case = padding_value snake_case = kwargs.pop('''padding_side''' , '''right''' ) snake_case = kwargs.pop('''return_attention_mask''' , A__ ) super().__init__(**A__ ) def UpperCamelCase ( self , A__ , A__ = True , A__ = None , A__ = False , A__ = None , A__ = None , A__ = None , ) -> BatchFeature: # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(A__ , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): snake_case = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( '''You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`''' F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) snake_case = processed_features[self.model_input_names[0]] snake_case = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(A__ ) == 0: if return_attention_mask: snake_case = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch snake_case = required_input[0] if isinstance(A__ , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. snake_case = 0 while len(required_input[index] ) == 0: index += 1 if index < len(A__ ): snake_case = required_input[index][0] if return_tensors is None: if is_tf_tensor(A__ ): snake_case = '''tf''' elif is_torch_tensor(A__ ): snake_case = '''pt''' elif isinstance(A__ , (int, float, list, tuple, np.ndarray) ): snake_case = '''np''' else: raise ValueError( F"""type of {first_element} unknown: {type(A__ )}. """ '''Should be one of a python, numpy, pytorch or tensorflow object.''' ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): snake_case = to_numpy(A__ ) else: snake_case = [to_numpy(A__ ) for v in value] # Convert padding_strategy in PaddingStrategy snake_case = self._get_padding_strategies(padding=A__ , max_length=A__ ) snake_case = processed_features[self.model_input_names[0]] snake_case = len(A__ ) if not all(len(A__ ) == batch_size for v in processed_features.values() ): raise ValueError('''Some items in the output dictionary have a different batch size than others.''' ) snake_case = [] for i in range(A__ ): snake_case = {k: v[i] for k, v in processed_features.items()} # truncation snake_case = self._truncate( A__ , max_length=A__ , pad_to_multiple_of=A__ , truncation=A__ , ) truncated_inputs.append(A__ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length snake_case = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) snake_case = PaddingStrategy.MAX_LENGTH snake_case = {} for i in range(A__ ): # padding snake_case = self._pad( truncated_inputs[i] , max_length=A__ , padding_strategy=A__ , pad_to_multiple_of=A__ , return_attention_mask=A__ , ) for key, value in outputs.items(): if key not in batch_outputs: snake_case = [] if value.dtype is np.dtype(np.floataa ): snake_case = value.astype(np.floataa ) batch_outputs[key].append(A__ ) return BatchFeature(A__ , tensor_type=A__ ) def UpperCamelCase ( self , A__ , A__ = None , A__ = PaddingStrategy.DO_NOT_PAD , A__ = None , A__ = None , ) -> dict: snake_case = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: snake_case = len(A__ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): snake_case = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of snake_case = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(A__ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: snake_case = np.ones(len(A__ ) , dtype=np.intaa ) if needs_to_be_padded: snake_case = max_length - len(A__ ) if self.padding_side == "right": if return_attention_mask: snake_case = np.pad( processed_features['''attention_mask'''] , (0, difference) ) snake_case = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) snake_case = np.pad( A__ , A__ , '''constant''' , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: snake_case = np.pad( processed_features['''attention_mask'''] , (difference, 0) ) snake_case = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) snake_case = np.pad( A__ , A__ , '''constant''' , constant_values=self.padding_value ) else: raise ValueError('''Invalid padding strategy:''' + str(self.padding_side ) ) return processed_features def UpperCamelCase ( self , A__ , A__ = None , A__ = None , A__ = None , ) -> Union[str, Any]: if not truncation: return processed_features elif truncation and max_length is None: raise ValueError('''When setting ``truncation=True``, make sure that ``max_length`` is defined.''' ) snake_case = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): snake_case = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of snake_case = len(A__ ) > max_length if needs_to_be_truncated: snake_case = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: snake_case = processed_features['''attention_mask'''][:max_length] return processed_features def UpperCamelCase ( self , A__=False , A__=None ) -> Union[str, Any]: # Get padding strategy if padding is not False: if padding is True: snake_case = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(A__ , A__ ): snake_case = PaddingStrategy(A__ ) elif isinstance(A__ , A__ ): snake_case = padding else: snake_case = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( '''Asking to pad but the feature_extractor does not have a padding value. Please select a value to use''' ''' as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.''' ) return padding_strategy
44
0
'''simple docstring''' import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _lowercase = logging.get_logger(__name__) def __UpperCamelCase ( a : Optional[Any] , a : str=False ) ->Optional[int]: snake_case = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""deit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""deit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""deit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""deit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""deit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""deit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""deit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""deit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""deit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""deit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''deit.embeddings.cls_token'''), ('''dist_token''', '''deit.embeddings.distillation_token'''), ('''patch_embed.proj.weight''', '''deit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''deit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''deit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" snake_case = [(pair[0], pair[1][4:]) if pair[1].startswith('''deit''' ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ('''norm.weight''', '''deit.layernorm.weight'''), ('''norm.bias''', '''deit.layernorm.bias'''), ('''head.weight''', '''cls_classifier.weight'''), ('''head.bias''', '''cls_classifier.bias'''), ('''head_dist.weight''', '''distillation_classifier.weight'''), ('''head_dist.bias''', '''distillation_classifier.bias'''), ] ) return rename_keys def __UpperCamelCase ( a : int , a : List[str] , a : Optional[int]=False ) ->Optional[Any]: for i in range(config.num_hidden_layers ): if base_model: snake_case = "" else: snake_case = "deit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) snake_case = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) snake_case = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict snake_case = in_proj_weight[ : config.hidden_size, : ] snake_case = in_proj_bias[: config.hidden_size] snake_case = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] snake_case = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] snake_case = in_proj_weight[ -config.hidden_size :, : ] snake_case = in_proj_bias[-config.hidden_size :] def __UpperCamelCase ( a : List[Any] , a : Any , a : List[str] ) ->List[str]: snake_case = dct.pop(_lowerCamelCase ) snake_case = val def __UpperCamelCase ( ) ->Optional[Any]: snake_case = "http://images.cocodataset.org/val2017/000000039769.jpg" snake_case = Image.open(requests.get(_lowerCamelCase , stream=_lowerCamelCase ).raw ) return im @torch.no_grad() def __UpperCamelCase ( a : Any , a : Dict ) ->str: snake_case = DeiTConfig() # all deit models have fine-tuned heads snake_case = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size snake_case = 1000 snake_case = "huggingface/label-files" snake_case = "imagenet-1k-id2label.json" snake_case = json.load(open(hf_hub_download(_lowerCamelCase , _lowerCamelCase , repo_type='''dataset''' ) , '''r''' ) ) snake_case = {int(_lowerCamelCase ): v for k, v in idalabel.items()} snake_case = idalabel snake_case = {v: k for k, v in idalabel.items()} snake_case = int(deit_name[-6:-4] ) snake_case = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith('''tiny''' ): snake_case = 192 snake_case = 768 snake_case = 12 snake_case = 3 elif deit_name[9:].startswith('''small''' ): snake_case = 384 snake_case = 1536 snake_case = 12 snake_case = 6 if deit_name[9:].startswith('''base''' ): pass elif deit_name[4:].startswith('''large''' ): snake_case = 1024 snake_case = 4096 snake_case = 24 snake_case = 16 # load original model from timm snake_case = timm.create_model(_lowerCamelCase , pretrained=_lowerCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys snake_case = timm_model.state_dict() snake_case = create_rename_keys(_lowerCamelCase , _lowerCamelCase ) for src, dest in rename_keys: rename_key(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) read_in_q_k_v(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # load HuggingFace model snake_case = DeiTForImageClassificationWithTeacher(_lowerCamelCase ).eval() model.load_state_dict(_lowerCamelCase ) # Check outputs on an image, prepared by DeiTImageProcessor snake_case = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 snake_case = DeiTImageProcessor(size=_lowerCamelCase , crop_size=config.image_size ) snake_case = image_processor(images=prepare_img() , return_tensors='''pt''' ) snake_case = encoding["pixel_values"] snake_case = model(_lowerCamelCase ) snake_case = timm_model(_lowerCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(_lowerCamelCase , outputs.logits , atol=1e-3 ) Path(_lowerCamelCase ).mkdir(exist_ok=_lowerCamelCase ) print(f"""Saving model {deit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_lowerCamelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_lowerCamelCase ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--deit_name', default='vit_deit_base_distilled_patch16_224', type=str, help='Name of the DeiT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _lowercase = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
705
'''simple docstring''' from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class _lowercase ( yaml.SafeLoader ): def UpperCamelCase ( self , A__ ) -> List[str]: snake_case = [self.constructed_objects[key_node] for key_node, _ in node.value] snake_case = [tuple(A__ ) if isinstance(A__ , A__ ) else key for key in keys] snake_case = Counter(A__ ) snake_case = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F"""Got duplicate yaml keys: {duplicate_keys}""" ) def UpperCamelCase ( self , A__ , A__=False ) -> List[Any]: snake_case = super().construct_mapping(A__ , deep=A__ ) self._check_no_duplicates_on_constructed_node(A__ ) return mapping def __UpperCamelCase ( a : str ) ->Tuple[Optional[str], str]: snake_case = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: snake_case = full_content[1:].index('''---''' ) + 1 snake_case = '''\n'''.join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(a ) class _lowercase ( __a ): # class attributes _UpperCAmelCase = {'''train_eval_index'''} # train-eval-index in the YAML metadata @classmethod def UpperCamelCase ( cls , A__ ) -> "DatasetMetadata": with open(A__ , encoding='''utf-8''' ) as readme_file: snake_case , snake_case = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(A__ ) else: return cls() def UpperCamelCase ( self , A__ ) -> str: if path.exists(): with open(A__ , encoding='''utf-8''' ) as readme_file: snake_case = readme_file.read() else: snake_case = None snake_case = self._to_readme(A__ ) with open(A__ , '''w''' , encoding='''utf-8''' ) as readme_file: readme_file.write(A__ ) def UpperCamelCase ( self , A__ = None ) -> str: if readme_content is not None: snake_case , snake_case = _split_yaml_from_readme(A__ ) snake_case = '''---\n''' + self.to_yaml_string() + '''---\n''' + content else: snake_case = '''---\n''' + self.to_yaml_string() + '''---\n''' return full_content @classmethod def UpperCamelCase ( cls , A__ ) -> "DatasetMetadata": snake_case = yaml.load(A__ , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields snake_case = { (key.replace('''-''' , '''_''' ) if key.replace('''-''' , '''_''' ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**A__ ) def UpperCamelCase ( self ) -> str: return yaml.safe_dump( { (key.replace('''_''' , '''-''' ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=A__ , allow_unicode=A__ , encoding='''utf-8''' , ).decode('''utf-8''' ) _lowercase = { 'image-classification': [], 'translation': [], 'image-segmentation': [], 'fill-mask': [], 'automatic-speech-recognition': [], 'token-classification': [], 'sentence-similarity': [], 'audio-classification': [], 'question-answering': [], 'summarization': [], 'zero-shot-classification': [], 'table-to-text': [], 'feature-extraction': [], 'other': [], 'multiple-choice': [], 'text-classification': [], 'text-to-image': [], 'text2text-generation': [], 'zero-shot-image-classification': [], 'tabular-classification': [], 'tabular-regression': [], 'image-to-image': [], 'tabular-to-text': [], 'unconditional-image-generation': [], 'text-retrieval': [], 'text-to-speech': [], 'object-detection': [], 'audio-to-audio': [], 'text-generation': [], 'conversational': [], 'table-question-answering': [], 'visual-question-answering': [], 'image-to-text': [], 'reinforcement-learning': [], 'voice-activity-detection': [], 'time-series-forecasting': [], 'document-question-answering': [], } if __name__ == "__main__": from argparse import ArgumentParser _lowercase = ArgumentParser(usage='Validate the yaml metadata block of a README.md file.') ap.add_argument('readme_filepath') _lowercase = ap.parse_args() _lowercase = Path(args.readme_filepath) _lowercase = DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
44
0
'''simple docstring''' import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_transfo_xl import CORPUS_NAME, VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() # We do this to be able to load python 2 datasets pickles # See e.g. https://stackoverflow.com/questions/2121874/python-pickling-after-changing-a-modules-directory/2121918#2121918 _lowercase = data_utils.TransfoXLTokenizer _lowercase = data_utils.TransfoXLCorpus _lowercase = data_utils _lowercase = data_utils def __UpperCamelCase ( a : int , a : Union[str, Any] , a : List[str] , a : Optional[int] ) ->Any: if transfo_xl_dataset_file: # Convert a pre-processed corpus (see original TensorFlow repo) with open(_SCREAMING_SNAKE_CASE , '''rb''' ) as fp: snake_case = pickle.load(_SCREAMING_SNAKE_CASE , encoding='''latin1''' ) # Save vocabulary and dataset cache as Dictionaries (should be better than pickles for the long-term) snake_case = pytorch_dump_folder_path + '''/''' + VOCAB_FILES_NAMES['''pretrained_vocab_file'''] print(f"""Save vocabulary to {pytorch_vocab_dump_path}""" ) snake_case = corpus.vocab.__dict__ torch.save(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) snake_case = corpus.__dict__ corpus_dict_no_vocab.pop('''vocab''' , _SCREAMING_SNAKE_CASE ) snake_case = pytorch_dump_folder_path + '''/''' + CORPUS_NAME print(f"""Save dataset to {pytorch_dataset_dump_path}""" ) torch.save(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if tf_checkpoint_path: # Convert a pre-trained TensorFlow model snake_case = os.path.abspath(_SCREAMING_SNAKE_CASE ) snake_case = os.path.abspath(_SCREAMING_SNAKE_CASE ) print(f"""Converting Transformer XL checkpoint from {tf_path} with config at {config_path}.""" ) # Initialise PyTorch model if transfo_xl_config_file == "": snake_case = TransfoXLConfig() else: snake_case = TransfoXLConfig.from_json_file(_SCREAMING_SNAKE_CASE ) print(f"""Building PyTorch model from configuration: {config}""" ) snake_case = TransfoXLLMHeadModel(_SCREAMING_SNAKE_CASE ) snake_case = load_tf_weights_in_transfo_xl(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Save pytorch-model snake_case = os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) snake_case = os.path.join(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) print(f"""Save PyTorch model to {os.path.abspath(_SCREAMING_SNAKE_CASE )}""" ) torch.save(model.state_dict() , _SCREAMING_SNAKE_CASE ) print(f"""Save configuration file to {os.path.abspath(_SCREAMING_SNAKE_CASE )}""" ) with open(_SCREAMING_SNAKE_CASE , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--tf_checkpoint_path', default='', type=str, help='An optional path to a TensorFlow checkpoint path to be converted.', ) parser.add_argument( '--transfo_xl_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained BERT model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--transfo_xl_dataset_file', default='', type=str, help='An optional dataset file to be converted in a vocabulary.', ) _lowercase = parser.parse_args() convert_transfo_xl_checkpoint_to_pytorch( args.tf_checkpoint_path, args.transfo_xl_config_file, args.pytorch_dump_folder_path, args.transfo_xl_dataset_file, )
706
'''simple docstring''' import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowercase ( __a , unittest.TestCase ): _UpperCAmelCase = CodeGenTokenizer _UpperCAmelCase = CodeGenTokenizerFast _UpperCAmelCase = True _UpperCAmelCase = {'''add_prefix_space''': True} _UpperCAmelCase = False def UpperCamelCase ( self ) -> Tuple: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt snake_case = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', '''<|endoftext|>''', ] snake_case = dict(zip(A__ , range(len(A__ ) ) ) ) snake_case = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] snake_case = {'''unk_token''': '''<unk>'''} snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(A__ ) ) def UpperCamelCase ( self , **A__ ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **A__ ) def UpperCamelCase ( self , **A__ ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **A__ ) def UpperCamelCase ( self , A__ ) -> Tuple: snake_case = '''lower newer''' snake_case = '''lower newer''' return input_text, output_text def UpperCamelCase ( self ) -> List[Any]: snake_case = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) snake_case = '''lower newer''' snake_case = ['''\u0120low''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] snake_case = tokenizer.tokenize(A__ , add_prefix_space=A__ ) self.assertListEqual(A__ , A__ ) snake_case = tokens + [tokenizer.unk_token] snake_case = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(A__ ) , A__ ) def UpperCamelCase ( self ) -> Optional[int]: if not self.test_rust_tokenizer: return snake_case = self.get_tokenizer() snake_case = self.get_rust_tokenizer(add_prefix_space=A__ ) snake_case = '''lower newer''' # Testing tokenization snake_case = tokenizer.tokenize(A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.tokenize(A__ ) self.assertListEqual(A__ , A__ ) # Testing conversion to ids without special tokens snake_case = tokenizer.encode(A__ , add_special_tokens=A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.encode(A__ , add_special_tokens=A__ ) self.assertListEqual(A__ , A__ ) # Testing conversion to ids with special tokens snake_case = self.get_rust_tokenizer(add_prefix_space=A__ ) snake_case = tokenizer.encode(A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.encode(A__ ) self.assertListEqual(A__ , A__ ) # Testing the unknown token snake_case = tokens + [rust_tokenizer.unk_token] snake_case = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(A__ ) , A__ ) def UpperCamelCase ( self , *A__ , **A__ ) -> List[str]: # It's very difficult to mix/test pretokenization with byte-level # And get both CodeGen and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def UpperCamelCase ( self , A__=15 ) -> Tuple: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): snake_case = self.rust_tokenizer_class.from_pretrained(A__ , **A__ ) # Simple input snake_case = '''This is a simple input''' snake_case = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case = ('''This is a simple input''', '''This is a pair''') snake_case = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(A__ , tokenizer_r.encode , A__ , max_length=A__ , padding='''max_length''' ) # Simple input self.assertRaises(A__ , tokenizer_r.encode_plus , A__ , max_length=A__ , padding='''max_length''' ) # Simple input self.assertRaises( A__ , tokenizer_r.batch_encode_plus , A__ , max_length=A__ , padding='''max_length''' , ) # Pair input self.assertRaises(A__ , tokenizer_r.encode , A__ , max_length=A__ , padding='''max_length''' ) # Pair input self.assertRaises(A__ , tokenizer_r.encode_plus , A__ , max_length=A__ , padding='''max_length''' ) # Pair input self.assertRaises( A__ , tokenizer_r.batch_encode_plus , A__ , max_length=A__ , padding='''max_length''' , ) def UpperCamelCase ( self ) -> Tuple: snake_case = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token='''<pad>''' ) # Simple input snake_case = '''This is a simple input''' snake_case = ['''This is a simple input looooooooong''', '''This is a simple input'''] snake_case = ('''This is a simple input''', '''This is a pair''') snake_case = [ ('''This is a simple input loooooong''', '''This is a simple input'''), ('''This is a simple pair loooooong''', '''This is a simple pair'''), ] snake_case = tokenizer.pad_token_id snake_case = tokenizer(A__ , padding='''max_length''' , max_length=30 , return_tensors='''np''' ) snake_case = tokenizer(A__ , padding=A__ , truncate=A__ , return_tensors='''np''' ) snake_case = tokenizer(*A__ , padding='''max_length''' , max_length=60 , return_tensors='''np''' ) snake_case = tokenizer(A__ , padding=A__ , truncate=A__ , return_tensors='''np''' ) # s # test single string max_length padding self.assertEqual(out_s['''input_ids'''].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s['''input_ids'''] ) self.assertTrue(0 in out_s['''attention_mask'''] ) # s2 # test automatic padding self.assertEqual(out_sa['''input_ids'''].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['''input_ids'''][0] ) self.assertFalse(0 in out_sa['''attention_mask'''][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['''input_ids'''][1] ) self.assertTrue(0 in out_sa['''attention_mask'''][1] ) # p # test single pair max_length padding self.assertEqual(out_p['''input_ids'''].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p['''input_ids'''] ) self.assertTrue(0 in out_p['''attention_mask'''] ) # p2 # test automatic padding pair self.assertEqual(out_pa['''input_ids'''].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['''input_ids'''][0] ) self.assertFalse(0 in out_pa['''attention_mask'''][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['''input_ids'''][1] ) self.assertTrue(0 in out_pa['''attention_mask'''][1] ) def UpperCamelCase ( self ) -> str: snake_case = '''$$$''' snake_case = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=A__ , add_bos_token=A__ ) snake_case = '''This is a simple input''' snake_case = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case = tokenizer.bos_token_id snake_case = tokenizer(A__ ) snake_case = tokenizer(A__ ) self.assertEqual(out_s.input_ids[0] , A__ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) snake_case = tokenizer.decode(out_s.input_ids ) snake_case = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , A__ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def UpperCamelCase ( self ) -> Any: snake_case = CodeGenTokenizer.from_pretrained('''Salesforce/codegen-350M-mono''' ) snake_case = '''\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#''' snake_case = '''\nif len_a > len_b: result = a\nelse: result = b''' snake_case = tokenizer.encode(A__ ) snake_case = ['''^#''', re.escape('''<|endoftext|>''' ), '''^\'\'\'''', '''^"""''', '''\n\n\n'''] snake_case = tokenizer.decode(A__ , truncate_before_pattern=A__ ) self.assertEqual(A__ , A__ ) def UpperCamelCase ( self ) -> Union[str, Any]: pass
44
0
'''simple docstring''' import math from datetime import datetime, timedelta def __UpperCamelCase ( a : int ) ->datetime: snake_case = year % 19 snake_case = year % 4 snake_case = year % 7 snake_case = math.floor(year / 100 ) snake_case = math.floor((13 + 8 * leap_day_inhibits) / 25 ) snake_case = leap_day_inhibits / 4 snake_case = ( 15 - lunar_orbit_correction + leap_day_inhibits - leap_day_reinstall_number ) % 30 snake_case = (4 + leap_day_inhibits - leap_day_reinstall_number) % 7 # days to be added to March 21 snake_case = (19 * metonic_cycle + secular_moon_shift) % 30 # PHM -> Paschal Full Moon snake_case = ( 2 * julian_leap_year + 4 * non_leap_year + 6 * days_to_add + century_starting_point ) % 7 if days_to_add == 29 and days_from_phm_to_sunday == 6: return datetime(a , 4 , 19 ) elif days_to_add == 28 and days_from_phm_to_sunday == 6: return datetime(a , 4 , 18 ) else: return datetime(a , 3 , 22 ) + timedelta( days=int(days_to_add + days_from_phm_to_sunday ) ) if __name__ == "__main__": for year in (1_994, 2_000, 2_010, 2_021, 2_023): _lowercase = 'will be' if year > datetime.now().year else 'was' print(f'Easter in {year} {tense} {gauss_easter(year)}')
707
'''simple docstring''' from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowercase : def __init__( self , A__ , A__=13 , A__=30 , A__=2 , A__=3 , A__=True , A__=True , A__=32 , A__=2 , A__=4 , A__=37 , A__="gelu" , A__=0.1 , A__=0.1 , A__=10 , A__=0.0_2 , A__=3 , A__=None , ) -> List[Any]: snake_case = parent snake_case = batch_size snake_case = image_size snake_case = patch_size snake_case = num_channels snake_case = is_training snake_case = use_labels snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_size snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = type_sequence_label_size snake_case = initializer_range snake_case = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) snake_case = (image_size // patch_size) ** 2 snake_case = num_patches + 1 def UpperCamelCase ( self ) -> int: snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case = None if self.use_labels: snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ) -> int: return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A__ , initializer_range=self.initializer_range , ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Union[str, Any]: snake_case = TFViTModel(config=A__ ) snake_case = model(A__ , training=A__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. snake_case = self.image_size // 2 snake_case = pixel_values[:, :, :image_size, :image_size] snake_case = model(A__ , interpolate_pos_encoding=A__ , training=A__ ) snake_case = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Optional[int]: snake_case = self.type_sequence_label_size snake_case = TFViTForImageClassification(A__ ) snake_case = model(A__ , labels=A__ , training=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. snake_case = self.image_size // 2 snake_case = pixel_values[:, :, :image_size, :image_size] snake_case = model(A__ , interpolate_pos_encoding=A__ , training=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images snake_case = 1 snake_case = TFViTForImageClassification(A__ ) snake_case = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case = model(A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.prepare_config_and_inputs() snake_case , snake_case , snake_case = config_and_inputs snake_case = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class _lowercase ( __a , __a , unittest.TestCase ): _UpperCAmelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () _UpperCAmelCase = ( {'''feature-extraction''': TFViTModel, '''image-classification''': TFViTForImageClassification} if is_tf_available() else {} ) _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False def UpperCamelCase ( self ) -> List[Any]: snake_case = TFViTModelTester(self ) snake_case = ConfigTester(self , config_class=A__ , has_text_modality=A__ , hidden_size=37 ) def UpperCamelCase ( self ) -> int: self.config_tester.run_common_tests() @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def UpperCamelCase ( self ) -> int: pass @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def UpperCamelCase ( self ) -> str: pass def UpperCamelCase ( self ) -> Union[str, Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(A__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) snake_case = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A__ , tf.keras.layers.Layer ) ) def UpperCamelCase ( self ) -> List[Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(A__ ) snake_case = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case = [*signature.parameters.keys()] snake_case = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A__ ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A__ ) def UpperCamelCase ( self ) -> Optional[Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A__ ) @slow def UpperCamelCase ( self ) -> Any: snake_case = TFViTModel.from_pretrained('''google/vit-base-patch16-224''' ) self.assertIsNotNone(A__ ) def __UpperCamelCase ( ) ->Any: snake_case = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class _lowercase ( unittest.TestCase ): @cached_property def UpperCamelCase ( self ) -> Optional[int]: return ViTImageProcessor.from_pretrained('''google/vit-base-patch16-224''' ) if is_vision_available() else None @slow def UpperCamelCase ( self ) -> Dict: snake_case = TFViTForImageClassification.from_pretrained('''google/vit-base-patch16-224''' ) snake_case = self.default_image_processor snake_case = prepare_img() snake_case = image_processor(images=A__ , return_tensors='''tf''' ) # forward pass snake_case = model(**A__ ) # verify the logits snake_case = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , A__ ) snake_case = tf.constant([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ) tf.debugging.assert_near(outputs.logits[0, :3] , A__ , atol=1e-4 )
44
0
'''simple docstring''' import string def __UpperCamelCase ( a : str ) ->Optional[Any]: snake_case = '' for i in sequence: snake_case = ord(lowerCAmelCase__ ) if 65 <= extract <= 90: output += chr(155 - extract ) elif 97 <= extract <= 122: output += chr(219 - extract ) else: output += i return output def __UpperCamelCase ( a : str ) ->str: snake_case = string.ascii_letters snake_case = string.ascii_lowercase[::-1] + string.ascii_uppercase[::-1] return "".join( letters_reversed[letters.index(lowerCAmelCase__ )] if c in letters else c for c in sequence ) def __UpperCamelCase ( ) ->Optional[Any]: from timeit import timeit print('''Running performance benchmarks...''' ) snake_case = 'from string import printable ; from __main__ import atbash, atbash_slow' print(f"""> atbash_slow(): {timeit('atbash_slow(printable)' , setup=lowerCAmelCase__ )} seconds""" ) print(f"""> atbash(): {timeit('atbash(printable)' , setup=lowerCAmelCase__ )} seconds""" ) if __name__ == "__main__": for example in ("ABCDEFGH", "123GGjj", "testStringtest", "with space"): print(f'{example} encrypted in atbash: {atbash(example)}') benchmark()
708
'''simple docstring''' import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from datasets.iterable_dataset import IterableDataset from datasets.load import dataset_module_factory, import_main_class from datasets.utils.file_utils import cached_path _lowercase = [ {'dataset': 'wikipedia', 'config_name': '20220301.de'}, {'dataset': 'wikipedia', 'config_name': '20220301.en'}, {'dataset': 'wikipedia', 'config_name': '20220301.fr'}, {'dataset': 'wikipedia', 'config_name': '20220301.frr'}, {'dataset': 'wikipedia', 'config_name': '20220301.it'}, {'dataset': 'wikipedia', 'config_name': '20220301.simple'}, {'dataset': 'snli', 'config_name': 'plain_text'}, {'dataset': 'eli5', 'config_name': 'LFQA_reddit'}, {'dataset': 'wiki40b', 'config_name': 'en'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.compressed'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.no_index'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.multiset.no_index'}, {'dataset': 'natural_questions', 'config_name': 'default'}, ] def __UpperCamelCase ( a : Dict=True ) ->str: if with_config: return [ { "testcase_name": d["dataset"] + "/" + d["config_name"], "dataset": d["dataset"], "config_name": d["config_name"], } for d in DATASETS_ON_HF_GCP ] else: return [ {"testcase_name": dataset, "dataset": dataset} for dataset in {d["dataset"] for d in DATASETS_ON_HF_GCP} ] @parameterized.named_parameters(list_datasets_on_hf_gcp_parameters(with_config=__a ) ) class _lowercase ( __a ): _UpperCAmelCase = None _UpperCAmelCase = None def UpperCamelCase ( self , A__ , A__ ) -> str: with TemporaryDirectory() as tmp_dir: snake_case = dataset_module_factory(A__ , cache_dir=A__ ) snake_case = import_main_class(dataset_module.module_path , dataset=A__ ) snake_case = builder_cls( cache_dir=A__ , config_name=A__ , hash=dataset_module.hash , ) snake_case = '''/'''.join( [ HF_GCP_BASE_URL, builder_instance._relative_data_dir(with_hash=A__ ).replace(os.sep , '''/''' ), config.DATASET_INFO_FILENAME, ] ) snake_case = cached_path(A__ , cache_dir=A__ ) self.assertTrue(os.path.exists(A__ ) ) @pytest.mark.integration def __UpperCamelCase ( a : List[str] ) ->Any: snake_case = tmp_path_factory.mktemp('''test_hf_gcp''' ) / '''test_wikipedia_simple''' snake_case = dataset_module_factory('''wikipedia''' , cache_dir=a ) snake_case = import_main_class(dataset_module.module_path ) snake_case = builder_cls( cache_dir=a , config_name='''20220301.frr''' , hash=dataset_module.hash , ) # use the HF cloud storage, not the original download_and_prepare that uses apache-beam snake_case = None builder_instance.download_and_prepare() snake_case = builder_instance.as_dataset() assert ds @pytest.mark.integration def __UpperCamelCase ( a : Any ) ->Union[str, Any]: snake_case = dataset_module_factory('''wikipedia''' , cache_dir=a ) snake_case = import_main_class(dataset_module.module_path , dataset=a ) snake_case = builder_cls( cache_dir=a , config_name='''20220301.frr''' , hash=dataset_module.hash , ) snake_case = builder_instance.as_streaming_dataset() assert ds assert isinstance(a , a ) assert "train" in ds assert isinstance(ds['''train'''] , a ) assert next(iter(ds['''train'''] ) )
44
0
'''simple docstring''' from ...processing_utils import ProcessorMixin class _lowercase ( a__ ): _UpperCAmelCase = "SpeechT5FeatureExtractor" _UpperCAmelCase = "SpeechT5Tokenizer" def __init__( self , A__ , A__ ) -> Optional[Any]: super().__init__(lowerCamelCase_ , lowerCamelCase_ ) def __call__( self , *A__ , **A__ ) -> Dict: snake_case = kwargs.pop('''audio''' , lowerCamelCase_ ) snake_case = kwargs.pop('''text''' , lowerCamelCase_ ) snake_case = kwargs.pop('''text_target''' , lowerCamelCase_ ) snake_case = kwargs.pop('''audio_target''' , lowerCamelCase_ ) snake_case = kwargs.pop('''sampling_rate''' , lowerCamelCase_ ) if audio is not None and text is not None: raise ValueError( '''Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?''' ) if audio_target is not None and text_target is not None: raise ValueError( '''Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?''' ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( '''You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process.''' ) if audio is not None: snake_case = self.feature_extractor(lowerCamelCase_ , *lowerCamelCase_ , sampling_rate=lowerCamelCase_ , **lowerCamelCase_ ) elif text is not None: snake_case = self.tokenizer(lowerCamelCase_ , **lowerCamelCase_ ) else: snake_case = None if audio_target is not None: snake_case = self.feature_extractor(audio_target=lowerCamelCase_ , *lowerCamelCase_ , sampling_rate=lowerCamelCase_ , **lowerCamelCase_ ) snake_case = targets['''input_values'''] elif text_target is not None: snake_case = self.tokenizer(lowerCamelCase_ , **lowerCamelCase_ ) snake_case = targets['''input_ids'''] else: snake_case = None if inputs is None: return targets if targets is not None: snake_case = labels snake_case = targets.get('''attention_mask''' ) if decoder_attention_mask is not None: snake_case = decoder_attention_mask return inputs def UpperCamelCase ( self , *A__ , **A__ ) -> Union[str, Any]: snake_case = kwargs.pop('''input_values''' , lowerCamelCase_ ) snake_case = kwargs.pop('''input_ids''' , lowerCamelCase_ ) snake_case = kwargs.pop('''labels''' , lowerCamelCase_ ) if input_values is not None and input_ids is not None: raise ValueError('''Cannot process both `input_values` and `input_ids` inputs.''' ) if input_values is None and input_ids is None and labels is None: raise ValueError( '''You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded.''' ) if input_values is not None: snake_case = self.feature_extractor.pad(lowerCamelCase_ , *lowerCamelCase_ , **lowerCamelCase_ ) elif input_ids is not None: snake_case = self.tokenizer.pad(lowerCamelCase_ , **lowerCamelCase_ ) else: snake_case = None if labels is not None: if "input_ids" in labels or (isinstance(lowerCamelCase_ , lowerCamelCase_ ) and "input_ids" in labels[0]): snake_case = self.tokenizer.pad(lowerCamelCase_ , **lowerCamelCase_ ) snake_case = targets['''input_ids'''] else: snake_case = self.feature_extractor.feature_size snake_case = self.feature_extractor.num_mel_bins snake_case = self.feature_extractor.pad(lowerCamelCase_ , *lowerCamelCase_ , **lowerCamelCase_ ) snake_case = feature_size_hack snake_case = targets['''input_values'''] else: snake_case = None if inputs is None: return targets if targets is not None: snake_case = labels snake_case = targets.get('''attention_mask''' ) if decoder_attention_mask is not None: snake_case = decoder_attention_mask return inputs def UpperCamelCase ( self , *A__ , **A__ ) -> Dict: return self.tokenizer.batch_decode(*lowerCamelCase_ , **lowerCamelCase_ ) def UpperCamelCase ( self , *A__ , **A__ ) -> Union[str, Any]: return self.tokenizer.decode(*lowerCamelCase_ , **lowerCamelCase_ )
709
'''simple docstring''' def __UpperCamelCase ( a : int , a : int ) ->int: while b: snake_case , snake_case = b, a % b return a def __UpperCamelCase ( a : int , a : int ) ->int: return a if b == 0 else euclidean_gcd_recursive(a , a % b ) def __UpperCamelCase ( ) ->Optional[Any]: print(f"""euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}""" ) print(f"""euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}""" ) print(f"""euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}""" ) print(f"""euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}""" ) print(f"""euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}""" ) print(f"""euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}""" ) print(f"""euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}""" ) print(f"""euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}""" ) if __name__ == "__main__": main()
44
0
'''simple docstring''' import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def __UpperCamelCase ( a : Optional[Any] ) ->Dict: snake_case = VideoMAEConfig() set_architecture_configs(a , a ) if "finetuned" not in model_name: snake_case = False if "finetuned" in model_name: snake_case = '''huggingface/label-files''' if "kinetics" in model_name: snake_case = 400 snake_case = '''kinetics400-id2label.json''' elif "ssv2" in model_name: snake_case = 174 snake_case = '''something-something-v2-id2label.json''' else: raise ValueError('''Model name should either contain \'kinetics\' or \'ssv2\' in case it\'s fine-tuned.''' ) snake_case = json.load(open(hf_hub_download(a , a , repo_type='''dataset''' ) , '''r''' ) ) snake_case = {int(a ): v for k, v in idalabel.items()} snake_case = idalabel snake_case = {v: k for k, v in idalabel.items()} return config def __UpperCamelCase ( a : Any , a : List[Any] ) ->int: if "small" in model_name: snake_case = 384 snake_case = 1536 snake_case = 12 snake_case = 16 snake_case = 12 snake_case = 3 snake_case = 192 snake_case = 768 elif "large" in model_name: snake_case = 1024 snake_case = 4096 snake_case = 24 snake_case = 16 snake_case = 12 snake_case = 8 snake_case = 512 snake_case = 2048 elif "huge" in model_name: snake_case = 1280 snake_case = 5120 snake_case = 32 snake_case = 16 snake_case = 12 snake_case = 8 snake_case = 640 snake_case = 2560 elif "base" not in model_name: raise ValueError('''Model name should include either \"small\", \"base\", \"large\", or \"huge\"''' ) def __UpperCamelCase ( a : Union[str, Any] ) ->Optional[Any]: if "encoder." in name: snake_case = name.replace('''encoder.''' , '''''' ) if "cls_token" in name: snake_case = name.replace('''cls_token''' , '''videomae.embeddings.cls_token''' ) if "decoder_pos_embed" in name: snake_case = name.replace('''decoder_pos_embed''' , '''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: snake_case = name.replace('''pos_embed''' , '''videomae.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: snake_case = name.replace('''patch_embed.proj''' , '''videomae.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: snake_case = name.replace('''patch_embed.norm''' , '''videomae.embeddings.norm''' ) if "decoder.blocks" in name: snake_case = name.replace('''decoder.blocks''' , '''decoder.decoder_layers''' ) if "blocks" in name: snake_case = name.replace('''blocks''' , '''videomae.encoder.layer''' ) if "attn.proj" in name: snake_case = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name and "bias" not in name: snake_case = name.replace('''attn''' , '''attention.self''' ) if "attn" in name: snake_case = name.replace('''attn''' , '''attention.attention''' ) if "norm1" in name: snake_case = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: snake_case = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: snake_case = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: snake_case = name.replace('''mlp.fc2''' , '''output.dense''' ) if "decoder_embed" in name: snake_case = name.replace('''decoder_embed''' , '''decoder.decoder_embed''' ) if "decoder_norm" in name: snake_case = name.replace('''decoder_norm''' , '''decoder.decoder_norm''' ) if "decoder_pred" in name: snake_case = name.replace('''decoder_pred''' , '''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name and "fc" not in name: snake_case = name.replace('''norm.weight''' , '''videomae.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name and "fc" not in name: snake_case = name.replace('''norm.bias''' , '''videomae.layernorm.bias''' ) if "head" in name and "decoder" not in name: snake_case = name.replace('''head''' , '''classifier''' ) return name def __UpperCamelCase ( a : Any , a : List[str] ) ->Tuple: for key in orig_state_dict.copy().keys(): snake_case = orig_state_dict.pop(a ) if key.startswith('''encoder.''' ): snake_case = key.replace('''encoder.''' , '''''' ) if "qkv" in key: snake_case = key.split('''.''' ) if key.startswith('''decoder.blocks''' ): snake_case = config.decoder_hidden_size snake_case = int(key_split[2] ) snake_case = '''decoder.decoder_layers.''' if "weight" in key: snake_case = val[:dim, :] snake_case = val[dim : dim * 2, :] snake_case = val[-dim:, :] else: snake_case = config.hidden_size snake_case = int(key_split[1] ) snake_case = '''videomae.encoder.layer.''' if "weight" in key: snake_case = val[:dim, :] snake_case = val[dim : dim * 2, :] snake_case = val[-dim:, :] else: snake_case = val return orig_state_dict def __UpperCamelCase ( ) ->List[str]: snake_case = hf_hub_download( repo_id='''hf-internal-testing/spaghetti-video''' , filename='''eating_spaghetti.npy''' , repo_type='''dataset''' ) snake_case = np.load(a ) return list(a ) def __UpperCamelCase ( a : Union[str, Any] , a : Dict , a : List[Any] , a : Any ) ->List[Any]: snake_case = get_videomae_config(a ) if "finetuned" in model_name: snake_case = VideoMAEForVideoClassification(a ) else: snake_case = VideoMAEForPreTraining(a ) # download original checkpoint, hosted on Google Drive snake_case = '''pytorch_model.bin''' gdown.cached_download(a , a , quiet=a ) snake_case = torch.load(a , map_location='''cpu''' ) if "model" in files: snake_case = files['''model'''] else: snake_case = files['''module'''] snake_case = convert_state_dict(a , a ) model.load_state_dict(a ) model.eval() # verify model on basic input snake_case = VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) snake_case = prepare_video() snake_case = image_processor(a , return_tensors='''pt''' ) if "finetuned" not in model_name: snake_case = hf_hub_download(repo_id='''hf-internal-testing/bool-masked-pos''' , filename='''bool_masked_pos.pt''' ) snake_case = torch.load(a ) snake_case = model(**a ) snake_case = outputs.logits snake_case = [ '''videomae-small-finetuned-kinetics''', '''videomae-small-finetuned-ssv2''', # Kinetics-400 checkpoints (short = pretrained only for 800 epochs instead of 1600) '''videomae-base-short''', '''videomae-base-short-finetuned-kinetics''', '''videomae-base''', '''videomae-base-finetuned-kinetics''', '''videomae-large''', '''videomae-large-finetuned-kinetics''', '''videomae-huge-finetuned-kinetics''', # Something-Something-v2 checkpoints (short = pretrained only for 800 epochs instead of 2400) '''videomae-base-short-ssv2''', '''videomae-base-short-finetuned-ssv2''', '''videomae-base-ssv2''', '''videomae-base-finetuned-ssv2''', ] # NOTE: logits were tested with image_mean and image_std equal to [0.5, 0.5, 0.5] and [0.5, 0.5, 0.5] if model_name == "videomae-small-finetuned-kinetics": snake_case = torch.Size([1, 400] ) snake_case = torch.tensor([-0.9291, -0.4061, -0.9307] ) elif model_name == "videomae-small-finetuned-ssv2": snake_case = torch.Size([1, 174] ) snake_case = torch.tensor([0.2671, -0.4689, -0.8235] ) elif model_name == "videomae-base": snake_case = torch.Size([1, 1408, 1536] ) snake_case = torch.tensor([[0.7739, 0.7968, 0.7089], [0.6701, 0.7487, 0.6209], [0.4287, 0.5158, 0.4773]] ) elif model_name == "videomae-base-short": snake_case = torch.Size([1, 1408, 1536] ) snake_case = torch.tensor([[0.7994, 0.9612, 0.8508], [0.7401, 0.8958, 0.8302], [0.5862, 0.7468, 0.7325]] ) # we verified the loss both for normalized and unnormalized targets for this one snake_case = torch.tensor([0.5142] ) if config.norm_pix_loss else torch.tensor([0.6469] ) elif model_name == "videomae-large": snake_case = torch.Size([1, 1408, 1536] ) snake_case = torch.tensor([[0.7149, 0.7997, 0.6966], [0.6768, 0.7869, 0.6948], [0.5139, 0.6221, 0.5605]] ) elif model_name == "videomae-large-finetuned-kinetics": snake_case = torch.Size([1, 400] ) snake_case = torch.tensor([0.0771, 0.0011, -0.3625] ) elif model_name == "videomae-huge-finetuned-kinetics": snake_case = torch.Size([1, 400] ) snake_case = torch.tensor([0.2433, 0.1632, -0.4894] ) elif model_name == "videomae-base-short-finetuned-kinetics": snake_case = torch.Size([1, 400] ) snake_case = torch.tensor([0.6588, 0.0990, -0.2493] ) elif model_name == "videomae-base-finetuned-kinetics": snake_case = torch.Size([1, 400] ) snake_case = torch.tensor([0.3669, -0.0688, -0.2421] ) elif model_name == "videomae-base-short-ssv2": snake_case = torch.Size([1, 1408, 1536] ) snake_case = torch.tensor([[0.4712, 0.5296, 0.5786], [0.2278, 0.2729, 0.4026], [0.0352, 0.0730, 0.2506]] ) elif model_name == "videomae-base-short-finetuned-ssv2": snake_case = torch.Size([1, 174] ) snake_case = torch.tensor([-0.0537, -0.1539, -0.3266] ) elif model_name == "videomae-base-ssv2": snake_case = torch.Size([1, 1408, 1536] ) snake_case = torch.tensor([[0.8131, 0.8727, 0.8546], [0.7366, 0.9377, 0.8870], [0.5935, 0.8874, 0.8564]] ) elif model_name == "videomae-base-finetuned-ssv2": snake_case = torch.Size([1, 174] ) snake_case = torch.tensor([0.1961, -0.8337, -0.6389] ) else: raise ValueError(f"""Model name not supported. Should be one of {model_names}""" ) # verify logits assert logits.shape == expected_shape if "finetuned" in model_name: assert torch.allclose(logits[0, :3] , a , atol=1e-4 ) else: print('''Logits:''' , logits[0, :3, :3] ) assert torch.allclose(logits[0, :3, :3] , a , atol=1e-4 ) print('''Logits ok!''' ) # verify loss, if applicable if model_name == "videomae-base-short": snake_case = outputs.loss assert torch.allclose(a , a , atol=1e-4 ) print('''Loss ok!''' ) if pytorch_dump_folder_path is not None: print(f"""Saving model and image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(a ) model.save_pretrained(a ) if push_to_hub: print('''Pushing to the hub...''' ) model.push_to_hub(a , organization='''nielsr''' ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://drive.google.com/u/1/uc?id=1tEhLyskjb755TJ65ptsrafUG2llSwQE1&amp;export=download&amp;confirm=t&amp;uuid=aa3276eb-fb7e-482a-adec-dc7171df14c4', type=str, help=( 'URL of the original PyTorch checkpoint (on Google Drive) you\'d like to convert. Should be a direct' ' download link.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default='/Users/nielsrogge/Documents/VideoMAE/Test', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--model_name', default='videomae-base', type=str, help='Name of the model.') parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) _lowercase = parser.parse_args() convert_videomae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
710
'''simple docstring''' import argparse import copy def __UpperCamelCase ( a : Union[str, Any] ) ->Tuple: snake_case = {} with open(a ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: snake_case = [] _list.append([line.split()[1], line.split()[2]] ) snake_case = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: snake_case = [] _list.append([line.split()[0], line.split()[2]] ) snake_case = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def __UpperCamelCase ( a : Dict , a : Tuple ) ->int: with open(a ) as f: snake_case = f.read(1 ) snake_case = start_node snake_case = [] snake_case = start_node snake_case = 0 while visiting not in first_solution: snake_case = 1_0000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(a ) and k[0] not in first_solution: snake_case = k[1] snake_case = k[0] first_solution.append(a ) snake_case = distance_of_first_solution + int(a ) snake_case = best_node first_solution.append(a ) snake_case = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 snake_case = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_0000 ) return first_solution, distance_of_first_solution def __UpperCamelCase ( a : Optional[int] , a : str ) ->str: snake_case = [] for n in solution[1:-1]: snake_case = solution.index(a ) for kn in solution[1:-1]: snake_case = solution.index(a ) if n == kn: continue snake_case = copy.deepcopy(a ) snake_case = kn snake_case = n snake_case = 0 for k in _tmp[:-1]: snake_case = _tmp[_tmp.index(a ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: snake_case = distance + int(i[1] ) _tmp.append(a ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) snake_case = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda a : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def __UpperCamelCase ( a : Any , a : Optional[Any] , a : int , a : Optional[int] , a : Union[str, Any] ) ->List[Any]: snake_case = 1 snake_case = first_solution snake_case = [] snake_case = distance_of_first_solution snake_case = solution while count <= iters: snake_case = find_neighborhood(a , a ) snake_case = 0 snake_case = neighborhood[index_of_best_solution] snake_case = len(a ) - 1 snake_case = False while not found: snake_case = 0 while i < len(a ): if best_solution[i] != solution[i]: snake_case = best_solution[i] snake_case = solution[i] break snake_case = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) snake_case = True snake_case = best_solution[:-1] snake_case = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: snake_case = cost snake_case = solution else: snake_case = index_of_best_solution + 1 snake_case = neighborhood[index_of_best_solution] if len(a ) >= size: tabu_list.pop(0 ) snake_case = count + 1 return best_solution_ever, best_cost def __UpperCamelCase ( a : Union[str, Any]=None ) ->Optional[Any]: snake_case = generate_neighbours(args.File ) snake_case , snake_case = generate_first_solution( args.File , a ) snake_case , snake_case = tabu_search( a , a , a , args.Iterations , args.Size , ) print(f"""Best solution: {best_sol}, with total distance: {best_cost}.""" ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser(description='Tabu Search') parser.add_argument( '-f', '--File', type=str, help='Path to the file containing the data', required=True, ) parser.add_argument( '-i', '--Iterations', type=int, help='How many iterations the algorithm should perform', required=True, ) parser.add_argument( '-s', '--Size', type=int, help='Size of the tabu list', required=True ) # Pass the arguments to main method main(parser.parse_args())
44
0
'''simple docstring''' from __future__ import annotations def __UpperCamelCase ( a : int ) ->list[int]: snake_case = [True] * limit snake_case = False snake_case = False snake_case = True for i in range(3 , int(limit**0.5 + 1 ) , 2 ): snake_case = i * 2 while index < limit: snake_case = False snake_case = index + i snake_case = [2] for i in range(3 , __lowerCAmelCase , 2 ): if is_prime[i]: primes.append(__lowerCAmelCase ) return primes def __UpperCamelCase ( a : Tuple = 100_0000 ) ->int: snake_case = prime_sieve(__lowerCAmelCase ) snake_case = 0 snake_case = 0 for i in range(len(__lowerCAmelCase ) ): for j in range(i + length , len(__lowerCAmelCase ) ): snake_case = sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: snake_case = j - i snake_case = sol return largest if __name__ == "__main__": print(f'{solution() = }')
711
'''simple docstring''' from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .notes_encoder import SpectrogramNotesEncoder from .continous_encoder import SpectrogramContEncoder from .pipeline_spectrogram_diffusion import ( SpectrogramContEncoder, SpectrogramDiffusionPipeline, TaFilmDecoder, ) try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .midi_utils import MidiProcessor
44
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowercase = { 'configuration_clipseg': [ 'CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP', 'CLIPSegConfig', 'CLIPSegTextConfig', 'CLIPSegVisionConfig', ], 'processing_clipseg': ['CLIPSegProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ 'CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST', 'CLIPSegModel', 'CLIPSegPreTrainedModel', 'CLIPSegTextModel', 'CLIPSegVisionModel', 'CLIPSegForImageSegmentation', ] if TYPE_CHECKING: from .configuration_clipseg import ( CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPSegConfig, CLIPSegTextConfig, CLIPSegVisionConfig, ) from .processing_clipseg import CLIPSegProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clipseg import ( CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPSegForImageSegmentation, CLIPSegModel, CLIPSegPreTrainedModel, CLIPSegTextModel, CLIPSegVisionModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
712
'''simple docstring''' from ...processing_utils import ProcessorMixin class _lowercase ( __a ): _UpperCAmelCase = '''WhisperFeatureExtractor''' _UpperCAmelCase = '''WhisperTokenizer''' def __init__( self , A__ , A__ ) -> Optional[Any]: super().__init__(A__ , A__ ) snake_case = self.feature_extractor snake_case = False def UpperCamelCase ( self , A__=None , A__=None , A__=True ) -> Union[str, Any]: return self.tokenizer.get_decoder_prompt_ids(task=A__ , language=A__ , no_timestamps=A__ ) def __call__( self , *A__ , **A__ ) -> Dict: # For backward compatibility if self._in_target_context_manager: return self.current_processor(*A__ , **A__ ) snake_case = kwargs.pop('''audio''' , A__ ) snake_case = kwargs.pop('''sampling_rate''' , A__ ) snake_case = kwargs.pop('''text''' , A__ ) if len(A__ ) > 0: snake_case = args[0] snake_case = args[1:] if audio is None and text is None: raise ValueError('''You need to specify either an `audio` or `text` input to process.''' ) if audio is not None: snake_case = self.feature_extractor(A__ , *A__ , sampling_rate=A__ , **A__ ) if text is not None: snake_case = self.tokenizer(A__ , **A__ ) if text is None: return inputs elif audio is None: return encodings else: snake_case = encodings['''input_ids'''] return inputs def UpperCamelCase ( self , *A__ , **A__ ) -> Optional[Any]: return self.tokenizer.batch_decode(*A__ , **A__ ) def UpperCamelCase ( self , *A__ , **A__ ) -> str: return self.tokenizer.decode(*A__ , **A__ ) def UpperCamelCase ( self , A__ , A__="np" ) -> Optional[Any]: return self.tokenizer.get_prompt_ids(A__ , return_tensors=A__ )
44
0
'''simple docstring''' from diffusers.utils.testing_utils import require_onnxruntime @require_onnxruntime class _lowercase : pass
713
'''simple docstring''' import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class _lowercase ( __a ): _UpperCAmelCase = '''char''' _UpperCAmelCase = '''bpe''' _UpperCAmelCase = '''wp''' _lowercase = (DecodeType.CHARACTER, DecodeType.BPE, DecodeType.WORDPIECE) class _lowercase ( __a ): _UpperCAmelCase = ['''image_processor''', '''char_tokenizer'''] _UpperCAmelCase = '''ViTImageProcessor''' _UpperCAmelCase = '''MgpstrTokenizer''' def __init__( self , A__=None , A__=None , **A__ ) -> List[Any]: snake_case = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , A__ , ) snake_case = kwargs.pop('''feature_extractor''' ) snake_case = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) snake_case = tokenizer snake_case = AutoTokenizer.from_pretrained('''gpt2''' ) snake_case = AutoTokenizer.from_pretrained('''bert-base-uncased''' ) super().__init__(A__ , A__ ) def __call__( self , A__=None , A__=None , A__=None , **A__ ) -> List[str]: if images is None and text is None: raise ValueError('''You need to specify either an `images` or `text` input to process.''' ) if images is not None: snake_case = self.image_processor(A__ , return_tensors=A__ , **A__ ) if text is not None: snake_case = self.char_tokenizer(A__ , return_tensors=A__ , **A__ ) if text is None: return inputs elif images is None: return encodings else: snake_case = encodings['''input_ids'''] return inputs def UpperCamelCase ( self , A__ ) -> Dict: snake_case , snake_case , snake_case = sequences snake_case = char_preds.size(0 ) snake_case , snake_case = self._decode_helper(A__ , '''char''' ) snake_case , snake_case = self._decode_helper(A__ , '''bpe''' ) snake_case , snake_case = self._decode_helper(A__ , '''wp''' ) snake_case = [] snake_case = [] for i in range(A__ ): snake_case = [char_scores[i], bpe_scores[i], wp_scores[i]] snake_case = [char_strs[i], bpe_strs[i], wp_strs[i]] snake_case = scores.index(max(A__ ) ) final_strs.append(strs[max_score_index] ) final_scores.append(scores[max_score_index] ) snake_case = {} snake_case = final_strs snake_case = final_scores snake_case = char_strs snake_case = bpe_strs snake_case = wp_strs return out def UpperCamelCase ( self , A__ , A__ ) -> Optional[Any]: if format == DecodeType.CHARACTER: snake_case = self.char_decode snake_case = 1 snake_case = '''[s]''' elif format == DecodeType.BPE: snake_case = self.bpe_decode snake_case = 2 snake_case = '''#''' elif format == DecodeType.WORDPIECE: snake_case = self.wp_decode snake_case = 1_02 snake_case = '''[SEP]''' else: raise ValueError(F"""Format {format} is not supported.""" ) snake_case , snake_case = [], [] snake_case = pred_logits.size(0 ) snake_case = pred_logits.size(1 ) snake_case , snake_case = pred_logits.topk(1 , dim=-1 , largest=A__ , sorted=A__ ) snake_case = preds_index.view(-1 , A__ )[:, 1:] snake_case = decoder(A__ ) snake_case , snake_case = torch.nn.functional.softmax(A__ , dim=2 ).max(dim=2 ) snake_case = preds_max_prob[:, 1:] for index in range(A__ ): snake_case = preds_str[index].find(A__ ) snake_case = preds_str[index][:pred_eos] snake_case = preds_index[index].cpu().tolist() snake_case = pred_index.index(A__ ) if eos_token in pred_index else -1 snake_case = preds_max_prob[index][: pred_eos_index + 1] snake_case = pred_max_prob.cumprod(dim=0 )[-1] if pred_max_prob.nelement() != 0 else 0.0 dec_strs.append(A__ ) conf_scores.append(A__ ) return dec_strs, conf_scores def UpperCamelCase ( self , A__ ) -> int: snake_case = [seq.replace(''' ''' , '''''' ) for seq in self.char_tokenizer.batch_decode(A__ )] return decode_strs def UpperCamelCase ( self , A__ ) -> List[str]: return self.bpe_tokenizer.batch_decode(A__ ) def UpperCamelCase ( self , A__ ) -> Union[str, Any]: snake_case = [seq.replace(''' ''' , '''''' ) for seq in self.wp_tokenizer.batch_decode(A__ )] return decode_strs
44
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _lowercase = {'configuration_speech_encoder_decoder': ['SpeechEncoderDecoderConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ['SpeechEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ['FlaxSpeechEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_speech_encoder_decoder import SpeechEncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_encoder_decoder import SpeechEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_speech_encoder_decoder import FlaxSpeechEncoderDecoderModel else: import sys _lowercase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
714
'''simple docstring''' import os from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen, xsplitext from ..table import array_cast from ..utils.py_utils import no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: from .features import FeatureType _lowercase , _lowercase , _lowercase = False, False, False @dataclass class _lowercase : _UpperCAmelCase = None _UpperCAmelCase = True _UpperCAmelCase = True _UpperCAmelCase = None # Automatically constructed _UpperCAmelCase = "dict" _UpperCAmelCase = pa.struct({'''bytes''': pa.binary(), '''path''': pa.string()} ) _UpperCAmelCase = field(default='''Audio''' , init=__a , repr=__a ) def __call__( self ) -> Optional[Any]: return self.pa_type def UpperCamelCase ( self , A__ ) -> dict: try: import soundfile as sf # soundfile is a dependency of librosa, needed to decode audio files. except ImportError as err: raise ImportError('''To support encoding audio data, please install \'soundfile\'.''' ) from err if isinstance(A__ , A__ ): return {"bytes": None, "path": value} elif isinstance(A__ , A__ ): return {"bytes": value, "path": None} elif "array" in value: # convert the audio array to wav bytes snake_case = BytesIO() sf.write(A__ , value['''array'''] , value['''sampling_rate'''] , format='''wav''' ) return {"bytes": buffer.getvalue(), "path": None} elif value.get('''path''' ) is not None and os.path.isfile(value['''path'''] ): # we set "bytes": None to not duplicate the data if they're already available locally if value["path"].endswith('''pcm''' ): # "PCM" only has raw audio bytes if value.get('''sampling_rate''' ) is None: # At least, If you want to convert "PCM-byte" to "WAV-byte", you have to know sampling rate raise KeyError('''To use PCM files, please specify a \'sampling_rate\' in Audio object''' ) if value.get('''bytes''' ): # If we already had PCM-byte, we don`t have to make "read file, make bytes" (just use it!) snake_case = np.frombuffer(value['''bytes'''] , dtype=np.intaa ).astype(np.floataa ) / 3_27_67 else: snake_case = np.memmap(value['''path'''] , dtype='''h''' , mode='''r''' ).astype(np.floataa ) / 3_27_67 snake_case = BytesIO(bytes() ) sf.write(A__ , A__ , value['''sampling_rate'''] , format='''wav''' ) return {"bytes": buffer.getvalue(), "path": None} else: return {"bytes": None, "path": value.get('''path''' )} elif value.get('''bytes''' ) is not None or value.get('''path''' ) is not None: # store the audio bytes, and path is used to infer the audio format using the file extension return {"bytes": value.get('''bytes''' ), "path": value.get('''path''' )} else: raise ValueError( F"""An audio sample should have one of 'path' or 'bytes' but they are missing or None in {value}.""" ) def UpperCamelCase ( self , A__ , A__ = None ) -> dict: if not self.decode: raise RuntimeError('''Decoding is disabled for this feature. Please use Audio(decode=True) instead.''' ) snake_case , snake_case = (value['''path'''], BytesIO(value['''bytes'''] )) if value['''bytes'''] is not None else (value['''path'''], None) if path is None and file is None: raise ValueError(F"""An audio sample should have one of 'path' or 'bytes' but both are None in {value}.""" ) try: import librosa import soundfile as sf except ImportError as err: raise ImportError('''To support decoding audio files, please install \'librosa\' and \'soundfile\'.''' ) from err snake_case = xsplitext(A__ )[1][1:].lower() if path is not None else None if not config.IS_OPUS_SUPPORTED and audio_format == "opus": raise RuntimeError( '''Decoding \'opus\' files requires system library \'libsndfile\'>=1.0.31, ''' '''You can try to update `soundfile` python library: `pip install "soundfile>=0.12.1"`. ''' ) elif not config.IS_MP3_SUPPORTED and audio_format == "mp3": raise RuntimeError( '''Decoding \'mp3\' files requires system library \'libsndfile\'>=1.1.0, ''' '''You can try to update `soundfile` python library: `pip install "soundfile>=0.12.1"`. ''' ) if file is None: snake_case = token_per_repo_id or {} snake_case = path.split('''::''' )[-1] try: snake_case = string_to_dict(A__ , config.HUB_DATASETS_URL )['''repo_id'''] snake_case = token_per_repo_id[repo_id] except (ValueError, KeyError): snake_case = None with xopen(A__ , '''rb''' , use_auth_token=A__ ) as f: snake_case , snake_case = sf.read(A__ ) else: snake_case , snake_case = sf.read(A__ ) snake_case = array.T if self.mono: snake_case = librosa.to_mono(A__ ) if self.sampling_rate and self.sampling_rate != sampling_rate: snake_case = librosa.resample(A__ , orig_sr=A__ , target_sr=self.sampling_rate ) snake_case = self.sampling_rate return {"path": path, "array": array, "sampling_rate": sampling_rate} def UpperCamelCase ( self ) -> Union["FeatureType", Dict[str, "FeatureType"]]: from .features import Value if self.decode: raise ValueError('''Cannot flatten a decoded Audio feature.''' ) return { "bytes": Value('''binary''' ), "path": Value('''string''' ), } def UpperCamelCase ( self , A__ ) -> pa.StructArray: if pa.types.is_string(storage.type ): snake_case = pa.array([None] * len(A__ ) , type=pa.binary() ) snake_case = pa.StructArray.from_arrays([bytes_array, storage] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): snake_case = pa.array([None] * len(A__ ) , type=pa.string() ) snake_case = pa.StructArray.from_arrays([storage, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ) and storage.type.get_all_field_indices('''array''' ): snake_case = pa.array([Audio().encode_example(A__ ) if x is not None else None for x in storage.to_pylist()] ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index('''bytes''' ) >= 0: snake_case = storage.field('''bytes''' ) else: snake_case = pa.array([None] * len(A__ ) , type=pa.binary() ) if storage.type.get_field_index('''path''' ) >= 0: snake_case = storage.field('''path''' ) else: snake_case = pa.array([None] * len(A__ ) , type=pa.string() ) snake_case = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=storage.is_null() ) return array_cast(A__ , self.pa_type ) def UpperCamelCase ( self , A__ ) -> pa.StructArray: @no_op_if_value_is_null def path_to_bytes(A__ ): with xopen(A__ , '''rb''' ) as f: snake_case = f.read() return bytes_ snake_case = pa.array( [ (path_to_bytes(x['''path'''] ) if x['''bytes'''] is None else x['''bytes''']) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) snake_case = pa.array( [os.path.basename(A__ ) if path is not None else None for path in storage.field('''path''' ).to_pylist()] , type=pa.string() , ) snake_case = pa.StructArray.from_arrays([bytes_array, path_array] , ['''bytes''', '''path'''] , mask=bytes_array.is_null() ) return array_cast(A__ , self.pa_type )
44
0
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
715
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class _lowercase : @staticmethod def UpperCamelCase ( *A__ , **A__ ) -> List[Any]: pass def __UpperCamelCase ( a : Image ) ->str: snake_case = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class _lowercase ( unittest.TestCase ): _UpperCAmelCase = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def UpperCamelCase ( self , A__ , A__ , A__ ) -> Union[str, Any]: snake_case = DepthEstimationPipeline(model=A__ , image_processor=A__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def UpperCamelCase ( self , A__ , A__ ) -> List[Any]: snake_case = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} , A__ ) import datasets snake_case = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' , '''image''' , split='''test''' ) snake_case = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] , A__ , ) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def UpperCamelCase ( self ) -> Optional[Any]: pass @slow @require_torch def UpperCamelCase ( self ) -> Dict: snake_case = '''Intel/dpt-large''' snake_case = pipeline('''depth-estimation''' , model=A__ ) snake_case = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) snake_case = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) , 2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) , 2.6_6_2 ) @require_torch def UpperCamelCase ( self ) -> Any: # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
44
0
'''simple docstring''' import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() _lowercase = logging.get_logger(__name__) _lowercase = """https://openaipublic.azureedge.net/jukebox/models/""" _lowercase = { """jukebox-1b-lyrics""": [ """5b/vqvae.pth.tar""", """5b/prior_level_0.pth.tar""", """5b/prior_level_1.pth.tar""", """1b_lyrics/prior_level_2.pth.tar""", ], """jukebox-5b-lyrics""": [ """5b/vqvae.pth.tar""", """5b/prior_level_0.pth.tar""", """5b/prior_level_1.pth.tar""", """5b_lyrics/prior_level_2.pth.tar""", ], } def __UpperCamelCase ( a : Tuple ) ->int: if key.endswith('''.model.1.bias''' ) and len(key.split('''.''' ) ) > 10: snake_case = key.replace('''.model.1.bias''' , '''.conv1d_1.bias''' ) elif key.endswith('''.model.1.weight''' ) and len(key.split('''.''' ) ) > 10: snake_case = key.replace('''.model.1.weight''' , '''.conv1d_1.weight''' ) elif key.endswith('''.model.3.bias''' ) and len(key.split('''.''' ) ) > 10: snake_case = key.replace('''.model.3.bias''' , '''.conv1d_2.bias''' ) elif key.endswith('''.model.3.weight''' ) and len(key.split('''.''' ) ) > 10: snake_case = key.replace('''.model.3.weight''' , '''.conv1d_2.weight''' ) if "conditioner_blocks.0." in key: snake_case = key.replace('''conditioner_blocks.0''' , '''conditioner_blocks''' ) if "prime_prior" in key: snake_case = key.replace('''prime_prior''' , '''encoder''' ) if ".emb." in key and "total" not in key and "absolute" not in key and "relative" not in key: snake_case = key.replace('''.emb.''' , '''.''' ) if key.endswith('''k''' ): # replace vqvae.X.k with vqvae.X.codebook return key.replace('''.k''' , '''.codebook''' ) if "y_emb." in key: return key.replace('''y_emb.''' , '''metadata_embedding.''' ) if "x_emb.emb." in key: snake_case = key.replace('''0.x_emb.emb''' , '''embed_tokens''' ) if "prime_state_ln" in key: return key.replace('''prime_state_ln''' , '''encoder.final_layer_norm''' ) if ".ln" in key: return key.replace('''.ln''' , '''.layer_norm''' ) if "_ln" in key: return key.replace('''_ln''' , '''_layer_norm''' ) if "prime_state_proj" in key: return key.replace('''prime_state_proj''' , '''encoder.proj_in''' ) if "prime_x_out" in key: return key.replace('''prime_x_out''' , '''encoder.lm_head''' ) if "prior.x_out" in key: return key.replace('''x_out''' , '''fc_proj_out''' ) if "x_emb" in key: return key.replace('''x_emb''' , '''embed_tokens''' ) return key def __UpperCamelCase ( a : Tuple , a : List[str] , a : Optional[Any] , a : int ) ->Union[str, Any]: snake_case = {} import re snake_case = re.compile(R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)''' ) snake_case = re.compile( R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) snake_case = re.compile(R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)''' ) snake_case = re.compile(R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)''' ) snake_case = re.compile( R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) snake_case = re.compile(R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)''' ) snake_case = re.compile(R'''conditioner_blocks.(\d*).cond.model.(\d*).(\d).(bias|weight)''' ) snake_case = re.compile( R'''conditioner_blocks.(\d*).cond.model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) snake_case = re.compile(R'''conditioner_blocks.(\d*).cond.model.(\d*).(bias|weight)''' ) for original_key, value in state_dict.items(): # rename vqvae.encoder keys if re_encoder_block_conv_in.fullmatch(a ): snake_case = re_encoder_block_conv_in.match(a ) snake_case = regex_match.groups() snake_case = int(groups[2] ) * 2 + int(groups[3] ) snake_case = f"""encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.{groups[-1]}""" snake_case = re_encoder_block_conv_in.sub(a , a ) elif re_encoder_block_resnet.fullmatch(a ): snake_case = re_encoder_block_resnet.match(a ) snake_case = regex_match.groups() snake_case = int(groups[2] ) * 2 + int(groups[3] ) snake_case = {'''1''': 1, '''3''': 2}[groups[-2]] snake_case = f"""encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.""" snake_case = f"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" snake_case = prefix + resnet_block snake_case = re_encoder_block_resnet.sub(a , a ) elif re_encoder_block_proj_out.fullmatch(a ): snake_case = re_encoder_block_proj_out.match(a ) snake_case = regex_match.groups() snake_case = f"""encoders.{groups[0]}.level_blocks.{groups[1]}.proj_out.{groups[-1]}""" snake_case = re_encoder_block_proj_out.sub(a , a ) # rename vqvae.decoder keys elif re_decoder_block_conv_out.fullmatch(a ): snake_case = re_decoder_block_conv_out.match(a ) snake_case = regex_match.groups() snake_case = int(groups[2] ) * 2 + int(groups[3] ) - 2 snake_case = f"""decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.{groups[-1]}""" snake_case = re_decoder_block_conv_out.sub(a , a ) elif re_decoder_block_resnet.fullmatch(a ): snake_case = re_decoder_block_resnet.match(a ) snake_case = regex_match.groups() snake_case = int(groups[2] ) * 2 + int(groups[3] ) - 2 snake_case = {'''1''': 1, '''3''': 2}[groups[-2]] snake_case = f"""decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.""" snake_case = f"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" snake_case = prefix + resnet_block snake_case = re_decoder_block_resnet.sub(a , a ) elif re_decoder_block_proj_in.fullmatch(a ): snake_case = re_decoder_block_proj_in.match(a ) snake_case = regex_match.groups() snake_case = f"""decoders.{groups[0]}.level_blocks.{groups[1]}.proj_in.{groups[-1]}""" snake_case = re_decoder_block_proj_in.sub(a , a ) # rename prior cond.model to upsampler.upsample_block and resnet elif re_prior_cond_conv_out.fullmatch(a ): snake_case = re_prior_cond_conv_out.match(a ) snake_case = regex_match.groups() snake_case = int(groups[1] ) * 2 + int(groups[2] ) - 2 snake_case = f"""conditioner_blocks.upsampler.upsample_block.{block_index}.{groups[-1]}""" snake_case = re_prior_cond_conv_out.sub(a , a ) elif re_prior_cond_resnet.fullmatch(a ): snake_case = re_prior_cond_resnet.match(a ) snake_case = regex_match.groups() snake_case = int(groups[1] ) * 2 + int(groups[2] ) - 2 snake_case = {'''1''': 1, '''3''': 2}[groups[-2]] snake_case = f"""conditioner_blocks.upsampler.upsample_block.{block_index}.""" snake_case = f"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" snake_case = prefix + resnet_block snake_case = re_prior_cond_resnet.sub(a , a ) elif re_prior_cond_proj_in.fullmatch(a ): snake_case = re_prior_cond_proj_in.match(a ) snake_case = regex_match.groups() snake_case = f"""conditioner_blocks.upsampler.proj_in.{groups[-1]}""" snake_case = re_prior_cond_proj_in.sub(a , a ) # keep original key else: snake_case = original_key snake_case = replace_key(a ) if f"""{key_prefix}.{key}""" not in model_state_dict or key is None: print(f"""failed converting {original_key} to {key}, does not match""" ) # handle missmatched shape elif value.shape != model_state_dict[f"""{key_prefix}.{key}"""].shape: snake_case = model_state_dict[f"""{key_prefix}.{key}"""] print(f"""{original_key}-> {key} : \nshape {val.shape} and { value.shape}, do not match""" ) snake_case = original_key snake_case = original_key snake_case = value return new_dict @torch.no_grad() def __UpperCamelCase ( a : List[str]=None , a : Tuple=None ) ->Tuple: for file in MODEL_MAPPING[model_name]: if not os.path.isfile(f"""{pytorch_dump_folder_path}/{file.split('/' )[-1]}""" ): snake_case = requests.get(f"""{PREFIX}{file}""" , allow_redirects=a ) os.makedirs(f"""{pytorch_dump_folder_path}/""" , exist_ok=a ) open(f"""{pytorch_dump_folder_path}/{file.split('/' )[-1]}""" , '''wb''' ).write(r.content ) snake_case = MODEL_MAPPING[model_name.split('''/''' )[-1]] snake_case = JukeboxConfig.from_pretrained(a ) snake_case = JukeboxModel(a ) snake_case = [] snake_case = {} for i, dict_name in enumerate(a ): snake_case = torch.load(f"""{pytorch_dump_folder_path}/{dict_name.split('/' )[-1]}""" )['''model'''] snake_case = {} for k in old_dic.keys(): if k.endswith('''.b''' ): snake_case = old_dic[k] elif k.endswith('''.w''' ): snake_case = old_dic[k] elif "level_2" not in dict_name and "cond.model." in k: snake_case = old_dic[k] else: snake_case = old_dic[k] snake_case = '''vqvae''' if i == 0 else f"""priors.{3 - i}""" snake_case = fix_jukebox_keys(a , model.state_dict() , a , a ) weight_dict.append(a ) snake_case = weight_dict.pop(0 ) model.vqvae.load_state_dict(a ) for i in range(len(a ) ): model.priors[i].load_state_dict(weight_dict[2 - i] ) Path(a ).mkdir(exist_ok=a ) with open(f"""{pytorch_dump_folder_path}/mapping.json""" , '''w''' ) as txtfile: json.dump(a , a ) print(f"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(a ) return weight_dict if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='jukebox-5b-lyrics', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='jukebox-5b-lyrics-converted', type=str, help='Path to the output PyTorch model directory.', ) _lowercase = parser.parse_args() convert_openai_checkpoint(args.model_name, args.pytorch_dump_folder_path)
716
'''simple docstring''' import argparse import torch from torch import nn from transformers import SpeechaTextConfig, SpeechaTextForConditionalGeneration def __UpperCamelCase ( a : Optional[int] ) ->Dict: snake_case = [ '''encoder.version''', '''decoder.version''', '''model.encoder.version''', '''model.decoder.version''', '''decoder.output_projection.weight''', '''_float_tensor''', '''encoder.embed_positions._float_tensor''', '''decoder.embed_positions._float_tensor''', ] for k in ignore_keys: state_dict.pop(a , a ) def __UpperCamelCase ( a : Optional[Any] ) ->int: snake_case = list(s_dict.keys() ) for key in keys: if "transformer_layers" in key: snake_case = s_dict.pop(a ) elif "subsample" in key: snake_case = s_dict.pop(a ) def __UpperCamelCase ( a : Optional[int] ) ->Optional[int]: snake_case , snake_case = emb.weight.shape snake_case = nn.Linear(a , a , bias=a ) snake_case = emb.weight.data return lin_layer def __UpperCamelCase ( a : Any , a : Tuple ) ->Tuple: snake_case = torch.load(a , map_location='''cpu''' ) snake_case = mam_aaa['''args'''] snake_case = mam_aaa['''model'''] snake_case = state_dict['''decoder.output_projection.weight'''] remove_ignore_keys_(a ) rename_keys(a ) snake_case = state_dict['''decoder.embed_tokens.weight'''].shape[0] snake_case = args.share_decoder_input_output_embed snake_case = [int(a ) for i in args.conv_kernel_sizes.split(''',''' )] snake_case = SpeechaTextConfig( vocab_size=a , max_source_positions=args.max_source_positions , max_target_positions=args.max_target_positions , encoder_layers=args.encoder_layers , decoder_layers=args.decoder_layers , encoder_attention_heads=args.encoder_attention_heads , decoder_attention_heads=args.decoder_attention_heads , encoder_ffn_dim=args.encoder_ffn_embed_dim , decoder_ffn_dim=args.decoder_ffn_embed_dim , d_model=args.encoder_embed_dim , dropout=args.dropout , attention_dropout=args.attention_dropout , activation_dropout=args.activation_dropout , activation_function='''relu''' , num_conv_layers=len(a ) , conv_channels=args.conv_channels , conv_kernel_sizes=a , input_feat_per_channel=args.input_feat_per_channel , input_channels=args.input_channels , tie_word_embeddings=a , num_beams=5 , max_length=200 , use_cache=a , decoder_start_token_id=2 , early_stopping=a , ) snake_case = SpeechaTextForConditionalGeneration(a ) snake_case , snake_case = model.model.load_state_dict(a , strict=a ) if len(a ) > 0 and not set(a ) <= { "encoder.embed_positions.weights", "decoder.embed_positions.weights", }: raise ValueError( '''Only `encoder.embed_positions.weights` and `decoder.embed_positions.weights` are allowed to be missing,''' f""" but all the following weights are missing {missing}""" ) if tie_embeds: snake_case = make_linear_from_emb(model.model.decoder.embed_tokens ) else: snake_case = lm_head_weights model.save_pretrained(a ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument('--fairseq_path', type=str, help='Path to the fairseq model (.pt) file.') parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') _lowercase = parser.parse_args() convert_fairseq_sat_checkpoint_to_tfms(args.fairseq_path, args.pytorch_dump_folder_path)
44
0
'''simple docstring''' import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelerate.test_utils import require_multi_gpu class _lowercase ( nn.Module ): def __init__( self ) -> str: super().__init__() snake_case = nn.Linear(3 , 4 ) snake_case = nn.BatchNormad(4 ) snake_case = nn.Linear(4 , 5 ) def UpperCamelCase ( self , A__ ) -> Tuple: return self.lineara(self.batchnorm(self.lineara(lowerCamelCase__ ) ) ) class _lowercase ( __lowerCAmelCase ): def UpperCamelCase ( self , A__ , *A__ , **A__ ) -> Any: return (args[0] + 1,) + args[1:], kwargs class _lowercase ( __lowerCAmelCase ): def UpperCamelCase ( self , A__ , A__ ) -> Any: return output + 1 class _lowercase ( unittest.TestCase ): def UpperCamelCase ( self ) -> Tuple: snake_case = ModelForTest() snake_case = ModelHook() add_hook_to_module(lowerCamelCase__ , lowerCamelCase__ ) self.assertEqual(test_model._hf_hook , lowerCamelCase__ ) self.assertTrue(hasattr(lowerCamelCase__ , '''_old_forward''' ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , '''forward''' ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ['''x'''] ) remove_hook_from_module(lowerCamelCase__ ) self.assertFalse(hasattr(lowerCamelCase__ , '''_hf_hook''' ) ) self.assertFalse(hasattr(lowerCamelCase__ , '''_old_forward''' ) ) def UpperCamelCase ( self ) -> str: snake_case = ModelForTest() snake_case = ModelHook() add_hook_to_module(lowerCamelCase__ , lowerCamelCase__ ) add_hook_to_module(lowerCamelCase__ , lowerCamelCase__ , append=lowerCamelCase__ ) self.assertEqual(isinstance(test_model._hf_hook , lowerCamelCase__ ) , lowerCamelCase__ ) self.assertEqual(len(test_model._hf_hook.hooks ) , 2 ) self.assertTrue(hasattr(lowerCamelCase__ , '''_old_forward''' ) ) # Check adding the hook did not change the name or the signature self.assertEqual(test_model.forward.__name__ , '''forward''' ) self.assertListEqual(list(inspect.signature(test_model.forward ).parameters ) , ['''x'''] ) remove_hook_from_module(lowerCamelCase__ ) self.assertFalse(hasattr(lowerCamelCase__ , '''_hf_hook''' ) ) self.assertFalse(hasattr(lowerCamelCase__ , '''_old_forward''' ) ) def UpperCamelCase ( self ) -> int: snake_case = ModelForTest() snake_case = torch.randn(2 , 3 ) snake_case = test_model(x + 1 ) snake_case = test_model(x + 2 ) snake_case = PreForwardHook() add_hook_to_module(lowerCamelCase__ , lowerCamelCase__ ) snake_case = test_model(lowerCamelCase__ ) self.assertTrue(torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain snake_case = PreForwardHook() add_hook_to_module(lowerCamelCase__ , lowerCamelCase__ ) snake_case = test_model(lowerCamelCase__ ) self.assertTrue(torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-5 ) ) # You need to use the sequential hook to chain two or more hooks snake_case = SequentialHook(PreForwardHook() , PreForwardHook() ) add_hook_to_module(lowerCamelCase__ , lowerCamelCase__ ) snake_case = test_model(lowerCamelCase__ ) assert torch.allclose(lowerCamelCase__ , lowerCamelCase__ , atol=1e-5 ) def UpperCamelCase ( self ) -> Optional[int]: snake_case = ModelForTest() snake_case = torch.randn(2 , 3 ) snake_case = test_model(lowerCamelCase__ ) snake_case = PostForwardHook() add_hook_to_module(lowerCamelCase__ , lowerCamelCase__ ) snake_case = test_model(lowerCamelCase__ ) self.assertTrue(torch.allclose(lowerCamelCase__ , output + 1 , atol=1e-5 ) ) # Attaching a hook to a model when it already has one replaces, does not chain snake_case = PostForwardHook() add_hook_to_module(lowerCamelCase__ , lowerCamelCase__ ) snake_case = test_model(lowerCamelCase__ ) self.assertTrue(torch.allclose(lowerCamelCase__ , output + 1 , atol=1e-5 ) ) # You need to use the sequential hook to chain two or more hooks snake_case = SequentialHook(PostForwardHook() , PostForwardHook() ) add_hook_to_module(lowerCamelCase__ , lowerCamelCase__ ) snake_case = test_model(lowerCamelCase__ ) assert torch.allclose(lowerCamelCase__ , output + 2 , atol=1e-5 ) def UpperCamelCase ( self ) -> Dict: snake_case = ModelForTest() snake_case = torch.randn(2 , 3 ) snake_case = test_model(lowerCamelCase__ ) snake_case = PostForwardHook() add_hook_to_module(lowerCamelCase__ , lowerCamelCase__ ) snake_case = test_model(lowerCamelCase__ ) self.assertTrue(torch.allclose(lowerCamelCase__ , output + 1 ) ) self.assertTrue(outputa.requires_grad ) snake_case = True snake_case = test_model(lowerCamelCase__ ) self.assertFalse(outputa.requires_grad ) @require_multi_gpu def UpperCamelCase ( self ) -> Dict: snake_case = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) # This will move each submodule on different devices add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(execution_device=0 ) ) add_hook_to_module(model.lineara , AlignDevicesHook(execution_device=1 ) ) self.assertEqual(model.lineara.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.weight.device , torch.device(0 ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device(0 ) ) self.assertEqual(model.lineara.weight.device , torch.device(1 ) ) # We can still make a forward pass. The input does not need to be on any particular device snake_case = torch.randn(2 , 3 ) snake_case = model(lowerCamelCase__ ) self.assertEqual(output.device , torch.device(1 ) ) # We can add a general hook to put back output on same device as input. add_hook_to_module(lowerCamelCase__ , AlignDevicesHook(io_same_device=lowerCamelCase__ ) ) snake_case = torch.randn(2 , 3 ).to(0 ) snake_case = model(lowerCamelCase__ ) self.assertEqual(output.device , torch.device(0 ) ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) # This will move each submodule on different devices snake_case = {'''execution_device''': 0 if torch.cuda.is_available() else '''cpu''', '''offload''': True} add_hook_to_module(model.lineara , AlignDevicesHook(**lowerCamelCase__ ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**lowerCamelCase__ ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**lowerCamelCase__ ) ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''meta''' ) ) # Buffers are not included in the offload by default, so are on the execution device snake_case = torch.device(hook_kwargs['''execution_device'''] ) self.assertEqual(model.batchnorm.running_mean.device , lowerCamelCase__ ) snake_case = torch.randn(2 , 3 ) snake_case = model(lowerCamelCase__ ) self.assertEqual(output.device , lowerCamelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) # Now test with buffers included in the offload snake_case = { '''execution_device''': 0 if torch.cuda.is_available() else '''cpu''', '''offload''': True, '''offload_buffers''': True, } add_hook_to_module(model.lineara , AlignDevicesHook(**lowerCamelCase__ ) ) add_hook_to_module(model.batchnorm , AlignDevicesHook(**lowerCamelCase__ ) ) add_hook_to_module(model.lineara , AlignDevicesHook(**lowerCamelCase__ ) ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device('''meta''' ) ) snake_case = torch.randn(2 , 3 ) snake_case = model(lowerCamelCase__ ) self.assertEqual(output.device , lowerCamelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_module(model.lineara ) remove_hook_from_module(model.batchnorm ) remove_hook_from_module(model.lineara ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) def UpperCamelCase ( self ) -> Optional[int]: snake_case = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) # This will move each submodule on different devices snake_case = 0 if torch.cuda.is_available() else '''cpu''' attach_align_device_hook(lowerCamelCase__ , execution_device=lowerCamelCase__ , offload=lowerCamelCase__ ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''meta''' ) ) # Buffers are not included in the offload by default, so are on the execution device snake_case = torch.device(lowerCamelCase__ ) self.assertEqual(model.batchnorm.running_mean.device , lowerCamelCase__ ) snake_case = torch.randn(2 , 3 ) snake_case = model(lowerCamelCase__ ) self.assertEqual(output.device , lowerCamelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(lowerCamelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) # Now test with buffers included in the offload attach_align_device_hook(lowerCamelCase__ , execution_device=lowerCamelCase__ , offload=lowerCamelCase__ , offload_buffers=lowerCamelCase__ ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device('''meta''' ) ) snake_case = torch.randn(2 , 3 ) snake_case = model(lowerCamelCase__ ) self.assertEqual(output.device , lowerCamelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(lowerCamelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) def UpperCamelCase ( self ) -> List[Any]: snake_case = ModelForTest() # Everything is on CPU self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) # This will move each submodule on different devices snake_case = 0 if torch.cuda.is_available() else '''cpu''' attach_align_device_hook( lowerCamelCase__ , execution_device=lowerCamelCase__ , offload=lowerCamelCase__ , weights_map=model.state_dict() ) # Parameters have been offloaded, so on the meta device self.assertEqual(model.lineara.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''meta''' ) ) # Buffers are not included in the offload by default, so are on the execution device snake_case = torch.device(lowerCamelCase__ ) self.assertEqual(model.batchnorm.running_mean.device , lowerCamelCase__ ) snake_case = torch.randn(2 , 3 ) snake_case = model(lowerCamelCase__ ) self.assertEqual(output.device , lowerCamelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(lowerCamelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) # Now test with buffers included in the offload attach_align_device_hook( lowerCamelCase__ , execution_device=lowerCamelCase__ , offload=lowerCamelCase__ , weights_map=model.state_dict() , offload_buffers=lowerCamelCase__ , ) # Parameters have been offloaded, so on the meta device, buffers included self.assertEqual(model.lineara.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''meta''' ) ) self.assertEqual(model.batchnorm.running_mean.device , torch.device('''meta''' ) ) snake_case = torch.randn(2 , 3 ) snake_case = model(lowerCamelCase__ ) self.assertEqual(output.device , lowerCamelCase__ ) # Removing hooks loads back the weights in the model. remove_hook_from_submodules(lowerCamelCase__ ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.batchnorm.weight.device , torch.device('''cpu''' ) ) self.assertEqual(model.lineara.weight.device , torch.device('''cpu''' ) )
717
'''simple docstring''' from ..utils import DummyObject, requires_backends class _lowercase ( metaclass=__a ): _UpperCAmelCase = ['''transformers''', '''torch''', '''note_seq'''] def __init__( self , *A__ , **A__ ) -> Union[str, Any]: requires_backends(self , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ) -> Optional[Any]: requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def UpperCamelCase ( cls , *A__ , **A__ ) -> Any: requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] )
44
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowercase = { 'configuration_megatron_bert': ['MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MegatronBertConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ 'MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'MegatronBertForCausalLM', 'MegatronBertForMaskedLM', 'MegatronBertForMultipleChoice', 'MegatronBertForNextSentencePrediction', 'MegatronBertForPreTraining', 'MegatronBertForQuestionAnswering', 'MegatronBertForSequenceClassification', 'MegatronBertForTokenClassification', 'MegatronBertModel', 'MegatronBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_megatron_bert import MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MegatronBertConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_megatron_bert import ( MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, MegatronBertPreTrainedModel, ) else: import sys _lowercase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
718
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator class _lowercase : def __init__( self , A__ ) -> None: snake_case = value snake_case = None snake_case = None class _lowercase : def __init__( self , A__ ) -> None: snake_case = tree def UpperCamelCase ( self , A__ ) -> int: if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self ) -> Iterator[int]: yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
44
0
'''simple docstring''' import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets _lowercase = """\ @inproceedings{kakwani2020indicnlpsuite, title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}}, author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar}, year={2020}, booktitle={Findings of EMNLP}, } """ _lowercase = """\ IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te. """ _lowercase = """ Compute IndicGLUE evaluation metric associated to each IndicGLUE dataset. Args: predictions: list of predictions to score (as int64), except for 'cvit-mkb-clsr' where each prediction is a vector (of float32). references: list of ground truth labels corresponding to the predictions (as int64), except for 'cvit-mkb-clsr' where each reference is a vector (of float32). Returns: depending on the IndicGLUE subset, one or several of: \"accuracy\": Accuracy \"f1\": F1 score \"precision\": Precision@10 Examples: >>> indic_glue_metric = datasets.load_metric('indic_glue', 'wnli') # 'wnli' or any of [\"copa\", \"sna\", \"csqa\", \"wstp\", \"inltkh\", \"bbca\", \"iitp-mr\", \"iitp-pr\", \"actsa-sc\", \"md\"] >>> references = [0, 1] >>> predictions = [0, 1] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> indic_glue_metric = datasets.load_metric('indic_glue', 'wiki-ner') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> indic_glue_metric = datasets.load_metric('indic_glue', 'cvit-mkb-clsr') >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]] >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'precision@10': 1.0} """ def __UpperCamelCase ( a : int , a : Optional[int] ) ->int: return float((preds == labels).mean() ) def __UpperCamelCase ( a : str , a : Union[str, Any] ) ->str: snake_case = simple_accuracy(UpperCamelCase__ , UpperCamelCase__ ) snake_case = float(fa_score(y_true=UpperCamelCase__ , y_pred=UpperCamelCase__ ) ) return { "accuracy": acc, "f1": fa, } def __UpperCamelCase ( a : Any , a : Dict ) ->Dict: snake_case = np.array(UpperCamelCase__ ) snake_case = np.array(UpperCamelCase__ ) snake_case = en_sentvecs.shape[0] # mean centering snake_case = en_sentvecs - np.mean(UpperCamelCase__ , axis=0 ) snake_case = in_sentvecs - np.mean(UpperCamelCase__ , axis=0 ) snake_case = cdist(UpperCamelCase__ , UpperCamelCase__ , '''cosine''' ) snake_case = np.array(range(UpperCamelCase__ ) ) snake_case = sim.argsort(axis=1 )[:, :10] snake_case = np.any(preds == actual[:, None] , axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowercase ( datasets.Metric ): def UpperCamelCase ( self ) -> Optional[int]: if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( '''You should supply a configuration name selected in ''' '''["wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", ''' '''"cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", ''' '''"wiki-ner"]''' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''int64''' ) if self.config_name != '''cvit-mkb-clsr''' else datasets.Sequence(datasets.Value('''float32''' ) ), '''references''': datasets.Value('''int64''' ) if self.config_name != '''cvit-mkb-clsr''' else datasets.Sequence(datasets.Value('''float32''' ) ), } ) , codebase_urls=[] , reference_urls=[] , format='''numpy''' if self.config_name != '''cvit-mkb-clsr''' else None , ) def UpperCamelCase ( self , A__ , A__ ) -> Tuple: if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(A__ , A__ )} elif self.config_name in ["wiki-ner"]: return acc_and_fa(A__ , A__ ) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(A__ , A__ )} else: raise KeyError( '''You should supply a configuration name selected in ''' '''["wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", ''' '''"cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", ''' '''"wiki-ner"]''' )
719
'''simple docstring''' import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() _lowercase = logging.get_logger(__name__) _lowercase = [ ('bert.bert', 'visual_bert'), ('bert.cls', 'cls'), ('bert.classifier', 'cls'), ('token_type_embeddings_visual', 'visual_token_type_embeddings'), ('position_embeddings_visual', 'visual_position_embeddings'), ('projection', 'visual_projection'), ] _lowercase = [ 'nlvr2_coco_pre_trained.th', 'nlvr2_fine_tuned.th', 'nlvr2_pre_trained.th', 'vcr_coco_pre_train.th', 'vcr_fine_tune.th', 'vcr_pre_train.th', 'vqa_coco_pre_trained.th', 'vqa_fine_tuned.th', 'vqa_pre_trained.th', ] def __UpperCamelCase ( a : List[str] ) ->Optional[int]: snake_case = torch.load(a , map_location='''cpu''' ) return sd def __UpperCamelCase ( a : Optional[int] , a : Union[str, Any] , a : int=rename_keys_prefix ) ->Tuple: snake_case = OrderedDict() snake_case = torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue snake_case = key for name_pair in rename_keys_prefix: snake_case = new_key.replace(name_pair[0] , name_pair[1] ) snake_case = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately snake_case = new_d['''cls.predictions.bias'''] return new_d @torch.no_grad() def __UpperCamelCase ( a : Optional[int] , a : int ) ->Union[str, Any]: assert ( checkpoint_path.split('''/''' )[-1] in ACCEPTABLE_CHECKPOINTS ), f"""The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.""" # Get Config if "pre" in checkpoint_path: snake_case = '''pretraining''' if "vcr" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 512} elif "vqa_advanced" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048} elif "vqa" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048} elif "nlvr" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 1024} else: raise NotImplementedError(f"""No implementation found for `{checkpoint_path}`.""" ) else: if "vcr" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 512} snake_case = '''multichoice''' elif "vqa_advanced" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048} snake_case = '''vqa_advanced''' elif "vqa" in checkpoint_path: snake_case = {'''visual_embedding_dim''': 2048, '''num_labels''': 3129} snake_case = '''vqa''' elif "nlvr" in checkpoint_path: snake_case = { '''visual_embedding_dim''': 1024, '''num_labels''': 2, } snake_case = '''nlvr''' snake_case = VisualBertConfig(**a ) # Load State Dict snake_case = load_state_dict(a ) snake_case = get_new_dict(a , a ) if model_type == "pretraining": snake_case = VisualBertForPreTraining(a ) elif model_type == "vqa": snake_case = VisualBertForQuestionAnswering(a ) elif model_type == "nlvr": snake_case = VisualBertForVisualReasoning(a ) elif model_type == "multichoice": snake_case = VisualBertForMultipleChoice(a ) model.load_state_dict(a ) # Save Checkpoints Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument('orig_checkpoint_path', type=str, help='A path to .th on local filesystem.') parser.add_argument('pytorch_dump_folder_path', type=str, help='Path to the output PyTorch model.') _lowercase = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
44
0
'''simple docstring''' import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate.utils import patch_environment @require_huggingface_suite class _lowercase ( unittest.TestCase ): def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = inspect.getfile(accelerate.test_utils ) snake_case = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ['''scripts''', '''external_deps''', '''test_metrics.py'''] ) from accelerate.test_utils.scripts.external_deps import test_metrics # noqa: F401 snake_case = test_metrics @require_cpu def UpperCamelCase ( self ) -> Optional[int]: debug_launcher(self.test_metrics.main , num_processes=1 ) @require_cpu def UpperCamelCase ( self ) -> List[Any]: debug_launcher(self.test_metrics.main ) @require_single_gpu def UpperCamelCase ( self ) -> Union[str, Any]: self.test_metrics.main() @require_multi_gpu def UpperCamelCase ( self ) -> Tuple: print(F"""Found {torch.cuda.device_count()} devices.""" ) snake_case = ['''torchrun''', F"""--nproc_per_node={torch.cuda.device_count()}""", self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(A__ , env=os.environ.copy() )
720
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() _lowercase = logging.get_logger(__name__) def __UpperCamelCase ( a : Dict , a : Optional[int] , a : Dict , a : Dict ) ->Union[str, Any]: snake_case = original_name.split('''.''' )[0] snake_case = key.split('''.''' ) snake_case = int(key_list[key_list.index(a ) - 2] ) snake_case = int(key_list[key_list.index(a ) - 1] ) snake_case = orig_block_num - offset snake_case = key.replace(f"""{orig_block_num}.{layer_num}.{original_name}""" , f"""block.{new_block_num}.{layer_num}.{new_name}""" ) return key def __UpperCamelCase ( a : Tuple ) ->Dict: snake_case = OrderedDict() snake_case , snake_case = 0, 0 for key, value in state_dict.items(): if key.startswith('''network''' ): snake_case = key.replace('''network''' , '''poolformer.encoder''' ) if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith('''bias''' ) and "patch_embed" not in key: patch_emb_offset += 1 snake_case = key[: key.find('''proj''' )] snake_case = key.replace(a , f"""patch_embeddings.{total_embed_found}.""" ) snake_case = key.replace('''proj''' , '''projection''' ) if key.endswith('''bias''' ): total_embed_found += 1 if "patch_embeddings" in key: snake_case = '''poolformer.encoder.''' + key if "mlp.fc1" in key: snake_case = replace_key_with_offset(a , a , '''mlp.fc1''' , '''output.conv1''' ) if "mlp.fc2" in key: snake_case = replace_key_with_offset(a , a , '''mlp.fc2''' , '''output.conv2''' ) if "norm1" in key: snake_case = replace_key_with_offset(a , a , '''norm1''' , '''before_norm''' ) if "norm2" in key: snake_case = replace_key_with_offset(a , a , '''norm2''' , '''after_norm''' ) if "layer_scale_1" in key: snake_case = replace_key_with_offset(a , a , '''layer_scale_1''' , '''layer_scale_1''' ) if "layer_scale_2" in key: snake_case = replace_key_with_offset(a , a , '''layer_scale_2''' , '''layer_scale_2''' ) if "head" in key: snake_case = key.replace('''head''' , '''classifier''' ) snake_case = value return new_state_dict def __UpperCamelCase ( ) ->Optional[int]: snake_case = '''http://images.cocodataset.org/val2017/000000039769.jpg''' snake_case = Image.open(requests.get(a , stream=a ).raw ) return image @torch.no_grad() def __UpperCamelCase ( a : Dict , a : Optional[Any] , a : Tuple ) ->List[str]: snake_case = PoolFormerConfig() # set attributes based on model_name snake_case = '''huggingface/label-files''' snake_case = model_name[-3:] snake_case = 1000 snake_case = '''imagenet-1k-id2label.json''' snake_case = (1, 1000) # set config attributes snake_case = json.load(open(hf_hub_download(a , a , repo_type='''dataset''' ) , '''r''' ) ) snake_case = {int(a ): v for k, v in idalabel.items()} snake_case = idalabel snake_case = {v: k for k, v in idalabel.items()} if size == "s12": snake_case = [2, 2, 6, 2] snake_case = [64, 128, 320, 512] snake_case = 4.0 snake_case = 0.9 elif size == "s24": snake_case = [4, 4, 12, 4] snake_case = [64, 128, 320, 512] snake_case = 4.0 snake_case = 0.9 elif size == "s36": snake_case = [6, 6, 18, 6] snake_case = [64, 128, 320, 512] snake_case = 4.0 snake_case = 1e-6 snake_case = 0.9 elif size == "m36": snake_case = [6, 6, 18, 6] snake_case = [96, 192, 384, 768] snake_case = 4.0 snake_case = 1e-6 snake_case = 0.95 elif size == "m48": snake_case = [8, 8, 24, 8] snake_case = [96, 192, 384, 768] snake_case = 4.0 snake_case = 1e-6 snake_case = 0.95 else: raise ValueError(f"""Size {size} not supported""" ) # load image processor snake_case = PoolFormerImageProcessor(crop_pct=a ) # Prepare image snake_case = prepare_img() snake_case = image_processor(images=a , return_tensors='''pt''' ).pixel_values logger.info(f"""Converting model {model_name}...""" ) # load original state dict snake_case = torch.load(a , map_location=torch.device('''cpu''' ) ) # rename keys snake_case = rename_keys(a ) # create HuggingFace model and load state dict snake_case = PoolFormerForImageClassification(a ) model.load_state_dict(a ) model.eval() # Define image processor snake_case = PoolFormerImageProcessor(crop_pct=a ) snake_case = image_processor(images=prepare_img() , return_tensors='''pt''' ).pixel_values # forward pass snake_case = model(a ) snake_case = outputs.logits # define expected logit slices for different models if size == "s12": snake_case = torch.tensor([-0.3045, -0.6758, -0.4869] ) elif size == "s24": snake_case = torch.tensor([0.4402, -0.1374, -0.8045] ) elif size == "s36": snake_case = torch.tensor([-0.6080, -0.5133, -0.5898] ) elif size == "m36": snake_case = torch.tensor([0.3952, 0.2263, -1.2668] ) elif size == "m48": snake_case = torch.tensor([0.1167, -0.0656, -0.3423] ) else: raise ValueError(f"""Size {size} not supported""" ) # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] , a , atol=1e-2 ) # finally, save model and image processor logger.info(f"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(a ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() parser.add_argument( '--model_name', default='poolformer_s12', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--checkpoint_path', default=None, type=str, help='Path to the original PyTorch checkpoint (.pth file).' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) _lowercase = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
44
0
def __UpperCamelCase ( a : Union[str, Any] ) ->List[str]: snake_case = len(a ) snake_case = sum(a ) snake_case = [[False for x in range(s + 1 )] for y in range(n + 1 )] for i in range(1 , n + 1 ): snake_case = True for i in range(1 , s + 1 ): snake_case = False for i in range(1 , n + 1 ): for j in range(1 , s + 1 ): snake_case = dp[i][j - 1] if arr[i - 1] <= j: snake_case = dp[i][j] or dp[i - 1][j - arr[i - 1]] for j in range(int(s / 2 ) , -1 , -1 ): if dp[n][j] is True: snake_case = s - 2 * j break return diff
721
'''simple docstring''' import argparse import json import logging import os import sys from unittest.mock import patch from transformers.testing_utils import TestCasePlus, get_gpu_count, slow _lowercase = [ os.path.join(os.path.dirname(__file__), dirname) for dirname in [ 'text-classification', 'language-modeling', 'summarization', 'token-classification', 'question-answering', ] ] sys.path.extend(SRC_DIRS) if SRC_DIRS is not None: import run_clm_flax import run_flax_glue import run_flax_ner import run_mlm_flax import run_qa import run_summarization_flax import run_ta_mlm_flax logging.basicConfig(level=logging.DEBUG) _lowercase = logging.getLogger() def __UpperCamelCase ( ) ->Tuple: snake_case = argparse.ArgumentParser() parser.add_argument('''-f''' ) snake_case = parser.parse_args() return args.f def __UpperCamelCase ( a : Dict , a : Tuple="eval" ) ->List[Any]: snake_case = os.path.join(a , f"""{split}_results.json""" ) if os.path.exists(a ): with open(a , '''r''' ) as f: return json.load(a ) raise ValueError(f"""can't find {path}""" ) _lowercase = logging.StreamHandler(sys.stdout) logger.addHandler(stream_handler) class _lowercase ( __a ): def UpperCamelCase ( self ) -> List[str]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_glue.py --model_name_or_path distilbert-base-uncased --output_dir {tmp_dir} --train_file ./tests/fixtures/tests_samples/MRPC/train.csv --validation_file ./tests/fixtures/tests_samples/MRPC/dev.csv --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --learning_rate=1e-4 --eval_steps=2 --warmup_steps=2 --seed=42 --max_seq_length=128 """.split() with patch.object(A__ , '''argv''' , A__ ): run_flax_glue.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.7_5 ) @slow def UpperCamelCase ( self ) -> List[Any]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_clm_flax.py --model_name_or_path distilgpt2 --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --block_size 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(A__ , '''argv''' , A__ ): run_clm_flax.main() snake_case = get_results(A__ ) self.assertLess(result['''eval_perplexity'''] , 1_00 ) @slow def UpperCamelCase ( self ) -> int: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_summarization.py --model_name_or_path t5-small --train_file tests/fixtures/tests_samples/xsum/sample.json --validation_file tests/fixtures/tests_samples/xsum/sample.json --test_file tests/fixtures/tests_samples/xsum/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=8 --do_train --do_eval --do_predict --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 --predict_with_generate """.split() with patch.object(A__ , '''argv''' , A__ ): run_summarization_flax.main() snake_case = get_results(A__ , split='''test''' ) self.assertGreaterEqual(result['''test_rouge1'''] , 10 ) self.assertGreaterEqual(result['''test_rouge2'''] , 2 ) self.assertGreaterEqual(result['''test_rougeL'''] , 7 ) self.assertGreaterEqual(result['''test_rougeLsum'''] , 7 ) @slow def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_mlm.py --model_name_or_path distilroberta-base --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --output_dir {tmp_dir} --overwrite_output_dir --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --logging_steps 2 --eval_steps 2 --do_train --do_eval --num_train_epochs=1 """.split() with patch.object(A__ , '''argv''' , A__ ): run_mlm_flax.main() snake_case = get_results(A__ ) self.assertLess(result['''eval_perplexity'''] , 42 ) @slow def UpperCamelCase ( self ) -> Dict: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_t5_mlm_flax.py --model_name_or_path t5-small --train_file ./tests/fixtures/sample_text.txt --validation_file ./tests/fixtures/sample_text.txt --do_train --do_eval --max_seq_length 128 --per_device_train_batch_size 4 --per_device_eval_batch_size 4 --num_train_epochs 2 --logging_steps 2 --eval_steps 2 --output_dir {tmp_dir} --overwrite_output_dir """.split() with patch.object(A__ , '''argv''' , A__ ): run_ta_mlm_flax.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.4_2 ) @slow def UpperCamelCase ( self ) -> int: # with so little data distributed training needs more epochs to get the score on par with 0/1 gpu snake_case = 7 if get_gpu_count() > 1 else 2 snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_flax_ner.py --model_name_or_path bert-base-uncased --train_file tests/fixtures/tests_samples/conll/sample.json --validation_file tests/fixtures/tests_samples/conll/sample.json --output_dir {tmp_dir} --overwrite_output_dir --do_train --do_eval --warmup_steps=2 --learning_rate=2e-4 --logging_steps 2 --eval_steps 2 --per_device_train_batch_size=2 --per_device_eval_batch_size=2 --num_train_epochs={epochs} --seed 7 """.split() with patch.object(A__ , '''argv''' , A__ ): run_flax_ner.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_accuracy'''] , 0.7_5 ) self.assertGreaterEqual(result['''eval_f1'''] , 0.3 ) @slow def UpperCamelCase ( self ) -> Any: snake_case = self.get_auto_remove_tmp_dir() snake_case = F""" run_qa.py --model_name_or_path bert-base-uncased --version_2_with_negative --train_file tests/fixtures/tests_samples/SQUAD/sample.json --validation_file tests/fixtures/tests_samples/SQUAD/sample.json --output_dir {tmp_dir} --overwrite_output_dir --num_train_epochs=3 --warmup_steps=2 --do_train --do_eval --logging_steps 2 --eval_steps 2 --learning_rate=2e-4 --per_device_train_batch_size=2 --per_device_eval_batch_size=1 """.split() with patch.object(A__ , '''argv''' , A__ ): run_qa.main() snake_case = get_results(A__ ) self.assertGreaterEqual(result['''eval_f1'''] , 30 ) self.assertGreaterEqual(result['''eval_exact'''] , 30 )
44
0
'''simple docstring''' import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class _lowercase ( __a ): def UpperCamelCase ( self , A__ ) -> int: with open(A__ , encoding='''utf-8''' ) as input_file: snake_case = re.compile(R'''(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)''' ) snake_case = input_file.read() snake_case = regexp.search(A__ ) return match def UpperCamelCase ( self , A__ ) -> List[Any]: with open(A__ , encoding='''utf-8''' ) as input_file: snake_case = re.compile(R'''#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()''' , re.DOTALL ) snake_case = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` snake_case = regexp.finditer(A__ ) snake_case = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def UpperCamelCase ( self ) -> Dict: snake_case = Path('''./datasets''' ) snake_case = list(dataset_paths.absolute().glob('''**/*.py''' ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(A__ ) ): raise AssertionError(F"""open(...) must use utf-8 encoding in {dataset}""" ) def UpperCamelCase ( self ) -> Optional[Any]: snake_case = Path('''./datasets''' ) snake_case = list(dataset_paths.absolute().glob('''**/*.py''' ) ) for dataset in dataset_files: if self._no_print_statements(str(A__ ) ): raise AssertionError(F"""print statement found in {dataset}. Use datasets.logger/logging instead.""" )
700
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS _lowercase = logging.get_logger(__name__) _lowercase = { 'linear': get_linear_schedule_with_warmup, 'cosine': get_cosine_schedule_with_warmup, 'cosine_w_restarts': get_cosine_with_hard_restarts_schedule_with_warmup, 'polynomial': get_polynomial_decay_schedule_with_warmup, 'constant': get_constant_schedule, 'constant_w_warmup': get_constant_schedule_with_warmup, } class _lowercase ( __a ): def __init__( self , A__=None , A__=None , *A__ , **A__ ) -> Union[str, Any]: super().__init__(*A__ , **A__ ) if config is None: assert isinstance(self.model , A__ ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F""" {self.model.__class__}""" ) snake_case = self.model.config else: snake_case = config snake_case = data_args snake_case = self.config.tgt_vocab_size if isinstance(self.config , A__ ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"""The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for""" ''' padding..''' ) if self.args.label_smoothing == 0: snake_case = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss snake_case = label_smoothed_nll_loss def UpperCamelCase ( self , A__ ) -> Tuple: if self.optimizer is None: snake_case = ['''bias''', '''LayerNorm.weight'''] snake_case = [ { '''params''': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], '''weight_decay''': self.args.weight_decay, }, { '''params''': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0, }, ] snake_case = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: snake_case = Adafactor snake_case = {'''scale_parameter''': False, '''relative_step''': False} else: snake_case = AdamW snake_case = { '''betas''': (self.args.adam_betaa, self.args.adam_betaa), '''eps''': self.args.adam_epsilon, } snake_case = self.args.learning_rate if self.sharded_ddp: snake_case = OSS( params=A__ , optim=A__ , **A__ , ) else: snake_case = optimizer_cls(A__ , **A__ ) if self.lr_scheduler is None: snake_case = self._get_lr_scheduler(A__ ) else: # ignoring --lr_scheduler logger.warning('''scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.''' ) def UpperCamelCase ( self , A__ ) -> Tuple: snake_case = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": snake_case = schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": snake_case = schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: snake_case = schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=A__ ) return scheduler def UpperCamelCase ( self ) -> Optional[torch.utils.data.Sampler]: if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> List[Any]: if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token snake_case = model(**A__ , use_cache=A__ )[0] snake_case = self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models snake_case , snake_case = model(**A__ , labels=A__ , use_cache=A__ )[:2] else: # compute label smoothed loss snake_case = model(**A__ , use_cache=A__ )[0] snake_case = torch.nn.functional.log_softmax(A__ , dim=-1 ) snake_case , snake_case = self.loss_fn(A__ , A__ , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def UpperCamelCase ( self , A__ , A__ ) -> Any: snake_case = inputs.pop('''labels''' ) snake_case , snake_case = self._compute_loss(A__ , A__ , A__ ) return loss def UpperCamelCase ( self , A__ , A__ , A__ , A__ = None , ) -> Tuple[Optional[float], Optional[torch.Tensor], Optional[torch.Tensor]]: snake_case = self._prepare_inputs(A__ ) snake_case = { '''max_length''': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, '''num_beams''': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: snake_case = self.model.generate( inputs['''input_ids'''] , attention_mask=inputs['''attention_mask'''] , **A__ , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: snake_case = self._pad_tensors_to_max_len(A__ , gen_kwargs['''max_length'''] ) snake_case = inputs.pop('''labels''' ) with torch.no_grad(): # compute loss on predict data snake_case , snake_case = self._compute_loss(A__ , A__ , A__ ) snake_case = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) snake_case = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: snake_case = self._pad_tensors_to_max_len(A__ , gen_kwargs['''max_length'''] ) return (loss, logits, labels) def UpperCamelCase ( self , A__ , A__ ) -> List[str]: # If PAD token is not defined at least EOS token has to be defined snake_case = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( '''Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be''' F""" padded to `max_length`={max_length}""" ) snake_case = pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) snake_case = tensor return padded_tensor
44
0
'''simple docstring''' from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": _lowercase = input('Enter image url: ').strip() print(f'Downloading image from {url} ...') _lowercase = BeautifulSoup(requests.get(url).content, 'html.parser') # The image URL is in the content field of the first meta tag with property og:image _lowercase = soup.find('meta', {'property': 'og:image'})['content'] _lowercase = requests.get(image_url).content _lowercase = f'{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg' with open(file_name, 'wb') as fp: fp.write(image_data) print(f'Done. Image saved to disk as {file_name}.')
701
'''simple docstring''' import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def __UpperCamelCase ( a : List[str] ) ->str: snake_case = [] for line in lines: snake_case = re.sub(R'''#.*''' , '''''' , a ) # remove comments if line: filtered_lines.append(a ) snake_case = '''\n'''.join(a ) # Make a hash from all this code snake_case = full_str.encode('''utf-8''' ) return shaaaa(a ).hexdigest() # get importable module names and hash for caching _lowercase = { 'csv': (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), 'json': (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), 'pandas': (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), 'parquet': (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), 'arrow': (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), 'text': (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), 'imagefolder': (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), 'audiofolder': (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions _lowercase = { '.csv': ('csv', {}), '.tsv': ('csv', {'sep': '\t'}), '.json': ('json', {}), '.jsonl': ('json', {}), '.parquet': ('parquet', {}), '.arrow': ('arrow', {}), '.txt': ('text', {}), } _EXTENSION_TO_MODULE.update({ext: ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _lowercase = {'imagefolder', 'audiofolder'} # Used to filter data files based on extensions given a module name _lowercase = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append('.zip') _MODULE_TO_EXTENSIONS["audiofolder"].append('.zip')
44
0
'''simple docstring''' import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def __UpperCamelCase ( a : Any , a : Any=0.999 , a : Any="cosine" , ) ->List[str]: if alpha_transform_type == "cosine": def alpha_bar_fn(a : Union[str, Any] ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(a : Tuple ): return math.exp(t * -12.0 ) else: raise ValueError(f"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) snake_case = [] for i in range(a ): snake_case = i / num_diffusion_timesteps snake_case = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(a ) / alpha_bar_fn(a ) , a ) ) return torch.tensor(a , dtype=torch.floataa ) class _lowercase ( __a , __a ): _UpperCAmelCase = [e.name for e in KarrasDiffusionSchedulers] _UpperCAmelCase = 2 @register_to_config def __init__( self , A__ = 10_00 , A__ = 0.0_0_0_8_5 , A__ = 0.0_1_2 , A__ = "linear" , A__ = None , A__ = "epsilon" , A__ = "linspace" , A__ = 0 , ): if trained_betas is not None: snake_case = torch.tensor(A__ , dtype=torch.floataa ) elif beta_schedule == "linear": snake_case = torch.linspace(A__ , A__ , A__ , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. snake_case = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , A__ , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule snake_case = betas_for_alpha_bar(A__ ) else: raise NotImplementedError(F"""{beta_schedule} does is not implemented for {self.__class__}""" ) snake_case = 1.0 - self.betas snake_case = torch.cumprod(self.alphas , dim=0 ) # set all values self.set_timesteps(A__ , A__ , A__ ) def UpperCamelCase ( self , A__ , A__=None ): if schedule_timesteps is None: snake_case = self.timesteps snake_case = (schedule_timesteps == timestep).nonzero() # The sigma index that is taken for the **very** first `step` # is always the second index (or the last index if there is only 1) # This way we can ensure we don't accidentally skip a sigma in # case we start in the middle of the denoising schedule (e.g. for image-to-image) if len(self._index_counter ) == 0: snake_case = 1 if len(A__ ) > 1 else 0 else: snake_case = timestep.cpu().item() if torch.is_tensor(A__ ) else timestep snake_case = self._index_counter[timestep_int] return indices[pos].item() @property def UpperCamelCase ( self ): # standard deviation of the initial noise distribution if self.config.timestep_spacing in ["linspace", "trailing"]: return self.sigmas.max() return (self.sigmas.max() ** 2 + 1) ** 0.5 def UpperCamelCase ( self , A__ , A__ , ): snake_case = self.index_for_timestep(A__ ) if self.state_in_first_order: snake_case = self.sigmas[step_index] else: snake_case = self.sigmas_interpol[step_index] snake_case = sample / ((sigma**2 + 1) ** 0.5) return sample def UpperCamelCase ( self , A__ , A__ = None , A__ = None , ): snake_case = num_inference_steps snake_case = num_train_timesteps or self.config.num_train_timesteps # "linspace", "leading", "trailing" corresponds to annotation of Table 2. of https://arxiv.org/abs/2305.08891 if self.config.timestep_spacing == "linspace": snake_case = np.linspace(0 , num_train_timesteps - 1 , A__ , dtype=A__ )[::-1].copy() elif self.config.timestep_spacing == "leading": snake_case = num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 snake_case = (np.arange(0 , A__ ) * step_ratio).round()[::-1].copy().astype(A__ ) timesteps += self.config.steps_offset elif self.config.timestep_spacing == "trailing": snake_case = num_train_timesteps / self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 snake_case = (np.arange(A__ , 0 , -step_ratio )).round().copy().astype(A__ ) timesteps -= 1 else: raise ValueError( F"""{self.config.timestep_spacing} is not supported. Please make sure to choose one of 'linspace', 'leading' or 'trailing'.""" ) snake_case = np.array(((1 - self.alphas_cumprod) / self.alphas_cumprod) ** 0.5 ) snake_case = torch.from_numpy(np.log(A__ ) ).to(A__ ) snake_case = np.interp(A__ , np.arange(0 , len(A__ ) ) , A__ ) snake_case = np.concatenate([sigmas, [0.0]] ).astype(np.floataa ) snake_case = torch.from_numpy(A__ ).to(device=A__ ) # interpolate sigmas snake_case = sigmas.log().lerp(sigmas.roll(1 ).log() , 0.5 ).exp() snake_case = torch.cat([sigmas[:1], sigmas[1:].repeat_interleave(2 ), sigmas[-1:]] ) snake_case = torch.cat( [sigmas_interpol[:1], sigmas_interpol[1:].repeat_interleave(2 ), sigmas_interpol[-1:]] ) if str(A__ ).startswith('''mps''' ): # mps does not support float64 snake_case = torch.from_numpy(A__ ).to(A__ , dtype=torch.floataa ) else: snake_case = torch.from_numpy(A__ ).to(A__ ) # interpolate timesteps snake_case = self.sigma_to_t(A__ ).to(A__ , dtype=timesteps.dtype ) snake_case = torch.stack((timesteps_interpol[1:-1, None], timesteps[1:, None]) , dim=-1 ).flatten() snake_case = torch.cat([timesteps[:1], interleaved_timesteps] ) snake_case = None # for exp beta schedules, such as the one for `pipeline_shap_e.py` # we need an index counter snake_case = defaultdict(A__ ) def UpperCamelCase ( self , A__ ): # get log sigma snake_case = sigma.log() # get distribution snake_case = log_sigma - self.log_sigmas[:, None] # get sigmas range snake_case = dists.ge(0 ).cumsum(dim=0 ).argmax(dim=0 ).clamp(max=self.log_sigmas.shape[0] - 2 ) snake_case = low_idx + 1 snake_case = self.log_sigmas[low_idx] snake_case = self.log_sigmas[high_idx] # interpolate sigmas snake_case = (low - log_sigma) / (low - high) snake_case = w.clamp(0 , 1 ) # transform interpolation to time range snake_case = (1 - w) * low_idx + w * high_idx snake_case = t.view(sigma.shape ) return t @property def UpperCamelCase ( self ): return self.sample is None def UpperCamelCase ( self , A__ , A__ , A__ , A__ = True , ): snake_case = self.index_for_timestep(A__ ) # advance index counter by 1 snake_case = timestep.cpu().item() if torch.is_tensor(A__ ) else timestep self._index_counter[timestep_int] += 1 if self.state_in_first_order: snake_case = self.sigmas[step_index] snake_case = self.sigmas_interpol[step_index + 1] snake_case = self.sigmas[step_index + 1] else: # 2nd order / KDPM2's method snake_case = self.sigmas[step_index - 1] snake_case = self.sigmas_interpol[step_index] snake_case = self.sigmas[step_index] # currently only gamma=0 is supported. This usually works best anyways. # We can support gamma in the future but then need to scale the timestep before # passing it to the model which requires a change in API snake_case = 0 snake_case = sigma * (gamma + 1) # Note: sigma_hat == sigma for now # 1. compute predicted original sample (x_0) from sigma-scaled predicted noise if self.config.prediction_type == "epsilon": snake_case = sigma_hat if self.state_in_first_order else sigma_interpol snake_case = sample - sigma_input * model_output elif self.config.prediction_type == "v_prediction": snake_case = sigma_hat if self.state_in_first_order else sigma_interpol snake_case = model_output * (-sigma_input / (sigma_input**2 + 1) ** 0.5) + ( sample / (sigma_input**2 + 1) ) elif self.config.prediction_type == "sample": raise NotImplementedError('''prediction_type not implemented yet: sample''' ) else: raise ValueError( F"""prediction_type given as {self.config.prediction_type} must be one of `epsilon`, or `v_prediction`""" ) if self.state_in_first_order: # 2. Convert to an ODE derivative for 1st order snake_case = (sample - pred_original_sample) / sigma_hat # 3. delta timestep snake_case = sigma_interpol - sigma_hat # store for 2nd order step snake_case = sample else: # DPM-Solver-2 # 2. Convert to an ODE derivative for 2nd order snake_case = (sample - pred_original_sample) / sigma_interpol # 3. delta timestep snake_case = sigma_next - sigma_hat snake_case = self.sample snake_case = None snake_case = sample + derivative * dt if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=A__ ) def UpperCamelCase ( self , A__ , A__ , A__ , ): # Make sure sigmas and timesteps have the same device and dtype as original_samples snake_case = self.sigmas.to(device=original_samples.device , dtype=original_samples.dtype ) if original_samples.device.type == "mps" and torch.is_floating_point(A__ ): # mps does not support float64 snake_case = self.timesteps.to(original_samples.device , dtype=torch.floataa ) snake_case = timesteps.to(original_samples.device , dtype=torch.floataa ) else: snake_case = self.timesteps.to(original_samples.device ) snake_case = timesteps.to(original_samples.device ) snake_case = [self.index_for_timestep(A__ , A__ ) for t in timesteps] snake_case = sigmas[step_indices].flatten() while len(sigma.shape ) < len(original_samples.shape ): snake_case = sigma.unsqueeze(-1 ) snake_case = original_samples + noise * sigma return noisy_samples def __len__( self ): return self.config.num_train_timesteps
702
'''simple docstring''' _lowercase = { 'Pillow': 'Pillow', 'accelerate': 'accelerate>=0.11.0', 'compel': 'compel==0.1.8', 'black': 'black~=23.1', 'datasets': 'datasets', 'filelock': 'filelock', 'flax': 'flax>=0.4.1', 'hf-doc-builder': 'hf-doc-builder>=0.3.0', 'huggingface-hub': 'huggingface-hub>=0.13.2', 'requests-mock': 'requests-mock==1.10.0', 'importlib_metadata': 'importlib_metadata', 'invisible-watermark': 'invisible-watermark', 'isort': 'isort>=5.5.4', 'jax': 'jax>=0.2.8,!=0.3.2', 'jaxlib': 'jaxlib>=0.1.65', 'Jinja2': 'Jinja2', 'k-diffusion': 'k-diffusion>=0.0.12', 'torchsde': 'torchsde', 'note_seq': 'note_seq', 'librosa': 'librosa', 'numpy': 'numpy', 'omegaconf': 'omegaconf', 'parameterized': 'parameterized', 'protobuf': 'protobuf>=3.20.3,<4', 'pytest': 'pytest', 'pytest-timeout': 'pytest-timeout', 'pytest-xdist': 'pytest-xdist', 'ruff': 'ruff>=0.0.241', 'safetensors': 'safetensors', 'sentencepiece': 'sentencepiece>=0.1.91,!=0.1.92', 'scipy': 'scipy', 'onnx': 'onnx', 'regex': 'regex!=2019.12.17', 'requests': 'requests', 'tensorboard': 'tensorboard', 'torch': 'torch>=1.4', 'torchvision': 'torchvision', 'transformers': 'transformers>=4.25.1', 'urllib3': 'urllib3<=2.0.0', }
44
0
'''simple docstring''' import json import os from pathlib import Path import pytest from datasets.download.download_config import DownloadConfig from datasets.download.download_manager import DownloadManager from datasets.utils.file_utils import hash_url_to_filename _lowercase = 'http://www.mocksite.com/file1.txt' _lowercase = '"text": ["foo", "foo"]' _lowercase = '6d8ce9aa78a471c7477201efbeabd3bb01ac2e7d100a6dc024ba1608361f90a8' class _lowercase : _UpperCAmelCase = 200 _UpperCAmelCase = {'''Content-Length''': '''100'''} _UpperCAmelCase = {} def UpperCamelCase ( self , **A__ ) -> str: return [bytes(A__ , '''utf-8''' )] def __UpperCamelCase ( *a : List[str] , **a : List[str] ) ->List[Any]: return MockResponse() @pytest.mark.parametrize('''urls_type''' , [str, list, dict] ) def __UpperCamelCase ( a : Tuple , a : int , a : List[str] ) ->Optional[Any]: import requests monkeypatch.setattr(a , '''request''' , a ) snake_case = URL if issubclass(a , a ): snake_case = url elif issubclass(a , a ): snake_case = [url] elif issubclass(a , a ): snake_case = {'''train''': url} snake_case = '''dummy''' snake_case = '''downloads''' snake_case = tmp_path snake_case = DownloadConfig( cache_dir=os.path.join(a , a ) , use_etag=a , ) snake_case = DownloadManager(dataset_name=a , download_config=a ) snake_case = dl_manager.download(a ) snake_case = urls for downloaded_paths in [downloaded_paths]: if isinstance(a , a ): snake_case = [downloaded_paths] snake_case = [urls] elif isinstance(a , a ): assert "train" in downloaded_paths.keys() snake_case = downloaded_paths.values() snake_case = urls.values() assert downloaded_paths for downloaded_path, input_url in zip(a , a ): assert downloaded_path == dl_manager.downloaded_paths[input_url] snake_case = Path(a ) snake_case = downloaded_path.parts assert parts[-1] == HASH assert parts[-2] == cache_subdir assert downloaded_path.exists() snake_case = downloaded_path.read_text() assert content == CONTENT snake_case = downloaded_path.with_suffix('''.json''' ) assert metadata_downloaded_path.exists() snake_case = json.loads(metadata_downloaded_path.read_text() ) assert metadata_content == {"url": URL, "etag": None} @pytest.mark.parametrize('''paths_type''' , [str, list, dict] ) def __UpperCamelCase ( a : List[str] , a : Optional[Any] , a : Optional[int] ) ->Tuple: snake_case = str(a ) if issubclass(a , a ): snake_case = filename elif issubclass(a , a ): snake_case = [filename] elif issubclass(a , a ): snake_case = {'''train''': filename} snake_case = '''dummy''' snake_case = xz_file.parent snake_case = '''extracted''' snake_case = DownloadConfig( cache_dir=a , use_etag=a , ) snake_case = DownloadManager(dataset_name=a , download_config=a ) snake_case = dl_manager.extract(a ) snake_case = paths for extracted_paths in [extracted_paths]: if isinstance(a , a ): snake_case = [extracted_paths] snake_case = [paths] elif isinstance(a , a ): assert "train" in extracted_paths.keys() snake_case = extracted_paths.values() snake_case = paths.values() assert extracted_paths for extracted_path, input_path in zip(a , a ): assert extracted_path == dl_manager.extracted_paths[input_path] snake_case = Path(a ) snake_case = extracted_path.parts assert parts[-1] == hash_url_to_filename(a , etag=a ) assert parts[-2] == extracted_subdir assert extracted_path.exists() snake_case = extracted_path.read_text() snake_case = text_file.read_text() assert extracted_file_content == expected_file_content def __UpperCamelCase ( a : Any , a : Dict ) ->Tuple: assert path.endswith('''.jsonl''' ) for num_items, line in enumerate(a , start=1 ): snake_case = json.loads(line.decode('''utf-8''' ) ) assert item.keys() == {"col_1", "col_2", "col_3"} assert num_items == 4 @pytest.mark.parametrize('''archive_jsonl''' , ['''tar_jsonl_path''', '''zip_jsonl_path'''] ) def __UpperCamelCase ( a : Dict , a : Tuple ) ->List[Any]: snake_case = request.getfixturevalue(a ) snake_case = DownloadManager() for num_jsonl, (path, file) in enumerate(dl_manager.iter_archive(a ) , start=1 ): _test_jsonl(a , a ) assert num_jsonl == 2 @pytest.mark.parametrize('''archive_nested_jsonl''' , ['''tar_nested_jsonl_path''', '''zip_nested_jsonl_path'''] ) def __UpperCamelCase ( a : Tuple , a : Union[str, Any] ) ->Any: snake_case = request.getfixturevalue(a ) snake_case = DownloadManager() for num_tar, (path, file) in enumerate(dl_manager.iter_archive(a ) , start=1 ): for num_jsonl, (subpath, subfile) in enumerate(dl_manager.iter_archive(a ) , start=1 ): _test_jsonl(a , a ) assert num_tar == 1 assert num_jsonl == 2 def __UpperCamelCase ( a : List[str] ) ->Optional[Any]: snake_case = DownloadManager() for num_file, file in enumerate(dl_manager.iter_files(a ) , start=1 ): assert os.path.basename(a ) == ("test.txt" if num_file == 1 else "train.txt") assert num_file == 2
703
'''simple docstring''' import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class _lowercase ( __a , __a , unittest.TestCase ): _UpperCAmelCase = IFInpaintingSuperResolutionPipeline _UpperCAmelCase = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} _UpperCAmelCase = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({'''original_image'''} ) _UpperCAmelCase = PipelineTesterMixin.required_optional_params - {'''latents'''} def UpperCamelCase ( self ) -> int: return self._get_superresolution_dummy_components() def UpperCamelCase ( self , A__ , A__=0 ) -> Union[str, Any]: if str(A__ ).startswith('''mps''' ): snake_case = torch.manual_seed(A__ ) else: snake_case = torch.Generator(device=A__ ).manual_seed(A__ ) snake_case = floats_tensor((1, 3, 16, 16) , rng=random.Random(A__ ) ).to(A__ ) snake_case = floats_tensor((1, 3, 32, 32) , rng=random.Random(A__ ) ).to(A__ ) snake_case = floats_tensor((1, 3, 32, 32) , rng=random.Random(A__ ) ).to(A__ ) snake_case = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCamelCase ( self ) -> List[Any]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def UpperCamelCase ( self ) -> Optional[Any]: self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def UpperCamelCase ( self ) -> List[str]: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1 ) def UpperCamelCase ( self ) -> int: self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def UpperCamelCase ( self ) -> Optional[Any]: self._test_save_load_local() def UpperCamelCase ( self ) -> Dict: self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
44
0
'''simple docstring''' from abc import ABC, abstractmethod from typing import List, Optional class _lowercase ( __a ): def __init__( self ) -> List[str]: # test for the above condition self.test() def UpperCamelCase ( self ) -> Dict: snake_case = 0 snake_case = False while not completed: if counter == 1: self.reset() snake_case = self.advance() if not self.does_advance(A__ ): raise Exception( '''Custom Constraint is not defined correctly. self.does_advance(self.advance()) must be true.''' ) snake_case , snake_case , snake_case = self.update(A__ ) counter += 1 if counter > 1_00_00: raise Exception('''update() does not fulfill the constraint.''' ) if self.remaining() != 0: raise Exception('''Custom Constraint is not defined correctly.''' ) @abstractmethod def UpperCamelCase ( self ) -> List[str]: raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def UpperCamelCase ( self , A__ ) -> int: raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def UpperCamelCase ( self , A__ ) -> Tuple: raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def UpperCamelCase ( self ) -> Any: raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def UpperCamelCase ( self ) -> Tuple: raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) @abstractmethod def UpperCamelCase ( self , A__=False ) -> Optional[Any]: raise NotImplementedError( F"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) class _lowercase ( __a ): def __init__( self , A__ ) -> Dict: super(A__ , self ).__init__() if not isinstance(A__ , A__ ) or len(A__ ) == 0: raise ValueError(F"""`token_ids` has to be a non-empty list, but is {token_ids}.""" ) if any((not isinstance(A__ , A__ ) or token_id < 0) for token_id in token_ids ): raise ValueError(F"""Each list in `token_ids` has to be a list of positive integers, but is {token_ids}.""" ) snake_case = token_ids snake_case = len(self.token_ids ) snake_case = -1 # the index of the currently fulfilled step snake_case = False def UpperCamelCase ( self ) -> Any: if self.completed: return None return self.token_ids[self.fulfilled_idx + 1] def UpperCamelCase ( self , A__ ) -> Optional[int]: if not isinstance(A__ , A__ ): raise ValueError(F"""`token_id` has to be an `int`, but is {token_id} of type {type(A__ )}""" ) if self.completed: return False return token_id == self.token_ids[self.fulfilled_idx + 1] def UpperCamelCase ( self , A__ ) -> int: if not isinstance(A__ , A__ ): raise ValueError(F"""`token_id` has to be an `int`, but is {token_id} of type {type(A__ )}""" ) snake_case = False snake_case = False snake_case = False if self.does_advance(A__ ): self.fulfilled_idx += 1 snake_case = True if self.fulfilled_idx == (self.seqlen - 1): snake_case = True snake_case = completed else: # failed to make progress. snake_case = True self.reset() return stepped, completed, reset def UpperCamelCase ( self ) -> int: snake_case = False snake_case = 0 def UpperCamelCase ( self ) -> Tuple: return self.seqlen - (self.fulfilled_idx + 1) def UpperCamelCase ( self , A__=False ) -> int: snake_case = PhrasalConstraint(self.token_ids ) if stateful: snake_case = self.seqlen snake_case = self.fulfilled_idx snake_case = self.completed return new_constraint class _lowercase : def __init__( self , A__ , A__=True ) -> Optional[Any]: snake_case = max([len(A__ ) for one in nested_token_ids] ) snake_case = {} for token_ids in nested_token_ids: snake_case = root for tidx, token_id in enumerate(A__ ): if token_id not in level: snake_case = {} snake_case = level[token_id] if no_subsets and self.has_subsets(A__ , A__ ): raise ValueError( '''Each list in `nested_token_ids` can\'t be a complete subset of another list, but is''' F""" {nested_token_ids}.""" ) snake_case = root def UpperCamelCase ( self , A__ ) -> List[Any]: snake_case = self.trie for current_token in current_seq: snake_case = start[current_token] snake_case = list(start.keys() ) return next_tokens def UpperCamelCase ( self , A__ ) -> Any: snake_case = self.next_tokens(A__ ) return len(A__ ) == 0 def UpperCamelCase ( self , A__ ) -> Union[str, Any]: snake_case = list(root.values() ) if len(A__ ) == 0: return 1 else: return sum([self.count_leaves(A__ ) for nn in next_nodes] ) def UpperCamelCase ( self , A__ , A__ ) -> Optional[int]: snake_case = self.count_leaves(A__ ) return len(A__ ) != leaf_count class _lowercase ( __a ): def __init__( self , A__ ) -> str: super(A__ , self ).__init__() if not isinstance(A__ , A__ ) or len(A__ ) == 0: raise ValueError(F"""`nested_token_ids` has to be a non-empty list, but is {nested_token_ids}.""" ) if any(not isinstance(A__ , A__ ) for token_ids in nested_token_ids ): raise ValueError(F"""`nested_token_ids` has to be a list of lists, but is {nested_token_ids}.""" ) if any( any((not isinstance(A__ , A__ ) or token_id < 0) for token_id in token_ids ) for token_ids in nested_token_ids ): raise ValueError( F"""Each list in `nested_token_ids` has to be a list of positive integers, but is {nested_token_ids}.""" ) snake_case = DisjunctiveTrie(A__ ) snake_case = nested_token_ids snake_case = self.trie.max_height snake_case = [] snake_case = False def UpperCamelCase ( self ) -> List[Any]: snake_case = self.trie.next_tokens(self.current_seq ) if len(A__ ) == 0: return None else: return token_list def UpperCamelCase ( self , A__ ) -> List[str]: if not isinstance(A__ , A__ ): raise ValueError(F"""`token_id` is supposed to be type `int`, but is {token_id} of type {type(A__ )}""" ) snake_case = self.trie.next_tokens(self.current_seq ) return token_id in next_tokens def UpperCamelCase ( self , A__ ) -> Union[str, Any]: if not isinstance(A__ , A__ ): raise ValueError(F"""`token_id` is supposed to be type `int`, but is {token_id} of type {type(A__ )}""" ) snake_case = False snake_case = False snake_case = False if self.does_advance(A__ ): self.current_seq.append(A__ ) snake_case = True else: snake_case = True self.reset() snake_case = self.trie.reached_leaf(self.current_seq ) snake_case = completed return stepped, completed, reset def UpperCamelCase ( self ) -> List[Any]: snake_case = False snake_case = [] def UpperCamelCase ( self ) -> Tuple: if self.completed: # since this can be completed without reaching max height return 0 else: return self.seqlen - len(self.current_seq ) def UpperCamelCase ( self , A__=False ) -> Optional[int]: snake_case = DisjunctiveConstraint(self.token_ids ) if stateful: snake_case = self.seqlen snake_case = self.current_seq snake_case = self.completed return new_constraint class _lowercase : def __init__( self , A__ ) -> Union[str, Any]: snake_case = constraints # max # of steps required to fulfill a given constraint snake_case = max([c.seqlen for c in constraints] ) snake_case = len(A__ ) snake_case = False self.init_state() def UpperCamelCase ( self ) -> Optional[Any]: snake_case = [] snake_case = None snake_case = [constraint.copy(stateful=A__ ) for constraint in self.constraints] def UpperCamelCase ( self ) -> Any: snake_case = 0 if self.inprogress_constraint: # extra points for having a constraint mid-fulfilled add += self.max_seqlen - self.inprogress_constraint.remaining() return (len(self.complete_constraints ) * self.max_seqlen) + add def UpperCamelCase ( self ) -> Optional[int]: snake_case = [] if self.inprogress_constraint is None: for constraint in self.pending_constraints: # "pending" == "unfulfilled yet" snake_case = constraint.advance() if isinstance(A__ , A__ ): token_list.append(A__ ) elif isinstance(A__ , A__ ): token_list.extend(A__ ) else: snake_case = self.inprogress_constraint.advance() if isinstance(A__ , A__ ): token_list.append(A__ ) elif isinstance(A__ , A__ ): token_list.extend(A__ ) if len(A__ ) == 0: return None else: return token_list def UpperCamelCase ( self , A__ ) -> Dict: self.init_state() if token_ids is not None: for token in token_ids: # completes or steps **one** constraint snake_case , snake_case = self.add(A__ ) # the entire list of constraints are fulfilled if self.completed: break def UpperCamelCase ( self , A__ ) -> Tuple: if not isinstance(A__ , A__ ): raise ValueError(F"""`token_id` should be an `int`, but is `{token_id}`.""" ) snake_case , snake_case = False, False if self.completed: snake_case = True snake_case = False return complete, stepped if self.inprogress_constraint is not None: # In the middle of fulfilling a constraint. If the `token_id` *does* makes an incremental progress to current # job, simply update the state snake_case , snake_case , snake_case = self.inprogress_constraint.update(A__ ) if reset: # 1. If the next token breaks the progress, then we must restart. # e.g. constraint = "I love pies" and sequence so far is "I love" but `token_id` == "books". # But that doesn't mean we self.init_state(), since we only reset the state for this particular # constraint, not the full list of constraints. self.pending_constraints.append(self.inprogress_constraint.copy(stateful=A__ ) ) snake_case = None if complete: # 2. If the next token completes the constraint, move it to completed list, set # inprogress to None. If there are no pending constraints either, then this full list of constraints # is complete. self.complete_constraints.append(self.inprogress_constraint ) snake_case = None if len(self.pending_constraints ) == 0: # we're done! snake_case = True else: # Not in the middle of fulfilling a constraint. So does this `token_id` helps us step towards any of our list # of constraints? for cidx, pending_constraint in enumerate(self.pending_constraints ): if pending_constraint.does_advance(A__ ): snake_case , snake_case , snake_case = pending_constraint.update(A__ ) if not stepped: raise Exception( '''`constraint.update(token_id)` is not yielding incremental progress, ''' '''even though `constraint.does_advance(token_id)` is true.''' ) if complete: self.complete_constraints.append(A__ ) snake_case = None if not complete and stepped: snake_case = pending_constraint if complete or stepped: # If we made any progress at all, then it's at least not a "pending constraint". snake_case = ( self.pending_constraints[:cidx] + self.pending_constraints[cidx + 1 :] ) if len(self.pending_constraints ) == 0 and self.inprogress_constraint is None: # If there's no longer any pending after this and no inprogress either, then we must be # complete. snake_case = True break # prevent accidentally stepping through multiple constraints with just one token. return complete, stepped def UpperCamelCase ( self , A__=True ) -> Union[str, Any]: snake_case = ConstraintListState(self.constraints ) # we actually never though self.constraints objects # throughout this process. So it's at initialization state. if stateful: snake_case = [ constraint.copy(stateful=A__ ) for constraint in self.complete_constraints ] if self.inprogress_constraint is not None: snake_case = self.inprogress_constraint.copy(stateful=A__ ) snake_case = [constraint.copy() for constraint in self.pending_constraints] return new_state
704
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy _lowercase = logging.get_logger(__name__) class _lowercase ( __a ): def __init__( self , A__ , A__ , A__ , **A__ ) -> Union[str, Any]: snake_case = feature_size snake_case = sampling_rate snake_case = padding_value snake_case = kwargs.pop('''padding_side''' , '''right''' ) snake_case = kwargs.pop('''return_attention_mask''' , A__ ) super().__init__(**A__ ) def UpperCamelCase ( self , A__ , A__ = True , A__ = None , A__ = False , A__ = None , A__ = None , A__ = None , ) -> BatchFeature: # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(A__ , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): snake_case = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( '''You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`''' F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) snake_case = processed_features[self.model_input_names[0]] snake_case = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(A__ ) == 0: if return_attention_mask: snake_case = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch snake_case = required_input[0] if isinstance(A__ , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. snake_case = 0 while len(required_input[index] ) == 0: index += 1 if index < len(A__ ): snake_case = required_input[index][0] if return_tensors is None: if is_tf_tensor(A__ ): snake_case = '''tf''' elif is_torch_tensor(A__ ): snake_case = '''pt''' elif isinstance(A__ , (int, float, list, tuple, np.ndarray) ): snake_case = '''np''' else: raise ValueError( F"""type of {first_element} unknown: {type(A__ )}. """ '''Should be one of a python, numpy, pytorch or tensorflow object.''' ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): snake_case = to_numpy(A__ ) else: snake_case = [to_numpy(A__ ) for v in value] # Convert padding_strategy in PaddingStrategy snake_case = self._get_padding_strategies(padding=A__ , max_length=A__ ) snake_case = processed_features[self.model_input_names[0]] snake_case = len(A__ ) if not all(len(A__ ) == batch_size for v in processed_features.values() ): raise ValueError('''Some items in the output dictionary have a different batch size than others.''' ) snake_case = [] for i in range(A__ ): snake_case = {k: v[i] for k, v in processed_features.items()} # truncation snake_case = self._truncate( A__ , max_length=A__ , pad_to_multiple_of=A__ , truncation=A__ , ) truncated_inputs.append(A__ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length snake_case = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) snake_case = PaddingStrategy.MAX_LENGTH snake_case = {} for i in range(A__ ): # padding snake_case = self._pad( truncated_inputs[i] , max_length=A__ , padding_strategy=A__ , pad_to_multiple_of=A__ , return_attention_mask=A__ , ) for key, value in outputs.items(): if key not in batch_outputs: snake_case = [] if value.dtype is np.dtype(np.floataa ): snake_case = value.astype(np.floataa ) batch_outputs[key].append(A__ ) return BatchFeature(A__ , tensor_type=A__ ) def UpperCamelCase ( self , A__ , A__ = None , A__ = PaddingStrategy.DO_NOT_PAD , A__ = None , A__ = None , ) -> dict: snake_case = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: snake_case = len(A__ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): snake_case = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of snake_case = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(A__ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: snake_case = np.ones(len(A__ ) , dtype=np.intaa ) if needs_to_be_padded: snake_case = max_length - len(A__ ) if self.padding_side == "right": if return_attention_mask: snake_case = np.pad( processed_features['''attention_mask'''] , (0, difference) ) snake_case = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) snake_case = np.pad( A__ , A__ , '''constant''' , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: snake_case = np.pad( processed_features['''attention_mask'''] , (difference, 0) ) snake_case = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) snake_case = np.pad( A__ , A__ , '''constant''' , constant_values=self.padding_value ) else: raise ValueError('''Invalid padding strategy:''' + str(self.padding_side ) ) return processed_features def UpperCamelCase ( self , A__ , A__ = None , A__ = None , A__ = None , ) -> Union[str, Any]: if not truncation: return processed_features elif truncation and max_length is None: raise ValueError('''When setting ``truncation=True``, make sure that ``max_length`` is defined.''' ) snake_case = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): snake_case = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of snake_case = len(A__ ) > max_length if needs_to_be_truncated: snake_case = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: snake_case = processed_features['''attention_mask'''][:max_length] return processed_features def UpperCamelCase ( self , A__=False , A__=None ) -> Union[str, Any]: # Get padding strategy if padding is not False: if padding is True: snake_case = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(A__ , A__ ): snake_case = PaddingStrategy(A__ ) elif isinstance(A__ , A__ ): snake_case = padding else: snake_case = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( '''Asking to pad but the feature_extractor does not have a padding value. Please select a value to use''' ''' as `padding_value`. For example: `feature_extractor.padding_value = 0.0`.''' ) return padding_strategy
44
0
'''simple docstring''' import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _lowercase = 16 _lowercase = 32 def __UpperCamelCase ( a : Accelerator , a : int = 16 ) ->str: snake_case = AutoTokenizer.from_pretrained('''bert-base-cased''' ) snake_case = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(a : Union[str, Any] ): # max_length=None => use the model max length (it's actually the default) snake_case = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=a , max_length=a ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): snake_case = datasets.map( a , batched=a , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library snake_case = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(a : int ): # On TPU it's best to pad everything to the same length or training will be very slow. snake_case = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": snake_case = 16 elif accelerator.mixed_precision != "no": snake_case = 8 else: snake_case = None return tokenizer.pad( a , padding='''longest''' , max_length=a , pad_to_multiple_of=a , return_tensors='''pt''' , ) # Instantiate dataloaders. snake_case = DataLoader( tokenized_datasets['''train'''] , shuffle=a , collate_fn=a , batch_size=a ) snake_case = DataLoader( tokenized_datasets['''validation'''] , shuffle=a , collate_fn=a , batch_size=a ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders _lowercase = mocked_dataloaders # noqa: F811 def __UpperCamelCase ( a : List[str] , a : int ) ->int: # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , a ) == "1": snake_case = 2 # Initialize accelerator snake_case = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs snake_case = config['''lr'''] snake_case = int(config['''num_epochs'''] ) snake_case = int(config['''seed'''] ) snake_case = int(config['''batch_size'''] ) snake_case = evaluate.load('''glue''' , '''mrpc''' ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=a ) def inner_training_loop(a : Tuple ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(a ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) snake_case = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=a ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). snake_case = model.to(accelerator.device ) # Instantiate optimizer snake_case = AdamW(params=model.parameters() , lr=a ) snake_case , snake_case = get_dataloaders(a , a ) # Instantiate scheduler snake_case = get_linear_schedule_with_warmup( optimizer=a , num_warmup_steps=100 , num_training_steps=(len(a ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. snake_case , snake_case , snake_case , snake_case , snake_case = accelerator.prepare( a , a , a , a , a ) # Now we train the model for epoch in range(a ): model.train() for step, batch in enumerate(a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) snake_case = model(**a ) snake_case = outputs.loss accelerator.backward(a ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(a ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): snake_case = model(**a ) snake_case = outputs.logits.argmax(dim=-1 ) snake_case , snake_case = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=a , references=a , ) snake_case = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , a ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def __UpperCamelCase ( ) ->Any: snake_case = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=a , default=a , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) snake_case = parser.parse_args() snake_case = {'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(a , a ) if __name__ == "__main__": main()
705
'''simple docstring''' from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class _lowercase ( yaml.SafeLoader ): def UpperCamelCase ( self , A__ ) -> List[str]: snake_case = [self.constructed_objects[key_node] for key_node, _ in node.value] snake_case = [tuple(A__ ) if isinstance(A__ , A__ ) else key for key in keys] snake_case = Counter(A__ ) snake_case = [key for key in counter if counter[key] > 1] if duplicate_keys: raise TypeError(F"""Got duplicate yaml keys: {duplicate_keys}""" ) def UpperCamelCase ( self , A__ , A__=False ) -> List[Any]: snake_case = super().construct_mapping(A__ , deep=A__ ) self._check_no_duplicates_on_constructed_node(A__ ) return mapping def __UpperCamelCase ( a : str ) ->Tuple[Optional[str], str]: snake_case = list(readme_content.splitlines() ) if full_content and full_content[0] == "---" and "---" in full_content[1:]: snake_case = full_content[1:].index('''---''' ) + 1 snake_case = '''\n'''.join(full_content[1:sep_idx] ) return yamlblock, "\n".join(full_content[sep_idx + 1 :] ) return None, "\n".join(a ) class _lowercase ( __a ): # class attributes _UpperCAmelCase = {'''train_eval_index'''} # train-eval-index in the YAML metadata @classmethod def UpperCamelCase ( cls , A__ ) -> "DatasetMetadata": with open(A__ , encoding='''utf-8''' ) as readme_file: snake_case , snake_case = _split_yaml_from_readme(readme_file.read() ) if yaml_string is not None: return cls.from_yaml_string(A__ ) else: return cls() def UpperCamelCase ( self , A__ ) -> str: if path.exists(): with open(A__ , encoding='''utf-8''' ) as readme_file: snake_case = readme_file.read() else: snake_case = None snake_case = self._to_readme(A__ ) with open(A__ , '''w''' , encoding='''utf-8''' ) as readme_file: readme_file.write(A__ ) def UpperCamelCase ( self , A__ = None ) -> str: if readme_content is not None: snake_case , snake_case = _split_yaml_from_readme(A__ ) snake_case = '''---\n''' + self.to_yaml_string() + '''---\n''' + content else: snake_case = '''---\n''' + self.to_yaml_string() + '''---\n''' return full_content @classmethod def UpperCamelCase ( cls , A__ ) -> "DatasetMetadata": snake_case = yaml.load(A__ , Loader=_NoDuplicateSafeLoader ) or {} # Convert the YAML keys to DatasetMetadata fields snake_case = { (key.replace('''-''' , '''_''' ) if key.replace('''-''' , '''_''' ) in cls._FIELDS_WITH_DASHES else key): value for key, value in metadata_dict.items() } return cls(**A__ ) def UpperCamelCase ( self ) -> str: return yaml.safe_dump( { (key.replace('''_''' , '''-''' ) if key in self._FIELDS_WITH_DASHES else key): value for key, value in self.items() } , sort_keys=A__ , allow_unicode=A__ , encoding='''utf-8''' , ).decode('''utf-8''' ) _lowercase = { 'image-classification': [], 'translation': [], 'image-segmentation': [], 'fill-mask': [], 'automatic-speech-recognition': [], 'token-classification': [], 'sentence-similarity': [], 'audio-classification': [], 'question-answering': [], 'summarization': [], 'zero-shot-classification': [], 'table-to-text': [], 'feature-extraction': [], 'other': [], 'multiple-choice': [], 'text-classification': [], 'text-to-image': [], 'text2text-generation': [], 'zero-shot-image-classification': [], 'tabular-classification': [], 'tabular-regression': [], 'image-to-image': [], 'tabular-to-text': [], 'unconditional-image-generation': [], 'text-retrieval': [], 'text-to-speech': [], 'object-detection': [], 'audio-to-audio': [], 'text-generation': [], 'conversational': [], 'table-question-answering': [], 'visual-question-answering': [], 'image-to-text': [], 'reinforcement-learning': [], 'voice-activity-detection': [], 'time-series-forecasting': [], 'document-question-answering': [], } if __name__ == "__main__": from argparse import ArgumentParser _lowercase = ArgumentParser(usage='Validate the yaml metadata block of a README.md file.') ap.add_argument('readme_filepath') _lowercase = ap.parse_args() _lowercase = Path(args.readme_filepath) _lowercase = DatasetMetadata.from_readme(readme_filepath) print(dataset_metadata) dataset_metadata.to_readme(readme_filepath)
44
0
'''simple docstring''' import argparse import torch from transformers import GPTaConfig, GPTaModel, load_tf_weights_in_gpta from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def __UpperCamelCase ( a : Tuple , a : str , a : Any ) ->Union[str, Any]: # Construct model if gpta_config_file == "": snake_case = GPTaConfig() else: snake_case = GPTaConfig.from_json_file(a ) snake_case = GPTaModel(a ) # Load weights from numpy load_tf_weights_in_gpta(a , a , a ) # Save pytorch-model snake_case = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME snake_case = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(f"""Save PyTorch model to {pytorch_weights_dump_path}""" ) torch.save(model.state_dict() , a ) print(f"""Save configuration file to {pytorch_config_dump_path}""" ) with open(a , '''w''' , encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser() # Required parameters parser.add_argument( '--gpt2_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--gpt2_config_file', default='', type=str, help=( 'An optional config json file corresponding to the pre-trained OpenAI model. \n' 'This specifies the model architecture.' ), ) _lowercase = parser.parse_args() convert_gpta_checkpoint_to_pytorch(args.gpta_checkpoint_path, args.gpta_config_file, args.pytorch_dump_folder_path)
706
'''simple docstring''' import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _lowercase ( __a , unittest.TestCase ): _UpperCAmelCase = CodeGenTokenizer _UpperCAmelCase = CodeGenTokenizerFast _UpperCAmelCase = True _UpperCAmelCase = {'''add_prefix_space''': True} _UpperCAmelCase = False def UpperCamelCase ( self ) -> Tuple: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt snake_case = [ '''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''\u0120''', '''\u0120l''', '''\u0120n''', '''\u0120lo''', '''\u0120low''', '''er''', '''\u0120lowest''', '''\u0120newer''', '''\u0120wider''', '''<unk>''', '''<|endoftext|>''', ] snake_case = dict(zip(A__ , range(len(A__ ) ) ) ) snake_case = ['''#version: 0.2''', '''\u0120 l''', '''\u0120l o''', '''\u0120lo w''', '''e r''', ''''''] snake_case = {'''unk_token''': '''<unk>'''} snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(A__ ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(A__ ) ) def UpperCamelCase ( self , **A__ ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return CodeGenTokenizer.from_pretrained(self.tmpdirname , **A__ ) def UpperCamelCase ( self , **A__ ) -> Union[str, Any]: kwargs.update(self.special_tokens_map ) return CodeGenTokenizerFast.from_pretrained(self.tmpdirname , **A__ ) def UpperCamelCase ( self , A__ ) -> Tuple: snake_case = '''lower newer''' snake_case = '''lower newer''' return input_text, output_text def UpperCamelCase ( self ) -> List[Any]: snake_case = CodeGenTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) snake_case = '''lower newer''' snake_case = ['''\u0120low''', '''er''', '''\u0120''', '''n''', '''e''', '''w''', '''er'''] snake_case = tokenizer.tokenize(A__ , add_prefix_space=A__ ) self.assertListEqual(A__ , A__ ) snake_case = tokens + [tokenizer.unk_token] snake_case = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(tokenizer.convert_tokens_to_ids(A__ ) , A__ ) def UpperCamelCase ( self ) -> Optional[int]: if not self.test_rust_tokenizer: return snake_case = self.get_tokenizer() snake_case = self.get_rust_tokenizer(add_prefix_space=A__ ) snake_case = '''lower newer''' # Testing tokenization snake_case = tokenizer.tokenize(A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.tokenize(A__ ) self.assertListEqual(A__ , A__ ) # Testing conversion to ids without special tokens snake_case = tokenizer.encode(A__ , add_special_tokens=A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.encode(A__ , add_special_tokens=A__ ) self.assertListEqual(A__ , A__ ) # Testing conversion to ids with special tokens snake_case = self.get_rust_tokenizer(add_prefix_space=A__ ) snake_case = tokenizer.encode(A__ , add_prefix_space=A__ ) snake_case = rust_tokenizer.encode(A__ ) self.assertListEqual(A__ , A__ ) # Testing the unknown token snake_case = tokens + [rust_tokenizer.unk_token] snake_case = [14, 15, 10, 9, 3, 2, 15, 19] self.assertListEqual(rust_tokenizer.convert_tokens_to_ids(A__ ) , A__ ) def UpperCamelCase ( self , *A__ , **A__ ) -> List[str]: # It's very difficult to mix/test pretokenization with byte-level # And get both CodeGen and Roberta to work at the same time (mostly an issue of adding a space before the string) pass def UpperCamelCase ( self , A__=15 ) -> Tuple: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): snake_case = self.rust_tokenizer_class.from_pretrained(A__ , **A__ ) # Simple input snake_case = '''This is a simple input''' snake_case = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case = ('''This is a simple input''', '''This is a pair''') snake_case = [ ('''This is a simple input 1''', '''This is a simple input 2'''), ('''This is a simple pair 1''', '''This is a simple pair 2'''), ] # Simple input tests self.assertRaises(A__ , tokenizer_r.encode , A__ , max_length=A__ , padding='''max_length''' ) # Simple input self.assertRaises(A__ , tokenizer_r.encode_plus , A__ , max_length=A__ , padding='''max_length''' ) # Simple input self.assertRaises( A__ , tokenizer_r.batch_encode_plus , A__ , max_length=A__ , padding='''max_length''' , ) # Pair input self.assertRaises(A__ , tokenizer_r.encode , A__ , max_length=A__ , padding='''max_length''' ) # Pair input self.assertRaises(A__ , tokenizer_r.encode_plus , A__ , max_length=A__ , padding='''max_length''' ) # Pair input self.assertRaises( A__ , tokenizer_r.batch_encode_plus , A__ , max_length=A__ , padding='''max_length''' , ) def UpperCamelCase ( self ) -> Tuple: snake_case = CodeGenTokenizer.from_pretrained(self.tmpdirname , pad_token='''<pad>''' ) # Simple input snake_case = '''This is a simple input''' snake_case = ['''This is a simple input looooooooong''', '''This is a simple input'''] snake_case = ('''This is a simple input''', '''This is a pair''') snake_case = [ ('''This is a simple input loooooong''', '''This is a simple input'''), ('''This is a simple pair loooooong''', '''This is a simple pair'''), ] snake_case = tokenizer.pad_token_id snake_case = tokenizer(A__ , padding='''max_length''' , max_length=30 , return_tensors='''np''' ) snake_case = tokenizer(A__ , padding=A__ , truncate=A__ , return_tensors='''np''' ) snake_case = tokenizer(*A__ , padding='''max_length''' , max_length=60 , return_tensors='''np''' ) snake_case = tokenizer(A__ , padding=A__ , truncate=A__ , return_tensors='''np''' ) # s # test single string max_length padding self.assertEqual(out_s['''input_ids'''].shape[-1] , 30 ) self.assertTrue(pad_token_id in out_s['''input_ids'''] ) self.assertTrue(0 in out_s['''attention_mask'''] ) # s2 # test automatic padding self.assertEqual(out_sa['''input_ids'''].shape[-1] , 33 ) # long slice doesn't have padding self.assertFalse(pad_token_id in out_sa['''input_ids'''][0] ) self.assertFalse(0 in out_sa['''attention_mask'''][0] ) # short slice does have padding self.assertTrue(pad_token_id in out_sa['''input_ids'''][1] ) self.assertTrue(0 in out_sa['''attention_mask'''][1] ) # p # test single pair max_length padding self.assertEqual(out_p['''input_ids'''].shape[-1] , 60 ) self.assertTrue(pad_token_id in out_p['''input_ids'''] ) self.assertTrue(0 in out_p['''attention_mask'''] ) # p2 # test automatic padding pair self.assertEqual(out_pa['''input_ids'''].shape[-1] , 52 ) # long slice pair doesn't have padding self.assertFalse(pad_token_id in out_pa['''input_ids'''][0] ) self.assertFalse(0 in out_pa['''attention_mask'''][0] ) # short slice pair does have padding self.assertTrue(pad_token_id in out_pa['''input_ids'''][1] ) self.assertTrue(0 in out_pa['''attention_mask'''][1] ) def UpperCamelCase ( self ) -> str: snake_case = '''$$$''' snake_case = CodeGenTokenizer.from_pretrained(self.tmpdirname , bos_token=A__ , add_bos_token=A__ ) snake_case = '''This is a simple input''' snake_case = ['''This is a simple input 1''', '''This is a simple input 2'''] snake_case = tokenizer.bos_token_id snake_case = tokenizer(A__ ) snake_case = tokenizer(A__ ) self.assertEqual(out_s.input_ids[0] , A__ ) self.assertTrue(all(o[0] == bos_token_id for o in out_sa.input_ids ) ) snake_case = tokenizer.decode(out_s.input_ids ) snake_case = tokenizer.batch_decode(out_sa.input_ids ) self.assertEqual(decode_s.split()[0] , A__ ) self.assertTrue(all(d.split()[0] == bos_token for d in decode_sa ) ) @slow def UpperCamelCase ( self ) -> Any: snake_case = CodeGenTokenizer.from_pretrained('''Salesforce/codegen-350M-mono''' ) snake_case = '''\nif len_a > len_b:\n result = a\nelse:\n result = b\n\n\n\n#''' snake_case = '''\nif len_a > len_b: result = a\nelse: result = b''' snake_case = tokenizer.encode(A__ ) snake_case = ['''^#''', re.escape('''<|endoftext|>''' ), '''^\'\'\'''', '''^"""''', '''\n\n\n'''] snake_case = tokenizer.decode(A__ , truncate_before_pattern=A__ ) self.assertEqual(A__ , A__ ) def UpperCamelCase ( self ) -> Union[str, Any]: pass
44
0
'''simple docstring''' import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from datasets.iterable_dataset import IterableDataset from datasets.load import dataset_module_factory, import_main_class from datasets.utils.file_utils import cached_path _lowercase = [ {'dataset': 'wikipedia', 'config_name': '20220301.de'}, {'dataset': 'wikipedia', 'config_name': '20220301.en'}, {'dataset': 'wikipedia', 'config_name': '20220301.fr'}, {'dataset': 'wikipedia', 'config_name': '20220301.frr'}, {'dataset': 'wikipedia', 'config_name': '20220301.it'}, {'dataset': 'wikipedia', 'config_name': '20220301.simple'}, {'dataset': 'snli', 'config_name': 'plain_text'}, {'dataset': 'eli5', 'config_name': 'LFQA_reddit'}, {'dataset': 'wiki40b', 'config_name': 'en'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.compressed'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.no_index'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.multiset.no_index'}, {'dataset': 'natural_questions', 'config_name': 'default'}, ] def __UpperCamelCase ( a : Dict=True ) ->str: if with_config: return [ { "testcase_name": d["dataset"] + "/" + d["config_name"], "dataset": d["dataset"], "config_name": d["config_name"], } for d in DATASETS_ON_HF_GCP ] else: return [ {"testcase_name": dataset, "dataset": dataset} for dataset in {d["dataset"] for d in DATASETS_ON_HF_GCP} ] @parameterized.named_parameters(list_datasets_on_hf_gcp_parameters(with_config=__a ) ) class _lowercase ( __a ): _UpperCAmelCase = None _UpperCAmelCase = None def UpperCamelCase ( self , A__ , A__ ) -> str: with TemporaryDirectory() as tmp_dir: snake_case = dataset_module_factory(A__ , cache_dir=A__ ) snake_case = import_main_class(dataset_module.module_path , dataset=A__ ) snake_case = builder_cls( cache_dir=A__ , config_name=A__ , hash=dataset_module.hash , ) snake_case = '''/'''.join( [ HF_GCP_BASE_URL, builder_instance._relative_data_dir(with_hash=A__ ).replace(os.sep , '''/''' ), config.DATASET_INFO_FILENAME, ] ) snake_case = cached_path(A__ , cache_dir=A__ ) self.assertTrue(os.path.exists(A__ ) ) @pytest.mark.integration def __UpperCamelCase ( a : List[str] ) ->Any: snake_case = tmp_path_factory.mktemp('''test_hf_gcp''' ) / '''test_wikipedia_simple''' snake_case = dataset_module_factory('''wikipedia''' , cache_dir=a ) snake_case = import_main_class(dataset_module.module_path ) snake_case = builder_cls( cache_dir=a , config_name='''20220301.frr''' , hash=dataset_module.hash , ) # use the HF cloud storage, not the original download_and_prepare that uses apache-beam snake_case = None builder_instance.download_and_prepare() snake_case = builder_instance.as_dataset() assert ds @pytest.mark.integration def __UpperCamelCase ( a : Any ) ->Union[str, Any]: snake_case = dataset_module_factory('''wikipedia''' , cache_dir=a ) snake_case = import_main_class(dataset_module.module_path , dataset=a ) snake_case = builder_cls( cache_dir=a , config_name='''20220301.frr''' , hash=dataset_module.hash , ) snake_case = builder_instance.as_streaming_dataset() assert ds assert isinstance(a , a ) assert "train" in ds assert isinstance(ds['''train'''] , a ) assert next(iter(ds['''train'''] ) )
707
'''simple docstring''' from __future__ import annotations import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFViTForImageClassification, TFViTModel if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowercase : def __init__( self , A__ , A__=13 , A__=30 , A__=2 , A__=3 , A__=True , A__=True , A__=32 , A__=2 , A__=4 , A__=37 , A__="gelu" , A__=0.1 , A__=0.1 , A__=10 , A__=0.0_2 , A__=3 , A__=None , ) -> List[Any]: snake_case = parent snake_case = batch_size snake_case = image_size snake_case = patch_size snake_case = num_channels snake_case = is_training snake_case = use_labels snake_case = hidden_size snake_case = num_hidden_layers snake_case = num_attention_heads snake_case = intermediate_size snake_case = hidden_act snake_case = hidden_dropout_prob snake_case = attention_probs_dropout_prob snake_case = type_sequence_label_size snake_case = initializer_range snake_case = scope # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) snake_case = (image_size // patch_size) ** 2 snake_case = num_patches + 1 def UpperCamelCase ( self ) -> int: snake_case = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case = None if self.use_labels: snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case = self.get_config() return config, pixel_values, labels def UpperCamelCase ( self ) -> int: return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A__ , initializer_range=self.initializer_range , ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Union[str, Any]: snake_case = TFViTModel(config=A__ ) snake_case = model(A__ , training=A__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # Test with an image with different size than the one specified in config. snake_case = self.image_size // 2 snake_case = pixel_values[:, :, :image_size, :image_size] snake_case = model(A__ , interpolate_pos_encoding=A__ , training=A__ ) snake_case = (image_size // self.patch_size) ** 2 + 1 self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, seq_length, self.hidden_size) ) def UpperCamelCase ( self , A__ , A__ , A__ ) -> Optional[int]: snake_case = self.type_sequence_label_size snake_case = TFViTForImageClassification(A__ ) snake_case = model(A__ , labels=A__ , training=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # Test with an image with different size than the one specified in config. snake_case = self.image_size // 2 snake_case = pixel_values[:, :, :image_size, :image_size] snake_case = model(A__ , interpolate_pos_encoding=A__ , training=A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images snake_case = 1 snake_case = TFViTForImageClassification(A__ ) snake_case = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) snake_case = model(A__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.prepare_config_and_inputs() snake_case , snake_case , snake_case = config_and_inputs snake_case = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class _lowercase ( __a , __a , unittest.TestCase ): _UpperCAmelCase = (TFViTModel, TFViTForImageClassification) if is_tf_available() else () _UpperCAmelCase = ( {'''feature-extraction''': TFViTModel, '''image-classification''': TFViTForImageClassification} if is_tf_available() else {} ) _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False def UpperCamelCase ( self ) -> List[Any]: snake_case = TFViTModelTester(self ) snake_case = ConfigTester(self , config_class=A__ , has_text_modality=A__ , hidden_size=37 ) def UpperCamelCase ( self ) -> int: self.config_tester.run_common_tests() @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def UpperCamelCase ( self ) -> int: pass @unittest.skip(reason='''ViT does not use inputs_embeds''' ) def UpperCamelCase ( self ) -> str: pass def UpperCamelCase ( self ) -> Union[str, Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(A__ ) self.assertIsInstance(model.get_input_embeddings() , (tf.keras.layers.Layer) ) snake_case = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A__ , tf.keras.layers.Layer ) ) def UpperCamelCase ( self ) -> List[Any]: snake_case , snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case = model_class(A__ ) snake_case = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case = [*signature.parameters.keys()] snake_case = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , A__ ) def UpperCamelCase ( self ) -> Union[str, Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A__ ) def UpperCamelCase ( self ) -> Optional[Any]: snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*A__ ) @slow def UpperCamelCase ( self ) -> Any: snake_case = TFViTModel.from_pretrained('''google/vit-base-patch16-224''' ) self.assertIsNotNone(A__ ) def __UpperCamelCase ( ) ->Any: snake_case = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class _lowercase ( unittest.TestCase ): @cached_property def UpperCamelCase ( self ) -> Optional[int]: return ViTImageProcessor.from_pretrained('''google/vit-base-patch16-224''' ) if is_vision_available() else None @slow def UpperCamelCase ( self ) -> Dict: snake_case = TFViTForImageClassification.from_pretrained('''google/vit-base-patch16-224''' ) snake_case = self.default_image_processor snake_case = prepare_img() snake_case = image_processor(images=A__ , return_tensors='''tf''' ) # forward pass snake_case = model(**A__ ) # verify the logits snake_case = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , A__ ) snake_case = tf.constant([-0.2_7_4_4, 0.8_2_1_5, -0.0_8_3_6] ) tf.debugging.assert_near(outputs.logits[0, :3] , A__ , atol=1e-4 )
44
0
'''simple docstring''' import copy from dataclasses import dataclass from pathlib import Path from typing import Dict, Optional, Union @dataclass class _lowercase : _UpperCAmelCase = None _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = None _UpperCAmelCase = None _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = False _UpperCAmelCase = True _UpperCAmelCase = None _UpperCAmelCase = 1 _UpperCAmelCase = None _UpperCAmelCase = False _UpperCAmelCase = None _UpperCAmelCase = None def UpperCamelCase ( self ) -> "DownloadConfig": return self.__class__(**{k: copy.deepcopy(A__ ) for k, v in self.__dict__.items()} )
708
'''simple docstring''' import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from datasets.iterable_dataset import IterableDataset from datasets.load import dataset_module_factory, import_main_class from datasets.utils.file_utils import cached_path _lowercase = [ {'dataset': 'wikipedia', 'config_name': '20220301.de'}, {'dataset': 'wikipedia', 'config_name': '20220301.en'}, {'dataset': 'wikipedia', 'config_name': '20220301.fr'}, {'dataset': 'wikipedia', 'config_name': '20220301.frr'}, {'dataset': 'wikipedia', 'config_name': '20220301.it'}, {'dataset': 'wikipedia', 'config_name': '20220301.simple'}, {'dataset': 'snli', 'config_name': 'plain_text'}, {'dataset': 'eli5', 'config_name': 'LFQA_reddit'}, {'dataset': 'wiki40b', 'config_name': 'en'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.compressed'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.nq.no_index'}, {'dataset': 'wiki_dpr', 'config_name': 'psgs_w100.multiset.no_index'}, {'dataset': 'natural_questions', 'config_name': 'default'}, ] def __UpperCamelCase ( a : Dict=True ) ->str: if with_config: return [ { "testcase_name": d["dataset"] + "/" + d["config_name"], "dataset": d["dataset"], "config_name": d["config_name"], } for d in DATASETS_ON_HF_GCP ] else: return [ {"testcase_name": dataset, "dataset": dataset} for dataset in {d["dataset"] for d in DATASETS_ON_HF_GCP} ] @parameterized.named_parameters(list_datasets_on_hf_gcp_parameters(with_config=__a ) ) class _lowercase ( __a ): _UpperCAmelCase = None _UpperCAmelCase = None def UpperCamelCase ( self , A__ , A__ ) -> str: with TemporaryDirectory() as tmp_dir: snake_case = dataset_module_factory(A__ , cache_dir=A__ ) snake_case = import_main_class(dataset_module.module_path , dataset=A__ ) snake_case = builder_cls( cache_dir=A__ , config_name=A__ , hash=dataset_module.hash , ) snake_case = '''/'''.join( [ HF_GCP_BASE_URL, builder_instance._relative_data_dir(with_hash=A__ ).replace(os.sep , '''/''' ), config.DATASET_INFO_FILENAME, ] ) snake_case = cached_path(A__ , cache_dir=A__ ) self.assertTrue(os.path.exists(A__ ) ) @pytest.mark.integration def __UpperCamelCase ( a : List[str] ) ->Any: snake_case = tmp_path_factory.mktemp('''test_hf_gcp''' ) / '''test_wikipedia_simple''' snake_case = dataset_module_factory('''wikipedia''' , cache_dir=a ) snake_case = import_main_class(dataset_module.module_path ) snake_case = builder_cls( cache_dir=a , config_name='''20220301.frr''' , hash=dataset_module.hash , ) # use the HF cloud storage, not the original download_and_prepare that uses apache-beam snake_case = None builder_instance.download_and_prepare() snake_case = builder_instance.as_dataset() assert ds @pytest.mark.integration def __UpperCamelCase ( a : Any ) ->Union[str, Any]: snake_case = dataset_module_factory('''wikipedia''' , cache_dir=a ) snake_case = import_main_class(dataset_module.module_path , dataset=a ) snake_case = builder_cls( cache_dir=a , config_name='''20220301.frr''' , hash=dataset_module.hash , ) snake_case = builder_instance.as_streaming_dataset() assert ds assert isinstance(a , a ) assert "train" in ds assert isinstance(ds['''train'''] , a ) assert next(iter(ds['''train'''] ) )
44
0
'''simple docstring''' import os from datetime import datetime as dt from github import Github _lowercase = [ 'good first issue', 'feature request', 'wip', ] def __UpperCamelCase ( ) ->List[Any]: snake_case = Github(os.environ['''GITHUB_TOKEN'''] ) snake_case = g.get_repo('''huggingface/accelerate''' ) snake_case = repo.get_issues(state='''open''' ) for issue in open_issues: snake_case = sorted([comment for comment in issue.get_comments()] , key=lambda a : i.created_at , reverse=a ) snake_case = comments[0] if len(a ) > 0 else None snake_case = dt.utcnow() snake_case = (current_time - issue.updated_at).days snake_case = (current_time - issue.created_at).days if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and days_since_updated > 7 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Close issue since it has been 7 days of inactivity since bot mention. issue.edit(state='''closed''' ) elif ( days_since_updated > 23 and days_since_creation >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Add stale comment issue.create_comment( '''This issue has been automatically marked as stale because it has not had ''' '''recent activity. If you think this still needs to be addressed ''' '''please comment on this thread.\n\nPlease note that issues that do not follow the ''' '''[contributing guidelines](https://github.com/huggingface/accelerate/blob/main/CONTRIBUTING.md) ''' '''are likely to be ignored.''' ) if __name__ == "__main__": main()
709
'''simple docstring''' def __UpperCamelCase ( a : int , a : int ) ->int: while b: snake_case , snake_case = b, a % b return a def __UpperCamelCase ( a : int , a : int ) ->int: return a if b == 0 else euclidean_gcd_recursive(a , a % b ) def __UpperCamelCase ( ) ->Optional[Any]: print(f"""euclidean_gcd(3, 5) = {euclidean_gcd(3 , 5 )}""" ) print(f"""euclidean_gcd(5, 3) = {euclidean_gcd(5 , 3 )}""" ) print(f"""euclidean_gcd(1, 3) = {euclidean_gcd(1 , 3 )}""" ) print(f"""euclidean_gcd(3, 6) = {euclidean_gcd(3 , 6 )}""" ) print(f"""euclidean_gcd(6, 3) = {euclidean_gcd(6 , 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3 , 5 )}""" ) print(f"""euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5 , 3 )}""" ) print(f"""euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1 , 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3 , 6 )}""" ) print(f"""euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6 , 3 )}""" ) if __name__ == "__main__": main()
44
0
'''simple docstring''' _lowercase = '\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' _lowercase = [{'type': 'code', 'content': INSTALL_CONTENT}] _lowercase = { '{processor_class}': 'FakeProcessorClass', '{model_class}': 'FakeModelClass', '{object_class}': 'FakeObjectClass', }
710
'''simple docstring''' import argparse import copy def __UpperCamelCase ( a : Union[str, Any] ) ->Tuple: snake_case = {} with open(a ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: snake_case = [] _list.append([line.split()[1], line.split()[2]] ) snake_case = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: snake_case = [] _list.append([line.split()[0], line.split()[2]] ) snake_case = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def __UpperCamelCase ( a : Dict , a : Tuple ) ->int: with open(a ) as f: snake_case = f.read(1 ) snake_case = start_node snake_case = [] snake_case = start_node snake_case = 0 while visiting not in first_solution: snake_case = 1_0000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(a ) and k[0] not in first_solution: snake_case = k[1] snake_case = k[0] first_solution.append(a ) snake_case = distance_of_first_solution + int(a ) snake_case = best_node first_solution.append(a ) snake_case = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 snake_case = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_0000 ) return first_solution, distance_of_first_solution def __UpperCamelCase ( a : Optional[int] , a : str ) ->str: snake_case = [] for n in solution[1:-1]: snake_case = solution.index(a ) for kn in solution[1:-1]: snake_case = solution.index(a ) if n == kn: continue snake_case = copy.deepcopy(a ) snake_case = kn snake_case = n snake_case = 0 for k in _tmp[:-1]: snake_case = _tmp[_tmp.index(a ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: snake_case = distance + int(i[1] ) _tmp.append(a ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) snake_case = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda a : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def __UpperCamelCase ( a : Any , a : Optional[Any] , a : int , a : Optional[int] , a : Union[str, Any] ) ->List[Any]: snake_case = 1 snake_case = first_solution snake_case = [] snake_case = distance_of_first_solution snake_case = solution while count <= iters: snake_case = find_neighborhood(a , a ) snake_case = 0 snake_case = neighborhood[index_of_best_solution] snake_case = len(a ) - 1 snake_case = False while not found: snake_case = 0 while i < len(a ): if best_solution[i] != solution[i]: snake_case = best_solution[i] snake_case = solution[i] break snake_case = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) snake_case = True snake_case = best_solution[:-1] snake_case = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: snake_case = cost snake_case = solution else: snake_case = index_of_best_solution + 1 snake_case = neighborhood[index_of_best_solution] if len(a ) >= size: tabu_list.pop(0 ) snake_case = count + 1 return best_solution_ever, best_cost def __UpperCamelCase ( a : Union[str, Any]=None ) ->Optional[Any]: snake_case = generate_neighbours(args.File ) snake_case , snake_case = generate_first_solution( args.File , a ) snake_case , snake_case = tabu_search( a , a , a , args.Iterations , args.Size , ) print(f"""Best solution: {best_sol}, with total distance: {best_cost}.""" ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser(description='Tabu Search') parser.add_argument( '-f', '--File', type=str, help='Path to the file containing the data', required=True, ) parser.add_argument( '-i', '--Iterations', type=int, help='How many iterations the algorithm should perform', required=True, ) parser.add_argument( '-s', '--Size', type=int, help='Size of the tabu list', required=True ) # Pass the arguments to main method main(parser.parse_args())
44
0