input
stringlengths
33
5k
output
stringlengths
32
5k
import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast logger = datasets.utils.logging.get_logger(__name__) @dataclass class ParquetConfig(datasets.BuilderConfig): """BuilderConfig for Parquet.""" batch_size: int = 10_000 columns: Optional[List[str]] = None features: Optional[datasets.Features] = None class Parquet(datasets.ArrowBasedBuilder): BUILDER_CONFIG_CLASS = ParquetConfig def _info(self): return datasets.DatasetInfo(features=self.config.features) def _split_generators(self, dl_manager): """We handle string, list and dicts in datafiles""" if not self.config.data_files: raise ValueError(f"At least one data file must be specified, but got data_files={self.config.data_files}") data_files = dl_manager.download_and_extract(self.config.data_files) if isinstance(data_files, (str, list, tuple)): files = data_files if isinstance(files, str): files = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive files = [dl_manager.iter_files(file) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"files": files})] splits = [] for split_name, files in data_files.items(): if isinstance(files, str): files = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive files = [dl_manager.iter_files(file) for file in files] # Infer features if they are stored in the arrow schema if self.info.features is None: for file in itertools.chain.from_iterable(files): with open(file, "rb") as f: features = datasets.Features.from_arrow_schema(pq.read_schema(f)) if self.config.columns is not None: features = datasets.Features( {col: feat for col, feat in features.items() if col in self.config.columns} ) self.info.features = features break splits.append(datasets.SplitGenerator(name=split_name, gen_kwargs={"files": files})) return splits def _cast_table(self, pa_table: pa.Table) -> pa.Table: if self.info.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example pa_table = table_cast(pa_table, self.info.features.arrow_schema) return pa_table def _generate_tables(self, files): if self.config.features is not None and self.config.columns is not None: if sorted(field.name for field in self.info.features.arrow_schema) != sorted(self.config.columns): raise ValueError( f"Tried to load parquet data with columns '{self.config.columns}' with mismatching features '{self.info.features}'" ) for file_idx, file in enumerate(itertools.chain.from_iterable(files)): with open(file, "rb") as f: parquet_file = pq.ParquetFile(f) try: for batch_idx, record_batch in enumerate( parquet_file.iter_batches(batch_size=self.config.batch_size, columns=self.config.columns) ): pa_table = pa.Table.from_batches([record_batch]) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield f"{file_idx}_{batch_idx}", self._cast_table(pa_table) except ValueError as e: logger.error(f"Failed to read file '{file}' with error {type(e)}: {e}") raise
import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast logger = datasets.utils.logging.get_logger(__name__) @dataclass class ParquetConfig(datasets.BuilderConfig): """BuilderConfig for Parquet.""" batch_size: int = 10_000 columns: Optional[List[str]] = None features: Optional[datasets.Features] = None class Parquet(datasets.ArrowBasedBuilder): BUILDER_CONFIG_CLASS = ParquetConfig def _info(self): return datasets.DatasetInfo(features=self.config.features) def _split_generators(self, dl_manager): """We handle string, list and dicts in datafiles""" if not self.config.data_files: raise ValueError(f"At least one data file must be specified, but got data_files={self.config.data_files}") data_files = dl_manager.download_and_extract(self.config.data_files) if isinstance(data_files, (str, list, tuple)): files = data_files if isinstance(files, str): files = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive files = [dl_manager.iter_files(file) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"files": files})] splits = [] for split_name, files in data_files.items(): if isinstance(files, str): files = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive files = [dl_manager.iter_files(file) for file in files] # Infer features if they are stored in the arrow schema if self.info.features is None: for file in itertools.chain.from_iterable(files): with open(file, "rb") as f: features = datasets.Features.from_arrow_schema(pq.read_schema(f)) if self.config.columns is not None: self.info.features = datasets.Features( {col: feat for col, feat in features.items() if col in self.config.columns} ) self.info.features = features break splits.append(datasets.SplitGenerator(name=split_name, gen_kwargs={"files": files})) return splits def _cast_table(self, pa_table: pa.Table) -> pa.Table: if self.info.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example pa_table = table_cast(pa_table, self.info.features.arrow_schema) return pa_table def _generate_tables(self, files): if self.config.features is not None and self.config.columns is not None: if sorted(field.name for field in self.info.features.arrow_schema) != sorted(self.config.columns): raise ValueError( f"Tried to load parquet data with columns '{self.config.columns}' with mismatching features '{self.info.features}'" ) for file_idx, file in enumerate(itertools.chain.from_iterable(files)): with open(file, "rb") as f: parquet_file = pq.ParquetFile(f) try: for batch_idx, record_batch in enumerate( parquet_file.iter_batches(batch_size=self.config.batch_size, columns=self.config.columns) ): pa_table = pa.Table.from_batches([record_batch]) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield f"{file_idx}_{batch_idx}", self._cast_table(pa_table) except ValueError as e: logger.error(f"Failed to read file '{file}' with error {type(e)}: {e}") raise
import warnings from sys import platform from typing import Optional import torch import torchaudio from torchaudio.io import StreamWriter dict_format = { torch.uint8: "u8", torch.int16: "s16", torch.int32: "s32", torch.int64: "s64", torch.float32: "flt", torch.float64: "dbl", } @torchaudio._extension.fail_if_no_ffmpeg def play_audio( waveform: torch.Tensor, sample_rate: Optional[float], device: Optional[str] = None, ) -> None: """Plays audio through specified or available output device. This function is currently only supported on MacOS, which has access to "audiotoolbox" output device that can play up to two audio channels. Args: waveform: Tensor containing the audio to play. Expected shape: `(time, num_channels)`. sample_rate: Sample rate of the audio to play. device: Output device to use. If None, the default device is used. """ if platform == "darwin": device = device or "audiotoolbox" path = "-" else: raise ValueError(f"This function only supports MacOS, but current OS is {platform}") available_devices = list(torchaudio.utils.ffmpeg_utils.get_output_devices().keys()) if device not in available_devices: raise ValueError(f"Device {device} is not available. Available devices are: {available_devices}") if waveform.dtype not in dict_format: raise ValueError(f"Unsupported type {waveform.dtype}. The list of supported types is: {dict_format.keys()}") format = dict_format[waveform.dtype] if waveform.ndim != 2: raise ValueError(f"Expected 2D tensor with shape `(time, num_channels)`, got {waveform.ndim}D tensor instead") time, num_channels = waveform.size() if num_channels > 2: warnings.warn( f"Expected up to 2 channels, got {num_channels} channels instead. Only the first 2 channels will be played." ) # Write to speaker device s = StreamWriter(dst=path, format=device) s.add_audio_stream(sample_rate, num_channels, format=format) # write audio to the device block_size = 256 with s.open(): for i in range(0, time, block_size): s.write_audio_chunk(0, waveform[i : i + block_size, :])
import warnings from sys import platform from typing import Optional import torch import torchaudio from torchaudio.io import StreamWriter dict_format = { torch.uint8: "u8", torch.int16: "s16", torch.int32: "s32", torch.int64: "s64", torch.float32: "flt", torch.float64: "dbl", } def play_audio( waveform: torch.Tensor, sample_rate: Optional[float], device: Optional[str] = None, ) -> None: """Plays audio through specified or available output device. This function is currently only supported on MacOS, which has access to "audiotoolbox" output device that can play up to two audio channels. Args: waveform: Tensor containing the audio to play. Expected shape: `(time, num_channels)`. sample_rate: Sample rate of the audio to play. device: Output device to use. If None, the default device is used. """ if platform == "darwin": device = device or "audiotoolbox" path = "-" else: raise ValueError(f"This function only supports MacOS, but current OS is {platform}") available_devices = list(torchaudio.utils.ffmpeg_utils.get_output_devices().keys()) if device not in available_devices: raise ValueError(f"Device {device} is not available. Available devices are: {available_devices}") if waveform.dtype not in dict_format: raise ValueError(f"Unsupported type {waveform.dtype}. The list of supported types is: {dict_format.keys()}") format = dict_format[waveform.dtype] if waveform.ndim != 2: raise ValueError(f"Expected 2D tensor with shape `(time, num_channels)`, got {waveform.ndim}D tensor instead") time, num_channels = waveform.size() if num_channels > 2: warnings.warn( f"Expected up to 2 channels, got {num_channels} channels instead. Only the first 2 channels will be played." ) # Write to speaker device s = StreamWriter(dst=path, format=device) s.add_audio_stream(sample_rate, num_channels, format=format) # write audio to the device block_size = 256 with s.open(): for i in range(0, time, block_size): s.write_audio_chunk(0, waveform[i : i + block_size, :])
# Copyright (c) OpenMMLab. All rights reserved. import argparse import tempfile from collections import OrderedDict import torch from mmengine import Config def parse_config(config_strings): temp_file = tempfile.NamedTemporaryFile() config_path = f'{temp_file.name}.py' with open(config_path, 'w') as f: f.write(config_strings) config = Config.fromfile(config_path) # check whether it is SSD if config.model.bbox_head.type != 'SSDHead': raise AssertionError('This is not a SSD model.') def convert(in_file, out_file): checkpoint = torch.load(in_file) in_state_dict = checkpoint.pop('state_dict') out_state_dict = OrderedDict() meta_info = checkpoint['meta'] parse_config('#' + meta_info['config']) for key, value in in_state_dict.items(): if 'extra' in key: layer_idx = int(key.split('.')[2]) new_key = 'neck.extra_layers.{}.{}.conv.'.format( layer_idx // 2, layer_idx % 2) + key.split('.')[-1] elif 'l2_norm' in key: new_key = 'neck.l2_norm.weight' elif 'bbox_head' in key: new_key = key[:21] + '.0' + key[21:] else: new_key = key out_state_dict[new_key] = value checkpoint['state_dict'] = out_state_dict if torch.__version__ >= '1.6': torch.save(checkpoint, out_file, _use_new_zipfile_serialization=False) else: torch.save(checkpoint, out_file) def main(): parser = argparse.ArgumentParser(description='Upgrade SSD version') parser.add_argument('in_file', help='input checkpoint file') parser.add_argument('out_file', help='output checkpoint file') args = parser.parse_args() convert(args.in_file, args.out_file) if __name__ == '__main__': main()
# Copyright (c) OpenMMLab. All rights reserved. import argparse import tempfile from collections import OrderedDict import torch from mmcv import Config def parse_config(config_strings): temp_file = tempfile.NamedTemporaryFile() config_path = f'{temp_file.name}.py' with open(config_path, 'w') as f: f.write(config_strings) config = Config.fromfile(config_path) # check whether it is SSD if config.model.bbox_head.type != 'SSDHead': raise AssertionError('This is not a SSD model.') def convert(in_file, out_file): checkpoint = torch.load(in_file) in_state_dict = checkpoint.pop('state_dict') out_state_dict = OrderedDict() meta_info = checkpoint['meta'] parse_config('#' + meta_info['config']) for key, value in in_state_dict.items(): if 'extra' in key: layer_idx = int(key.split('.')[2]) new_key = 'neck.extra_layers.{}.{}.conv.'.format( layer_idx // 2, layer_idx % 2) + key.split('.')[-1] elif 'l2_norm' in key: new_key = 'neck.l2_norm.weight' elif 'bbox_head' in key: new_key = key[:21] + '.0' + key[21:] else: new_key = key out_state_dict[new_key] = value checkpoint['state_dict'] = out_state_dict if torch.__version__ >= '1.6': torch.save(checkpoint, out_file, _use_new_zipfile_serialization=False) else: torch.save(checkpoint, out_file) def main(): parser = argparse.ArgumentParser(description='Upgrade SSD version') parser.add_argument('in_file', help='input checkpoint file') parser.add_argument('out_file', help='output checkpoint file') args = parser.parse_args() convert(args.in_file, args.out_file) if __name__ == '__main__': main()
import pytest import torch import torchaudio class GreedyCTCDecoder(torch.nn.Module): def __init__(self, labels, blank: int = 0): super().__init__() self.blank = blank self.labels = labels def forward(self, logits: torch.Tensor) -> str: """Given a sequence logits over labels, get the best path string Args: logits (Tensor): Logit tensors. Shape `[num_seq, num_label]`. Returns: str: The resulting transcript """ best_path = torch.argmax(logits, dim=-1) # [num_seq,] best_path = torch.unique_consecutive(best_path, dim=-1) hypothesis = [] for i in best_path: if i != self.blank: hypothesis.append(self.labels[i]) return "".join(hypothesis) @pytest.fixture def ctc_decoder(): return GreedyCTCDecoder _FILES = { "en": "Lab41-SRI-VOiCES-src-sp0307-ch127535-sg0042.flac", "de": "20090505-0900-PLENARY-16-de_20090505-21_56_00_8.flac", "en2": "20120613-0900-PLENARY-8-en_20120613-13_46_50_3.flac", "es": "20130207-0900-PLENARY-7-es_20130207-13_02_05_5.flac", "fr": "20121212-0900-PLENARY-5-fr_20121212-11_37_04_10.flac", "it": "20170516-0900-PLENARY-16-it_20170516-18_56_31_1.flac", } @pytest.fixture def sample_speech(tmp_path, lang): if lang not in _FILES: raise NotImplementedError(f"Unexpected lang: {lang}") filename = _FILES[lang] path = tmp_path.parent / filename if not path.exists(): torchaudio.utils.download_asset(f"test-assets/{filename}", path=path) return path def pytest_addoption(parser): parser.addoption( "--use-tmp-hub-dir", action="store_true", help=( "When provided, tests will use temporary directory as Torch Hub directory. " "Downloaded models will be deleted after each test." ), ) @pytest.fixture(autouse=True) def temp_hub_dir(tmpdir, pytestconfig): if not pytestconfig.getoption("use_tmp_hub_dir"): yield else: org_dir = torch.hub.get_dir() torch.hub.set_dir(tmpdir) yield torch.hub.set_dir(org_dir) @pytest.fixture() def emissions(): path = torchaudio.utils.download_asset("test-assets/emissions-8555-28447-0012.pt") return torch.load(path)
import pytest import torch import torchaudio class GreedyCTCDecoder(torch.nn.Module): def __init__(self, labels, blank: int = 0): super().__init__() self.blank = blank self.labels = labels def forward(self, logits: torch.Tensor) -> str: """Given a sequence logits over labels, get the best path string Args: logits (Tensor): Logit tensors. Shape `[num_seq, num_label]`. Returns: str: The resulting transcript """ best_path = torch.argmax(logits, dim=-1) # [num_seq,] best_path = torch.unique_consecutive(best_path, dim=-1) hypothesis = [] for i in best_path: if i != self.blank: hypothesis.append(self.labels[i]) return "".join(hypothesis) @pytest.fixture def ctc_decoder(): return GreedyCTCDecoder _FILES = { "en": "Lab41-SRI-VOiCES-src-sp0307-ch127535-sg0042.flac", "de": "20090505-0900-PLENARY-16-de_20090505-21_56_00_8.flac", "en2": "20120613-0900-PLENARY-8-en_20120613-13_46_50_3.flac", "es": "20130207-0900-PLENARY-7-es_20130207-13_02_05_5.flac", "fr": "20121212-0900-PLENARY-5-fr_20121212-11_37_04_10.flac", "it": "20170516-0900-PLENARY-16-it_20170516-18_56_31_1.flac", } @pytest.fixture def sample_speech(tmp_path, lang): if lang not in _FILES: raise NotImplementedError(f"Unexpected lang: {lang}") filename = _FILES[lang] path = tmp_path.parent / filename if not path.exists(): torchaudio.utils.download_asset(f"test-assets/{filename}", path=path) return path def pytest_addoption(parser): parser.addoption( "--use-tmp-hub-dir", action="store_true", help=( "When provided, tests will use temporary directory as Torch Hub directory. " "Downloaded models will be deleted after each test." ), ) @pytest.fixture(autouse=True) def temp_hub_dir(tmpdir, pytestconfig): if not pytestconfig.getoption("use_tmp_hub_dir"): yield else: org_dir = torch.hub.get_dir() torch.hub.set_dir(tmpdir) yield torch.hub.set_dir(org_dir)
from dataclasses import dataclass from typing import List, Optional, Tuple import torch from torch import Tensor from torchaudio._extension import fail_if_no_align __all__ = [] @fail_if_no_align def forced_align( log_probs: Tensor, targets: Tensor, input_lengths: Optional[Tensor] = None, target_lengths: Optional[Tensor] = None, blank: int = 0, ) -> Tuple[Tensor, Tensor]: r"""Align a CTC label sequence to an emission. .. devices:: CPU CUDA .. properties:: TorchScript Args: log_probs (Tensor): log probability of CTC emission output. Tensor of shape `(B, T, C)`. where `B` is the batch size, `T` is the input length, `C` is the number of characters in alphabet including blank. targets (Tensor): Target sequence. Tensor of shape `(B, L)`, where `L` is the target length. input_lengths (Tensor or None, optional): Lengths of the inputs (max value must each be <= `T`). 1-D Tensor of shape `(B,)`. target_lengths (Tensor or None, optional): Lengths of the targets. 1-D Tensor of shape `(B,)`. blank_id (int, optional): The index of blank symbol in CTC emission. (Default: 0) Returns: Tuple(Tensor, Tensor): Tensor: Label for each time step in the alignment path computed using forced alignment. Tensor: Log probability scores of the labels for each time step. Note: The sequence length of `log_probs` must satisfy: .. math:: L_{\text{log\_probs}} \ge L_{\text{label}} + N_{\text{repeat}} where :math:`N_{\text{repeat}}` is the number of consecutively repeated tokens. For example, in str `"aabbc"`, the number of repeats are `2`. Note: The current version only supports ``batch_size==1``. """ if blank in targets: raise ValueError(f"targets Tensor shouldn't contain blank index. Found {targets}.") if torch.max(targets) >= log_probs.shape[-1]: raise ValueError("targets values must be less than the CTC dimension") if input_lengths is None: batch_size, length = log_probs.size(0), log_probs.size(1) input_lengths = torch.full((batch_size,), length, dtype=torch.int64, device=log_probs.device) if target_lengths is None: batch_size, length = targets.size(0), targets.size(1) target_lengths = torch.full((batch_size,), length, dtype=torch.int64, device=targets.device) # For TorchScript compatibility assert input_lengths is not None assert target_lengths is not None paths, scores = torch.ops.torchaudio.forced_align(log_probs, targets, input_lengths, target_lengths, blank) return paths, scores @dataclass class TokenSpan: """TokenSpan() Token with time stamps and score. Returned by :py:func:`merge_tokens`. """ token: int """The token""" start: int """The start time (inclusive) in emission time axis.""" end: int """The end time (exclusive) in emission time axis.""" score: float """The score of the this token.""" def __len__(self) -> int: """Returns the time span""" return self.end - self.start def merge_tokens(tokens: Tensor, scores: Tensor, blank: int = 0) -> List[TokenSpan]: """Removes repeated tokens and blank tokens from the given CTC token sequence. Args: tokens (Tensor): Alignment tokens (unbatched) returned from :py:func:`forced_align`. Shape: `(time, )`. scores (Tensor): Alignment scores (unbatched) returned from :py:func:`forced_align`. Shape: `(time, )`. When computing the token-size score, the given score is averaged across the corresponding time span. Returns: list of TokenSpan Example: >>> aligned_tokens, scores = forced_align(emission, targets, input_lengths, target_lengths) >>> token_spans = merge_tokens(aligned_tokens[0], scores[0]) """ if tokens.ndim != 1 or scores.ndim != 1: raise ValueError("`tokens` and `scores` must be 1D Tensor.") if len(tokens) != len(scores): raise ValueError("`tokens` and `scores` must be the same length.") diff = torch.diff( tokens, prepend=torch.tensor([-1], device=tokens.device), append=torch.tensor([-1], device=tokens.device) ) changes_wo_blank = torch.nonzero((diff != 0)).squeeze().tolist() tokens = tokens.tolist() spans = [ TokenSpan(token=token, start=start, end=end, score=scores[start:end].mean().item()) for start, end in zip(changes_wo_blank[:-1], changes_wo_blank[1:]) if (token := tokens[start]) != blank ] return spans
from dataclasses import dataclass from typing import List, Optional, Tuple import torch from torch import Tensor from torchaudio._extension import fail_if_no_align __all__ = [] @fail_if_no_align def forced_align( log_probs: Tensor, targets: Tensor, input_lengths: Optional[Tensor] = None, target_lengths: Optional[Tensor] = None, blank: int = 0, ) -> Tuple[Tensor, Tensor]: r"""Align a CTC label sequence to an emission. .. devices:: CPU CUDA .. properties:: TorchScript Args: log_probs (Tensor): log probability of CTC emission output. Tensor of shape `(B, T, C)`. where `B` is the batch size, `T` is the input length, `C` is the number of characters in alphabet including blank. targets (Tensor): Target sequence. Tensor of shape `(B, L)`, where `L` is the target length. input_lengths (Tensor or None, optional): Lengths of the inputs (max value must each be <= `T`). 1-D Tensor of shape `(B,)`. target_lengths (Tensor or None, optional): Lengths of the targets. 1-D Tensor of shape `(B,)`. blank_id (int, optional): The index of blank symbol in CTC emission. (Default: 0) Returns: Tuple(Tensor, Tensor): Tensor: Label for each time step in the alignment path computed using forced alignment. Tensor: Log probability scores of the labels for each time step. Note: The sequence length of `log_probs` must satisfy: .. math:: L_{\text{log\_probs}} \ge L_{\text{label}} + N_{\text{repeat}} where :math:`N_{\text{repeat}}` is the number of consecutively repeated tokens. For example, in str `"aabbc"`, the number of repeats are `2`. Note: The current version only supports ``batch_size==1``. """ if blank in targets: raise ValueError(f"targets Tensor shouldn't contain blank index. Found {targets}.") if torch.max(targets) >= log_probs.shape[-1]: raise ValueError("targets values must be less than the CTC dimension") if input_lengths is None: batch_size, length = log_probs.size(0), log_probs.size(1) input_lengths = torch.full((batch_size,), length, dtype=torch.int64, device=log_probs.device) if target_lengths is None: batch_size, length = targets.size(0), targets.size(1) target_lengths = torch.full((batch_size,), length, dtype=torch.int64, device=targets.device) # For TorchScript compatibility assert input_lengths is not None assert target_lengths is not None paths, scores = torch.ops.torchaudio.forced_align(log_probs, targets, input_lengths, target_lengths, blank) return paths, scores @dataclass class TokenSpan: """TokenSpan() Token with time stamps and score. Returned by :py:func:`merge_tokens`. """ token: int """The token""" start: int """The start time (inclusive) in emission time axis.""" end: int """The end time (exclusive) in emission time axis.""" score: float """The score of the this token.""" def __len__(self) -> int: """Returns the time span""" return self.end - self.start def merge_tokens(tokens: Tensor, scores: Tensor, blank: int = 0) -> List[TokenSpan]: """Removes repeated tokens and blank tokens from the given CTC token sequence. Args: tokens (Tensor): Alignment tokens (unbatched) returned from :py:func:`forced_align`. Shape: `(time, )`. scores (Tensor): Alignment scores (unbatched) returned from :py:func:`forced_align`. Shape: `(time, )`. When computing the token-size score, the given score is averaged across the corresponding time span. Returns: list of TokenSpan Example: >>> aligned_tokens, scores = forced_align(emission, targets, input_lengths, target_lengths) >>> token_spans = merge_tokens(aligned_tokens[0], scores[0]) """ if tokens.ndim != 1 or scores.ndim != 1: raise ValueError("`tokens` and `scores` must be 1D Tensor.") if len(tokens) != len(scores): raise ValueError("`tokens` and `scores` must be the same length.") t_prev = blank i = start = -1 spans = [] for t, token in enumerate(tokens): if token != t_prev: if t_prev != blank: spans.append(TokenSpan(t_prev.item(), start, t, scores[start:t].mean().item())) if token != blank: i += 1 start = t t_prev = token if t_prev != blank: spans.append(TokenSpan(t_prev.item(), start, len(tokens), scores[start:].mean().item())) return spans
from docarray.typing.url.any_url import AnyUrl from docarray.typing.url.image_url import ImageUrl __all__ = ['ImageUrl', 'AnyUrl']
from .image_url import ImageUrl __all__ = ['ImageUrl']
# Copyright (c) OpenMMLab. All rights reserved. import mmcv import mmengine from mmengine.utils import digit_version from .version import __version__, version_info mmcv_minimum_version = '2.0.0rc4' mmcv_maximum_version = '2.1.0' mmcv_version = digit_version(mmcv.__version__) mmengine_minimum_version = '0.7.1' mmengine_maximum_version = '1.0.0' mmengine_version = digit_version(mmengine.__version__) assert (mmcv_version >= digit_version(mmcv_minimum_version) and mmcv_version < digit_version(mmcv_maximum_version)), \ f'MMCV=={mmcv.__version__} is used but incompatible. ' \ f'Please install mmcv>={mmcv_minimum_version}, <{mmcv_maximum_version}.' assert (mmengine_version >= digit_version(mmengine_minimum_version) and mmengine_version < digit_version(mmengine_maximum_version)), \ f'MMEngine=={mmengine.__version__} is used but incompatible. ' \ f'Please install mmengine>={mmengine_minimum_version}, ' \ f'<{mmengine_maximum_version}.' __all__ = ['__version__', 'version_info', 'digit_version']
# Copyright (c) OpenMMLab. All rights reserved. import mmcv import mmengine from mmengine.utils import digit_version from .version import __version__, version_info mmcv_minimum_version = '2.0.0rc4' mmcv_maximum_version = '2.1.0' mmcv_version = digit_version(mmcv.__version__) mmengine_minimum_version = '0.6.0' mmengine_maximum_version = '1.0.0' mmengine_version = digit_version(mmengine.__version__) assert (mmcv_version >= digit_version(mmcv_minimum_version) and mmcv_version < digit_version(mmcv_maximum_version)), \ f'MMCV=={mmcv.__version__} is used but incompatible. ' \ f'Please install mmcv>={mmcv_minimum_version}, <{mmcv_maximum_version}.' assert (mmengine_version >= digit_version(mmengine_minimum_version) and mmengine_version < digit_version(mmengine_maximum_version)), \ f'MMEngine=={mmengine.__version__} is used but incompatible. ' \ f'Please install mmengine>={mmengine_minimum_version}, ' \ f'<{mmengine_maximum_version}.' __all__ = ['__version__', 'version_info', 'digit_version']
"""Map-reduce chain. Splits up a document, sends the smaller parts to the LLM with one prompt, then combines the results with another one. """ from __future__ import annotations from collections.abc import Mapping from typing import Any, Optional from langchain_core._api import deprecated from langchain_core.callbacks import CallbackManagerForChainRun, Callbacks from langchain_core.documents import Document from langchain_core.language_models import BaseLanguageModel from langchain_core.prompts import BasePromptTemplate from langchain_text_splitters import TextSplitter from pydantic import ConfigDict from langchain.chains import ReduceDocumentsChain from langchain.chains.base import Chain from langchain.chains.combine_documents.base import BaseCombineDocumentsChain from langchain.chains.combine_documents.map_reduce import MapReduceDocumentsChain from langchain.chains.combine_documents.stuff import StuffDocumentsChain from langchain.chains.llm import LLMChain @deprecated( since="0.2.13", removal="1.0", message=( "Refer to migration guide here for a recommended implementation using " "LangGraph: https://python.langchain.com/docs/versions/migrating_chains/map_reduce_chain/" ". See also LangGraph guides for map-reduce: " "https://langchain-ai.github.io/langgraph/how-tos/map-reduce/." ), ) class MapReduceChain(Chain): """Map-reduce chain.""" combine_documents_chain: BaseCombineDocumentsChain """Chain to use to combine documents.""" text_splitter: TextSplitter """Text splitter to use.""" input_key: str = "input_text" #: :meta private: output_key: str = "output_text" #: :meta private: @classmethod def from_params( cls, llm: BaseLanguageModel, prompt: BasePromptTemplate, text_splitter: TextSplitter, callbacks: Callbacks = None, combine_chain_kwargs: Optional[Mapping[str, Any]] = None, reduce_chain_kwargs: Optional[Mapping[str, Any]] = None, **kwargs: Any, ) -> MapReduceChain: """Construct a map-reduce chain that uses the chain for map and reduce.""" llm_chain = LLMChain(llm=llm, prompt=prompt, callbacks=callbacks) stuff_chain = StuffDocumentsChain( llm_chain=llm_chain, callbacks=callbacks, **(reduce_chain_kwargs if reduce_chain_kwargs else {}), ) reduce_documents_chain = ReduceDocumentsChain( combine_documents_chain=stuff_chain ) combine_documents_chain = MapReduceDocumentsChain( llm_chain=llm_chain, reduce_documents_chain=reduce_documents_chain, callbacks=callbacks, **(combine_chain_kwargs if combine_chain_kwargs else {}), ) return cls( combine_documents_chain=combine_documents_chain, text_splitter=text_splitter, callbacks=callbacks, **kwargs, ) model_config = ConfigDict( arbitrary_types_allowed=True, extra="forbid", ) @property def input_keys(self) -> list[str]: """Expect input key. :meta private: """ return [self.input_key] @property def output_keys(self) -> list[str]: """Return output key. :meta private: """ return [self.output_key] def _call( self, inputs: dict[str, str], run_manager: Optional[CallbackManagerForChainRun] = None, ) -> dict[str, str]: _run_manager = run_manager or CallbackManagerForChainRun.get_noop_manager() # Split the larger text into smaller chunks. doc_text = inputs.pop(self.input_key) texts = self.text_splitter.split_text(doc_text) docs = [Document(page_content=text) for text in texts] _inputs: dict[str, Any] = { **inputs, self.combine_documents_chain.input_key: docs, } outputs = self.combine_documents_chain.run( _inputs, callbacks=_run_manager.get_child() ) return {self.output_key: outputs}
"""Map-reduce chain. Splits up a document, sends the smaller parts to the LLM with one prompt, then combines the results with another one. """ from __future__ import annotations from collections.abc import Mapping from typing import Any, Optional from langchain_core._api import deprecated from langchain_core.callbacks import CallbackManagerForChainRun, Callbacks from langchain_core.documents import Document from langchain_core.language_models import BaseLanguageModel from langchain_core.prompts import BasePromptTemplate from langchain_text_splitters import TextSplitter from pydantic import ConfigDict from langchain.chains import ReduceDocumentsChain from langchain.chains.base import Chain from langchain.chains.combine_documents.base import BaseCombineDocumentsChain from langchain.chains.combine_documents.map_reduce import MapReduceDocumentsChain from langchain.chains.combine_documents.stuff import StuffDocumentsChain from langchain.chains.llm import LLMChain @deprecated( since="0.2.13", removal="1.0", message=( "Refer to migration guide here for a recommended implementation using " "LangGraph: https://python.langchain.com/docs/versions/migrating_chains/map_reduce_chain/" # noqa: E501 ". See also LangGraph guides for map-reduce: " "https://langchain-ai.github.io/langgraph/how-tos/map-reduce/." ), ) class MapReduceChain(Chain): """Map-reduce chain.""" combine_documents_chain: BaseCombineDocumentsChain """Chain to use to combine documents.""" text_splitter: TextSplitter """Text splitter to use.""" input_key: str = "input_text" #: :meta private: output_key: str = "output_text" #: :meta private: @classmethod def from_params( cls, llm: BaseLanguageModel, prompt: BasePromptTemplate, text_splitter: TextSplitter, callbacks: Callbacks = None, combine_chain_kwargs: Optional[Mapping[str, Any]] = None, reduce_chain_kwargs: Optional[Mapping[str, Any]] = None, **kwargs: Any, ) -> MapReduceChain: """Construct a map-reduce chain that uses the chain for map and reduce.""" llm_chain = LLMChain(llm=llm, prompt=prompt, callbacks=callbacks) stuff_chain = StuffDocumentsChain( llm_chain=llm_chain, callbacks=callbacks, **(reduce_chain_kwargs if reduce_chain_kwargs else {}), ) reduce_documents_chain = ReduceDocumentsChain( combine_documents_chain=stuff_chain ) combine_documents_chain = MapReduceDocumentsChain( llm_chain=llm_chain, reduce_documents_chain=reduce_documents_chain, callbacks=callbacks, **(combine_chain_kwargs if combine_chain_kwargs else {}), ) return cls( combine_documents_chain=combine_documents_chain, text_splitter=text_splitter, callbacks=callbacks, **kwargs, ) model_config = ConfigDict( arbitrary_types_allowed=True, extra="forbid", ) @property def input_keys(self) -> list[str]: """Expect input key. :meta private: """ return [self.input_key] @property def output_keys(self) -> list[str]: """Return output key. :meta private: """ return [self.output_key] def _call( self, inputs: dict[str, str], run_manager: Optional[CallbackManagerForChainRun] = None, ) -> dict[str, str]: _run_manager = run_manager or CallbackManagerForChainRun.get_noop_manager() # Split the larger text into smaller chunks. doc_text = inputs.pop(self.input_key) texts = self.text_splitter.split_text(doc_text) docs = [Document(page_content=text) for text in texts] _inputs: dict[str, Any] = { **inputs, self.combine_documents_chain.input_key: docs, } outputs = self.combine_documents_chain.run( _inputs, callbacks=_run_manager.get_child() ) return {self.output_key: outputs}
from torchaudio._internal.module_utils import dropping_support from ._alignment import forced_align as _forced_align, merge_tokens, TokenSpan from .filtering import ( allpass_biquad, band_biquad, bandpass_biquad, bandreject_biquad, bass_biquad, biquad, contrast, dcshift, deemph_biquad, dither, equalizer_biquad, filtfilt, flanger, gain, highpass_biquad, lfilter, lowpass_biquad, overdrive, phaser, riaa_biquad, treble_biquad, vad, ) forced_align = dropping_support(_forced_align) from .functional import ( add_noise, amplitude_to_DB, apply_beamforming, apply_codec, compute_deltas, convolve, create_dct, DB_to_amplitude, deemphasis, detect_pitch_frequency, edit_distance, fftconvolve, frechet_distance, griffinlim, inverse_spectrogram, linear_fbanks, loudness, mask_along_axis, mask_along_axis_iid, melscale_fbanks, mu_law_decoding, mu_law_encoding, mvdr_weights_rtf, mvdr_weights_souden, phase_vocoder, pitch_shift, preemphasis, psd, resample, rnnt_loss, rtf_evd, rtf_power, sliding_window_cmn, spectral_centroid, spectrogram, speed, ) __all__ = [ "amplitude_to_DB", "compute_deltas", "create_dct", "melscale_fbanks", "linear_fbanks", "DB_to_amplitude", "loudness", "detect_pitch_frequency", "griffinlim", "mask_along_axis", "mask_along_axis_iid", "mu_law_encoding", "mu_law_decoding", "phase_vocoder", "sliding_window_cmn", "spectrogram", "inverse_spectrogram", "spectral_centroid", "allpass_biquad", "band_biquad", "bandpass_biquad", "bandreject_biquad", "bass_biquad", "biquad", "contrast", "dither", "dcshift", "deemph_biquad", "equalizer_biquad", "filtfilt", "flanger", "forced_align", "merge_tokens", "TokenSpan", "gain", "highpass_biquad", "lfilter", "lowpass_biquad", "overdrive", "phaser", "riaa_biquad", "treble_biquad", "vad", "apply_codec", "resample", "edit_distance", "pitch_shift", "rnnt_loss", "psd", "mvdr_weights_souden", "mvdr_weights_rtf", "rtf_evd", "rtf_power", "apply_beamforming", "fftconvolve", "convolve", "add_noise", "speed", "preemphasis", "deemphasis", "frechet_distance", ]
from torchaudio._internal.module_utils import dropping_support from ._alignment import forced_align as _forced_align, merge_tokens, TokenSpan from .filtering import ( allpass_biquad, band_biquad, bandpass_biquad, bandreject_biquad, bass_biquad, biquad, contrast, dcshift, deemph_biquad, dither, equalizer_biquad, filtfilt, flanger, gain, highpass_biquad, lfilter, lowpass_biquad, overdrive, phaser, riaa_biquad, treble_biquad, vad, ) forced_align = dropping_support(_forced_align) from .functional import ( add_noise, amplitude_to_DB, apply_beamforming, apply_codec, compute_deltas, convolve, create_dct, DB_to_amplitude, deemphasis, detect_pitch_frequency, edit_distance, fftconvolve, frechet_distance, griffinlim, inverse_spectrogram, linear_fbanks, loudness, mask_along_axis, mask_along_axis_iid, melscale_fbanks, mu_law_decoding, mu_law_encoding, mvdr_weights_rtf, mvdr_weights_souden, phase_vocoder, pitch_shift, preemphasis, psd, resample, rnnt_loss as _rnnt_loss, rtf_evd, rtf_power, sliding_window_cmn, spectral_centroid, spectrogram, speed, ) rnnt_loss = dropping_support(_rnnt_loss) __all__ = [ "amplitude_to_DB", "compute_deltas", "create_dct", "melscale_fbanks", "linear_fbanks", "DB_to_amplitude", "loudness", "detect_pitch_frequency", "griffinlim", "mask_along_axis", "mask_along_axis_iid", "mu_law_encoding", "mu_law_decoding", "phase_vocoder", "sliding_window_cmn", "spectrogram", "inverse_spectrogram", "spectral_centroid", "allpass_biquad", "band_biquad", "bandpass_biquad", "bandreject_biquad", "bass_biquad", "biquad", "contrast", "dither", "dcshift", "deemph_biquad", "equalizer_biquad", "filtfilt", "flanger", "forced_align", "merge_tokens", "TokenSpan", "gain", "highpass_biquad", "lfilter", "lowpass_biquad", "overdrive", "phaser", "riaa_biquad", "treble_biquad", "vad", "apply_codec", "resample", "edit_distance", "pitch_shift", "rnnt_loss", "psd", "mvdr_weights_souden", "mvdr_weights_rtf", "rtf_evd", "rtf_power", "apply_beamforming", "fftconvolve", "convolve", "add_noise", "speed", "preemphasis", "deemphasis", "frechet_distance", ]
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from typing import Optional, List, Dict import hnswlib import numpy as np from jina import Executor, requests, DocumentArray, Document from jina_commons import get_logger from jina_commons.indexers.dump import import_vectors class HnswlibSearcher(Executor): """Hnswlib powered vector indexer For more information about the Hnswlib supported parameters, please consult: - https://github.com/nmslib/hnswlib .. note:: Hnswlib package dependency is only required at the query time. """ def __init__( self, default_top_k: int = 10, metric: str = 'cosine', dump_path: Optional[str] = None, default_traversal_paths: Optional[List[str]] = None, is_distance: bool = False, ef_construction: int = 400, ef_query: int = 50, max_connection: int = 64, *args, **kwargs, ): """ Initialize an HnswlibSearcher :param default_top_k: get tok k vectors :param distance: distance can be 'l2', 'ip', or 'cosine' :param dump_path: the path to load ids and vecs :param traverse_path: traverse path on docs, e.g. ['r'], ['c'] :param reverse_score: True if add reversed distance as the `similarity` for match score, else return `distance` as score for match score. :param ef_construction: defines a construction time/accuracy trade-off :param ef_query: sets the query time accuracy/speed trade-off :param max_connection: defines tha maximum number of outgoing connections in the graph :param args: :param kwargs: """ super().__init__(*args, **kwargs) self.default_top_k = default_top_k self.metric = metric self.default_traversal_paths = default_traversal_paths or ['r'] self.is_distance = is_distance self.ef_construction = ef_construction self.ef_query = ef_query self.max_connection = max_connection self.logger = get_logger(self) dump_path = dump_path or kwargs.get('runtime_args', {}).get('dump_path', None) if dump_path is not None: self.logger.info('Start building "HnswlibSearcher" from dump data') ids, vecs = import_vectors(dump_path, str(self.runtime_args.pea_id)) self._ids = np.array(list(ids)) self._vecs = np.array(list(vecs)) num_dim = self._vecs.shape[1] self._indexer = hnswlib.Index(space=self.metric, dim=num_dim) self._indexer.init_index(max_elements=len(self._vecs), ef_construction=self.ef_construction, M=self.max_connection) self._doc_id_to_offset = {} self._load_index(self._ids, self._vecs) else: self.logger.warning( 'No data loaded in "HnswlibSearcher". Use .rolling_update() to re-initialize it...' ) def _load_index(self, ids, vecs): for idx, v in enumerate(vecs): self._indexer.add_items(v.astype(np.float32), idx) self._doc_id_to_offset[ids[idx]] = idx self._indexer.set_ef(self.ef_query) @requests(on='/search') def search(self, docs: Optional[DocumentArray], parameters: Dict, **kwargs): if docs is None: return if not hasattr(self, '_indexer'): self.logger.warning('Querying against an empty index') return top_k = parameters.get('top_k', self.default_top_k) traversal_paths = parameters.get( 'traversal_paths', self.default_traversal_paths ) for doc in docs.traverse_flat(traversal_paths): indices, dists = self._indexer.knn_query(doc.embedding, k=top_k) for idx, dist in zip(indices[0], dists[0]): match = Document(id=self._ids[idx], embedding=self._vecs[idx]) if self.is_distance: match.scores[self.metric] = dist else: if self.metric == 'cosine' or self.metric == 'ip': match.scores[self.metric] = 1 - dist else: match.scores[self.metric] = 1 / (1 + dist) doc.matches.append(match) @requests(on='/fill_embedding') def fill_embedding(self, docs: Optional[DocumentArray], **kwargs): if docs is None: return for doc in docs: doc_idx = self._doc_id_to_offset.get(doc.id) if doc_idx is not None: doc.embedding = np.array( self._indexer.get_items([int(doc_idx)])[0] ) else: self.logger.warning(f'Document {doc.id} not found in index')
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from typing import Optional, List, Dict import hnswlib import numpy as np from jina import Executor, requests, DocumentArray, Document from jina_commons import get_logger from jina_commons.indexers.dump import import_vectors class HnswlibSearcher(Executor): """Hnswlib powered vector indexer For more information about the Hnswlib supported parameters, please consult: - https://github.com/nmslib/hnswlib .. note:: Hnswlib package dependency is only required at the query time. """ def __init__( self, default_top_k: int = 10, metric: str = 'cosine', dump_path: Optional[str] = None, default_traversal_paths: Optional[List[str]] = None, is_distance: bool = False, ef_construction: int = 400, ef_query: int = 50, max_connection: int = 64, *args, **kwargs, ): """ Initialize an HnswlibSearcher :param default_top_k: get tok k vectors :param distance: distance can be 'l2', 'ip', or 'cosine' :param dump_path: the path to load ids and vecs :param traverse_path: traverse path on docs, e.g. ['r'], ['c'] :param reverse_score: True if add reversed distance as the `similarity` for match score, else return `distance` as score for match score. :param ef_construction: defines a construction time/accuracy trade-off :param ef_query: sets the query time accuracy/speed trade-off :param max_connection: defines tha maximum number of outgoing connections in the graph :param args: :param kwargs: """ super().__init__(*args, **kwargs) self.default_top_k = default_top_k self.metric = metric self.default_traversal_paths = default_traversal_paths or ['r'] self.is_distance = is_distance self.ef_construction = ef_construction self.ef_query = ef_query self.max_connection = max_connection self.logger = get_logger(self) dump_path = dump_path or kwargs.get('runtime_args', {}).get('dump_path', None) if dump_path is not None: self.logger.info('Start building "HnswlibSearcher" from dump data') ids, vecs = import_vectors(dump_path, str(self.runtime_args.pea_id)) self._ids = np.array(list(ids)) self._vecs = np.array(list(vecs)) num_dim = self._vecs.shape[1] self._indexer = hnswlib.Index(space=self.metric, dim=num_dim) self._indexer.init_index(max_elements=len(self._vecs), ef_construction=self.ef_construction, M=self.max_connection) self._doc_id_to_offset = {} self._load_index(self._ids, self._vecs) else: self.logger.warning( 'No data loaded in "HnswlibSearcher". Use .rolling_update() to re-initialize it...' ) def _load_index(self, ids, vecs): for idx, v in enumerate(vecs): self._indexer.add_items(v.astype(np.float32), idx) self._doc_id_to_offset[ids[idx]] = idx self._indexer.set_ef(self.ef_query) @requests(on='/search') def search(self, docs: Optional[DocumentArray], parameters: Dict, **kwargs): if docs is None: return if not hasattr(self, '_indexer'): self.logger.warning('Querying against an empty index') return top_k = parameters.get('top_k', self.default_top_k) traversal_paths = parameters.get( 'traversal_paths', self.default_traversal_paths ) for doc in docs.traverse_flat(traversal_paths): indices, dists = self._indexer.knn_query(doc.embedding, k=top_k) for idx, dist in zip(indices[0], dists[0]): match = Document(id=self._ids[idx], embedding=self._vecs[idx]) if self.is_distance: match.scores[self.metric] = dist else: if self.metric == 'cosine' or self.metric == 'ip': match.scores[self.metric] = 1 - dist else: match.scores[self.metric] = 1 / (1 + dist) doc.matches.append(match) @requests(on='/fill_embedding') def fill_embedding(self, docs: Optional[DocumentArray], **kwargs): if docs is None: return for doc in docs: doc.embedding = np.array( self._indexer.get_items([int(self._doc_id_to_offset[str(doc.id)])])[0] )
"""Init file of LlamaIndex.""" __version__ = "0.12.31" import logging from logging import NullHandler from typing import Callable, Optional try: # Force pants to install eval_type_backport on 3.9 import eval_type_backport # noqa # type: ignore except ImportError: pass # response from llama_index.core.base.response.schema import Response # import global eval handler from llama_index.core.callbacks.global_handlers import set_global_handler from llama_index.core.data_structs.struct_type import IndexStructType from llama_index.core.embeddings.mock_embed_model import MockEmbedding # indices # loading from llama_index.core.indices import ( ComposableGraph, DocumentSummaryIndex, GPTDocumentSummaryIndex, GPTKeywordTableIndex, GPTListIndex, GPTRAKEKeywordTableIndex, GPTSimpleKeywordTableIndex, GPTTreeIndex, GPTVectorStoreIndex, KeywordTableIndex, KnowledgeGraphIndex, ListIndex, PropertyGraphIndex, RAKEKeywordTableIndex, SimpleKeywordTableIndex, SummaryIndex, TreeIndex, VectorStoreIndex, load_graph_from_storage, load_index_from_storage, load_indices_from_storage, ) # structured from llama_index.core.indices.common.struct_store.base import ( SQLDocumentContextBuilder, ) # prompt helper from llama_index.core.indices.prompt_helper import PromptHelper # prompts from llama_index.core.prompts import ( BasePromptTemplate, ChatPromptTemplate, # backwards compatibility Prompt, PromptTemplate, SelectorPromptTemplate, ) from llama_index.core.readers import SimpleDirectoryReader, download_loader # Response Synthesizer from llama_index.core.response_synthesizers.factory import get_response_synthesizer from llama_index.core.schema import Document, QueryBundle from llama_index.core.service_context import ( ServiceContext, set_global_service_context, ) # global settings from llama_index.core.settings import Settings # storage from llama_index.core.storage.storage_context import StorageContext # sql wrapper from llama_index.core.utilities.sql_wrapper import SQLDatabase # global tokenizer from llama_index.core.utils import get_tokenizer, set_global_tokenizer # best practices for library logging: # https://docs.python.org/3/howto/logging.html#configuring-logging-for-a-library logging.getLogger(__name__).addHandler(NullHandler()) __all__ = [ "StorageContext", "ServiceContext", "ComposableGraph", # indices "SummaryIndex", "VectorStoreIndex", "SimpleKeywordTableIndex", "KeywordTableIndex", "RAKEKeywordTableIndex", "TreeIndex", "DocumentSummaryIndex", "KnowledgeGraphIndex", "PropertyGraphIndex", # indices - legacy names "GPTKeywordTableIndex", "GPTKnowledgeGraphIndex", "GPTSimpleKeywordTableIndex", "GPTRAKEKeywordTableIndex", "GPTListIndex", "ListIndex", "GPTTreeIndex", "GPTVectorStoreIndex", "GPTDocumentSummaryIndex", "Prompt", "PromptTemplate", "BasePromptTemplate", "ChatPromptTemplate", "SelectorPromptTemplate", "SummaryPrompt", "TreeInsertPrompt", "TreeSelectPrompt", "TreeSelectMultiplePrompt", "RefinePrompt", "QuestionAnswerPrompt", "KeywordExtractPrompt", "QueryKeywordExtractPrompt", "Response", "Document", "SimpleDirectoryReader", "MockEmbedding", "SQLDatabase", "SQLDocumentContextBuilder", "SQLContextBuilder", "PromptHelper", "IndexStructType", "download_loader", "load_graph_from_storage", "load_index_from_storage", "load_indices_from_storage", "QueryBundle", "get_response_synthesizer", "set_global_service_context", "set_global_handler", "set_global_tokenizer", "get_tokenizer", "Settings", ] # eval global toggle from llama_index.core.callbacks.base_handler import BaseCallbackHandler global_handler: Optional[BaseCallbackHandler] = None # NOTE: keep for backwards compatibility SQLContextBuilder = SQLDocumentContextBuilder # global tokenizer global_tokenizer: Optional[Callable[[str], list]] = None
"""Init file of LlamaIndex.""" __version__ = "0.12.30" import logging from logging import NullHandler from typing import Callable, Optional try: # Force pants to install eval_type_backport on 3.9 import eval_type_backport # noqa # type: ignore except ImportError: pass # response from llama_index.core.base.response.schema import Response # import global eval handler from llama_index.core.callbacks.global_handlers import set_global_handler from llama_index.core.data_structs.struct_type import IndexStructType from llama_index.core.embeddings.mock_embed_model import MockEmbedding # indices # loading from llama_index.core.indices import ( ComposableGraph, DocumentSummaryIndex, GPTDocumentSummaryIndex, GPTKeywordTableIndex, GPTListIndex, GPTRAKEKeywordTableIndex, GPTSimpleKeywordTableIndex, GPTTreeIndex, GPTVectorStoreIndex, KeywordTableIndex, KnowledgeGraphIndex, ListIndex, PropertyGraphIndex, RAKEKeywordTableIndex, SimpleKeywordTableIndex, SummaryIndex, TreeIndex, VectorStoreIndex, load_graph_from_storage, load_index_from_storage, load_indices_from_storage, ) # structured from llama_index.core.indices.common.struct_store.base import ( SQLDocumentContextBuilder, ) # prompt helper from llama_index.core.indices.prompt_helper import PromptHelper # prompts from llama_index.core.prompts import ( BasePromptTemplate, ChatPromptTemplate, # backwards compatibility Prompt, PromptTemplate, SelectorPromptTemplate, ) from llama_index.core.readers import SimpleDirectoryReader, download_loader # Response Synthesizer from llama_index.core.response_synthesizers.factory import get_response_synthesizer from llama_index.core.schema import Document, QueryBundle from llama_index.core.service_context import ( ServiceContext, set_global_service_context, ) # global settings from llama_index.core.settings import Settings # storage from llama_index.core.storage.storage_context import StorageContext # sql wrapper from llama_index.core.utilities.sql_wrapper import SQLDatabase # global tokenizer from llama_index.core.utils import get_tokenizer, set_global_tokenizer # best practices for library logging: # https://docs.python.org/3/howto/logging.html#configuring-logging-for-a-library logging.getLogger(__name__).addHandler(NullHandler()) __all__ = [ "StorageContext", "ServiceContext", "ComposableGraph", # indices "SummaryIndex", "VectorStoreIndex", "SimpleKeywordTableIndex", "KeywordTableIndex", "RAKEKeywordTableIndex", "TreeIndex", "DocumentSummaryIndex", "KnowledgeGraphIndex", "PropertyGraphIndex", # indices - legacy names "GPTKeywordTableIndex", "GPTKnowledgeGraphIndex", "GPTSimpleKeywordTableIndex", "GPTRAKEKeywordTableIndex", "GPTListIndex", "ListIndex", "GPTTreeIndex", "GPTVectorStoreIndex", "GPTDocumentSummaryIndex", "Prompt", "PromptTemplate", "BasePromptTemplate", "ChatPromptTemplate", "SelectorPromptTemplate", "SummaryPrompt", "TreeInsertPrompt", "TreeSelectPrompt", "TreeSelectMultiplePrompt", "RefinePrompt", "QuestionAnswerPrompt", "KeywordExtractPrompt", "QueryKeywordExtractPrompt", "Response", "Document", "SimpleDirectoryReader", "MockEmbedding", "SQLDatabase", "SQLDocumentContextBuilder", "SQLContextBuilder", "PromptHelper", "IndexStructType", "download_loader", "load_graph_from_storage", "load_index_from_storage", "load_indices_from_storage", "QueryBundle", "get_response_synthesizer", "set_global_service_context", "set_global_handler", "set_global_tokenizer", "get_tokenizer", "Settings", ] # eval global toggle from llama_index.core.callbacks.base_handler import BaseCallbackHandler global_handler: Optional[BaseCallbackHandler] = None # NOTE: keep for backwards compatibility SQLContextBuilder = SQLDocumentContextBuilder # global tokenizer global_tokenizer: Optional[Callable[[str], list]] = None
# flake8: noqa """Test SQL database wrapper with schema support. Using DuckDB as SQLite does not support schemas. """ import pytest from sqlalchemy import ( Column, Integer, MetaData, Sequence, String, Table, create_engine, event, insert, schema, ) import sqlalchemy as sa from packaging import version from langchain_community.utilities.sql_database import SQLDatabase metadata_obj = MetaData() event.listen(metadata_obj, "before_create", schema.CreateSchema("schema_a")) event.listen(metadata_obj, "before_create", schema.CreateSchema("schema_b")) user = Table( "user", metadata_obj, Column("user_id", Integer, Sequence("user_id_seq"), primary_key=True), Column("user_name", String, nullable=False), schema="schema_a", ) company = Table( "company", metadata_obj, Column("company_id", Integer, Sequence("company_id_seq"), primary_key=True), Column("company_location", String, nullable=False), schema="schema_b", ) @pytest.mark.xfail( version.parse(sa.__version__).major == 1, reason="SQLAlchemy 1.x issues" ) def test_table_info() -> None: """Test that table info is constructed properly.""" engine = create_engine("duckdb:///:memory:") metadata_obj.create_all(engine) db = SQLDatabase(engine, schema="schema_a", metadata=metadata_obj) output = db.table_info expected_output = """ CREATE TABLE schema_a."user" ( user_id INTEGER NOT NULL, user_name VARCHAR NOT NULL, PRIMARY KEY (user_id) ) /* 3 rows from user table: user_id user_name */ """ assert sorted(" ".join(output.split())) == sorted(" ".join(expected_output.split())) @pytest.mark.xfail( version.parse(sa.__version__).major == 1, reason="SQLAlchemy 1.x issues" ) def test_sql_database_run() -> None: """Test that commands can be run successfully and returned in correct format.""" engine = create_engine("duckdb:///:memory:") metadata_obj.create_all(engine) stmt = insert(user).values(user_id=13, user_name="Harrison") with engine.begin() as conn: conn.execute(stmt) with pytest.warns(Warning) as records: db = SQLDatabase(engine, schema="schema_a") # Metadata creation with duckdb raises 3 warnings at the moment about reflection. # As a stop-gap to increase strictness of pytest to fail on warnings, we'll # explicitly catch the warnings and assert that it's the one we expect. # We may need to revisit at a later stage and determine why a warning is being # raised here. for record in records: assert isinstance(record.message, Warning) assert any( record.message.args[0] # type: ignore[union-attr] == "duckdb-engine doesn't yet support reflection on indices" for record in records ) command = 'select user_name from "user" where user_id = 13' output = db.run(command) expected_output = "[('Harrison',)]" assert output == expected_output
# flake8: noqa """Test SQL database wrapper with schema support. Using DuckDB as SQLite does not support schemas. """ import pytest from sqlalchemy import ( Column, Integer, MetaData, Sequence, String, Table, create_engine, event, insert, schema, ) import sqlalchemy as sa from packaging import version from langchain_community.utilities.sql_database import SQLDatabase metadata_obj = MetaData() event.listen(metadata_obj, "before_create", schema.CreateSchema("schema_a")) event.listen(metadata_obj, "before_create", schema.CreateSchema("schema_b")) user = Table( "user", metadata_obj, Column("user_id", Integer, Sequence("user_id_seq"), primary_key=True), Column("user_name", String, nullable=False), schema="schema_a", ) company = Table( "company", metadata_obj, Column("company_id", Integer, Sequence("company_id_seq"), primary_key=True), Column("company_location", String, nullable=False), schema="schema_b", ) @pytest.mark.xfail( version.parse(sa.__version__).major == 1, reason="SQLAlchemy 1.x issues" ) def test_table_info() -> None: """Test that table info is constructed properly.""" engine = create_engine("duckdb:///:memory:") metadata_obj.create_all(engine) db = SQLDatabase(engine, schema="schema_a", metadata=metadata_obj) output = db.table_info expected_output = """ CREATE TABLE schema_a."user" ( user_id INTEGER NOT NULL, user_name VARCHAR NOT NULL, PRIMARY KEY (user_id) ) /* 3 rows from user table: user_id user_name */ """ assert sorted(" ".join(output.split())) == sorted(" ".join(expected_output.split())) @pytest.mark.xfail( version.parse(sa.__version__).major == 1, reason="SQLAlchemy 1.x issues" ) def test_sql_database_run() -> None: """Test that commands can be run successfully and returned in correct format.""" engine = create_engine("duckdb:///:memory:") metadata_obj.create_all(engine) stmt = insert(user).values(user_id=13, user_name="Harrison") with engine.begin() as conn: conn.execute(stmt) with pytest.warns(Warning) as records: db = SQLDatabase(engine, schema="schema_a") # Metadata creation with duckdb raises 3 warnings at the moment about reflection. # As a stop-gap to increase strictness of pytest to fail on warnings, we'll # explicitly catch the warnings and assert that it's the one we expect. # We may need to revisit at a later stage and determine why a warning is being # raised here. for record in records: assert isinstance(record.message, Warning) assert any( record.message.args[0] # type: ignore == "duckdb-engine doesn't yet support reflection on indices" for record in records ) command = 'select user_name from "user" where user_id = 13' output = db.run(command) expected_output = "[('Harrison',)]" assert output == expected_output
from llama_index.core.base.llms.types import ( LLMMetadata, ) from llama_index.core.bridge.pydantic import Field from llama_index.llms.openai_like.base import OpenAILike class OPEA(OpenAILike): """ Adapter for a OPEA LLM. Examples: `pip install llama-index-llms-opea` ```python from llama_index.llms.opea import OPEA llm = OPEA( model="meta-llama/Meta-Llama-3.1-8B-Instruct", api_base="http://localhost:8080/v1", ) ``` """ is_chat_model: bool = Field( default=True, description=LLMMetadata.model_fields["is_chat_model"].description, ) @classmethod def class_name(cls) -> str: return "OPEA"
from llama_index.core.base.llms.types import ( LLMMetadata, ) from llama_index.core.bridge.pydantic import Field from llama_index.llms.openai_like.base import OpenAILike class OPEA(OpenAILike): """Adapter for a OPEA LLM. Examples: `pip install llama-index-llms-opea` ```python from llama_index.llms.opea import OPEA llm = OPEA( model="meta-llama/Meta-Llama-3.1-8B-Instruct", api_base="http://localhost:8080/v1", ) ``` """ is_chat_model: bool = Field( default=True, description=LLMMetadata.model_fields["is_chat_model"].description, ) @classmethod def class_name(cls) -> str: return "OPEA"
import numpy as np import pytest from keras.src import layers from keras.src import models from keras.src import testing class MaskingTest(testing.TestCase): @pytest.mark.requires_trainable_backend def test_masking_basics(self): self.run_layer_test( layers.Masking, init_kwargs={"mask_value": 0.0}, input_shape=(2, 3, 2), expected_output_shape=(2, 3, 2), expected_num_trainable_weights=0, expected_num_non_trainable_weights=0, expected_num_seed_generators=0, expected_num_losses=0, supports_masking=True, assert_built_after_instantiation=True, ) @pytest.mark.requires_trainable_backend def test_masking_correctness(self): x = np.array( [ [[0.0, 0.0], [1.0, 2.0], [0.0, 0.0]], [[2.0, 2.0], [0.0, 0.0], [2.0, 1.0]], ] ) expected_mask = [[False, True, False], [True, False, True]] layer = layers.Masking(mask_value=0.0) self.assertAllClose(layer.compute_mask(x), expected_mask) test_obj = self class TestLayer(layers.Layer): def __init__(self, **kwargs): super().__init__(**kwargs) self.supports_masking = True def compute_output_shape(self, input_shape): return input_shape def call(self, inputs, mask=None): assert mask is not None test_obj.assertAllClose(mask, expected_mask) return inputs model = models.Sequential( [ layers.Masking(mask_value=0.0), TestLayer(), ] ) model(x)
import numpy as np import pytest from keras.src import layers from keras.src import models from keras.src import testing class MaskingTest(testing.TestCase): @pytest.mark.requires_trainable_backend def test_masking_basics(self): self.run_layer_test( layers.Masking, init_kwargs={"mask_value": 0.0}, input_shape=(2, 3, 2), expected_output_shape=(2, 3, 2), expected_num_trainable_weights=0, expected_num_non_trainable_weights=0, expected_num_seed_generators=0, expected_num_losses=0, supports_masking=True, ) @pytest.mark.requires_trainable_backend def test_masking_correctness(self): x = np.array( [ [[0.0, 0.0], [1.0, 2.0], [0.0, 0.0]], [[2.0, 2.0], [0.0, 0.0], [2.0, 1.0]], ] ) expected_mask = [[False, True, False], [True, False, True]] layer = layers.Masking(mask_value=0.0) self.assertAllClose(layer.compute_mask(x), expected_mask) test_obj = self class TestLayer(layers.Layer): def __init__(self, **kwargs): super().__init__(**kwargs) self.supports_masking = True def compute_output_shape(self, input_shape): return input_shape def call(self, inputs, mask=None): assert mask is not None test_obj.assertAllClose(mask, expected_mask) return inputs model = models.Sequential( [ layers.Masking(mask_value=0.0), TestLayer(), ] ) model(x)
""" This file loads sentences from a provided text file. It is expected, that the there is one sentence per line in that text file. SimCSE will be training using these sentences. Checkpoints are stored every 500 steps to the output folder. Usage: python train_simcse_from_file.py path/to/sentences.txt """ import gzip import logging import math import sys from datetime import datetime import tqdm from torch.utils.data import DataLoader from sentence_transformers import InputExample, LoggingHandler, SentenceTransformer, losses, models #### Just some code to print debug information to stdout logging.basicConfig( format="%(asctime)s - %(message)s", datefmt="%Y-%m-%d %H:%M:%S", level=logging.INFO, handlers=[LoggingHandler()] ) #### /print debug information to stdout # Training parameters model_name = "distilroberta-base" train_batch_size = 128 max_seq_length = 32 num_epochs = 1 # Input file path (a text file, each line a sentence) if len(sys.argv) < 2: print("Run this script with: python {} path/to/sentences.txt".format(sys.argv[0])) exit() filepath = sys.argv[1] # Save path to store our model output_name = "" if len(sys.argv) >= 3: output_name = "-" + sys.argv[2].replace(" ", "_").replace("/", "_").replace("\\", "_") model_output_path = "output/train_simcse{}-{}".format(output_name, datetime.now().strftime("%Y-%m-%d_%H-%M-%S")) # Use Huggingface/transformers model (like BERT, RoBERTa, XLNet, XLM-R) for mapping tokens to embeddings word_embedding_model = models.Transformer(model_name, max_seq_length=max_seq_length) # Apply mean pooling to get one fixed sized sentence vector pooling_model = models.Pooling(word_embedding_model.get_word_embedding_dimension()) model = SentenceTransformer(modules=[word_embedding_model, pooling_model]) ################# Read the train corpus ################# train_samples = [] with gzip.open(filepath, "rt", encoding="utf8") if filepath.endswith(".gz") else open( filepath, encoding="utf8" ) as fIn: for line in tqdm.tqdm(fIn, desc="Read file"): line = line.strip() if len(line) >= 10: train_samples.append(InputExample(texts=[line, line])) logging.info("Train sentences: {}".format(len(train_samples))) # We train our model using the MultipleNegativesRankingLoss train_dataloader = DataLoader(train_samples, shuffle=True, batch_size=train_batch_size, drop_last=True) train_loss = losses.MultipleNegativesRankingLoss(model) warmup_steps = math.ceil(len(train_dataloader) * num_epochs * 0.1) # 10% of train data for warm-up logging.info("Warmup-steps: {}".format(warmup_steps)) # Train the model model.fit( train_objectives=[(train_dataloader, train_loss)], epochs=num_epochs, warmup_steps=warmup_steps, optimizer_params={"lr": 5e-5}, checkpoint_path=model_output_path, show_progress_bar=True, use_amp=False, # Set to True, if your GPU supports FP16 cores )
""" This file loads sentences from a provided text file. It is expected, that the there is one sentence per line in that text file. SimCSE will be training using these sentences. Checkpoints are stored every 500 steps to the output folder. Usage: python train_simcse_from_file.py path/to/sentences.txt """ from torch.utils.data import DataLoader import math from sentence_transformers import models, losses from sentence_transformers import LoggingHandler, SentenceTransformer, InputExample import logging from datetime import datetime import gzip import sys import tqdm #### Just some code to print debug information to stdout logging.basicConfig( format="%(asctime)s - %(message)s", datefmt="%Y-%m-%d %H:%M:%S", level=logging.INFO, handlers=[LoggingHandler()] ) #### /print debug information to stdout # Training parameters model_name = "distilroberta-base" train_batch_size = 128 max_seq_length = 32 num_epochs = 1 # Input file path (a text file, each line a sentence) if len(sys.argv) < 2: print("Run this script with: python {} path/to/sentences.txt".format(sys.argv[0])) exit() filepath = sys.argv[1] # Save path to store our model output_name = "" if len(sys.argv) >= 3: output_name = "-" + sys.argv[2].replace(" ", "_").replace("/", "_").replace("\\", "_") model_output_path = "output/train_simcse{}-{}".format(output_name, datetime.now().strftime("%Y-%m-%d_%H-%M-%S")) # Use Huggingface/transformers model (like BERT, RoBERTa, XLNet, XLM-R) for mapping tokens to embeddings word_embedding_model = models.Transformer(model_name, max_seq_length=max_seq_length) # Apply mean pooling to get one fixed sized sentence vector pooling_model = models.Pooling(word_embedding_model.get_word_embedding_dimension()) model = SentenceTransformer(modules=[word_embedding_model, pooling_model]) ################# Read the train corpus ################# train_samples = [] with gzip.open(filepath, "rt", encoding="utf8") if filepath.endswith(".gz") else open( filepath, encoding="utf8" ) as fIn: for line in tqdm.tqdm(fIn, desc="Read file"): line = line.strip() if len(line) >= 10: train_samples.append(InputExample(texts=[line, line])) logging.info("Train sentences: {}".format(len(train_samples))) # We train our model using the MultipleNegativesRankingLoss train_dataloader = DataLoader(train_samples, shuffle=True, batch_size=train_batch_size, drop_last=True) train_loss = losses.MultipleNegativesRankingLoss(model) warmup_steps = math.ceil(len(train_dataloader) * num_epochs * 0.1) # 10% of train data for warm-up logging.info("Warmup-steps: {}".format(warmup_steps)) # Train the model model.fit( train_objectives=[(train_dataloader, train_loss)], epochs=num_epochs, warmup_steps=warmup_steps, optimizer_params={"lr": 5e-5}, checkpoint_path=model_output_path, show_progress_bar=True, use_amp=False, # Set to True, if your GPU supports FP16 cores )
# Copyright (c) OpenMMLab. All rights reserved. from .cityscapes_metric import CityScapesMetric from .coco_metric import CocoMetric from .coco_panoptic_metric import CocoPanopticMetric from .crowdhuman_metric import CrowdHumanMetric from .lvis_metric import LVISMetric from .openimages_metric import OpenImagesMetric from .voc_metric import VOCMetric __all__ = [ 'CityScapesMetric', 'CocoMetric', 'CocoPanopticMetric', 'OpenImagesMetric', 'VOCMetric', 'LVISMetric', 'CrowdHumanMetric' ]
# Copyright (c) OpenMMLab. All rights reserved. from .cityscapes_metric import CityScapesMetric from .coco_metric import CocoMetric from .coco_panoptic_metric import CocoPanopticMetric from .lvis_metric import LVISMetric from .openimages_metric import OpenImagesMetric from .voc_metric import VOCMetric __all__ = [ 'CityScapesMetric', 'CocoMetric', 'CocoPanopticMetric', 'OpenImagesMetric', 'VOCMetric', 'LVISMetric' ]
import numpy as np from docarray import BaseDoc, DocList from docarray.typing import NdArray from pydantic import Field from jina import Executor, requests class TextDoc(BaseDoc): text: str = Field(description="The text of the document", default="") class EmbeddingResponseModel(TextDoc): embeddings: NdArray = Field(description="The embedding of the texts", default=[]) class Config(BaseDoc.Config): allow_population_by_field_name = True arbitrary_types_allowed = True json_encoders = {NdArray: lambda v: v.tolist()} class SampleExecutor(Executor): @requests(on="/encode") def foo(self, docs: DocList[TextDoc], **kwargs) -> DocList[EmbeddingResponseModel]: ret = [] for doc in docs: ret.append( EmbeddingResponseModel( id=doc.id, text=doc.text, embeddings=np.random.random((1, 64)), ) ) return DocList[EmbeddingResponseModel](ret)
import numpy as np from docarray import BaseDoc, DocList from docarray.typing import NdArray from pydantic import Field from jina import Executor, requests class TextDoc(BaseDoc): text: str class EmbeddingResponseModel(BaseDoc): embeddings: NdArray = Field(description="The embedding of the texts", default=[]) class Config(BaseDoc.Config): allow_population_by_field_name = True arbitrary_types_allowed = True json_encoders = {NdArray: lambda v: v.tolist()} class SampleExecutor(Executor): @requests(on="/encode") def foo(self, docs: DocList[TextDoc], **kwargs) -> DocList[EmbeddingResponseModel]: ret = [] for doc in docs: ret.append( EmbeddingResponseModel(id=doc.id, embeddings=np.random.random((1, 64))) ) return DocList[EmbeddingResponseModel](ret)
import os import numpy as np import pytest import torch from pydantic import parse_obj_as from docarray import BaseDocument from docarray.typing.tensor.audio.audio_ndarray import AudioNdArray from docarray.typing.tensor.audio.audio_torch_tensor import AudioTorchTensor from docarray.utils.misc import is_tf_available tf_available = is_tf_available() if tf_available: import tensorflow as tf import tensorflow._api.v2.experimental.numpy as tnp from docarray.typing.tensor.audio import AudioTensorFlowTensor @pytest.mark.parametrize( 'tensor,cls_audio_tensor,cls_tensor', [ (torch.zeros(1000, 2), AudioTorchTensor, torch.Tensor), (np.zeros((1000, 2)), AudioNdArray, np.ndarray), ], ) def test_set_audio_tensor(tensor, cls_audio_tensor, cls_tensor): class MyAudioDoc(BaseDocument): tensor: cls_audio_tensor doc = MyAudioDoc(tensor=tensor) assert isinstance(doc.tensor, cls_audio_tensor) assert isinstance(doc.tensor, cls_tensor) assert (doc.tensor == tensor).all() @pytest.mark.tensorflow def test_set_audio_tensorflow_tensor(): class MyAudioDoc(BaseDocument): tensor: AudioTensorFlowTensor doc = MyAudioDoc(tensor=tf.zeros((1000, 2))) assert isinstance(doc.tensor, AudioTensorFlowTensor) assert isinstance(doc.tensor.tensor, tf.Tensor) assert tnp.allclose(doc.tensor.tensor, tf.zeros((1000, 2))) @pytest.mark.parametrize( 'cls_tensor,tensor', [ (AudioNdArray, np.zeros((1000, 2))), (AudioTorchTensor, torch.zeros(1000, 2)), (AudioTorchTensor, np.zeros((1000, 2))), ], ) def test_validation(cls_tensor, tensor): arr = parse_obj_as(cls_tensor, tensor) assert isinstance(arr, cls_tensor) @pytest.mark.tensorflow def test_validation_tensorflow(): arr = parse_obj_as(AudioTensorFlowTensor, tf.zeros((1000, 2))) assert isinstance(arr, AudioTensorFlowTensor) @pytest.mark.parametrize( 'cls_tensor,tensor', [ (AudioNdArray, torch.zeros(1000, 2)), (AudioNdArray, 'hello'), (AudioTorchTensor, 'hello'), ], ) def test_illegal_validation(cls_tensor, tensor): match = str(cls_tensor).split('.')[-1][:-2] with pytest.raises(ValueError, match=match): parse_obj_as(cls_tensor, tensor) @pytest.mark.proto @pytest.mark.parametrize( 'cls_tensor,tensor,proto_key', [ (AudioTorchTensor, torch.zeros(1000, 2), AudioTorchTensor._proto_type_name), (AudioNdArray, np.zeros((1000, 2)), AudioNdArray._proto_type_name), ], ) def test_proto_tensor(cls_tensor, tensor, proto_key): tensor = parse_obj_as(cls_tensor, tensor) proto = tensor._to_node_protobuf() assert proto_key in str(proto) @pytest.mark.tensorflow def test_proto_tensor_tensorflow(): tensor = parse_obj_as(AudioTensorFlowTensor, tf.zeros((1000, 2))) proto = tensor._to_node_protobuf() assert AudioTensorFlowTensor._proto_type_name in str(proto) @pytest.mark.parametrize( 'cls_tensor,tensor', [ (AudioTorchTensor, torch.zeros(1000, 2)), (AudioNdArray, np.zeros((1000, 2))), ], ) def test_save_audio_tensor_to_wav_file(cls_tensor, tensor, tmpdir): tmp_file = str(tmpdir / 'tmp.wav') audio_tensor = parse_obj_as(cls_tensor, tensor) audio_tensor.save(tmp_file) assert os.path.isfile(tmp_file) @pytest.mark.tensorflow def test_save_audio_tensorflow_tensor_to_wav_file(tmpdir): tmp_file = str(tmpdir / 'tmp.wav') audio_tensor = parse_obj_as(AudioTensorFlowTensor, tf.zeros((1000, 2))) audio_tensor.save(tmp_file) assert os.path.isfile(tmp_file)
import os import numpy as np import pytest import torch from pydantic import parse_obj_as from docarray import BaseDocument from docarray.typing.tensor.audio.audio_ndarray import AudioNdArray from docarray.typing.tensor.audio.audio_torch_tensor import AudioTorchTensor from docarray.utils.misc import is_tf_available tf_available = is_tf_available() if tf_available: import tensorflow as tf import tensorflow._api.v2.experimental.numpy as tnp from docarray.typing.tensor.audio import AudioTensorFlowTensor @pytest.mark.parametrize( 'tensor,cls_audio_tensor,cls_tensor', [ (torch.zeros(1000, 2), AudioTorchTensor, torch.Tensor), (np.zeros((1000, 2)), AudioNdArray, np.ndarray), ], ) def test_set_audio_tensor(tensor, cls_audio_tensor, cls_tensor): class MyAudioDoc(BaseDocument): tensor: cls_audio_tensor doc = MyAudioDoc(tensor=tensor) assert isinstance(doc.tensor, cls_audio_tensor) assert isinstance(doc.tensor, cls_tensor) assert (doc.tensor == tensor).all() @pytest.mark.tensorflow def test_set_audio_tensorflow_tensor(): class MyAudioDoc(BaseDocument): tensor: AudioTensorFlowTensor doc = MyAudioDoc(tensor=tf.zeros((1000, 2))) assert isinstance(doc.tensor, AudioTensorFlowTensor) assert isinstance(doc.tensor.tensor, tf.Tensor) assert tnp.allclose(doc.tensor.tensor, tf.zeros((1000, 2))) @pytest.mark.parametrize( 'cls_tensor,tensor', [ (AudioNdArray, np.zeros((1000, 2))), (AudioTorchTensor, torch.zeros(1000, 2)), (AudioTorchTensor, np.zeros((1000, 2))), ], ) def test_validation(cls_tensor, tensor): arr = parse_obj_as(cls_tensor, tensor) assert isinstance(arr, cls_tensor) @pytest.mark.tensorflow def test_validation_tensorflow(): arr = parse_obj_as(AudioTensorFlowTensor, tf.zeros((1000, 2))) assert isinstance(arr, AudioTensorFlowTensor) @pytest.mark.parametrize( 'cls_tensor,tensor', [ (AudioNdArray, torch.zeros(1000, 2)), (AudioNdArray, 'hello'), (AudioTorchTensor, 'hello'), ], ) def test_illegal_validation(cls_tensor, tensor): match = str(cls_tensor).split('.')[-1][:-2] with pytest.raises(ValueError, match=match): parse_obj_as(cls_tensor, tensor) @pytest.mark.proto @pytest.mark.parametrize( 'cls_tensor,tensor,proto_key', [ (AudioTorchTensor, torch.zeros(1000, 2), AudioTorchTensor._proto_type_name), (AudioNdArray, np.zeros((1000, 2)), AudioNdArray._proto_type_name), ], ) def test_proto_tensor(cls_tensor, tensor, proto_key): tensor = parse_obj_as(cls_tensor, tensor) proto = tensor._to_node_protobuf() assert proto_key in str(proto) @pytest.mark.tensorflow def test_proto_tensor_tensorflow(): tensor = parse_obj_as(AudioTensorFlowTensor, tf.zeros((1000, 2))) proto = tensor._to_node_protobuf() assert AudioTensorFlowTensor._proto_type_name in str(proto) @pytest.mark.parametrize( 'cls_tensor,tensor', [ (AudioTorchTensor, torch.zeros(1000, 2)), (AudioNdArray, np.zeros((1000, 2))), ], ) def test_save_audio_tensor_to_wav_file(cls_tensor, tensor, tmpdir): tmp_file = str(tmpdir / 'tmp.wav') audio_tensor = parse_obj_as(cls_tensor, tensor) audio_tensor.save_to_wav_file(tmp_file) assert os.path.isfile(tmp_file) @pytest.mark.tensorflow def test_save_audio_tensorflow_tensor_to_wav_file(tmpdir): tmp_file = str(tmpdir / 'tmp.wav') audio_tensor = parse_obj_as(AudioTensorFlowTensor, tf.zeros((1000, 2))) audio_tensor.save_to_wav_file(tmp_file) assert os.path.isfile(tmp_file)
_base_ = '../mask_rcnn/mask-rcnn_r50-caffe_fpn_ms-1x_coco.py' # model settings model = dict( type='PointRend', roi_head=dict( type='PointRendRoIHead', mask_roi_extractor=dict( type='GenericRoIExtractor', aggregation='concat', roi_layer=dict( _delete_=True, type='SimpleRoIAlign', output_size=14), out_channels=256, featmap_strides=[4]), mask_head=dict( _delete_=True, type='CoarseMaskHead', num_fcs=2, in_channels=256, conv_out_channels=256, fc_out_channels=1024, num_classes=80, loss_mask=dict( type='CrossEntropyLoss', use_mask=True, loss_weight=1.0)), point_head=dict( type='MaskPointHead', num_fcs=3, in_channels=256, fc_channels=256, num_classes=80, coarse_pred_each_layer=True, loss_point=dict( type='CrossEntropyLoss', use_mask=True, loss_weight=1.0))), # model training and testing settings train_cfg=dict( rcnn=dict( mask_size=7, num_points=14 * 14, oversample_ratio=3, importance_sample_ratio=0.75)), test_cfg=dict( rcnn=dict( subdivision_steps=5, subdivision_num_points=28 * 28, scale_factor=2)))
_base_ = '../mask_rcnn/mask_rcnn_r50_caffe_fpn_mstrain_1x_coco.py' # model settings model = dict( type='PointRend', roi_head=dict( type='PointRendRoIHead', mask_roi_extractor=dict( type='GenericRoIExtractor', aggregation='concat', roi_layer=dict( _delete_=True, type='SimpleRoIAlign', output_size=14), out_channels=256, featmap_strides=[4]), mask_head=dict( _delete_=True, type='CoarseMaskHead', num_fcs=2, in_channels=256, conv_out_channels=256, fc_out_channels=1024, num_classes=80, loss_mask=dict( type='CrossEntropyLoss', use_mask=True, loss_weight=1.0)), point_head=dict( type='MaskPointHead', num_fcs=3, in_channels=256, fc_channels=256, num_classes=80, coarse_pred_each_layer=True, loss_point=dict( type='CrossEntropyLoss', use_mask=True, loss_weight=1.0))), # model training and testing settings train_cfg=dict( rcnn=dict( mask_size=7, num_points=14 * 14, oversample_ratio=3, importance_sample_ratio=0.75)), test_cfg=dict( rcnn=dict( subdivision_steps=5, subdivision_num_points=28 * 28, scale_factor=2)))
import PIL.Image import pytest import torch import torchvision.transforms.v2._utils from common_utils import DEFAULT_SIZE, make_bounding_boxes, make_detection_masks, make_image from torchvision import tv_tensors from torchvision.transforms.v2._utils import has_all, has_any from torchvision.transforms.v2.functional import to_pil_image IMAGE = make_image(DEFAULT_SIZE, color_space="RGB") BOUNDING_BOX = make_bounding_boxes(DEFAULT_SIZE, format=tv_tensors.BoundingBoxFormat.XYXY) MASK = make_detection_masks(DEFAULT_SIZE) @pytest.mark.parametrize( ("sample", "types", "expected"), [ ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image,), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.BoundingBoxes,), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Mask,), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.Mask), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.BoundingBoxes, tv_tensors.Mask), True), ((MASK,), (tv_tensors.Image, tv_tensors.BoundingBoxes), False), ((BOUNDING_BOX,), (tv_tensors.Image, tv_tensors.Mask), False), ((IMAGE,), (tv_tensors.BoundingBoxes, tv_tensors.Mask), False), ( (IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), True, ), ((), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), False), ((IMAGE, BOUNDING_BOX, MASK), (lambda obj: isinstance(obj, tv_tensors.Image),), True), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: False,), False), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: True,), True), ((IMAGE,), (tv_tensors.Image, PIL.Image.Image, torchvision.transforms.v2._utils.is_pure_tensor), True), ( (torch.Tensor(IMAGE),), (tv_tensors.Image, PIL.Image.Image, torchvision.transforms.v2._utils.is_pure_tensor), True, ), ( (to_pil_image(IMAGE),), (tv_tensors.Image, PIL.Image.Image, torchvision.transforms.v2._utils.is_pure_tensor), True, ), ], ) def test_has_any(sample, types, expected): assert has_any(sample, *types) is expected @pytest.mark.parametrize( ("sample", "types", "expected"), [ ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image,), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.BoundingBoxes,), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Mask,), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.Mask), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.BoundingBoxes, tv_tensors.Mask), True), ( (IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), True, ), ((BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes), False), ((BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.Mask), False), ((IMAGE, MASK), (tv_tensors.BoundingBoxes, tv_tensors.Mask), False), ( (IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), True, ), ((BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), False), ((IMAGE, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), False), ((IMAGE, BOUNDING_BOX), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), False), ( (IMAGE, BOUNDING_BOX, MASK), (lambda obj: isinstance(obj, (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask)),), True, ), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: False,), False), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: True,), True), ], ) def test_has_all(sample, types, expected): assert has_all(sample, *types) is expected
import PIL.Image import pytest import torch import torchvision.transforms.v2._utils from common_utils import DEFAULT_SIZE, make_bounding_boxes, make_detection_mask, make_image from torchvision import tv_tensors from torchvision.transforms.v2._utils import has_all, has_any from torchvision.transforms.v2.functional import to_pil_image IMAGE = make_image(DEFAULT_SIZE, color_space="RGB") BOUNDING_BOX = make_bounding_boxes(DEFAULT_SIZE, format=tv_tensors.BoundingBoxFormat.XYXY) MASK = make_detection_mask(DEFAULT_SIZE) @pytest.mark.parametrize( ("sample", "types", "expected"), [ ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image,), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.BoundingBoxes,), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Mask,), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.Mask), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.BoundingBoxes, tv_tensors.Mask), True), ((MASK,), (tv_tensors.Image, tv_tensors.BoundingBoxes), False), ((BOUNDING_BOX,), (tv_tensors.Image, tv_tensors.Mask), False), ((IMAGE,), (tv_tensors.BoundingBoxes, tv_tensors.Mask), False), ( (IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), True, ), ((), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), False), ((IMAGE, BOUNDING_BOX, MASK), (lambda obj: isinstance(obj, tv_tensors.Image),), True), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: False,), False), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: True,), True), ((IMAGE,), (tv_tensors.Image, PIL.Image.Image, torchvision.transforms.v2._utils.is_pure_tensor), True), ( (torch.Tensor(IMAGE),), (tv_tensors.Image, PIL.Image.Image, torchvision.transforms.v2._utils.is_pure_tensor), True, ), ( (to_pil_image(IMAGE),), (tv_tensors.Image, PIL.Image.Image, torchvision.transforms.v2._utils.is_pure_tensor), True, ), ], ) def test_has_any(sample, types, expected): assert has_any(sample, *types) is expected @pytest.mark.parametrize( ("sample", "types", "expected"), [ ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image,), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.BoundingBoxes,), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Mask,), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.Mask), True), ((IMAGE, BOUNDING_BOX, MASK), (tv_tensors.BoundingBoxes, tv_tensors.Mask), True), ( (IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), True, ), ((BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes), False), ((BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.Mask), False), ((IMAGE, MASK), (tv_tensors.BoundingBoxes, tv_tensors.Mask), False), ( (IMAGE, BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), True, ), ((BOUNDING_BOX, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), False), ((IMAGE, MASK), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), False), ((IMAGE, BOUNDING_BOX), (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask), False), ( (IMAGE, BOUNDING_BOX, MASK), (lambda obj: isinstance(obj, (tv_tensors.Image, tv_tensors.BoundingBoxes, tv_tensors.Mask)),), True, ), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: False,), False), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: True,), True), ], ) def test_has_all(sample, types, expected): assert has_all(sample, *types) is expected
from .cmuarctic import CMUARCTIC from .cmudict import CMUDict from .commonvoice import COMMONVOICE from .dr_vctk import DR_VCTK from .fluentcommands import FluentSpeechCommands from .gtzan import GTZAN from .iemocap import IEMOCAP from .librilight_limited import LibriLightLimited from .librimix import LibriMix from .librispeech import LIBRISPEECH from .librispeech_biasing import LibriSpeechBiasing from .libritts import LIBRITTS from .ljspeech import LJSPEECH from .musdb_hq import MUSDB_HQ from .quesst14 import QUESST14 from .snips import Snips from .speechcommands import SPEECHCOMMANDS from .tedlium import TEDLIUM from .vctk import VCTK_092 from .voxceleb1 import VoxCeleb1Identification, VoxCeleb1Verification from .yesno import YESNO __all__ = [ "COMMONVOICE", "LIBRISPEECH", "LibriSpeechBiasing", "LibriLightLimited", "SPEECHCOMMANDS", "VCTK_092", "DR_VCTK", "YESNO", "LJSPEECH", "GTZAN", "CMUARCTIC", "CMUDict", "LibriMix", "LIBRITTS", "TEDLIUM", "QUESST14", "MUSDB_HQ", "FluentSpeechCommands", "VoxCeleb1Identification", "VoxCeleb1Verification", "IEMOCAP", "Snips", ]
from .cmuarctic import CMUARCTIC from .cmudict import CMUDict from .commonvoice import COMMONVOICE from .dr_vctk import DR_VCTK from .fluentcommands import FluentSpeechCommands from .gtzan import GTZAN from .iemocap import IEMOCAP from .librilight_limited import LibriLightLimited from .librimix import LibriMix from .librispeech import LIBRISPEECH from .libritts import LIBRITTS from .ljspeech import LJSPEECH from .musdb_hq import MUSDB_HQ from .quesst14 import QUESST14 from .snips import Snips from .speechcommands import SPEECHCOMMANDS from .tedlium import TEDLIUM from .vctk import VCTK_092 from .voxceleb1 import VoxCeleb1Identification, VoxCeleb1Verification from .yesno import YESNO __all__ = [ "COMMONVOICE", "LIBRISPEECH", "LibriLightLimited", "SPEECHCOMMANDS", "VCTK_092", "DR_VCTK", "YESNO", "LJSPEECH", "GTZAN", "CMUARCTIC", "CMUDict", "LibriMix", "LIBRITTS", "TEDLIUM", "QUESST14", "MUSDB_HQ", "FluentSpeechCommands", "VoxCeleb1Identification", "VoxCeleb1Verification", "IEMOCAP", "Snips", ]
"""Simple Reader for Memos.""" from typing import Dict, List from urllib.parse import urljoin from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class MemosReader(BaseReader): """ Memos reader. Reads content from an Memos. """ def __init__(self, host: str = "https://demo.usememos.com/") -> None: """Init params.""" self._memoUrl = urljoin(host, "api/memo") def load_data(self, params: Dict = {}) -> List[Document]: """ Load data from RSS feeds. Args: params (Dict): Filtering parameters. Returns: List[Document]: List of documents. """ import requests documents = [] realUrl = self._memoUrl if not params: realUrl = urljoin(self._memoUrl, "all", False) try: req = requests.get(realUrl, params) res = req.json() except ValueError: raise ValueError("Your Memo URL is not valid") if "data" not in res: raise ValueError("Invalid Memo response") memos = res["data"] for memo in memos: content = memo["content"] extra_info = { "creator": memo["creator"], "resource_list": memo["resourceList"], id: memo["id"], } documents.append(Document(text=content, extra_info=extra_info)) return documents
"""Simple Reader for Memos.""" from typing import Dict, List from urllib.parse import urljoin from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class MemosReader(BaseReader): """Memos reader. Reads content from an Memos. """ def __init__(self, host: str = "https://demo.usememos.com/") -> None: """Init params.""" self._memoUrl = urljoin(host, "api/memo") def load_data(self, params: Dict = {}) -> List[Document]: """Load data from RSS feeds. Args: params (Dict): Filtering parameters. Returns: List[Document]: List of documents. """ import requests documents = [] realUrl = self._memoUrl if not params: realUrl = urljoin(self._memoUrl, "all", False) try: req = requests.get(realUrl, params) res = req.json() except ValueError: raise ValueError("Your Memo URL is not valid") if "data" not in res: raise ValueError("Invalid Memo response") memos = res["data"] for memo in memos: content = memo["content"] extra_info = { "creator": memo["creator"], "resource_list": memo["resourceList"], id: memo["id"], } documents.append(Document(text=content, extra_info=extra_info)) return documents
#!/usr/bin/env python # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License """Utilities to handle file locking in `datasets`.""" import os from filelock import FileLock as FileLock_ from filelock import UnixFileLock from filelock import __version__ as _filelock_version from packaging import version class FileLock(FileLock_): """ A `filelock.FileLock` initializer that handles long paths. It also uses the current umask for lock files. """ MAX_FILENAME_LENGTH = 255 def __init__(self, lock_file, *args, **kwargs): # The "mode" argument is required if we want to use the current umask in filelock >= 3.10 # In previous previous it was already using the current umask. if "mode" not in kwargs and version.parse(_filelock_version) >= version.parse("3.10.0"): umask = os.umask(0o666) os.umask(umask) kwargs["mode"] = 0o666 & ~umask lock_file = self.hash_filename_if_too_long(lock_file) super().__init__(lock_file, *args, **kwargs) @classmethod def hash_filename_if_too_long(cls, path: str) -> str: path = os.path.abspath(os.path.expanduser(path)) filename = os.path.basename(path) max_filename_length = cls.MAX_FILENAME_LENGTH if issubclass(cls, UnixFileLock): max_filename_length = min(max_filename_length, os.statvfs(os.path.dirname(path)).f_namemax) if len(filename) > max_filename_length: dirname = os.path.dirname(path) hashed_filename = str(hash(filename)) new_filename = ( filename[: max_filename_length - len(hashed_filename) - 8] + "..." + hashed_filename + ".lock" ) return os.path.join(dirname, new_filename) else: return path
#!/usr/bin/env python # coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License """Utilities to handle file locking in `datasets`.""" import os from filelock import FileLock as FileLock_ from filelock import UnixFileLock from filelock import __version__ as _filelock_version from packaging import version class FileLock(FileLock_): """ A `filelock.FileLock` initializer that handles long paths. It also uses the current umask for lock files. """ MAX_FILENAME_LENGTH = 255 def __init__(self, lock_file, *args, **kwargs): # The "mode" argument is required if we want to use the current umask in filelock >= 3.10 # In previous previous it was already using the current umask. if "mode" not in kwargs and version.parse(_filelock_version) >= version.parse("3.10.0"): umask = os.umask(0o666) os.umask(umask) kwargs["mode"] = 0o666 & ~umask lock_file = self.hash_filename_if_too_long(lock_file) super().__init__(lock_file, *args, **kwargs) @classmethod def hash_filename_if_too_long(cls, path: str) -> str: path = os.path.abspath(os.path.expanduser(path)) filename = os.path.basename(path) max_filename_length = cls.MAX_FILENAME_LENGTH if issubclass(cls, UnixFileLock): max_filename_length = min(max_filename_length, os.statvfs(os.path.dirname(path)).f_namemax) if len(filename) > max_filename_length: dirname = os.path.dirname(path) hashed_filename = str(hash(filename)) new_filename = ( filename[: max_filename_length - len(hashed_filename) - 8] + "..." + hashed_filename + ".lock" ) return os.path.join(dirname, new_filename) else: return path
from typing import TYPE_CHECKING, Any, Callable, List, Optional, Type from llama_index.core.bridge.pydantic import BaseModel, ConfigDict from .errors import WorkflowValidationError from .utils import ( is_free_function, validate_step_signature, inspect_signature, ServiceDefinition, ) if TYPE_CHECKING: # pragma: no cover from .workflow import Workflow from .retry_policy import RetryPolicy class StepConfig(BaseModel): model_config = ConfigDict(arbitrary_types_allowed=True) accepted_events: List[Any] event_name: str return_types: List[Any] context_parameter: Optional[str] num_workers: int requested_services: List[ServiceDefinition] retry_policy: Optional[RetryPolicy] def step( *args: Any, workflow: Optional[Type["Workflow"]] = None, pass_context: bool = False, num_workers: int = 4, retry_policy: Optional[RetryPolicy] = None, ) -> Callable: """ Decorator used to mark methods and functions as workflow steps. Decorators are evaluated at import time, but we need to wait for starting the communication channels until runtime. For this reason, we temporarily store the list of events that will be consumed by this step in the function object itself. Args: workflow: Workflow class to which the decorated step will be added. Only needed when using the decorator on free functions instead of class methods. num_workers: The number of workers that will process events for the decorated step. The default value works most of the times. retry_policy: The policy used to retry a step that encountered an error while running. """ def decorator(func: Callable) -> Callable: if not isinstance(num_workers, int) or num_workers <= 0: raise WorkflowValidationError( "num_workers must be an integer greater than 0" ) # This will raise providing a message with the specific validation failure spec = inspect_signature(func) validate_step_signature(spec) event_name, accepted_events = next(iter(spec.accepted_events.items())) # store the configuration in the function object func.__step_config = StepConfig( # type: ignore[attr-defined] accepted_events=accepted_events, event_name=event_name, return_types=spec.return_types, context_parameter=spec.context_parameter, num_workers=num_workers, requested_services=spec.requested_services or [], retry_policy=retry_policy, ) # If this is a free function, call add_step() explicitly. if is_free_function(func.__qualname__): if workflow is None: msg = f"To decorate {func.__name__} please pass a workflow class to the @step decorator." raise WorkflowValidationError(msg) workflow.add_step(func) return func if len(args): # The decorator was used without parentheses, like `@step` func = args[0] decorator(func) return func return decorator
from typing import TYPE_CHECKING, Any, Callable, List, Optional, Type from llama_index.core.bridge.pydantic import BaseModel, ConfigDict from .errors import WorkflowValidationError from .utils import ( is_free_function, validate_step_signature, inspect_signature, ServiceDefinition, ) if TYPE_CHECKING: # pragma: no cover from .workflow import Workflow from .retry_policy import RetryPolicy class StepConfig(BaseModel): model_config = ConfigDict(arbitrary_types_allowed=True) accepted_events: List[Any] event_name: str return_types: List[Any] context_parameter: Optional[str] num_workers: int requested_services: List[ServiceDefinition] retry_policy: Optional[RetryPolicy] def step( *args: Any, workflow: Optional[Type["Workflow"]] = None, pass_context: bool = False, num_workers: int = 4, retry_policy: Optional[RetryPolicy] = None, ) -> Callable: """Decorator used to mark methods and functions as workflow steps. Decorators are evaluated at import time, but we need to wait for starting the communication channels until runtime. For this reason, we temporarily store the list of events that will be consumed by this step in the function object itself. Args: workflow: Workflow class to which the decorated step will be added. Only needed when using the decorator on free functions instead of class methods. num_workers: The number of workers that will process events for the decorated step. The default value works most of the times. retry_policy: The policy used to retry a step that encountered an error while running. """ def decorator(func: Callable) -> Callable: if not isinstance(num_workers, int) or num_workers <= 0: raise WorkflowValidationError( "num_workers must be an integer greater than 0" ) # This will raise providing a message with the specific validation failure spec = inspect_signature(func) validate_step_signature(spec) event_name, accepted_events = next(iter(spec.accepted_events.items())) # store the configuration in the function object func.__step_config = StepConfig( # type: ignore[attr-defined] accepted_events=accepted_events, event_name=event_name, return_types=spec.return_types, context_parameter=spec.context_parameter, num_workers=num_workers, requested_services=spec.requested_services or [], retry_policy=retry_policy, ) # If this is a free function, call add_step() explicitly. if is_free_function(func.__qualname__): if workflow is None: msg = f"To decorate {func.__name__} please pass a workflow class to the @step decorator." raise WorkflowValidationError(msg) workflow.add_step(func) return func if len(args): # The decorator was used without parentheses, like `@step` func = args[0] decorator(func) return func return decorator
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.quantizers import deserialize as deserialize from keras.src.quantizers import get as get from keras.src.quantizers import serialize as serialize from keras.src.quantizers.quantizers import AbsMaxQuantizer as AbsMaxQuantizer from keras.src.quantizers.quantizers import Quantizer as Quantizer from keras.src.quantizers.quantizers import abs_max_quantize as abs_max_quantize from keras.src.quantizers.quantizers import ( compute_float8_amax_history as compute_float8_amax_history, ) from keras.src.quantizers.quantizers import ( compute_float8_scale as compute_float8_scale, ) from keras.src.quantizers.quantizers import ( fake_quant_with_min_max_vars as fake_quant_with_min_max_vars, ) from keras.src.quantizers.quantizers import ( quantize_and_dequantize as quantize_and_dequantize, )
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.quantizers import deserialize from keras.src.quantizers import get from keras.src.quantizers import serialize from keras.src.quantizers.quantizers import AbsMaxQuantizer from keras.src.quantizers.quantizers import Quantizer from keras.src.quantizers.quantizers import abs_max_quantize from keras.src.quantizers.quantizers import compute_float8_amax_history from keras.src.quantizers.quantizers import compute_float8_scale from keras.src.quantizers.quantizers import fake_quant_with_min_max_vars from keras.src.quantizers.quantizers import quantize_and_dequantize
_base_ = '../_base_/default_runtime.py' # dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' # Example to use different file client # Method 1: simply set the data root and let the file I/O module # automatically infer from prefix (not support LMDB and Memcache yet) # data_root = 's3://openmmlab/datasets/detection/coco/' # Method 2: Use `backend_args`, `file_client_args` in versions before 3.0.0rc6 # backend_args = dict( # backend='petrel', # path_mapping=dict({ # './data/': 's3://openmmlab/datasets/detection/', # 'data/': 's3://openmmlab/datasets/detection/' # })) backend_args = None # Align with Detectron2 backend = 'pillow' train_pipeline = [ dict( type='LoadImageFromFile', backend_args=backend_args, imdecode_backend=backend), dict(type='LoadAnnotations', with_bbox=True), dict( type='RandomChoiceResize', scales=[(1333, 640), (1333, 672), (1333, 704), (1333, 736), (1333, 768), (1333, 800)], keep_ratio=True, backend=backend), dict(type='RandomFlip', prob=0.5), dict(type='PackDetInputs') ] test_pipeline = [ dict( type='LoadImageFromFile', backend_args=backend_args, imdecode_backend=backend), dict(type='Resize', scale=(1333, 800), keep_ratio=True, backend=backend), dict(type='LoadAnnotations', with_bbox=True), dict( type='PackDetInputs', meta_keys=('img_id', 'img_path', 'ori_shape', 'img_shape', 'scale_factor')) ] train_dataloader = dict( batch_size=2, num_workers=2, persistent_workers=True, pin_memory=True, sampler=dict(type='InfiniteSampler', shuffle=True), batch_sampler=dict(type='AspectRatioBatchSampler'), dataset=dict( type=dataset_type, data_root=data_root, ann_file='annotations/instances_train2017.json', data_prefix=dict(img='train2017/'), filter_cfg=dict(filter_empty_gt=True, min_size=32), pipeline=train_pipeline, backend_args=backend_args)) val_dataloader = dict( batch_size=1, num_workers=2, persistent_workers=True, drop_last=False, pin_memory=True, sampler=dict(type='DefaultSampler', shuffle=False), dataset=dict( type=dataset_type, data_root=data_root, ann_file='annotations/instances_val2017.json', data_prefix=dict(img='val2017/'), test_mode=True, pipeline=test_pipeline, backend_args=backend_args)) test_dataloader = val_dataloader val_evaluator = dict( type='CocoMetric', ann_file=data_root + 'annotations/instances_val2017.json', metric='bbox', format_only=False, backend_args=backend_args) test_evaluator = val_evaluator # training schedule for 90k max_iter = 90000 train_cfg = dict( type='IterBasedTrainLoop', max_iters=max_iter, val_interval=10000) val_cfg = dict(type='ValLoop') test_cfg = dict(type='TestLoop') # learning rate param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=1000), dict( type='MultiStepLR', begin=0, end=max_iter, by_epoch=False, milestones=[60000, 80000], gamma=0.1) ] # optimizer optim_wrapper = dict( type='OptimWrapper', optimizer=dict(type='SGD', lr=0.02, momentum=0.9, weight_decay=0.0001)) # Default setting for scaling LR automatically # - `enable` means enable scaling LR automatically # or not by default. # - `base_batch_size` = (8 GPUs) x (2 samples per GPU). auto_scale_lr = dict(enable=False, base_batch_size=16) default_hooks = dict(checkpoint=dict(by_epoch=False, interval=10000)) log_processor = dict(by_epoch=False)
_base_ = '../_base_/default_runtime.py' # dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' # file_client_args = dict( # backend='petrel', # path_mapping=dict({ # './data/': 's3://openmmlab/datasets/detection/', # 'data/': 's3://openmmlab/datasets/detection/' # })) file_client_args = dict(backend='disk') # Align with Detectron2 backend = 'pillow' train_pipeline = [ dict( type='LoadImageFromFile', file_client_args=file_client_args, imdecode_backend=backend), dict(type='LoadAnnotations', with_bbox=True), dict( type='RandomChoiceResize', scales=[(1333, 640), (1333, 672), (1333, 704), (1333, 736), (1333, 768), (1333, 800)], keep_ratio=True, backend=backend), dict(type='RandomFlip', prob=0.5), dict(type='PackDetInputs') ] test_pipeline = [ dict( type='LoadImageFromFile', file_client_args=file_client_args, imdecode_backend=backend), dict(type='Resize', scale=(1333, 800), keep_ratio=True, backend=backend), dict(type='LoadAnnotations', with_bbox=True), dict( type='PackDetInputs', meta_keys=('img_id', 'img_path', 'ori_shape', 'img_shape', 'scale_factor')) ] train_dataloader = dict( batch_size=2, num_workers=2, persistent_workers=True, pin_memory=True, sampler=dict(type='InfiniteSampler', shuffle=True), batch_sampler=dict(type='AspectRatioBatchSampler'), dataset=dict( type=dataset_type, data_root=data_root, ann_file='annotations/instances_train2017.json', data_prefix=dict(img='train2017/'), filter_cfg=dict(filter_empty_gt=True, min_size=32), pipeline=train_pipeline)) val_dataloader = dict( batch_size=1, num_workers=2, persistent_workers=True, drop_last=False, pin_memory=True, sampler=dict(type='DefaultSampler', shuffle=False), dataset=dict( type=dataset_type, data_root=data_root, ann_file='annotations/instances_val2017.json', data_prefix=dict(img='val2017/'), test_mode=True, pipeline=test_pipeline)) test_dataloader = val_dataloader val_evaluator = dict( type='CocoMetric', ann_file=data_root + 'annotations/instances_val2017.json', metric='bbox', format_only=False) test_evaluator = val_evaluator # training schedule for 90k max_iter = 90000 train_cfg = dict( type='IterBasedTrainLoop', max_iters=max_iter, val_interval=10000) val_cfg = dict(type='ValLoop') test_cfg = dict(type='TestLoop') # learning rate param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=1000), dict( type='MultiStepLR', begin=0, end=max_iter, by_epoch=False, milestones=[60000, 80000], gamma=0.1) ] # optimizer optim_wrapper = dict( type='OptimWrapper', optimizer=dict(type='SGD', lr=0.02, momentum=0.9, weight_decay=0.0001)) # Default setting for scaling LR automatically # - `enable` means enable scaling LR automatically # or not by default. # - `base_batch_size` = (8 GPUs) x (2 samples per GPU). auto_scale_lr = dict(enable=False, base_batch_size=16) default_hooks = dict(checkpoint=dict(by_epoch=False, interval=10000)) log_processor = dict(by_epoch=False)
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import ClickTool from langchain_community.tools.playwright.click import ClickToolInput # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handling optional imports. DEPRECATED_LOOKUP = { "ClickToolInput": "langchain_community.tools.playwright.click", "ClickTool": "langchain_community.tools", } _import_attribute = create_importer(__package__, deprecated_lookups=DEPRECATED_LOOKUP) def __getattr__(name: str) -> Any: """Look up attributes dynamically.""" return _import_attribute(name) __all__ = [ "ClickTool", "ClickToolInput", ]
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import ClickTool from langchain_community.tools.playwright.click import ClickToolInput # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handling optional imports. DEPRECATED_LOOKUP = { "ClickToolInput": "langchain_community.tools.playwright.click", "ClickTool": "langchain_community.tools", } _import_attribute = create_importer(__package__, deprecated_lookups=DEPRECATED_LOOKUP) def __getattr__(name: str) -> Any: """Look up attributes dynamically.""" return _import_attribute(name) __all__ = [ "ClickToolInput", "ClickTool", ]
"""Retrieval evaluators.""" from typing import List, Optional, Tuple from llama_index.core.base.base_retriever import BaseRetriever from llama_index.core.bridge.pydantic import Field, SerializeAsAny from llama_index.core.evaluation.retrieval.base import ( BaseRetrievalEvaluator, RetrievalEvalMode, ) from llama_index.core.indices.base_retriever import BaseRetriever from llama_index.core.postprocessor.types import BaseNodePostprocessor from llama_index.core.schema import ImageNode, TextNode class RetrieverEvaluator(BaseRetrievalEvaluator): """ Retriever evaluator. This module will evaluate a retriever using a set of metrics. Args: metrics (List[BaseRetrievalMetric]): Sequence of metrics to evaluate retriever: Retriever to evaluate. node_postprocessors (Optional[List[BaseNodePostprocessor]]): Post-processor to apply after retrieval. """ retriever: BaseRetriever = Field(..., description="Retriever to evaluate") node_postprocessors: Optional[List[SerializeAsAny[BaseNodePostprocessor]]] = Field( default=None, description="Optional post-processor" ) async def _aget_retrieved_ids_and_texts( self, query: str, mode: RetrievalEvalMode = RetrievalEvalMode.TEXT ) -> Tuple[List[str], List[str]]: """Get retrieved ids and texts, potentially applying a post-processor.""" retrieved_nodes = await self.retriever.aretrieve(query) if self.node_postprocessors: for node_postprocessor in self.node_postprocessors: retrieved_nodes = node_postprocessor.postprocess_nodes( retrieved_nodes, query_str=query ) return ( [node.node.node_id for node in retrieved_nodes], [node.text for node in retrieved_nodes], ) class MultiModalRetrieverEvaluator(BaseRetrievalEvaluator): """ Retriever evaluator. This module will evaluate a retriever using a set of metrics. Args: metrics (List[BaseRetrievalMetric]): Sequence of metrics to evaluate retriever: Retriever to evaluate. node_postprocessors (Optional[List[BaseNodePostprocessor]]): Post-processor to apply after retrieval. """ retriever: BaseRetriever = Field(..., description="Retriever to evaluate") node_postprocessors: Optional[List[SerializeAsAny[BaseNodePostprocessor]]] = Field( default=None, description="Optional post-processor" ) async def _aget_retrieved_ids_and_texts( self, query: str, mode: RetrievalEvalMode = RetrievalEvalMode.TEXT ) -> Tuple[List[str], List[str]]: """Get retrieved ids.""" retrieved_nodes = await self.retriever.aretrieve(query) image_nodes: List[ImageNode] = [] text_nodes: List[TextNode] = [] if self.node_postprocessors: for node_postprocessor in self.node_postprocessors: retrieved_nodes = node_postprocessor.postprocess_nodes( retrieved_nodes, query_str=query ) for scored_node in retrieved_nodes: node = scored_node.node if isinstance(node, ImageNode): image_nodes.append(node) if isinstance(node, TextNode): text_nodes.append(node) if mode == "text": return ( [node.node_id for node in text_nodes], [node.text for node in text_nodes], ) elif mode == "image": return ( [node.node_id for node in image_nodes], [node.text for node in image_nodes], ) else: raise ValueError("Unsupported mode.")
"""Retrieval evaluators.""" from typing import List, Optional, Tuple from llama_index.core.base.base_retriever import BaseRetriever from llama_index.core.bridge.pydantic import Field, SerializeAsAny from llama_index.core.evaluation.retrieval.base import ( BaseRetrievalEvaluator, RetrievalEvalMode, ) from llama_index.core.indices.base_retriever import BaseRetriever from llama_index.core.postprocessor.types import BaseNodePostprocessor from llama_index.core.schema import ImageNode, TextNode class RetrieverEvaluator(BaseRetrievalEvaluator): """Retriever evaluator. This module will evaluate a retriever using a set of metrics. Args: metrics (List[BaseRetrievalMetric]): Sequence of metrics to evaluate retriever: Retriever to evaluate. node_postprocessors (Optional[List[BaseNodePostprocessor]]): Post-processor to apply after retrieval. """ retriever: BaseRetriever = Field(..., description="Retriever to evaluate") node_postprocessors: Optional[List[SerializeAsAny[BaseNodePostprocessor]]] = Field( default=None, description="Optional post-processor" ) async def _aget_retrieved_ids_and_texts( self, query: str, mode: RetrievalEvalMode = RetrievalEvalMode.TEXT ) -> Tuple[List[str], List[str]]: """Get retrieved ids and texts, potentially applying a post-processor.""" retrieved_nodes = await self.retriever.aretrieve(query) if self.node_postprocessors: for node_postprocessor in self.node_postprocessors: retrieved_nodes = node_postprocessor.postprocess_nodes( retrieved_nodes, query_str=query ) return ( [node.node.node_id for node in retrieved_nodes], [node.text for node in retrieved_nodes], ) class MultiModalRetrieverEvaluator(BaseRetrievalEvaluator): """Retriever evaluator. This module will evaluate a retriever using a set of metrics. Args: metrics (List[BaseRetrievalMetric]): Sequence of metrics to evaluate retriever: Retriever to evaluate. node_postprocessors (Optional[List[BaseNodePostprocessor]]): Post-processor to apply after retrieval. """ retriever: BaseRetriever = Field(..., description="Retriever to evaluate") node_postprocessors: Optional[List[SerializeAsAny[BaseNodePostprocessor]]] = Field( default=None, description="Optional post-processor" ) async def _aget_retrieved_ids_and_texts( self, query: str, mode: RetrievalEvalMode = RetrievalEvalMode.TEXT ) -> Tuple[List[str], List[str]]: """Get retrieved ids.""" retrieved_nodes = await self.retriever.aretrieve(query) image_nodes: List[ImageNode] = [] text_nodes: List[TextNode] = [] if self.node_postprocessors: for node_postprocessor in self.node_postprocessors: retrieved_nodes = node_postprocessor.postprocess_nodes( retrieved_nodes, query_str=query ) for scored_node in retrieved_nodes: node = scored_node.node if isinstance(node, ImageNode): image_nodes.append(node) if isinstance(node, TextNode): text_nodes.append(node) if mode == "text": return ( [node.node_id for node in text_nodes], [node.text for node in text_nodes], ) elif mode == "image": return ( [node.node_id for node in image_nodes], [node.text for node in image_nodes], ) else: raise ValueError("Unsupported mode.")
from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class Translation: """`Feature` for translations with fixed languages per example. Here for compatibility with tfds. Args: languages (`dict`): A dictionary for each example mapping string language codes to string translations. Example: ```python >>> # At construction time: >>> datasets.features.Translation(languages=['en', 'fr', 'de']) >>> # During data generation: >>> yield { ... 'en': 'the cat', ... 'fr': 'le chat', ... 'de': 'die katze' ... } ``` """ languages: list[str] id: Optional[str] = field(default=None, repr=False) # Automatically constructed dtype: ClassVar[str] = "dict" pa_type: ClassVar[Any] = None _type: str = field(default="Translation", init=False, repr=False) def __call__(self): return pa.struct({lang: pa.string() for lang in sorted(self.languages)}) def flatten(self) -> Union["FeatureType", dict[str, "FeatureType"]]: """Flatten the Translation feature into a dictionary.""" from .features import Value return {k: Value("string") for k in sorted(self.languages)} @dataclass class TranslationVariableLanguages: """`Feature` for translations with variable languages per example. Here for compatibility with tfds. Args: languages (`dict`): A dictionary for each example mapping string language codes to one or more string translations. The languages present may vary from example to example. Returns: - `language` or `translation` (variable-length 1D `tf.Tensor` of `tf.string`): Language codes sorted in ascending order or plain text translations, sorted to align with language codes. Example: ```python >>> # At construction time: >>> datasets.features.TranslationVariableLanguages(languages=['en', 'fr', 'de']) >>> # During data generation: >>> yield { ... 'en': 'the cat', ... 'fr': ['le chat', 'la chatte,'] ... 'de': 'die katze' ... } >>> # Tensor returned : >>> { ... 'language': ['en', 'de', 'fr', 'fr'], ... 'translation': ['the cat', 'die katze', 'la chatte', 'le chat'], ... } ``` """ languages: Optional[list] = None num_languages: Optional[int] = None id: Optional[str] = field(default=None, repr=False) # Automatically constructed dtype: ClassVar[str] = "dict" pa_type: ClassVar[Any] = None _type: str = field(default="TranslationVariableLanguages", init=False, repr=False) def __post_init__(self): self.languages = sorted(set(self.languages)) if self.languages else None self.num_languages = len(self.languages) if self.languages else None def __call__(self): return pa.struct({"language": pa.list_(pa.string()), "translation": pa.list_(pa.string())}) def encode_example(self, translation_dict): lang_set = set(self.languages) if set(translation_dict) == {"language", "translation"}: return translation_dict elif self.languages and set(translation_dict) - lang_set: raise ValueError( f"Some languages in example ({', '.join(sorted(set(translation_dict) - lang_set))}) are not in valid set ({', '.join(lang_set)})." ) # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. translation_tuples = [] for lang, text in translation_dict.items(): if isinstance(text, str): translation_tuples.append((lang, text)) else: translation_tuples.extend([(lang, el) for el in text]) # Ensure translations are in ascending order by language code. languages, translations = zip(*sorted(translation_tuples)) return {"language": languages, "translation": translations} def flatten(self) -> Union["FeatureType", dict[str, "FeatureType"]]: """Flatten the TranslationVariableLanguages feature into a dictionary.""" from .features import Sequence, Value return { "language": Sequence(Value("string")), "translation": Sequence(Value("string")), }
from dataclasses import dataclass, field from typing import TYPE_CHECKING, Any, ClassVar, Optional, Union import pyarrow as pa if TYPE_CHECKING: from .features import FeatureType @dataclass class Translation: """`Feature` for translations with fixed languages per example. Here for compatibility with tfds. Args: languages (`dict`): A dictionary for each example mapping string language codes to string translations. Example: ```python >>> # At construction time: >>> datasets.features.Translation(languages=['en', 'fr', 'de']) >>> # During data generation: >>> yield { ... 'en': 'the cat', ... 'fr': 'le chat', ... 'de': 'die katze' ... } ``` """ languages: list[str] id: Optional[str] = None # Automatically constructed dtype: ClassVar[str] = "dict" pa_type: ClassVar[Any] = None _type: str = field(default="Translation", init=False, repr=False) def __call__(self): return pa.struct({lang: pa.string() for lang in sorted(self.languages)}) def flatten(self) -> Union["FeatureType", dict[str, "FeatureType"]]: """Flatten the Translation feature into a dictionary.""" from .features import Value return {k: Value("string") for k in sorted(self.languages)} @dataclass class TranslationVariableLanguages: """`Feature` for translations with variable languages per example. Here for compatibility with tfds. Args: languages (`dict`): A dictionary for each example mapping string language codes to one or more string translations. The languages present may vary from example to example. Returns: - `language` or `translation` (variable-length 1D `tf.Tensor` of `tf.string`): Language codes sorted in ascending order or plain text translations, sorted to align with language codes. Example: ```python >>> # At construction time: >>> datasets.features.TranslationVariableLanguages(languages=['en', 'fr', 'de']) >>> # During data generation: >>> yield { ... 'en': 'the cat', ... 'fr': ['le chat', 'la chatte,'] ... 'de': 'die katze' ... } >>> # Tensor returned : >>> { ... 'language': ['en', 'de', 'fr', 'fr'], ... 'translation': ['the cat', 'die katze', 'la chatte', 'le chat'], ... } ``` """ languages: Optional[list] = None num_languages: Optional[int] = None id: Optional[str] = None # Automatically constructed dtype: ClassVar[str] = "dict" pa_type: ClassVar[Any] = None _type: str = field(default="TranslationVariableLanguages", init=False, repr=False) def __post_init__(self): self.languages = sorted(set(self.languages)) if self.languages else None self.num_languages = len(self.languages) if self.languages else None def __call__(self): return pa.struct({"language": pa.list_(pa.string()), "translation": pa.list_(pa.string())}) def encode_example(self, translation_dict): lang_set = set(self.languages) if set(translation_dict) == {"language", "translation"}: return translation_dict elif self.languages and set(translation_dict) - lang_set: raise ValueError( f"Some languages in example ({', '.join(sorted(set(translation_dict) - lang_set))}) are not in valid set ({', '.join(lang_set)})." ) # Convert dictionary into tuples, splitting out cases where there are # multiple translations for a single language. translation_tuples = [] for lang, text in translation_dict.items(): if isinstance(text, str): translation_tuples.append((lang, text)) else: translation_tuples.extend([(lang, el) for el in text]) # Ensure translations are in ascending order by language code. languages, translations = zip(*sorted(translation_tuples)) return {"language": languages, "translation": translations} def flatten(self) -> Union["FeatureType", dict[str, "FeatureType"]]: """Flatten the TranslationVariableLanguages feature into a dictionary.""" from .features import Sequence, Value return { "language": Sequence(Value("string")), "translation": Sequence(Value("string")), }
# coding=utf-8 # Copyright 2025 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Fast Image processor class for Pvt.""" from ...image_processing_utils_fast import BaseImageProcessorFast from ...image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling from ...utils import auto_docstring @auto_docstring class PvtImageProcessorFast(BaseImageProcessorFast): resample = PILImageResampling.BILINEAR image_mean = IMAGENET_DEFAULT_MEAN image_std = IMAGENET_DEFAULT_STD size = {"height": 224, "width": 224} default_to_square = True crop_size = None do_resize = True do_center_crop = None do_rescale = True do_normalize = True do_convert_rgb = None model_input_names = ["pixel_values"] __all__ = ["PvtImageProcessorFast"]
# coding=utf-8 # Copyright 2025 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Fast Image processor class for Pvt.""" from ...image_processing_utils_fast import ( BASE_IMAGE_PROCESSOR_FAST_DOCSTRING, BaseImageProcessorFast, ) from ...image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling from ...utils import add_start_docstrings @add_start_docstrings( "Constructs a fast Pvt image processor.", BASE_IMAGE_PROCESSOR_FAST_DOCSTRING, ) class PvtImageProcessorFast(BaseImageProcessorFast): resample = PILImageResampling.BILINEAR image_mean = IMAGENET_DEFAULT_MEAN image_std = IMAGENET_DEFAULT_STD size = {"height": 224, "width": 224} default_to_square = True crop_size = None do_resize = True do_center_crop = None do_rescale = True do_normalize = True do_convert_rgb = None model_input_names = ["pixel_values"] __all__ = ["PvtImageProcessorFast"]
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '3.1.0' short_version = __version__ def parse_version_info(version_str): """Parse a version string into a tuple. Args: version_str (str): The version string. Returns: tuple[int | str]: The version info, e.g., "1.3.0" is parsed into (1, 3, 0), and "2.0.0rc1" is parsed into (2, 0, 0, 'rc1'). """ version_info = [] for x in version_str.split('.'): if x.isdigit(): version_info.append(int(x)) elif x.find('rc') != -1: patch_version = x.split('rc') version_info.append(int(patch_version[0])) version_info.append(f'rc{patch_version[1]}') return tuple(version_info) version_info = parse_version_info(__version__)
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '3.0.0' short_version = __version__ def parse_version_info(version_str): """Parse a version string into a tuple. Args: version_str (str): The version string. Returns: tuple[int | str]: The version info, e.g., "1.3.0" is parsed into (1, 3, 0), and "2.0.0rc1" is parsed into (2, 0, 0, 'rc1'). """ version_info = [] for x in version_str.split('.'): if x.isdigit(): version_info.append(int(x)) elif x.find('rc') != -1: patch_version = x.split('rc') version_info.append(int(patch_version[0])) version_info.append(f'rc{patch_version[1]}') return tuple(version_info) version_info = parse_version_info(__version__)
import random import numpy as np import torch from torchvision import transforms as T from torchvision.transforms import functional as F def pad_if_smaller(img, size, fill=0): min_size = min(img.size) if min_size < size: ow, oh = img.size padh = size - oh if oh < size else 0 padw = size - ow if ow < size else 0 img = F.pad(img, (0, 0, padw, padh), fill=fill) return img class Compose: def __init__(self, transforms): self.transforms = transforms def __call__(self, image, target): for t in self.transforms: image, target = t(image, target) return image, target class RandomResize: def __init__(self, min_size, max_size=None): self.min_size = min_size if max_size is None: max_size = min_size self.max_size = max_size def __call__(self, image, target): size = random.randint(self.min_size, self.max_size) image = F.resize(image, size, antialias=True) target = F.resize(target, size, interpolation=T.InterpolationMode.NEAREST) return image, target class RandomHorizontalFlip: def __init__(self, flip_prob): self.flip_prob = flip_prob def __call__(self, image, target): if random.random() < self.flip_prob: image = F.hflip(image) target = F.hflip(target) return image, target class RandomCrop: def __init__(self, size): self.size = size def __call__(self, image, target): image = pad_if_smaller(image, self.size) target = pad_if_smaller(target, self.size, fill=255) crop_params = T.RandomCrop.get_params(image, (self.size, self.size)) image = F.crop(image, *crop_params) target = F.crop(target, *crop_params) return image, target class CenterCrop: def __init__(self, size): self.size = size def __call__(self, image, target): image = F.center_crop(image, self.size) target = F.center_crop(target, self.size) return image, target class PILToTensor: def __call__(self, image, target): image = F.pil_to_tensor(image) target = torch.as_tensor(np.array(target), dtype=torch.int64) return image, target class ConvertImageDtype: def __init__(self, dtype): self.dtype = dtype def __call__(self, image, target): image = F.convert_image_dtype(image, self.dtype) return image, target class Normalize: def __init__(self, mean, std): self.mean = mean self.std = std def __call__(self, image, target): image = F.normalize(image, mean=self.mean, std=self.std) return image, target
import random import numpy as np import torch from torchvision import transforms as T from torchvision.transforms import functional as F def pad_if_smaller(img, size, fill=0): min_size = min(img.size) if min_size < size: ow, oh = img.size padh = size - oh if oh < size else 0 padw = size - ow if ow < size else 0 img = F.pad(img, (0, 0, padw, padh), fill=fill) return img class Compose: def __init__(self, transforms): self.transforms = transforms def __call__(self, image, target): for t in self.transforms: image, target = t(image, target) return image, target class RandomResize: def __init__(self, min_size, max_size=None): self.min_size = min_size if max_size is None: max_size = min_size self.max_size = max_size def __call__(self, image, target): size = random.randint(self.min_size, self.max_size) image = F.resize(image, size) target = F.resize(target, size, interpolation=T.InterpolationMode.NEAREST) return image, target class RandomHorizontalFlip: def __init__(self, flip_prob): self.flip_prob = flip_prob def __call__(self, image, target): if random.random() < self.flip_prob: image = F.hflip(image) target = F.hflip(target) return image, target class RandomCrop: def __init__(self, size): self.size = size def __call__(self, image, target): image = pad_if_smaller(image, self.size) target = pad_if_smaller(target, self.size, fill=255) crop_params = T.RandomCrop.get_params(image, (self.size, self.size)) image = F.crop(image, *crop_params) target = F.crop(target, *crop_params) return image, target class CenterCrop: def __init__(self, size): self.size = size def __call__(self, image, target): image = F.center_crop(image, self.size) target = F.center_crop(target, self.size) return image, target class PILToTensor: def __call__(self, image, target): image = F.pil_to_tensor(image) target = torch.as_tensor(np.array(target), dtype=torch.int64) return image, target class ConvertImageDtype: def __init__(self, dtype): self.dtype = dtype def __call__(self, image, target): image = F.convert_image_dtype(image, self.dtype) return image, target class Normalize: def __init__(self, mean, std): self.mean = mean self.std = std def __call__(self, image, target): image = F.normalize(image, mean=self.mean, std=self.std) return image, target
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '0.10.6' def parse_version_info(version_str): """Parse the version information. Args: version_str (str): version string like '0.1.0'. Returns: tuple: version information contains major, minor, micro version. """ version_info = [] for x in version_str.split('.'): if x.isdigit(): version_info.append(int(x)) elif x.find('rc') != -1: patch_version = x.split('rc') version_info.append(int(patch_version[0])) version_info.append(f'rc{patch_version[1]}') return tuple(version_info) version_info = parse_version_info(__version__)
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '0.10.5' def parse_version_info(version_str): """Parse the version information. Args: version_str (str): version string like '0.1.0'. Returns: tuple: version information contains major, minor, micro version. """ version_info = [] for x in version_str.split('.'): if x.isdigit(): version_info.append(int(x)) elif x.find('rc') != -1: patch_version = x.split('rc') version_info.append(int(patch_version[0])) version_info.append(f'rc{patch_version[1]}') return tuple(version_info) version_info = parse_version_info(__version__)
from __future__ import annotations import logging from typing import Literal import torch from torch import Tensor from sentence_transformers.models.InputModule import InputModule from .tokenizer import WhitespaceTokenizer logger = logging.getLogger(__name__) class BoW(InputModule): """Implements a Bag-of-Words (BoW) model to derive sentence embeddings. A weighting can be added to allow the generation of tf-idf vectors. The output vector has the size of the vocab. """ save_in_root: bool = False config_keys: list[str] = ["vocab", "word_weights", "unknown_word_weight", "cumulative_term_frequency"] def __init__( self, vocab: list[str], word_weights: dict[str, float] = {}, unknown_word_weight: float = 1, cumulative_term_frequency: bool = True, ): super().__init__() vocab = list(dict.fromkeys(vocab)) # Ensure vocab is unique self.vocab = vocab self.word_weights = word_weights self.unknown_word_weight = unknown_word_weight self.cumulative_term_frequency = cumulative_term_frequency # Maps wordIdx -> word weight self.weights = [] num_unknown_words = 0 for word in vocab: weight = unknown_word_weight if word in word_weights: weight = word_weights[word] elif word.lower() in word_weights: weight = word_weights[word.lower()] else: num_unknown_words += 1 self.weights.append(weight) logger.info( f"{num_unknown_words} out of {len(vocab)} words without a weighting value. Set weight to {unknown_word_weight}" ) self.tokenizer = WhitespaceTokenizer(vocab, stop_words=set(), do_lower_case=False) self.sentence_embedding_dimension = len(vocab) def forward(self, features: dict[str, Tensor]): # Nothing to do, everything is done in get_sentence_features return features def tokenize(self, texts: list[str], **kwargs) -> list[int]: tokenized = [self.tokenizer.tokenize(text, **kwargs) for text in texts] return self.get_sentence_features(tokenized) def get_sentence_embedding_dimension(self): return self.sentence_embedding_dimension def get_sentence_features( self, tokenized_texts: list[list[int]], pad_seq_length: int = 0 ) -> dict[Literal["sentence_embedding"], torch.Tensor]: vectors = [] for tokens in tokenized_texts: vector = torch.zeros(self.get_sentence_embedding_dimension(), dtype=torch.float32) for token in tokens: if self.cumulative_term_frequency: vector[token] += self.weights[token] else: vector[token] = self.weights[token] vectors.append(vector) return {"sentence_embedding": torch.stack(vectors)} def save(self, output_path: str, *args, safe_serialization: bool = True, **kwargs) -> None: self.save_config(output_path)
from __future__ import annotations import json import logging import os from typing import Literal import torch from torch import Tensor, nn from .tokenizer import WhitespaceTokenizer logger = logging.getLogger(__name__) class BoW(nn.Module): """Implements a Bag-of-Words (BoW) model to derive sentence embeddings. A weighting can be added to allow the generation of tf-idf vectors. The output vector has the size of the vocab. """ def __init__( self, vocab: list[str], word_weights: dict[str, float] = {}, unknown_word_weight: float = 1, cumulative_term_frequency: bool = True, ): super().__init__() vocab = list(set(vocab)) # Ensure vocab is unique self.config_keys = ["vocab", "word_weights", "unknown_word_weight", "cumulative_term_frequency"] self.vocab = vocab self.word_weights = word_weights self.unknown_word_weight = unknown_word_weight self.cumulative_term_frequency = cumulative_term_frequency # Maps wordIdx -> word weight self.weights = [] num_unknown_words = 0 for word in vocab: weight = unknown_word_weight if word in word_weights: weight = word_weights[word] elif word.lower() in word_weights: weight = word_weights[word.lower()] else: num_unknown_words += 1 self.weights.append(weight) logger.info( f"{num_unknown_words} out of {len(vocab)} words without a weighting value. Set weight to {unknown_word_weight}" ) self.tokenizer = WhitespaceTokenizer(vocab, stop_words=set(), do_lower_case=False) self.sentence_embedding_dimension = len(vocab) def forward(self, features: dict[str, Tensor]): # Nothing to do, everything is done in get_sentence_features return features def tokenize(self, texts: list[str], **kwargs) -> list[int]: tokenized = [self.tokenizer.tokenize(text, **kwargs) for text in texts] return self.get_sentence_features(tokenized) def get_sentence_embedding_dimension(self): return self.sentence_embedding_dimension def get_sentence_features( self, tokenized_texts: list[list[int]], pad_seq_length: int = 0 ) -> dict[Literal["sentence_embedding"], torch.Tensor]: vectors = [] for tokens in tokenized_texts: vector = torch.zeros(self.get_sentence_embedding_dimension(), dtype=torch.float32) for token in tokens: if self.cumulative_term_frequency: vector[token] += self.weights[token] else: vector[token] = self.weights[token] vectors.append(vector) return {"sentence_embedding": torch.stack(vectors)} def get_config_dict(self): return {key: self.__dict__[key] for key in self.config_keys} def save(self, output_path): with open(os.path.join(output_path, "config.json"), "w") as fOut: json.dump(self.get_config_dict(), fOut, indent=2) @staticmethod def load(input_path): with open(os.path.join(input_path, "config.json")) as fIn: config = json.load(fIn) return BoW(**config)
"""Tests for dask shared by different test modules.""" from typing import Literal import numpy as np import pandas as pd from dask import array as da from dask import dataframe as dd from distributed import Client import xgboost as xgb from xgboost.testing.updater import get_basescore def check_init_estimation_clf( tree_method: str, device: Literal["cpu", "cuda"], client: Client ) -> None: """Test init estimation for classsifier.""" from sklearn.datasets import make_classification X, y = make_classification(n_samples=4096 * 2, n_features=32, random_state=1994) clf = xgb.XGBClassifier( n_estimators=1, max_depth=1, tree_method=tree_method, device=device ) clf.fit(X, y) base_score = get_basescore(clf) dx = da.from_array(X).rechunk(chunks=(32, None)) dy = da.from_array(y).rechunk(chunks=(32,)) dclf = xgb.dask.DaskXGBClassifier( n_estimators=1, max_depth=1, tree_method=tree_method, device=device, ) dclf.client = client dclf.fit(dx, dy) dbase_score = get_basescore(dclf) np.testing.assert_allclose(base_score, dbase_score) def check_init_estimation_reg( tree_method: str, device: Literal["cpu", "cuda"], client: Client ) -> None: """Test init estimation for regressor.""" from sklearn.datasets import make_regression # pylint: disable=unbalanced-tuple-unpacking X, y = make_regression(n_samples=4096 * 2, n_features=32, random_state=1994) reg = xgb.XGBRegressor( n_estimators=1, max_depth=1, tree_method=tree_method, device=device ) reg.fit(X, y) base_score = get_basescore(reg) dx = da.from_array(X).rechunk(chunks=(32, None)) dy = da.from_array(y).rechunk(chunks=(32,)) dreg = xgb.dask.DaskXGBRegressor( n_estimators=1, max_depth=1, tree_method=tree_method, device=device ) dreg.client = client dreg.fit(dx, dy) dbase_score = get_basescore(dreg) np.testing.assert_allclose(base_score, dbase_score) def check_init_estimation( tree_method: str, device: Literal["cpu", "cuda"], client: Client ) -> None: """Test init estimation.""" check_init_estimation_reg(tree_method, device, client) check_init_estimation_clf(tree_method, device, client) def check_uneven_nan( client: Client, tree_method: str, device: Literal["cpu", "cuda"], n_workers: int ) -> None: """Issue #9271, not every worker has missing value.""" assert n_workers >= 2 with client.as_current(): clf = xgb.dask.DaskXGBClassifier(tree_method=tree_method, device=device) X = pd.DataFrame({"a": range(10000), "b": range(10000, 0, -1)}) y = pd.Series([*[0] * 5000, *[1] * 5000]) X.loc[:3000:1000, "a"] = np.nan client.wait_for_workers(n_workers=n_workers) clf.fit( dd.from_pandas(X, npartitions=n_workers), dd.from_pandas(y, npartitions=n_workers), )
"""Tests for dask shared by different test modules.""" import numpy as np import pandas as pd from dask import array as da from dask import dataframe as dd from distributed import Client import xgboost as xgb from xgboost.testing.updater import get_basescore def check_init_estimation_clf(tree_method: str, client: Client) -> None: """Test init estimation for classsifier.""" from sklearn.datasets import make_classification X, y = make_classification(n_samples=4096 * 2, n_features=32, random_state=1994) clf = xgb.XGBClassifier(n_estimators=1, max_depth=1, tree_method=tree_method) clf.fit(X, y) base_score = get_basescore(clf) dx = da.from_array(X).rechunk(chunks=(32, None)) dy = da.from_array(y).rechunk(chunks=(32,)) dclf = xgb.dask.DaskXGBClassifier( n_estimators=1, max_depth=1, tree_method=tree_method ) dclf.client = client dclf.fit(dx, dy) dbase_score = get_basescore(dclf) np.testing.assert_allclose(base_score, dbase_score) def check_init_estimation_reg(tree_method: str, client: Client) -> None: """Test init estimation for regressor.""" from sklearn.datasets import make_regression # pylint: disable=unbalanced-tuple-unpacking X, y = make_regression(n_samples=4096 * 2, n_features=32, random_state=1994) reg = xgb.XGBRegressor(n_estimators=1, max_depth=1, tree_method=tree_method) reg.fit(X, y) base_score = get_basescore(reg) dx = da.from_array(X).rechunk(chunks=(32, None)) dy = da.from_array(y).rechunk(chunks=(32,)) dreg = xgb.dask.DaskXGBRegressor( n_estimators=1, max_depth=1, tree_method=tree_method ) dreg.client = client dreg.fit(dx, dy) dbase_score = get_basescore(dreg) np.testing.assert_allclose(base_score, dbase_score) def check_init_estimation(tree_method: str, client: Client) -> None: """Test init estimation.""" check_init_estimation_reg(tree_method, client) check_init_estimation_clf(tree_method, client) def check_uneven_nan(client: Client, tree_method: str, n_workers: int) -> None: """Issue #9271, not every worker has missing value.""" assert n_workers >= 2 with client.as_current(): clf = xgb.dask.DaskXGBClassifier(tree_method=tree_method) X = pd.DataFrame({"a": range(10000), "b": range(10000, 0, -1)}) y = pd.Series([*[0] * 5000, *[1] * 5000]) X["a"][:3000:1000] = np.nan client.wait_for_workers(n_workers=n_workers) clf.fit( dd.from_pandas(X, npartitions=n_workers), dd.from_pandas(y, npartitions=n_workers), )
# Copyright (c) OpenMMLab. All rights reserved. from unittest.mock import MagicMock, Mock import torch from torch import nn from mmengine.hooks import OptimizerHook class TestOptimizerHook: def test_after_train_iter(self): class Model(nn.Module): def __init__(self): super().__init__() self.conv1 = nn.Conv2d( in_channels=1, out_channels=2, kernel_size=3, stride=1, padding=1, dilation=1) self.conv2 = nn.Conv2d( in_channels=2, out_channels=2, kernel_size=3, stride=1, padding=1, dilation=1) self.conv3 = nn.Conv2d( in_channels=1, out_channels=2, kernel_size=3, stride=1, padding=1, dilation=1) def forward(self, x): x1 = self.conv1(x) x2 = self.conv2(x1) return x1, x2 model = Model() x = torch.rand(1, 1, 3, 3) dummy_runner = MagicMock() dummy_runner.optim_wrapper.zero_grad = Mock(return_value=None) dummy_runner.optim_wrapper.step = Mock(return_value=None) dummy_runner.model = model dummy_runner.outputs = dict() dummy_runner.outputs['num_samples'] = 0 class DummyLogger(): def __init__(self): self.msg = '' def log(self, msg=None, **kwargs): self.msg += msg dummy_runner.logger = DummyLogger() optimizer_hook = OptimizerHook( dict(max_norm=2), detect_anomalous_params=True) dummy_runner.outputs['loss'] = model(x)[0].sum() dummy_runner.outputs['loss'].backward = Mock( wraps=dummy_runner.outputs['loss'].backward) optimizer_hook.detect_anomalous_parameters = Mock( wraps=optimizer_hook.detect_anomalous_parameters) optimizer_hook.clip_grads = Mock(wraps=optimizer_hook.clip_grads) optimizer_hook.after_train_iter(dummy_runner, 0) # assert the parameters of conv2 and conv3 are not in the # computational graph which is with x1.sum() as root. assert 'conv2.weight' in dummy_runner.logger.msg assert 'conv2.bias' in dummy_runner.logger.msg assert 'conv3.weight' in dummy_runner.logger.msg assert 'conv3.bias' in dummy_runner.logger.msg assert 'conv1.weight' not in dummy_runner.logger.msg assert 'conv1.bias' not in dummy_runner.logger.msg dummy_runner.optim_wrapper.step.assert_called() dummy_runner.outputs['loss'].backward.assert_called() optimizer_hook.clip_grads.assert_called() optimizer_hook.detect_anomalous_parameters.assert_called() dummy_runner.outputs['loss'] = model(x)[1].sum() dummy_runner.logger.msg = '' optimizer_hook.after_train_iter(dummy_runner, 0) # assert the parameters of conv3 are not in the computational graph assert 'conv3.weight' in dummy_runner.logger.msg assert 'conv3.bias' in dummy_runner.logger.msg assert 'conv2.weight' not in dummy_runner.logger.msg assert 'conv2.bias' not in dummy_runner.logger.msg assert 'conv1.weight' not in dummy_runner.logger.msg assert 'conv1.bias' not in dummy_runner.logger.msg # grad_clip is None and detect_anomalous_parameters is False optimizer_hook = OptimizerHook(detect_anomalous_params=False) optimizer_hook.detect_anomalous_parameters = Mock( wraps=optimizer_hook.detect_anomalous_parameters) optimizer_hook.clip_grads = Mock(wraps=optimizer_hook.clip_grads) dummy_runner.outputs['loss'] = model(x)[0].sum() dummy_runner.outputs['loss'].backward = Mock( wraps=dummy_runner.outputs['loss'].backward) optimizer_hook.after_train_iter(dummy_runner, 0) dummy_runner.optim_wrapper.step.assert_called() dummy_runner.outputs['loss'].backward.assert_called() optimizer_hook.clip_grads.assert_not_called() optimizer_hook.detect_anomalous_parameters.assert_not_called()
# Copyright (c) OpenMMLab. All rights reserved. from unittest.mock import MagicMock, Mock import torch from torch import nn from mmengine.hooks import OptimizerHook class TestOptimizerHook: def test_after_train_iter(self): class Model(nn.Module): def __init__(self): super().__init__() self.conv1 = nn.Conv2d( in_channels=1, out_channels=2, kernel_size=3, stride=1, padding=1, dilation=1) self.conv2 = nn.Conv2d( in_channels=2, out_channels=2, kernel_size=3, stride=1, padding=1, dilation=1) self.conv3 = nn.Conv2d( in_channels=1, out_channels=2, kernel_size=3, stride=1, padding=1, dilation=1) def forward(self, x): x1 = self.conv1(x) x2 = self.conv2(x1) return x1, x2 model = Model() x = torch.rand(1, 1, 3, 3) dummy_runner = MagicMock() dummy_runner.optimizer.zero_grad = Mock(return_value=None) dummy_runner.optimizer.step = Mock(return_value=None) dummy_runner.model = model dummy_runner.outputs = dict() dummy_runner.outputs['num_samples'] = 0 class DummyLogger(): def __init__(self): self.msg = '' def log(self, msg=None, **kwargs): self.msg += msg dummy_runner.logger = DummyLogger() optimizer_hook = OptimizerHook( dict(max_norm=2), detect_anomalous_params=True) dummy_runner.outputs['loss'] = model(x)[0].sum() dummy_runner.outputs['loss'].backward = Mock( wraps=dummy_runner.outputs['loss'].backward) optimizer_hook.detect_anomalous_parameters = Mock( wraps=optimizer_hook.detect_anomalous_parameters) optimizer_hook.clip_grads = Mock(wraps=optimizer_hook.clip_grads) optimizer_hook.after_train_iter(dummy_runner, 0) # assert the parameters of conv2 and conv3 are not in the # computational graph which is with x1.sum() as root. assert 'conv2.weight' in dummy_runner.logger.msg assert 'conv2.bias' in dummy_runner.logger.msg assert 'conv3.weight' in dummy_runner.logger.msg assert 'conv3.bias' in dummy_runner.logger.msg assert 'conv1.weight' not in dummy_runner.logger.msg assert 'conv1.bias' not in dummy_runner.logger.msg dummy_runner.optimizer.step.assert_called() dummy_runner.outputs['loss'].backward.assert_called() optimizer_hook.clip_grads.assert_called() optimizer_hook.detect_anomalous_parameters.assert_called() dummy_runner.outputs['loss'] = model(x)[1].sum() dummy_runner.logger.msg = '' optimizer_hook.after_train_iter(dummy_runner, 0) # assert the parameters of conv3 are not in the computational graph assert 'conv3.weight' in dummy_runner.logger.msg assert 'conv3.bias' in dummy_runner.logger.msg assert 'conv2.weight' not in dummy_runner.logger.msg assert 'conv2.bias' not in dummy_runner.logger.msg assert 'conv1.weight' not in dummy_runner.logger.msg assert 'conv1.bias' not in dummy_runner.logger.msg # grad_clip is None and detect_anomalous_parameters is False optimizer_hook = OptimizerHook(detect_anomalous_params=False) optimizer_hook.detect_anomalous_parameters = Mock( wraps=optimizer_hook.detect_anomalous_parameters) optimizer_hook.clip_grads = Mock(wraps=optimizer_hook.clip_grads) dummy_runner.outputs['loss'] = model(x)[0].sum() dummy_runner.outputs['loss'].backward = Mock( wraps=dummy_runner.outputs['loss'].backward) optimizer_hook.after_train_iter(dummy_runner, 0) dummy_runner.optimizer.step.assert_called() dummy_runner.outputs['loss'].backward.assert_called() optimizer_hook.clip_grads.assert_not_called() optimizer_hook.detect_anomalous_parameters.assert_not_called()
from __future__ import annotations # TODO: Consider renaming all evaluators to CrossEncoder..., e.g. CrossEncoderNanoBEIREvaluator, CrossEncoderClassificationEvaluator, etc. from .CEBinaryAccuracyEvaluator import CEBinaryAccuracyEvaluator from .CEBinaryClassificationEvaluator import CEBinaryClassificationEvaluator from .CEClassificationEvaluator import CEClassificationEvaluator from .CECorrelationEvaluator import CECorrelationEvaluator from .CEF1Evaluator import CEF1Evaluator from .CENanoBEIREvaluator import CENanoBEIREvaluator from .CERerankingEvaluator import CERerankingEvaluator from .CESoftmaxAccuracyEvaluator import CESoftmaxAccuracyEvaluator __all__ = [ "CEClassificationEvaluator", "CECorrelationEvaluator", "CERerankingEvaluator", "CENanoBEIREvaluator", "CEBinaryAccuracyEvaluator", # Deprecated "CEBinaryClassificationEvaluator", # Deprecated "CEF1Evaluator", # Deprecated "CESoftmaxAccuracyEvaluator", # Deprecated ]
from __future__ import annotations from .CEBinaryAccuracyEvaluator import CEBinaryAccuracyEvaluator from .CEBinaryClassificationEvaluator import CEBinaryClassificationEvaluator from .CECorrelationEvaluator import CECorrelationEvaluator from .CEF1Evaluator import CEF1Evaluator from .CERerankingEvaluator import CERerankingEvaluator from .CESoftmaxAccuracyEvaluator import CESoftmaxAccuracyEvaluator __all__ = [ "CEBinaryAccuracyEvaluator", "CEBinaryClassificationEvaluator", "CECorrelationEvaluator", "CEF1Evaluator", "CESoftmaxAccuracyEvaluator", "CERerankingEvaluator", ]
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .single_stage import SingleStageDetector @DETECTORS.register_module() class VFNet(SingleStageDetector): """Implementation of `VarifocalNet (VFNet).<https://arxiv.org/abs/2008.13367>`_""" def __init__(self, backbone, neck, bbox_head, train_cfg=None, test_cfg=None, pretrained=None, init_cfg=None): super(VFNet, self).__init__(backbone, neck, bbox_head, train_cfg, test_cfg, pretrained, init_cfg)
from ..builder import DETECTORS from .single_stage import SingleStageDetector @DETECTORS.register_module() class VFNet(SingleStageDetector): """Implementation of `VarifocalNet (VFNet).<https://arxiv.org/abs/2008.13367>`_""" def __init__(self, backbone, neck, bbox_head, train_cfg=None, test_cfg=None, pretrained=None, init_cfg=None): super(VFNet, self).__init__(backbone, neck, bbox_head, train_cfg, test_cfg, pretrained, init_cfg)
"""LLM Compiler agent pack.""" from typing import Any, Dict, List, Optional from llama_index.core.agent import AgentRunner from llama_index.core.callbacks import CallbackManager from llama_index.core.llama_pack.base import BaseLlamaPack from llama_index.core.llms.llm import LLM from llama_index.core.tools.types import BaseTool from llama_index.llms.openai import OpenAI from .step import LLMCompilerAgentWorker class LLMCompilerAgentPack(BaseLlamaPack): """ LLMCompilerAgent pack. Args: tools (List[BaseTool]): List of tools to use. llm (Optional[LLM]): LLM to use. """ def __init__( self, tools: List[BaseTool], llm: Optional[LLM] = None, callback_manager: Optional[CallbackManager] = None, agent_worker_kwargs: Optional[Dict[str, Any]] = None, agent_runner_kwargs: Optional[Dict[str, Any]] = None, ) -> None: """Init params.""" self.llm = llm or OpenAI(model="gpt-4") self.callback_manager = callback_manager or self.llm.callback_manager self.agent_worker = LLMCompilerAgentWorker.from_tools( tools, llm=llm, verbose=True, callback_manager=self.callback_manager, **(agent_worker_kwargs or {}) ) self.agent = AgentRunner( self.agent_worker, callback_manager=self.callback_manager, **(agent_runner_kwargs or {}) ) def get_modules(self) -> Dict[str, Any]: """Get modules.""" return { "llm": self.llm, "callback_manager": self.callback_manager, "agent_worker": self.agent_worker, "agent": self.agent, } def run(self, *args: Any, **kwargs: Any) -> Any: """Run the pipeline.""" return self.agent.chat(*args, **kwargs)
"""LLM Compiler agent pack.""" from typing import Any, Dict, List, Optional from llama_index.core.agent import AgentRunner from llama_index.core.callbacks import CallbackManager from llama_index.core.llama_pack.base import BaseLlamaPack from llama_index.core.llms.llm import LLM from llama_index.core.tools.types import BaseTool from llama_index.llms.openai import OpenAI from .step import LLMCompilerAgentWorker class LLMCompilerAgentPack(BaseLlamaPack): """LLMCompilerAgent pack. Args: tools (List[BaseTool]): List of tools to use. llm (Optional[LLM]): LLM to use. """ def __init__( self, tools: List[BaseTool], llm: Optional[LLM] = None, callback_manager: Optional[CallbackManager] = None, agent_worker_kwargs: Optional[Dict[str, Any]] = None, agent_runner_kwargs: Optional[Dict[str, Any]] = None, ) -> None: """Init params.""" self.llm = llm or OpenAI(model="gpt-4") self.callback_manager = callback_manager or self.llm.callback_manager self.agent_worker = LLMCompilerAgentWorker.from_tools( tools, llm=llm, verbose=True, callback_manager=self.callback_manager, **(agent_worker_kwargs or {}) ) self.agent = AgentRunner( self.agent_worker, callback_manager=self.callback_manager, **(agent_runner_kwargs or {}) ) def get_modules(self) -> Dict[str, Any]: """Get modules.""" return { "llm": self.llm, "callback_manager": self.callback_manager, "agent_worker": self.agent_worker, "agent": self.agent, } def run(self, *args: Any, **kwargs: Any) -> Any: """Run the pipeline.""" return self.agent.chat(*args, **kwargs)
"""Test PremChat model""" from typing import cast import pytest from langchain_core.messages import AIMessage, HumanMessage, SystemMessage, ToolMessage from pydantic import SecretStr from pytest import CaptureFixture from langchain_community.chat_models import ChatPremAI from langchain_community.chat_models.premai import ( SINGLE_TOOL_PROMPT_TEMPLATE, TOOL_PROMPT_HEADER, _messages_to_prompt_dict, ) @pytest.mark.requires("premai") def test_api_key_is_string() -> None: llm = ChatPremAI(premai_api_key="secret-api-key", project_id=8) # type: ignore[call-arg] assert isinstance(llm.premai_api_key, SecretStr) @pytest.mark.requires("premai") def test_api_key_masked_when_passed_via_constructor( capsys: CaptureFixture, ) -> None: llm = ChatPremAI(premai_api_key="secret-api-key", project_id=8) # type: ignore[call-arg] print(llm.premai_api_key, end="") # noqa: T201 captured = capsys.readouterr() assert captured.out == "**********" def test_messages_to_prompt_dict_with_valid_messages() -> None: system_message, result = _messages_to_prompt_dict( [ SystemMessage(content="System Prompt"), HumanMessage(content="User message #1"), AIMessage(content="AI message #1"), HumanMessage(content="User message #2"), AIMessage(content="AI message #2"), ToolMessage(content="Tool Message #1", tool_call_id="test_tool"), AIMessage(content="AI message #3"), ] ) expected_tool_message = SINGLE_TOOL_PROMPT_TEMPLATE.format( tool_id="test_tool", tool_response="Tool Message #1" ) expected = [ {"role": "user", "content": "User message #1"}, {"role": "assistant", "content": "AI message #1"}, {"role": "user", "content": "User message #2"}, {"role": "assistant", "content": "AI message #2"}, {"role": "assistant", "content": "AI message #3"}, {"role": "user", "content": TOOL_PROMPT_HEADER + expected_tool_message}, ] assert system_message == "System Prompt" assert result == expected @pytest.mark.requires("premai") def test_premai_initialization() -> None: for model in [ ChatPremAI(model="prem-ai-model", premai_api_key="xyz", project_id=8), # type: ignore[call-arg] ChatPremAI(model_name="prem-ai-model", api_key="xyz", project_id=8), # type: ignore[arg-type] ]: assert model.model == "prem-ai-model" assert model.temperature is None assert model.max_tokens is None assert model.max_retries == 1 assert cast(SecretStr, model.premai_api_key).get_secret_value() == "xyz"
"""Test PremChat model""" from typing import cast import pytest from langchain_core.messages import AIMessage, HumanMessage, SystemMessage, ToolMessage from pydantic import SecretStr from pytest import CaptureFixture from langchain_community.chat_models import ChatPremAI from langchain_community.chat_models.premai import ( SINGLE_TOOL_PROMPT_TEMPLATE, TOOL_PROMPT_HEADER, _messages_to_prompt_dict, ) @pytest.mark.requires("premai") def test_api_key_is_string() -> None: llm = ChatPremAI(premai_api_key="secret-api-key", project_id=8) # type: ignore[call-arg] assert isinstance(llm.premai_api_key, SecretStr) @pytest.mark.requires("premai") def test_api_key_masked_when_passed_via_constructor( capsys: CaptureFixture, ) -> None: llm = ChatPremAI(premai_api_key="secret-api-key", project_id=8) # type: ignore[call-arg] print(llm.premai_api_key, end="") # noqa: T201 captured = capsys.readouterr() assert captured.out == "**********" def test_messages_to_prompt_dict_with_valid_messages() -> None: system_message, result = _messages_to_prompt_dict( [ SystemMessage(content="System Prompt"), HumanMessage(content="User message #1"), AIMessage(content="AI message #1"), HumanMessage(content="User message #2"), AIMessage(content="AI message #2"), ToolMessage(content="Tool Message #1", tool_call_id="test_tool"), AIMessage(content="AI message #3"), ] ) expected_tool_message = SINGLE_TOOL_PROMPT_TEMPLATE.format( tool_id="test_tool", tool_response="Tool Message #1" ) expected = [ {"role": "user", "content": "User message #1"}, {"role": "assistant", "content": "AI message #1"}, {"role": "user", "content": "User message #2"}, {"role": "assistant", "content": "AI message #2"}, {"role": "assistant", "content": "AI message #3"}, {"role": "user", "content": TOOL_PROMPT_HEADER + expected_tool_message}, ] assert system_message == "System Prompt" assert result == expected @pytest.mark.requires("premai") def test_premai_initialization() -> None: for model in [ ChatPremAI(model="prem-ai-model", premai_api_key="xyz", project_id=8), # type: ignore[call-arg] ChatPremAI(model_name="prem-ai-model", api_key="xyz", project_id=8), # type: ignore[arg-type, call-arg] ]: assert model.model == "prem-ai-model" assert model.temperature is None assert model.max_tokens is None assert model.max_retries == 1 assert cast(SecretStr, model.premai_api_key).get_secret_value() == "xyz"
from __future__ import annotations try: from typing import Self except ImportError: from typing_extensions import Self import torch from torch import Tensor, nn from sentence_transformers.models.Module import Module class WeightedLayerPooling(Module): """Token embeddings are weighted mean of their different hidden layer representations""" config_keys: list[str] = ["word_embedding_dimension", "layer_start", "num_hidden_layers"] def __init__( self, word_embedding_dimension, num_hidden_layers: int = 12, layer_start: int = 4, layer_weights=None ): super().__init__() self.word_embedding_dimension = word_embedding_dimension self.layer_start = layer_start self.num_hidden_layers = num_hidden_layers self.layer_weights = ( layer_weights if layer_weights is not None else nn.Parameter(torch.tensor([1] * (num_hidden_layers + 1 - layer_start), dtype=torch.float)) ) def forward(self, features: dict[str, Tensor]): ft_all_layers = features["all_layer_embeddings"] all_layer_embedding = torch.stack(ft_all_layers) all_layer_embedding = all_layer_embedding[self.layer_start :, :, :, :] # Start from 4th layers output weight_factor = self.layer_weights.unsqueeze(-1).unsqueeze(-1).unsqueeze(-1).expand(all_layer_embedding.size()) weighted_average = (weight_factor * all_layer_embedding).sum(dim=0) / self.layer_weights.sum() features.update({"token_embeddings": weighted_average}) return features def get_word_embedding_dimension(self): return self.word_embedding_dimension def save(self, output_path: str, *args, safe_serialization: bool = True, **kwargs) -> None: self.save_config(output_path) self.save_torch_weights(output_path, safe_serialization=safe_serialization) @classmethod def load( cls, model_name_or_path: str, subfolder: str = "", token: bool | str | None = None, cache_folder: str | None = None, revision: str | None = None, local_files_only: bool = False, **kwargs, ) -> Self: hub_kwargs = { "subfolder": subfolder, "token": token, "cache_folder": cache_folder, "revision": revision, "local_files_only": local_files_only, } config = cls.load_config(model_name_or_path=model_name_or_path, **hub_kwargs) model = cls(**config) model = cls.load_torch_weights(model_name_or_path=model_name_or_path, model=model, **hub_kwargs) return model
from __future__ import annotations import json import os import torch from safetensors.torch import load_model as load_safetensors_model from safetensors.torch import save_model as save_safetensors_model from torch import Tensor, nn class WeightedLayerPooling(nn.Module): """Token embeddings are weighted mean of their different hidden layer representations""" def __init__( self, word_embedding_dimension, num_hidden_layers: int = 12, layer_start: int = 4, layer_weights=None ): super().__init__() self.config_keys = ["word_embedding_dimension", "layer_start", "num_hidden_layers"] self.word_embedding_dimension = word_embedding_dimension self.layer_start = layer_start self.num_hidden_layers = num_hidden_layers self.layer_weights = ( layer_weights if layer_weights is not None else nn.Parameter(torch.tensor([1] * (num_hidden_layers + 1 - layer_start), dtype=torch.float)) ) def forward(self, features: dict[str, Tensor]): ft_all_layers = features["all_layer_embeddings"] all_layer_embedding = torch.stack(ft_all_layers) all_layer_embedding = all_layer_embedding[self.layer_start :, :, :, :] # Start from 4th layers output weight_factor = self.layer_weights.unsqueeze(-1).unsqueeze(-1).unsqueeze(-1).expand(all_layer_embedding.size()) weighted_average = (weight_factor * all_layer_embedding).sum(dim=0) / self.layer_weights.sum() features.update({"token_embeddings": weighted_average}) return features def get_word_embedding_dimension(self): return self.word_embedding_dimension def get_config_dict(self): return {key: self.__dict__[key] for key in self.config_keys} def save(self, output_path: str, safe_serialization: bool = True): with open(os.path.join(output_path, "config.json"), "w") as fOut: json.dump(self.get_config_dict(), fOut, indent=2) if safe_serialization: save_safetensors_model(self, os.path.join(output_path, "model.safetensors")) else: torch.save(self.state_dict(), os.path.join(output_path, "pytorch_model.bin")) @staticmethod def load(input_path): with open(os.path.join(input_path, "config.json")) as fIn: config = json.load(fIn) model = WeightedLayerPooling(**config) if os.path.exists(os.path.join(input_path, "model.safetensors")): load_safetensors_model(model, os.path.join(input_path, "model.safetensors")) else: model.load_state_dict( torch.load( os.path.join(input_path, "pytorch_model.bin"), map_location=torch.device("cpu"), weights_only=True ) ) return model
import logging import re from typing import Any import uvicorn.config from colorama import Fore def remove_color_codes(s: str) -> str: return re.sub(r"\x1B(?:[@-Z\\-_]|\[[0-?]*[ -/]*[@-~])", "", s) def fmt_kwargs(kwargs: dict) -> str: return ", ".join(f"{n}={repr(v)}" for n, v in kwargs.items()) def print_attribute( title: str, value: Any, title_color: str = Fore.GREEN, value_color: str = "" ) -> None: logger = logging.getLogger() logger.info( str(value), extra={ "title": f"{title.rstrip(':')}:", "title_color": title_color, "color": value_color, }, ) def generate_uvicorn_config(): """ Generates a uvicorn logging config that silences uvicorn's default logging and tells it to use the native logging module. """ log_config = dict(uvicorn.config.LOGGING_CONFIG) log_config["loggers"]["uvicorn"] = {"handlers": []} log_config["loggers"]["uvicorn.error"] = {"handlers": []} log_config["loggers"]["uvicorn.access"] = {"handlers": []} return log_config
import logging import re from typing import Any import uvicorn.config from colorama import Fore def remove_color_codes(s: str) -> str: return re.sub(r"\x1B(?:[@-Z\\-_]|\[[0-?]*[ -/]*[@-~])", "", s) def fmt_kwargs(kwargs: dict) -> str: return ", ".join(f"{n}={repr(v)}" for n, v in kwargs.items()) def print_attribute( title: str, value: Any, title_color: str = Fore.GREEN, value_color: str = "" ) -> None: logger = logging.getLogger() logger.info( str(value), extra={ "title": f"{title.rstrip(':')}:", "title_color": title_color, "color": value_color, }, ) def generate_uvicorn_config(): """ Generates a uvicorn logging config that silences uvicorn's default logging and tells it to use the native logging module. """ log_config = dict(uvicorn.config.LOGGING_CONFIG) log_config["loggers"]["uvicorn"] = {"handlers": []} log_config["loggers"]["uvicorn.error"] = {"handlers": []} log_config["loggers"]["uvicorn.access"] = {"handlers": []} return log_config
import torch from torchaudio_unittest.common_utils import PytorchTestCase, skipIfNoCuda from .tacotron2_loss_impl import Tacotron2LossGradcheckTests, Tacotron2LossShapeTests, Tacotron2LossTorchscriptTests @skipIfNoCuda class TestTacotron2LossShapeFloat32CUDA(PytorchTestCase, Tacotron2LossShapeTests): dtype = torch.float32 device = torch.device("cuda") @skipIfNoCuda class TestTacotron2TorchsciptFloat32CUDA(PytorchTestCase, Tacotron2LossTorchscriptTests): dtype = torch.float32 device = torch.device("cuda") @skipIfNoCuda class TestTacotron2GradcheckFloat64CUDA(PytorchTestCase, Tacotron2LossGradcheckTests): dtype = torch.float64 # gradcheck needs a higher numerical accuracy device = torch.device("cuda")
import torch from torchaudio_unittest.common_utils import PytorchTestCase, skipIfNoCuda from .tacotron2_loss_impl import ( Tacotron2LossGradcheckTests, Tacotron2LossShapeTests, Tacotron2LossTorchscriptTests, ) @skipIfNoCuda class TestTacotron2LossShapeFloat32CUDA(PytorchTestCase, Tacotron2LossShapeTests): dtype = torch.float32 device = torch.device("cuda") @skipIfNoCuda class TestTacotron2TorchsciptFloat32CUDA(PytorchTestCase, Tacotron2LossTorchscriptTests): dtype = torch.float32 device = torch.device("cuda") @skipIfNoCuda class TestTacotron2GradcheckFloat64CUDA(PytorchTestCase, Tacotron2LossGradcheckTests): dtype = torch.float64 # gradcheck needs a higher numerical accuracy device = torch.device("cuda")
_base_ = '../faster_rcnn/faster-rcnn_r50_fpn_1x_coco.py' model = dict( backbone=dict( plugins=[ dict( cfg=dict( type='GeneralizedAttention', spatial_range=-1, num_heads=8, attention_type='1111', kv_stride=2), stages=(False, False, True, True), position='after_conv2') ], dcn=dict(type='DCN', deform_groups=1, fallback_on_stride=False), stage_with_dcn=(False, True, True, True)))
_base_ = '../faster_rcnn/faster_rcnn_r50_fpn_1x_coco.py' model = dict( backbone=dict( plugins=[ dict( cfg=dict( type='GeneralizedAttention', spatial_range=-1, num_heads=8, attention_type='1111', kv_stride=2), stages=(False, False, True, True), position='after_conv2') ], dcn=dict(type='DCN', deform_groups=1, fallback_on_stride=False), stage_with_dcn=(False, True, True, True)))
from ._dsp import oscillator_bank from .functional import add_noise, barkscale_fbanks, convolve, fftconvolve __all__ = [ "add_noise", "barkscale_fbanks", "convolve", "fftconvolve", "oscillator_bank", ]
from .functional import add_noise, barkscale_fbanks, convolve, fftconvolve __all__ = ["add_noise", "barkscale_fbanks", "convolve", "fftconvolve"]
_base_ = './mask-rcnn_x101-32x4d_fpn_gn-ws-all_2x_coco.py' # learning policy max_epochs = 24 train_cfg = dict(max_epochs=max_epochs) # learning rate param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[20, 23], gamma=0.1) ]
_base_ = './mask_rcnn_x101_32x4d_fpn_gn_ws-all_2x_coco.py' # learning policy max_epochs = 24 train_cfg = dict(max_epochs=max_epochs) # learning rate param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[20, 23], gamma=0.1) ]
from __future__ import annotations from collections.abc import Iterable from torch import Tensor from sentence_transformers import util from sentence_transformers.losses.CoSENTLoss import CoSENTLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseCoSENTLoss(CoSENTLoss): def __init__(self, model: SparseEncoder, scale: float = 20.0, similarity_fct=util.cos_sim) -> None: """ This class implements CoSENT (Cosine Sentence). It expects that each of the InputExamples consists of a pair of texts and a float valued label, representing the expected similarity score between the pair. It computes the following loss function: ``loss = logsum(1+exp(s(i,j)-s(k,l))+exp...)``, where ``(i,j)`` and ``(k,l)`` are any of the input pairs in the batch such that the expected similarity of ``(i,j)`` is greater than ``(k,l)``. The summation is over all possible pairs of input pairs in the batch that match this condition. Args: model: SparseEncoder similarity_fct: Function to compute the PAIRWISE similarity between embeddings. Default is ``util.pairwise_cos_sim``. scale: Output of similarity function is multiplied by scale value. Represents the inverse temperature. References: - For further details, see: https://kexue.fm/archives/8847 Requirements: - Need to be used in SpladeLoss or CSRLoss as a loss function. - Sentence pairs with corresponding similarity scores in range of the similarity function. Default is [-1,1]. Inputs: +--------------------------------+------------------------+ | Texts | Labels | +================================+========================+ | (sentence_A, sentence_B) pairs | float similarity score | +--------------------------------+------------------------+ Relations: - :class:`SparseAnglELoss` is SparseCoSENTLoss with ``pairwise_angle_sim`` as the metric, rather than ``pairwise_cos_sim``. Example: :: from datasets import Dataset from sentence_transformers.sparse_encoder import SparseEncoder, SparseEncoderTrainer, losses model = SparseEncoder("distilbert/distilbert-base-uncased") train_dataset = Dataset.from_dict( { "sentence1": ["It's nice weather outside today.", "He drove to work."], "sentence2": ["It's so sunny.", "She walked to the store."], "score": [1.0, 0.3], } ) loss = losses.SpladeLoss( model=model, loss=losses.SparseCoSENTLoss(model), corpus_regularizer_weight=5e-5, use_corpus_regularizer_only=True ) trainer = SparseEncoderTrainer(model=model, train_dataset=train_dataset, loss=loss) trainer.train() """ model.similarity_fn_name = "cosine" return super().__init__(model, scale=scale, similarity_fct=similarity_fct) def forward(self, sentence_features: Iterable[dict[str, Tensor]], labels: Tensor) -> Tensor: raise AttributeError("SparseCoSENTLoss should not be used alone. Use it with SpladeLoss or CSRLoss.")
from __future__ import annotations from collections.abc import Iterable from torch import Tensor from sentence_transformers import util from sentence_transformers.losses.CoSENTLoss import CoSENTLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseCoSENTLoss(CoSENTLoss): def __init__(self, model: SparseEncoder, scale: float = 20.0, similarity_fct=util.cos_sim) -> None: """ This class implements CoSENT (Cosine Sentence). It expects that each of the InputExamples consists of a pair of texts and a float valued label, representing the expected similarity score between the pair. It computes the following loss function: ``loss = logsum(1+exp(s(i,j)-s(k,l))+exp...)``, where ``(i,j)`` and ``(k,l)`` are any of the input pairs in the batch such that the expected similarity of ``(i,j)`` is greater than ``(k,l)``. The summation is over all possible pairs of input pairs in the batch that match this condition. Args: model: SparseEncoder similarity_fct: Function to compute the PAIRWISE similarity between embeddings. Default is ``util.pairwise_cos_sim``. scale: Output of similarity function is multiplied by scale value. Represents the inverse temperature. References: - For further details, see: https://kexue.fm/archives/8847 Requirements: - Need to be used in SpladeLoss or CSRLoss as a loss function. - Sentence pairs with corresponding similarity scores in range of the similarity function. Default is [-1,1]. Inputs: +--------------------------------+------------------------+ | Texts | Labels | +================================+========================+ | (sentence_A, sentence_B) pairs | float similarity score | +--------------------------------+------------------------+ Relations: - :class:`SparseAnglELoss` is SparseCoSENTLoss with ``pairwise_angle_sim`` as the metric, rather than ``pairwise_cos_sim``. Example: :: from datasets import Dataset from sentence_transformers.sparse_encoder import SparseEncoder, SparseEncoderTrainer, losses model = SparseEncoder("distilbert/distilbert-base-uncased") train_dataset = Dataset.from_dict( { "sentence1": ["It's nice weather outside today.", "He drove to work."], "sentence2": ["It's so sunny.", "She walked to the store."], "score": [1.0, 0.3], } ) loss = losses.SpladeLoss(model=model, loss=losses.SparseCoSENTLoss(model), lambda_corpus=5e-5, all_docs=True) trainer = SparseEncoderTrainer(model=model, train_dataset=train_dataset, loss=loss) trainer.train() """ model.similarity_fn_name = "cosine" return super().__init__(model, scale=scale, similarity_fct=similarity_fct) def forward(self, sentence_features: Iterable[dict[str, Tensor]], labels: Tensor) -> Tensor: raise AttributeError("SparseCoSENTLoss should not be used alone. Use it with SpladeLoss or CSRLoss.")
import warnings from abc import ABC from typing import Any, Optional from langchain_core._api import deprecated from langchain_core.chat_history import ( BaseChatMessageHistory, InMemoryChatMessageHistory, ) from langchain_core.memory import BaseMemory from langchain_core.messages import AIMessage, HumanMessage from pydantic import Field from langchain.memory.utils import get_prompt_input_key @deprecated( since="0.3.1", removal="1.0.0", message=( "Please see the migration guide at: " "https://python.langchain.com/docs/versions/migrating_memory/" ), ) class BaseChatMemory(BaseMemory, ABC): """Abstract base class for chat memory. **ATTENTION** This abstraction was created prior to when chat models had native tool calling capabilities. It does **NOT** support native tool calling capabilities for chat models and will fail SILENTLY if used with a chat model that has native tool calling. DO NOT USE THIS ABSTRACTION FOR NEW CODE. """ chat_memory: BaseChatMessageHistory = Field( default_factory=InMemoryChatMessageHistory ) output_key: Optional[str] = None input_key: Optional[str] = None return_messages: bool = False def _get_input_output( self, inputs: dict[str, Any], outputs: dict[str, str] ) -> tuple[str, str]: if self.input_key is None: prompt_input_key = get_prompt_input_key(inputs, self.memory_variables) else: prompt_input_key = self.input_key if self.output_key is None: if len(outputs) == 1: output_key = list(outputs.keys())[0] elif "output" in outputs: output_key = "output" warnings.warn( f"'{self.__class__.__name__}' got multiple output keys:" f" {outputs.keys()}. The default 'output' key is being used." f" If this is not desired, please manually set 'output_key'." ) else: raise ValueError( f"Got multiple output keys: {outputs.keys()}, cannot " f"determine which to store in memory. Please set the " f"'output_key' explicitly." ) else: output_key = self.output_key return inputs[prompt_input_key], outputs[output_key] def save_context(self, inputs: dict[str, Any], outputs: dict[str, str]) -> None: """Save context from this conversation to buffer.""" input_str, output_str = self._get_input_output(inputs, outputs) self.chat_memory.add_messages( [ HumanMessage(content=input_str), AIMessage(content=output_str), ] ) async def asave_context( self, inputs: dict[str, Any], outputs: dict[str, str] ) -> None: """Save context from this conversation to buffer.""" input_str, output_str = self._get_input_output(inputs, outputs) await self.chat_memory.aadd_messages( [ HumanMessage(content=input_str), AIMessage(content=output_str), ] ) def clear(self) -> None: """Clear memory contents.""" self.chat_memory.clear() async def aclear(self) -> None: """Clear memory contents.""" await self.chat_memory.aclear()
import warnings from abc import ABC from typing import Any, Dict, Optional, Tuple from langchain_core._api import deprecated from langchain_core.chat_history import ( BaseChatMessageHistory, InMemoryChatMessageHistory, ) from langchain_core.memory import BaseMemory from langchain_core.messages import AIMessage, HumanMessage from pydantic import Field from langchain.memory.utils import get_prompt_input_key @deprecated( since="0.3.1", removal="1.0.0", message=( "Please see the migration guide at: " "https://python.langchain.com/docs/versions/migrating_memory/" ), ) class BaseChatMemory(BaseMemory, ABC): """Abstract base class for chat memory. **ATTENTION** This abstraction was created prior to when chat models had native tool calling capabilities. It does **NOT** support native tool calling capabilities for chat models and will fail SILENTLY if used with a chat model that has native tool calling. DO NOT USE THIS ABSTRACTION FOR NEW CODE. """ chat_memory: BaseChatMessageHistory = Field( default_factory=InMemoryChatMessageHistory ) output_key: Optional[str] = None input_key: Optional[str] = None return_messages: bool = False def _get_input_output( self, inputs: Dict[str, Any], outputs: Dict[str, str] ) -> Tuple[str, str]: if self.input_key is None: prompt_input_key = get_prompt_input_key(inputs, self.memory_variables) else: prompt_input_key = self.input_key if self.output_key is None: if len(outputs) == 1: output_key = list(outputs.keys())[0] elif "output" in outputs: output_key = "output" warnings.warn( f"'{self.__class__.__name__}' got multiple output keys:" f" {outputs.keys()}. The default 'output' key is being used." f" If this is not desired, please manually set 'output_key'." ) else: raise ValueError( f"Got multiple output keys: {outputs.keys()}, cannot " f"determine which to store in memory. Please set the " f"'output_key' explicitly." ) else: output_key = self.output_key return inputs[prompt_input_key], outputs[output_key] def save_context(self, inputs: Dict[str, Any], outputs: Dict[str, str]) -> None: """Save context from this conversation to buffer.""" input_str, output_str = self._get_input_output(inputs, outputs) self.chat_memory.add_messages( [ HumanMessage(content=input_str), AIMessage(content=output_str), ] ) async def asave_context( self, inputs: Dict[str, Any], outputs: Dict[str, str] ) -> None: """Save context from this conversation to buffer.""" input_str, output_str = self._get_input_output(inputs, outputs) await self.chat_memory.aadd_messages( [ HumanMessage(content=input_str), AIMessage(content=output_str), ] ) def clear(self) -> None: """Clear memory contents.""" self.chat_memory.clear() async def aclear(self) -> None: """Clear memory contents.""" await self.chat_memory.aclear()
# Copyright 2024 The OpenXLA Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== from absl.testing import absltest import numpy as np from xla.codegen.testlib import _extension from xla.codegen.testlib import utilities as testlib_utilities from xla.python import xla_extension create_literal = testlib_utilities.create_literal_from_np class HloModuleParse(absltest.TestCase): def test_from_instruction(self): shape = xla_extension.Shape.array_shape(np.dtype(np.int32), (4,)) hlo_parameter = _extension.HloInstruction.create_parameter( 0, shape, "input" ) hlo_op = _extension.HloInstruction.create_variadic( shape, _extension.HloOpcode.sine, [hlo_parameter] ) hlo_module = _extension.HloModule(hlo_op.name() + "_module") hlo_module.add_entry_computation( _extension.build_hlo_computation(hlo_op, hlo_parameter) ) expected_parts = [ "HloModule sine_module,", "{", "%input = s32[4]{0} parameter(0)", "ROOT %sine = s32[4]{0} sine(%input)", "}", ] self.assertContainsInOrder( expected_parts, str(hlo_module), ) class LiteralFromNpTest(absltest.TestCase): def test_output_same_as_input(self): array = np.array([1, 2, 3, 4], dtype=np.int32) got = create_literal(array) np.testing.assert_array_equal(np.asarray(got), array) class DummyKernelRunnerTest(absltest.TestCase): def test_dummy_kernel(self): runner = _extension.DummyAddKernelRunner() in_arg1 = create_literal(np.array([1, 2, 3, 4], dtype=np.int32)) in_arg2 = create_literal(np.array([5, 6, 7, 8], dtype=np.int32)) out_arg = create_literal(np.array([0, 0, 0, 0], dtype=np.int32)) runner.call([in_arg1, in_arg2, out_arg]) np.testing.assert_array_equal( np.asarray(out_arg), np.asarray(in_arg1) + np.asarray(in_arg2) ) if __name__ == "__main__": absltest.main()
# Copyright 2024 The OpenXLA Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== from absl.testing import absltest import numpy as np from xla.codegen.testlib import _extension from xla.codegen.testlib import utilities as testlib_utilities create_literal = testlib_utilities.create_literal_from_np class LiteralFromNpTest(absltest.TestCase): def test_output_same_as_input(self): array = np.array([1, 2, 3, 4], dtype=np.int32) got = create_literal(array) np.testing.assert_array_equal(np.asarray(got), array) class DummyKernelRunnerTest(absltest.TestCase): def test_dummy_kernel(self): runner = _extension.DummyAddKernelRunner() in_arg1 = create_literal(np.array([1, 2, 3, 4], dtype=np.int32)) in_arg2 = create_literal(np.array([5, 6, 7, 8], dtype=np.int32)) out_arg = create_literal(np.array([0, 0, 0, 0], dtype=np.int32)) runner.call([in_arg1, in_arg2, out_arg]) np.testing.assert_array_equal( np.asarray(out_arg), np.asarray(in_arg1) + np.asarray(in_arg2) ) if __name__ == "__main__": absltest.main()
import numpy as np import pytest import torch from docarray import BaseDocument from docarray.base_document import AnyDocument from docarray.typing import ( AnyEmbedding, AnyUrl, ImageUrl, Mesh3DUrl, NdArray, PointCloud3DUrl, TextUrl, TorchTensor, ) @pytest.mark.proto def test_proto_all_types(): class Mymmdoc(BaseDocument): tensor: NdArray torch_tensor: TorchTensor embedding: AnyEmbedding any_url: AnyUrl image_url: ImageUrl text_url: TextUrl mesh_url: Mesh3DUrl point_cloud_url: PointCloud3DUrl doc = Mymmdoc( tensor=np.zeros((3, 224, 224)), torch_tensor=torch.zeros((3, 224, 224)), embedding=np.zeros((100, 1)), any_url='http://jina.ai', image_url='http://jina.ai/bla.jpg', text_url='http://jina.ai', mesh_url='http://jina.ai/mesh.obj', point_cloud_url='http://jina.ai/mesh.obj', ) new_doc = AnyDocument.from_protobuf(doc.to_protobuf()) for field, value in new_doc: if field == 'embedding': # embedding is a Union type, not supported by isinstance assert isinstance(value, np.ndarray) or isinstance(value, torch.Tensor) else: assert isinstance(value, doc._get_field_type(field)) @pytest.mark.tensorflow def test_proto_all_types_proto3(): import tensorflow as tf from docarray.typing import TensorFlowTensor class Mymmdoc(BaseDocument): tensor: NdArray torch_tensor: TorchTensor tf_tensor: TensorFlowTensor embedding: AnyEmbedding any_url: AnyUrl image_url: ImageUrl text_url: TextUrl mesh_url: Mesh3DUrl point_cloud_url: PointCloud3DUrl doc = Mymmdoc( tensor=np.zeros((3, 224, 224)), torch_tensor=torch.zeros((3, 224, 224)), tf_tensor=tf.zeros((3, 224, 224)), embedding=np.zeros((100, 1)), any_url='http://jina.ai', image_url='http://jina.ai/bla.jpg', text_url='http://jina.ai', mesh_url='http://jina.ai/mesh.obj', point_cloud_url='http://jina.ai/mesh.obj', ) new_doc = AnyDocument.from_protobuf(doc.to_protobuf()) for field, value in new_doc: if field == 'embedding': # embedding is a Union type, not supported by isinstance assert isinstance(value, np.ndarray) or isinstance(value, torch.Tensor) else: assert isinstance(value, doc._get_field_type(field))
import numpy as np import torch from docarray import BaseDocument from docarray.base_document import AnyDocument from docarray.typing import ( AnyEmbedding, AnyUrl, ImageUrl, Mesh3DUrl, NdArray, PointCloud3DUrl, TextUrl, TorchTensor, ) def test_proto_all_types(): class Mymmdoc(BaseDocument): tensor: NdArray torch_tensor: TorchTensor embedding: AnyEmbedding any_url: AnyUrl image_url: ImageUrl text_url: TextUrl mesh_url: Mesh3DUrl point_cloud_url: PointCloud3DUrl doc = Mymmdoc( tensor=np.zeros((3, 224, 224)), torch_tensor=torch.zeros((3, 224, 224)), embedding=np.zeros((100, 1)), any_url='http://jina.ai', image_url='http://jina.ai/bla.jpg', text_url='http://jina.ai', mesh_url='http://jina.ai/mesh.obj', point_cloud_url='http://jina.ai/mesh.obj', ) new_doc = AnyDocument.from_protobuf(doc.to_protobuf()) for field, value in new_doc: if field == 'embedding': # embedding is a Union type, not supported by isinstance assert isinstance(value, np.ndarray) or isinstance(value, torch.Tensor) else: assert isinstance(value, doc._get_field_type(field))
from typing import Any, Optional, Type, TypeVar, Union from docarray.base_document import BaseDocument from docarray.typing import TextUrl from docarray.typing.tensor.embedding import AnyEmbedding T = TypeVar('T', bound='Text') class Text(BaseDocument): """ Document for handling text. It can contain a TextUrl (`Text.url`), a str (`Text.text`), and an AnyEmbedding (`Text.embedding`). EXAMPLE USAGE: You can use this Document directly: .. code-block:: python from docarray.documents import Text # use it directly txt_doc = Text(url='http://www.jina.ai/') txt_doc.text = txt_doc.url.load() model = MyEmbeddingModel() txt_doc.embedding = model(txt_doc.text) You can extend this Document: .. code-block:: python from docarray.documents import Text from docarray.typing import AnyEmbedding from typing import Optional # extend it class MyText(Text): second_embedding: Optional[AnyEmbedding] txt_doc = MyText(url='http://www.jina.ai/') txt_doc.text = txt_doc.url.load() model = MyEmbeddingModel() txt_doc.embedding = model(txt_doc.text) txt_doc.second_embedding = model(txt_doc.text) You can use this Document for composition: .. code-block:: python from docarray import BaseDocument from docarray.documents import Image, Text # compose it class MultiModalDoc(BaseDocument): image_doc: Image text_doc: Text mmdoc = MultiModalDoc( image_doc=Image(url="http://www.jina.ai/image.jpg"), text_doc=Text(text="hello world, how are you doing?"), ) mmdoc.text_doc.text = mmdoc.text_doc.url.load() """ text: Optional[str] = None url: Optional[TextUrl] = None embedding: Optional[AnyEmbedding] = None @classmethod def validate( cls: Type[T], value: Union[str, Any], ) -> T: if isinstance(value, str): value = cls(text=value) return super().validate(value)
from typing import Optional from docarray.base_document import BaseDocument from docarray.typing import TextUrl from docarray.typing.tensor.embedding import AnyEmbedding class Text(BaseDocument): """ Document for handling text. It can contain a TextUrl (`Text.url`), a str (`Text.text`), and an AnyEmbedding (`Text.embedding`). EXAMPLE USAGE: You can use this Document directly: .. code-block:: python from docarray.documents import Text # use it directly txt_doc = Text(url='http://www.jina.ai/') txt_doc.text = txt_doc.url.load() model = MyEmbeddingModel() txt_doc.embedding = model(txt_doc.text) You can extend this Document: .. code-block:: python from docarray.documents import Text from docarray.typing import AnyEmbedding from typing import Optional # extend it class MyText(Text): second_embedding: Optional[AnyEmbedding] txt_doc = MyText(url='http://www.jina.ai/') txt_doc.text = txt_doc.url.load() model = MyEmbeddingModel() txt_doc.embedding = model(txt_doc.text) txt_doc.second_embedding = model(txt_doc.text) You can use this Document for composition: .. code-block:: python from docarray import BaseDocument from docarray.documents import Image, Text # compose it class MultiModalDoc(BaseDocument): image_doc: Image text_doc: Text mmdoc = MultiModalDoc( image_doc=Image(url="http://www.jina.ai/image.jpg"), text_doc=Text(text="hello world, how are you doing?"), ) mmdoc.text_doc.text = mmdoc.text_doc.url.load() """ text: Optional[str] = None url: Optional[TextUrl] = None embedding: Optional[AnyEmbedding] = None
# Copyright (c) OpenMMLab. All rights reserved. import torch.nn as nn from mmcv.cnn import ConvModule from mmcv.ops import MaskedConv2d from ..builder import HEADS from .guided_anchor_head import FeatureAdaption, GuidedAnchorHead @HEADS.register_module() class GARetinaHead(GuidedAnchorHead): """Guided-Anchor-based RetinaNet head.""" def __init__(self, num_classes, in_channels, stacked_convs=4, conv_cfg=None, norm_cfg=None, init_cfg=None, **kwargs): if init_cfg is None: init_cfg = dict( type='Normal', layer='Conv2d', std=0.01, override=[ dict( type='Normal', name='conv_loc', std=0.01, bias_prob=0.01), dict( type='Normal', name='retina_cls', std=0.01, bias_prob=0.01) ]) self.stacked_convs = stacked_convs self.conv_cfg = conv_cfg self.norm_cfg = norm_cfg super(GARetinaHead, self).__init__( num_classes, in_channels, init_cfg=init_cfg, **kwargs) def _init_layers(self): """Initialize layers of the head.""" self.relu = nn.ReLU(inplace=True) self.cls_convs = nn.ModuleList() self.reg_convs = nn.ModuleList() for i in range(self.stacked_convs): chn = self.in_channels if i == 0 else self.feat_channels self.cls_convs.append( ConvModule( chn, self.feat_channels, 3, stride=1, padding=1, conv_cfg=self.conv_cfg, norm_cfg=self.norm_cfg)) self.reg_convs.append( ConvModule( chn, self.feat_channels, 3, stride=1, padding=1, conv_cfg=self.conv_cfg, norm_cfg=self.norm_cfg)) self.conv_loc = nn.Conv2d(self.feat_channels, 1, 1) self.conv_shape = nn.Conv2d(self.feat_channels, self.num_anchors * 2, 1) self.feature_adaption_cls = FeatureAdaption( self.feat_channels, self.feat_channels, kernel_size=3, deform_groups=self.deform_groups) self.feature_adaption_reg = FeatureAdaption( self.feat_channels, self.feat_channels, kernel_size=3, deform_groups=self.deform_groups) self.retina_cls = MaskedConv2d( self.feat_channels, self.num_anchors * self.cls_out_channels, 3, padding=1) self.retina_reg = MaskedConv2d( self.feat_channels, self.num_anchors * 4, 3, padding=1) def forward_single(self, x): """Forward feature map of a single scale level.""" cls_feat = x reg_feat = x for cls_conv in self.cls_convs: cls_feat = cls_conv(cls_feat) for reg_conv in self.reg_convs: reg_feat = reg_conv(reg_feat) loc_pred = self.conv_loc(cls_feat) shape_pred = self.conv_shape(reg_feat) cls_feat = self.feature_adaption_cls(cls_feat, shape_pred) reg_feat = self.feature_adaption_reg(reg_feat, shape_pred) if not self.training: mask = loc_pred.sigmoid()[0] >= self.loc_filter_thr else: mask = None cls_score = self.retina_cls(cls_feat, mask) bbox_pred = self.retina_reg(reg_feat, mask) return cls_score, bbox_pred, shape_pred, loc_pred
import torch.nn as nn from mmcv.cnn import ConvModule from mmcv.ops import MaskedConv2d from ..builder import HEADS from .guided_anchor_head import FeatureAdaption, GuidedAnchorHead @HEADS.register_module() class GARetinaHead(GuidedAnchorHead): """Guided-Anchor-based RetinaNet head.""" def __init__(self, num_classes, in_channels, stacked_convs=4, conv_cfg=None, norm_cfg=None, init_cfg=None, **kwargs): if init_cfg is None: init_cfg = dict( type='Normal', layer='Conv2d', std=0.01, override=[ dict( type='Normal', name='conv_loc', std=0.01, bias_prob=0.01), dict( type='Normal', name='retina_cls', std=0.01, bias_prob=0.01) ]) self.stacked_convs = stacked_convs self.conv_cfg = conv_cfg self.norm_cfg = norm_cfg super(GARetinaHead, self).__init__( num_classes, in_channels, init_cfg=init_cfg, **kwargs) def _init_layers(self): """Initialize layers of the head.""" self.relu = nn.ReLU(inplace=True) self.cls_convs = nn.ModuleList() self.reg_convs = nn.ModuleList() for i in range(self.stacked_convs): chn = self.in_channels if i == 0 else self.feat_channels self.cls_convs.append( ConvModule( chn, self.feat_channels, 3, stride=1, padding=1, conv_cfg=self.conv_cfg, norm_cfg=self.norm_cfg)) self.reg_convs.append( ConvModule( chn, self.feat_channels, 3, stride=1, padding=1, conv_cfg=self.conv_cfg, norm_cfg=self.norm_cfg)) self.conv_loc = nn.Conv2d(self.feat_channels, 1, 1) self.conv_shape = nn.Conv2d(self.feat_channels, self.num_anchors * 2, 1) self.feature_adaption_cls = FeatureAdaption( self.feat_channels, self.feat_channels, kernel_size=3, deform_groups=self.deform_groups) self.feature_adaption_reg = FeatureAdaption( self.feat_channels, self.feat_channels, kernel_size=3, deform_groups=self.deform_groups) self.retina_cls = MaskedConv2d( self.feat_channels, self.num_anchors * self.cls_out_channels, 3, padding=1) self.retina_reg = MaskedConv2d( self.feat_channels, self.num_anchors * 4, 3, padding=1) def forward_single(self, x): """Forward feature map of a single scale level.""" cls_feat = x reg_feat = x for cls_conv in self.cls_convs: cls_feat = cls_conv(cls_feat) for reg_conv in self.reg_convs: reg_feat = reg_conv(reg_feat) loc_pred = self.conv_loc(cls_feat) shape_pred = self.conv_shape(reg_feat) cls_feat = self.feature_adaption_cls(cls_feat, shape_pred) reg_feat = self.feature_adaption_reg(reg_feat, shape_pred) if not self.training: mask = loc_pred.sigmoid()[0] >= self.loc_filter_thr else: mask = None cls_score = self.retina_cls(cls_feat, mask) bbox_pred = self.retina_reg(reg_feat, mask) return cls_score, bbox_pred, shape_pred, loc_pred
"""Init file of LlamaIndex.""" __version__ = "0.12.11" import logging from logging import NullHandler from typing import Callable, Optional try: # Force pants to install eval_type_backport on 3.9 import eval_type_backport # noqa # type: ignore except ImportError: pass # response from llama_index.core.base.response.schema import Response # import global eval handler from llama_index.core.callbacks.global_handlers import set_global_handler from llama_index.core.data_structs.struct_type import IndexStructType from llama_index.core.embeddings.mock_embed_model import MockEmbedding # indices # loading from llama_index.core.indices import ( ComposableGraph, DocumentSummaryIndex, GPTDocumentSummaryIndex, GPTKeywordTableIndex, GPTListIndex, GPTRAKEKeywordTableIndex, GPTSimpleKeywordTableIndex, GPTTreeIndex, GPTVectorStoreIndex, KeywordTableIndex, KnowledgeGraphIndex, ListIndex, PropertyGraphIndex, RAKEKeywordTableIndex, SimpleKeywordTableIndex, SummaryIndex, TreeIndex, VectorStoreIndex, load_graph_from_storage, load_index_from_storage, load_indices_from_storage, ) # structured from llama_index.core.indices.common.struct_store.base import ( SQLDocumentContextBuilder, ) # prompt helper from llama_index.core.indices.prompt_helper import PromptHelper # prompts from llama_index.core.prompts import ( BasePromptTemplate, ChatPromptTemplate, # backwards compatibility Prompt, PromptTemplate, SelectorPromptTemplate, ) from llama_index.core.readers import SimpleDirectoryReader, download_loader # Response Synthesizer from llama_index.core.response_synthesizers.factory import get_response_synthesizer from llama_index.core.schema import Document, QueryBundle from llama_index.core.service_context import ( ServiceContext, set_global_service_context, ) # global settings from llama_index.core.settings import Settings # storage from llama_index.core.storage.storage_context import StorageContext # sql wrapper from llama_index.core.utilities.sql_wrapper import SQLDatabase # global tokenizer from llama_index.core.utils import get_tokenizer, set_global_tokenizer # best practices for library logging: # https://docs.python.org/3/howto/logging.html#configuring-logging-for-a-library logging.getLogger(__name__).addHandler(NullHandler()) __all__ = [ "StorageContext", "ServiceContext", "ComposableGraph", # indices "SummaryIndex", "VectorStoreIndex", "SimpleKeywordTableIndex", "KeywordTableIndex", "RAKEKeywordTableIndex", "TreeIndex", "DocumentSummaryIndex", "KnowledgeGraphIndex", "PropertyGraphIndex", # indices - legacy names "GPTKeywordTableIndex", "GPTKnowledgeGraphIndex", "GPTSimpleKeywordTableIndex", "GPTRAKEKeywordTableIndex", "GPTListIndex", "ListIndex", "GPTTreeIndex", "GPTVectorStoreIndex", "GPTDocumentSummaryIndex", "Prompt", "PromptTemplate", "BasePromptTemplate", "ChatPromptTemplate", "SelectorPromptTemplate", "SummaryPrompt", "TreeInsertPrompt", "TreeSelectPrompt", "TreeSelectMultiplePrompt", "RefinePrompt", "QuestionAnswerPrompt", "KeywordExtractPrompt", "QueryKeywordExtractPrompt", "Response", "Document", "SimpleDirectoryReader", "VellumPredictor", "VellumPromptRegistry", "MockEmbedding", "SQLDatabase", "SQLDocumentContextBuilder", "SQLContextBuilder", "PromptHelper", "IndexStructType", "download_loader", "load_graph_from_storage", "load_index_from_storage", "load_indices_from_storage", "QueryBundle", "get_response_synthesizer", "set_global_service_context", "set_global_handler", "set_global_tokenizer", "get_tokenizer", "Settings", ] # eval global toggle from llama_index.core.callbacks.base_handler import BaseCallbackHandler global_handler: Optional[BaseCallbackHandler] = None # NOTE: keep for backwards compatibility SQLContextBuilder = SQLDocumentContextBuilder # global tokenizer global_tokenizer: Optional[Callable[[str], list]] = None
"""Init file of LlamaIndex.""" __version__ = "0.12.10" import logging from logging import NullHandler from typing import Callable, Optional try: # Force pants to install eval_type_backport on 3.9 import eval_type_backport # noqa # type: ignore except ImportError: pass # response from llama_index.core.base.response.schema import Response # import global eval handler from llama_index.core.callbacks.global_handlers import set_global_handler from llama_index.core.data_structs.struct_type import IndexStructType from llama_index.core.embeddings.mock_embed_model import MockEmbedding # indices # loading from llama_index.core.indices import ( ComposableGraph, DocumentSummaryIndex, GPTDocumentSummaryIndex, GPTKeywordTableIndex, GPTListIndex, GPTRAKEKeywordTableIndex, GPTSimpleKeywordTableIndex, GPTTreeIndex, GPTVectorStoreIndex, KeywordTableIndex, KnowledgeGraphIndex, ListIndex, PropertyGraphIndex, RAKEKeywordTableIndex, SimpleKeywordTableIndex, SummaryIndex, TreeIndex, VectorStoreIndex, load_graph_from_storage, load_index_from_storage, load_indices_from_storage, ) # structured from llama_index.core.indices.common.struct_store.base import ( SQLDocumentContextBuilder, ) # prompt helper from llama_index.core.indices.prompt_helper import PromptHelper # prompts from llama_index.core.prompts import ( BasePromptTemplate, ChatPromptTemplate, # backwards compatibility Prompt, PromptTemplate, SelectorPromptTemplate, ) from llama_index.core.readers import SimpleDirectoryReader, download_loader # Response Synthesizer from llama_index.core.response_synthesizers.factory import get_response_synthesizer from llama_index.core.schema import Document, QueryBundle from llama_index.core.service_context import ( ServiceContext, set_global_service_context, ) # global settings from llama_index.core.settings import Settings # storage from llama_index.core.storage.storage_context import StorageContext # sql wrapper from llama_index.core.utilities.sql_wrapper import SQLDatabase # global tokenizer from llama_index.core.utils import get_tokenizer, set_global_tokenizer # best practices for library logging: # https://docs.python.org/3/howto/logging.html#configuring-logging-for-a-library logging.getLogger(__name__).addHandler(NullHandler()) __all__ = [ "StorageContext", "ServiceContext", "ComposableGraph", # indices "SummaryIndex", "VectorStoreIndex", "SimpleKeywordTableIndex", "KeywordTableIndex", "RAKEKeywordTableIndex", "TreeIndex", "DocumentSummaryIndex", "KnowledgeGraphIndex", "PropertyGraphIndex", # indices - legacy names "GPTKeywordTableIndex", "GPTKnowledgeGraphIndex", "GPTSimpleKeywordTableIndex", "GPTRAKEKeywordTableIndex", "GPTListIndex", "ListIndex", "GPTTreeIndex", "GPTVectorStoreIndex", "GPTDocumentSummaryIndex", "Prompt", "PromptTemplate", "BasePromptTemplate", "ChatPromptTemplate", "SelectorPromptTemplate", "SummaryPrompt", "TreeInsertPrompt", "TreeSelectPrompt", "TreeSelectMultiplePrompt", "RefinePrompt", "QuestionAnswerPrompt", "KeywordExtractPrompt", "QueryKeywordExtractPrompt", "Response", "Document", "SimpleDirectoryReader", "VellumPredictor", "VellumPromptRegistry", "MockEmbedding", "SQLDatabase", "SQLDocumentContextBuilder", "SQLContextBuilder", "PromptHelper", "IndexStructType", "download_loader", "load_graph_from_storage", "load_index_from_storage", "load_indices_from_storage", "QueryBundle", "get_response_synthesizer", "set_global_service_context", "set_global_handler", "set_global_tokenizer", "get_tokenizer", "Settings", ] # eval global toggle from llama_index.core.callbacks.base_handler import BaseCallbackHandler global_handler: Optional[BaseCallbackHandler] = None # NOTE: keep for backwards compatibility SQLContextBuilder = SQLDocumentContextBuilder # global tokenizer global_tokenizer: Optional[Callable[[str], list]] = None
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] dataset_type = 'CocoDataset' data_root = 'data/coco/' # file_client_args = dict( # backend='petrel', # path_mapping=dict({ # './data/': 's3://openmmlab/datasets/detection/', # 'data/': 's3://openmmlab/datasets/detection/' # })) file_client_args = dict(backend='disk') # model settings preprocess_cfg = dict( mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True) model = dict( type='CenterNet', preprocess_cfg=preprocess_cfg, backbone=dict( type='ResNet', depth=18, norm_eval=False, norm_cfg=dict(type='BN'), init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet18')), neck=dict( type='CTResNetNeck', in_channels=512, num_deconv_filters=(256, 128, 64), num_deconv_kernels=(4, 4, 4), use_dcn=True), bbox_head=dict( type='CenterNetHead', num_classes=80, in_channels=64, feat_channels=64, loss_center_heatmap=dict(type='GaussianFocalLoss', loss_weight=1.0), loss_wh=dict(type='L1Loss', loss_weight=0.1), loss_offset=dict(type='L1Loss', loss_weight=1.0)), train_cfg=None, test_cfg=dict(topk=100, local_maximum_kernel=3, max_per_img=100)) train_pipeline = [ dict(type='LoadImageFromFile', file_client_args=file_client_args), dict(type='LoadAnnotations', with_bbox=True), dict( type='PhotoMetricDistortion', brightness_delta=32, contrast_range=(0.5, 1.5), saturation_range=(0.5, 1.5), hue_delta=18), dict( type='RandomCenterCropPad', # The cropped images are padded into squares during training, # but may be less than crop_size. crop_size=(512, 512), ratios=(0.6, 0.7, 0.8, 0.9, 1.0, 1.1, 1.2, 1.3), mean=[0, 0, 0], std=[1, 1, 1], to_rgb=True, test_pad_mode=None), # Make sure the output is always crop_size. dict(type='Resize', scale=(512, 512), keep_ratio=True), dict(type='RandomFlip', prob=0.5), dict(type='PackDetInputs') ] test_pipeline = [ dict( type='LoadImageFromFile', to_float32=True, file_client_args=file_client_args), # don't need Resize dict( type='RandomCenterCropPad', ratios=None, border=None, mean=[0, 0, 0], std=[1, 1, 1], to_rgb=True, test_mode=True, test_pad_mode=['logical_or', 31], test_pad_add_pix=1), dict( type='PackDetInputs', meta_keys=('img_id', 'img_path', 'ori_shape', 'img_shape', 'border')) ] # Use RepeatDataset to speed up training train_dataloader = dict( batch_size=16, num_workers=4, persistent_workers=True, sampler=dict(type='DefaultSampler', shuffle=True), dataset=dict( _delete_=True, type='RepeatDataset', times=5, dataset=dict( type=dataset_type, data_root=data_root, ann_file='annotations/instances_train2017.json', data_prefix=dict(img='train2017/'), filter_cfg=dict(filter_empty_gt=True, min_size=32), pipeline=train_pipeline))) val_dataloader = dict(dataset=dict(pipeline=test_pipeline)) test_dataloader = val_dataloader # optimizer # Based on the default settings of modern detectors, the SGD effect is better # than the Adam in the source code, so we use SGD default settings and # if you use adam+lr5e-4, the map is 29.1. default_hooks = dict( optimizer=dict( _delete_=True, type='OptimizerHook', grad_clip=dict(max_norm=35, norm_type=2))) max_epochs = 28 # learning policy # Based on the default settings of modern detectors, we added warmup settings. param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=1000), dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[18, 24], # the real step is [18*5, 24*5] gamma=0.1) ] train_cfg = dict(max_epochs=max_epochs) # the real epoch is 28*5=140 # NOTE: `auto_scale_lr` is for automatically scaling LR, # USER SHOULD NOT CHANGE ITS VALUES. # base_batch_size = (8 GPUs) x (16 samples per GPU) auto_scale_lr = dict(base_batch_size=128)
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( type='CenterNet', backbone=dict( type='ResNet', depth=18, norm_eval=False, norm_cfg=dict(type='BN'), init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet18')), neck=dict( type='CTResNetNeck', in_channel=512, num_deconv_filters=(256, 128, 64), num_deconv_kernels=(4, 4, 4), use_dcn=True), bbox_head=dict( type='CenterNetHead', num_classes=80, in_channel=64, feat_channel=64, loss_center_heatmap=dict(type='GaussianFocalLoss', loss_weight=1.0), loss_wh=dict(type='L1Loss', loss_weight=0.1), loss_offset=dict(type='L1Loss', loss_weight=1.0)), train_cfg=None, test_cfg=dict(topk=100, local_maximum_kernel=3, max_per_img=100)) # We fixed the incorrect img_norm_cfg problem in the source code. img_norm_cfg = dict( mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True) train_pipeline = [ dict(type='LoadImageFromFile', to_float32=True, color_type='color'), dict(type='LoadAnnotations', with_bbox=True), dict( type='PhotoMetricDistortion', brightness_delta=32, contrast_range=(0.5, 1.5), saturation_range=(0.5, 1.5), hue_delta=18), dict( type='RandomCenterCropPad', crop_size=(512, 512), ratios=(0.6, 0.7, 0.8, 0.9, 1.0, 1.1, 1.2, 1.3), mean=[0, 0, 0], std=[1, 1, 1], to_rgb=True, test_pad_mode=None), dict(type='Resize', img_scale=(512, 512), keep_ratio=True), dict(type='RandomFlip', flip_ratio=0.5), dict(type='Normalize', **img_norm_cfg), dict(type='DefaultFormatBundle'), dict(type='Collect', keys=['img', 'gt_bboxes', 'gt_labels']) ] test_pipeline = [ dict(type='LoadImageFromFile', to_float32=True), dict( type='MultiScaleFlipAug', scale_factor=1.0, flip=False, transforms=[ dict(type='Resize', keep_ratio=True), dict( type='RandomCenterCropPad', ratios=None, border=None, mean=[0, 0, 0], std=[1, 1, 1], to_rgb=True, test_mode=True, test_pad_mode=['logical_or', 31], test_pad_add_pix=1), dict(type='RandomFlip'), dict(type='Normalize', **img_norm_cfg), dict(type='DefaultFormatBundle'), dict( type='Collect', meta_keys=('filename', 'ori_filename', 'ori_shape', 'img_shape', 'pad_shape', 'scale_factor', 'flip', 'flip_direction', 'img_norm_cfg', 'border'), keys=['img']) ]) ] dataset_type = 'CocoDataset' data_root = 'data/coco/' # Use RepeatDataset to speed up training data = dict( samples_per_gpu=16, workers_per_gpu=4, train=dict( _delete_=True, type='RepeatDataset', times=5, dataset=dict( type=dataset_type, ann_file=data_root + 'annotations/instances_train2017.json', img_prefix=data_root + 'train2017/', pipeline=train_pipeline)), val=dict(pipeline=test_pipeline), test=dict(pipeline=test_pipeline)) # optimizer # Based on the default settings of modern detectors, the SGD effect is better # than the Adam in the source code, so we use SGD default settings and # if you use adam+lr5e-4, the map is 29.1. optimizer_config = dict( _delete_=True, grad_clip=dict(max_norm=35, norm_type=2)) # learning policy # Based on the default settings of modern detectors, we added warmup settings. lr_config = dict( policy='step', warmup='linear', warmup_iters=1000, warmup_ratio=1.0 / 1000, step=[18, 24]) # the real step is [18*5, 24*5] runner = dict(max_epochs=28) # the real epoch is 28*5=140 # NOTE: `auto_scale_lr` is for automatically scaling LR, # USER SHOULD NOT CHANGE ITS VALUES. # base_batch_size = (8 GPUs) x (16 samples per GPU) auto_scale_lr = dict(base_batch_size=128)
import PIL.Image import pytest import torch import torchvision.transforms.v2.utils from common_utils import DEFAULT_SIZE, make_bounding_box, make_detection_mask, make_image from torchvision import datapoints from torchvision.transforms.v2.functional import to_pil_image from torchvision.transforms.v2.utils import has_all, has_any IMAGE = make_image(DEFAULT_SIZE, color_space="RGB") BOUNDING_BOX = make_bounding_box(DEFAULT_SIZE, format=datapoints.BoundingBoxFormat.XYXY) MASK = make_detection_mask(DEFAULT_SIZE) @pytest.mark.parametrize( ("sample", "types", "expected"), [ ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Image,), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.BoundingBoxes,), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Mask,), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.Mask), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.BoundingBoxes, datapoints.Mask), True), ((MASK,), (datapoints.Image, datapoints.BoundingBoxes), False), ((BOUNDING_BOX,), (datapoints.Image, datapoints.Mask), False), ((IMAGE,), (datapoints.BoundingBoxes, datapoints.Mask), False), ( (IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), True, ), ((), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), False), ((IMAGE, BOUNDING_BOX, MASK), (lambda obj: isinstance(obj, datapoints.Image),), True), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: False,), False), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: True,), True), ((IMAGE,), (datapoints.Image, PIL.Image.Image, torchvision.transforms.v2.utils.is_simple_tensor), True), ( (torch.Tensor(IMAGE),), (datapoints.Image, PIL.Image.Image, torchvision.transforms.v2.utils.is_simple_tensor), True, ), ( (to_pil_image(IMAGE),), (datapoints.Image, PIL.Image.Image, torchvision.transforms.v2.utils.is_simple_tensor), True, ), ], ) def test_has_any(sample, types, expected): assert has_any(sample, *types) is expected @pytest.mark.parametrize( ("sample", "types", "expected"), [ ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Image,), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.BoundingBoxes,), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Mask,), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.Mask), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.BoundingBoxes, datapoints.Mask), True), ( (IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), True, ), ((BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes), False), ((BOUNDING_BOX, MASK), (datapoints.Image, datapoints.Mask), False), ((IMAGE, MASK), (datapoints.BoundingBoxes, datapoints.Mask), False), ( (IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), True, ), ((BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), False), ((IMAGE, MASK), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), False), ((IMAGE, BOUNDING_BOX), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), False), ( (IMAGE, BOUNDING_BOX, MASK), (lambda obj: isinstance(obj, (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask)),), True, ), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: False,), False), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: True,), True), ], ) def test_has_all(sample, types, expected): assert has_all(sample, *types) is expected
import PIL.Image import pytest import torch import torchvision.transforms.v2.utils from common_utils import DEFAULT_SIZE, make_bounding_box, make_detection_mask, make_image from torchvision import datapoints from torchvision.transforms.v2.functional import to_image_pil from torchvision.transforms.v2.utils import has_all, has_any IMAGE = make_image(DEFAULT_SIZE, color_space="RGB") BOUNDING_BOX = make_bounding_box(DEFAULT_SIZE, format=datapoints.BoundingBoxFormat.XYXY) MASK = make_detection_mask(DEFAULT_SIZE) @pytest.mark.parametrize( ("sample", "types", "expected"), [ ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Image,), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.BoundingBoxes,), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Mask,), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.Mask), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.BoundingBoxes, datapoints.Mask), True), ((MASK,), (datapoints.Image, datapoints.BoundingBoxes), False), ((BOUNDING_BOX,), (datapoints.Image, datapoints.Mask), False), ((IMAGE,), (datapoints.BoundingBoxes, datapoints.Mask), False), ( (IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), True, ), ((), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), False), ((IMAGE, BOUNDING_BOX, MASK), (lambda obj: isinstance(obj, datapoints.Image),), True), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: False,), False), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: True,), True), ((IMAGE,), (datapoints.Image, PIL.Image.Image, torchvision.transforms.v2.utils.is_simple_tensor), True), ( (torch.Tensor(IMAGE),), (datapoints.Image, PIL.Image.Image, torchvision.transforms.v2.utils.is_simple_tensor), True, ), ( (to_image_pil(IMAGE),), (datapoints.Image, PIL.Image.Image, torchvision.transforms.v2.utils.is_simple_tensor), True, ), ], ) def test_has_any(sample, types, expected): assert has_any(sample, *types) is expected @pytest.mark.parametrize( ("sample", "types", "expected"), [ ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Image,), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.BoundingBoxes,), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Mask,), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.Mask), True), ((IMAGE, BOUNDING_BOX, MASK), (datapoints.BoundingBoxes, datapoints.Mask), True), ( (IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), True, ), ((BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes), False), ((BOUNDING_BOX, MASK), (datapoints.Image, datapoints.Mask), False), ((IMAGE, MASK), (datapoints.BoundingBoxes, datapoints.Mask), False), ( (IMAGE, BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), True, ), ((BOUNDING_BOX, MASK), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), False), ((IMAGE, MASK), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), False), ((IMAGE, BOUNDING_BOX), (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask), False), ( (IMAGE, BOUNDING_BOX, MASK), (lambda obj: isinstance(obj, (datapoints.Image, datapoints.BoundingBoxes, datapoints.Mask)),), True, ), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: False,), False), ((IMAGE, BOUNDING_BOX, MASK), (lambda _: True,), True), ], ) def test_has_all(sample, types, expected): assert has_all(sample, *types) is expected
"""Hubspot reader.""" from typing import List from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class HubspotReader(BaseReader): """ Hubspot reader. Reads data from a Hubspot account. Args: access_token(str): Hubspot API key. """ def __init__(self, access_token: str) -> None: """Initialize Hubspot reader.""" self.access_token = access_token def load_data(self) -> List[Document]: """ Load deals, contacts and companies data from Hubspot. Returns: List[Document]: List of documents, where each document represensts a list of Hubspot objects """ from hubspot import HubSpot api_client = HubSpot(access_token=self.access_token) all_deals = api_client.crm.deals.get_all() all_contacts = api_client.crm.contacts.get_all() all_companies = api_client.crm.companies.get_all() return [ Document( text=f"{all_deals}".replace("\n", ""), extra_info={"type": "deals"} ), Document( text=f"{all_contacts}".replace("\n", ""), extra_info={"type": "contacts"}, ), Document( text=f"{all_companies}".replace("\n", ""), extra_info={"type": "companies"}, ), ]
"""Hubspot reader.""" from typing import List from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class HubspotReader(BaseReader): """Hubspot reader. Reads data from a Hubspot account. Args: access_token(str): Hubspot API key. """ def __init__(self, access_token: str) -> None: """Initialize Hubspot reader.""" self.access_token = access_token def load_data(self) -> List[Document]: """Load deals, contacts and companies data from Hubspot. Returns: List[Document]: List of documents, where each document represensts a list of Hubspot objects """ from hubspot import HubSpot api_client = HubSpot(access_token=self.access_token) all_deals = api_client.crm.deals.get_all() all_contacts = api_client.crm.contacts.get_all() all_companies = api_client.crm.companies.get_all() return [ Document( text=f"{all_deals}".replace("\n", ""), extra_info={"type": "deals"} ), Document( text=f"{all_contacts}".replace("\n", ""), extra_info={"type": "contacts"}, ), Document( text=f"{all_companies}".replace("\n", ""), extra_info={"type": "companies"}, ), ]
"""LangChain **Runnable** and the **LangChain Expression Language (LCEL)**. The LangChain Expression Language (LCEL) offers a declarative method to build production-grade programs that harness the power of LLMs. Programs created using LCEL and LangChain Runnables inherently support synchronous, asynchronous, batch, and streaming operations. Support for **async** allows servers hosting LCEL based programs to scale better for higher concurrent loads. **Streaming** of intermediate outputs as they're being generated allows for creating more responsive UX. This module contains schema and implementation of LangChain Runnables primitives. """ from langchain_core.runnables.base import ( Runnable, RunnableBinding, RunnableGenerator, RunnableLambda, RunnableMap, RunnableParallel, RunnableSequence, RunnableSerializable, ) from langchain_core.runnables.branch import RunnableBranch from langchain_core.runnables.config import RunnableConfig, patch_config from langchain_core.runnables.fallbacks import RunnableWithFallbacks from langchain_core.runnables.passthrough import RunnablePassthrough from langchain_core.runnables.router import RouterInput, RouterRunnable from langchain_core.runnables.utils import ( ConfigurableField, ConfigurableFieldMultiOption, ConfigurableFieldSingleOption, ) __all__ = [ "ConfigurableField", "ConfigurableFieldMultiOption", "ConfigurableFieldSingleOption", "RouterInput", "RouterRunnable", "Runnable", "RunnableBinding", "RunnableBranch", "RunnableConfig", "RunnableGenerator", "RunnableLambda", "RunnableMap", "RunnableParallel", "RunnablePassthrough", "RunnableSequence", "RunnableSerializable", "RunnableWithFallbacks", "patch_config", ]
"""LangChain **Runnable** and the **LangChain Expression Language (LCEL)**. The LangChain Expression Language (LCEL) offers a declarative method to build production-grade programs that harness the power of LLMs. Programs created using LCEL and LangChain Runnables inherently support synchronous, asynchronous, batch, and streaming operations. Support for **async** allows servers hosting LCEL based programs to scale better for higher concurrent loads. **Streaming** of intermediate outputs as they're being generated allows for creating more responsive UX. This module contains schema and implementation of LangChain Runnables primitives. """ from langchain_core.runnables.base import ( Runnable, RunnableBinding, RunnableGenerator, RunnableLambda, RunnableMap, RunnableParallel, RunnableSequence, RunnableSerializable, ) from langchain_core.runnables.branch import RunnableBranch from langchain_core.runnables.config import RunnableConfig, patch_config from langchain_core.runnables.fallbacks import RunnableWithFallbacks from langchain_core.runnables.passthrough import RunnablePassthrough from langchain_core.runnables.router import RouterInput, RouterRunnable from langchain_core.runnables.utils import ( ConfigurableField, ConfigurableFieldMultiOption, ConfigurableFieldSingleOption, ) __all__ = [ "ConfigurableField", "ConfigurableFieldSingleOption", "ConfigurableFieldMultiOption", "patch_config", "RouterInput", "RouterRunnable", "Runnable", "RunnableSerializable", "RunnableBinding", "RunnableBranch", "RunnableConfig", "RunnableGenerator", "RunnableLambda", "RunnableMap", "RunnableParallel", "RunnablePassthrough", "RunnableSequence", "RunnableWithFallbacks", ]
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings as _warnings import docarray as _docarray if _sys.version_info < (3, 7, 0): raise OSError(f'Jina requires Python >= 3.7, but yours is {_sys.version_info}') def _warning_on_one_line(message, category, filename, lineno, *args, **kwargs): return '\033[1;33m%s: %s\033[0m \033[1;30m(raised from %s:%s)\033[0m\n' % ( category.__name__, message, filename, lineno, ) def _ignore_google_warnings(): import warnings warnings.filterwarnings( 'ignore', category=DeprecationWarning, message='Deprecated call to `pkg_resources.declare_namespace(\'google\')`.', append=True, ) _warnings.formatwarning = _warning_on_one_line _warnings.simplefilter('always', DeprecationWarning, append=True) _ignore_google_warnings() # fix fork error on MacOS but seems no effect? must do EXPORT manually before jina start _os.environ['OBJC_DISABLE_INITIALIZE_FORK_SAFETY'] = 'YES' # JINA_MP_START_METHOD has higher priority than os-patch _start_method = _os.environ.get('JINA_MP_START_METHOD', None) if _start_method and _start_method.lower() in {'fork', 'spawn', 'forkserver'}: from multiprocessing import set_start_method as _set_start_method try: _set_start_method(_start_method.lower()) _warnings.warn( f'multiprocessing start method is set to `{_start_method.lower()}`' ) except Exception as e: _warnings.warn( f'failed to set multiprocessing start_method to `{_start_method.lower()}`: {e!r}' ) elif _sys.version_info >= (3, 8, 0) and _platform.system() == 'Darwin': # DO SOME OS-WISE PATCHES # temporary fix for python 3.8 on macos where the default start is set to "spawn" # https://docs.python.org/3/library/multiprocessing.html#contexts-and-start-methods from multiprocessing import set_start_method as _set_start_method try: _set_start_method('fork') _warnings.warn(f'multiprocessing start method is set to `fork`') except Exception as e: _warnings.warn(f'failed to set multiprocessing start_method to `fork`: {e!r}') # do not change this line manually # this is managed by git tag and updated on every release # NOTE: this represents the NEXT release version __version__ = '3.23.6' # do not change this line manually # this is managed by proto/build-proto.sh and updated on every execution __proto_version__ = '0.1.27' try: __docarray_version__ = _docarray.__version__ except AttributeError as e: raise RuntimeError( '`docarray` dependency is not installed correctly, please reinstall with `pip install -U --force-reinstall docarray`' ) try: _signal.signal(_signal.SIGINT, _signal.default_int_handler) except Exception as exc: _warnings.warn(f'failed to set default signal handler: {exc!r}`') def _set_nofile(nofile_atleast=4096): """ Set nofile soft limit to at least 4096, useful for running matlplotlib/seaborn on parallel executing plot generators vs. Ubuntu default ulimit -n 1024 or OS X El Captian 256 temporary setting extinguishing with Python session. :param nofile_atleast: nofile soft limit :return: nofile soft limit and nofile hard limit """ try: import resource as res except ImportError: # Windows res = None if res is None: return (None,) * 2 soft, ohard = res.getrlimit(res.RLIMIT_NOFILE) hard = ohard if soft < nofile_atleast: soft = nofile_atleast if hard < soft: hard = soft try: res.setrlimit(res.RLIMIT_NOFILE, (soft, hard)) except (ValueError, res.error): try: hard = soft print(f'trouble with max limit, retrying with soft,hard {soft},{hard}') res.setrlimit(res.RLIMIT_NOFILE, (soft, hard)) except Exception: print('failed to set ulimit, giving up') soft, hard = res.getrlimit(res.RLIMIT_NOFILE) return soft, hard _set_nofile() # ONLY FIRST CLASS CITIZENS ARE ALLOWED HERE, namely Document, Executor Flow # Document from jina._docarray import Document, DocumentArray # Client from jina.clients import Client # Deployment from jina.orchestrate.deployments import Deployment from jina.orchestrate.flow.asyncio import AsyncFlow # Flow from jina.orchestrate.flow.base import Flow # Executor from jina.serve.executors import BaseExecutor as Executor from jina.serve.executors.decorators import dynamic_batching, monitor, requests # Custom Gateway from jina.serve.runtimes.gateway.gateway import Gateway
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings as _warnings import docarray as _docarray if _sys.version_info < (3, 7, 0): raise OSError(f'Jina requires Python >= 3.7, but yours is {_sys.version_info}') def _warning_on_one_line(message, category, filename, lineno, *args, **kwargs): return '\033[1;33m%s: %s\033[0m \033[1;30m(raised from %s:%s)\033[0m\n' % ( category.__name__, message, filename, lineno, ) def _ignore_google_warnings(): import warnings warnings.filterwarnings( 'ignore', category=DeprecationWarning, message='Deprecated call to `pkg_resources.declare_namespace(\'google\')`.', append=True ) _warnings.formatwarning = _warning_on_one_line _warnings.simplefilter('always', DeprecationWarning, append=True) _ignore_google_warnings() # fix fork error on MacOS but seems no effect? must do EXPORT manually before jina start _os.environ['OBJC_DISABLE_INITIALIZE_FORK_SAFETY'] = 'YES' # JINA_MP_START_METHOD has higher priority than os-patch _start_method = _os.environ.get('JINA_MP_START_METHOD', None) if _start_method and _start_method.lower() in {'fork', 'spawn', 'forkserver'}: from multiprocessing import set_start_method as _set_start_method try: _set_start_method(_start_method.lower()) _warnings.warn( f'multiprocessing start method is set to `{_start_method.lower()}`' ) except Exception as e: _warnings.warn( f'failed to set multiprocessing start_method to `{_start_method.lower()}`: {e!r}' ) elif _sys.version_info >= (3, 8, 0) and _platform.system() == 'Darwin': # DO SOME OS-WISE PATCHES # temporary fix for python 3.8 on macos where the default start is set to "spawn" # https://docs.python.org/3/library/multiprocessing.html#contexts-and-start-methods from multiprocessing import set_start_method as _set_start_method try: _set_start_method('fork') _warnings.warn(f'multiprocessing start method is set to `fork`') except Exception as e: _warnings.warn(f'failed to set multiprocessing start_method to `fork`: {e!r}') # do not change this line manually # this is managed by git tag and updated on every release # NOTE: this represents the NEXT release version __version__ = '3.23.6' # do not change this line manually # this is managed by proto/build-proto.sh and updated on every execution __proto_version__ = '0.1.27' try: __docarray_version__ = _docarray.__version__ except AttributeError as e: raise RuntimeError( '`docarray` dependency is not installed correctly, please reinstall with `pip install -U --force-reinstall docarray`' ) try: _signal.signal(_signal.SIGINT, _signal.default_int_handler) except Exception as exc: _warnings.warn(f'failed to set default signal handler: {exc!r}`') def _set_nofile(nofile_atleast=4096): """ Set nofile soft limit to at least 4096, useful for running matlplotlib/seaborn on parallel executing plot generators vs. Ubuntu default ulimit -n 1024 or OS X El Captian 256 temporary setting extinguishing with Python session. :param nofile_atleast: nofile soft limit :return: nofile soft limit and nofile hard limit """ try: import resource as res except ImportError: # Windows res = None if res is None: return (None,) * 2 soft, ohard = res.getrlimit(res.RLIMIT_NOFILE) hard = ohard if soft < nofile_atleast: soft = nofile_atleast if hard < soft: hard = soft try: res.setrlimit(res.RLIMIT_NOFILE, (soft, hard)) except (ValueError, res.error): try: hard = soft print(f'trouble with max limit, retrying with soft,hard {soft},{hard}') res.setrlimit(res.RLIMIT_NOFILE, (soft, hard)) except Exception: print('failed to set ulimit, giving up') soft, hard = res.getrlimit(res.RLIMIT_NOFILE) return soft, hard _set_nofile() # ONLY FIRST CLASS CITIZENS ARE ALLOWED HERE, namely Document, Executor Flow # Document from jina._docarray import Document, DocumentArray # Client from jina.clients import Client # Deployment from jina.orchestrate.deployments import Deployment from jina.orchestrate.flow.asyncio import AsyncFlow # Flow from jina.orchestrate.flow.base import Flow # Executor from jina.serve.executors import BaseExecutor as Executor from jina.serve.executors.decorators import dynamic_batching, monitor, requests # Custom Gateway from jina.serve.runtimes.gateway.gateway import Gateway
_base_ = '../grounding_dino_swin-t_pretrain_obj365.py' data_root = 'data/cityscapes/' class_name = ('person', 'rider', 'car', 'truck', 'bus', 'train', 'motorcycle', 'bicycle') palette = [(220, 20, 60), (255, 0, 0), (0, 0, 142), (0, 0, 70), (0, 60, 100), (0, 80, 100), (0, 0, 230), (119, 11, 32)] metainfo = dict(classes=class_name, palette=palette) train_pipeline = [ dict(type='LoadImageFromFile'), dict(type='LoadAnnotations', with_bbox=True), dict(type='RandomFlip', prob=0.5), dict( type='RandomChoice', transforms=[ [ dict( type='RandomChoiceResize', scales=[(480, 1333), (512, 1333), (544, 1333), (576, 1333), (608, 1333), (640, 1333), (672, 1333), (704, 1333), (736, 1333), (768, 1333), (800, 1333)], keep_ratio=True) ], [ dict( type='RandomChoiceResize', # The radio of all image in train dataset < 7 # follow the original implement scales=[(400, 4200), (500, 4200), (600, 4200)], keep_ratio=True), dict( type='RandomCrop', crop_type='absolute_range', crop_size=(384, 600), allow_negative_crop=True), dict( type='RandomChoiceResize', scales=[(480, 1333), (512, 1333), (544, 1333), (576, 1333), (608, 1333), (640, 1333), (672, 1333), (704, 1333), (736, 1333), (768, 1333), (800, 1333)], keep_ratio=True) ] ]), dict( type='PackDetInputs', meta_keys=('img_id', 'img_path', 'ori_shape', 'img_shape', 'scale_factor', 'flip', 'flip_direction', 'text', 'custom_entities')) ] train_dataloader = dict( sampler=dict(_delete_=True, type='DefaultSampler', shuffle=True), batch_sampler=dict(type='AspectRatioBatchSampler'), dataset=dict( _delete_=True, type='RepeatDataset', times=10, dataset=dict( type='CocoDataset', data_root=data_root, metainfo=metainfo, filter_cfg=dict(filter_empty_gt=False, min_size=32), pipeline=train_pipeline, return_classes=True, data_prefix=dict(img='leftImg8bit/train/'), ann_file='annotations/instancesonly_filtered_gtFine_train.json'))) val_dataloader = dict( dataset=dict( metainfo=metainfo, data_root=data_root, return_classes=True, ann_file='annotations/instancesonly_filtered_gtFine_val.json', data_prefix=dict(img='leftImg8bit/val/'))) test_dataloader = val_dataloader val_evaluator = dict( type='CocoMetric', ann_file=data_root + 'annotations/instancesonly_filtered_gtFine_val.json', metric='bbox', format_only=False) test_evaluator = val_evaluator optim_wrapper = dict( _delete_=True, type='OptimWrapper', optimizer=dict(type='AdamW', lr=0.0001, weight_decay=0.0001), clip_grad=dict(max_norm=0.1, norm_type=2), paramwise_cfg=dict(custom_keys={ 'absolute_pos_embed': dict(decay_mult=0.), 'backbone': dict(lr_mult=0.1) })) # learning policy max_epochs = 5 param_scheduler = [ dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[4], gamma=0.1) ] train_cfg = dict(max_epochs=max_epochs, val_interval=1) default_hooks = dict(checkpoint=dict(max_keep_ckpts=1, save_best='auto')) load_from = 'https://download.openmmlab.com/mmdetection/v3.0/mm_grounding_dino/grounding_dino_swin-t_pretrain_obj365_goldg_grit9m_v3det/grounding_dino_swin-t_pretrain_obj365_goldg_grit9m_v3det_20231204_095047-b448804b.pth' # noqa
_base_ = '../grounding_dino_swin-t_pretrain_obj365.py' data_root = 'data/cityscapes/' class_name = ('person', 'rider', 'car', 'truck', 'bus', 'train', 'motorcycle', 'bicycle') palette = [(220, 20, 60), (255, 0, 0), (0, 0, 142), (0, 0, 70), (0, 60, 100), (0, 80, 100), (0, 0, 230), (119, 11, 32)] metainfo = dict(classes=class_name, palette=palette) train_pipeline = [ dict(type='LoadImageFromFile'), dict(type='LoadAnnotations', with_bbox=True), dict(type='RandomFlip', prob=0.5), dict( type='RandomChoice', transforms=[ [ dict( type='RandomChoiceResize', scales=[(480, 1333), (512, 1333), (544, 1333), (576, 1333), (608, 1333), (640, 1333), (672, 1333), (704, 1333), (736, 1333), (768, 1333), (800, 1333)], keep_ratio=True) ], [ dict( type='RandomChoiceResize', # The radio of all image in train dataset < 7 # follow the original implement scales=[(400, 4200), (500, 4200), (600, 4200)], keep_ratio=True), dict( type='RandomCrop', crop_type='absolute_range', crop_size=(384, 600), allow_negative_crop=True), dict( type='RandomChoiceResize', scales=[(480, 1333), (512, 1333), (544, 1333), (576, 1333), (608, 1333), (640, 1333), (672, 1333), (704, 1333), (736, 1333), (768, 1333), (800, 1333)], keep_ratio=True) ] ]), dict( type='PackDetInputs', meta_keys=('img_id', 'img_path', 'ori_shape', 'img_shape', 'scale_factor', 'flip', 'flip_direction', 'text', 'custom_entities')) ] train_dataloader = dict( sampler=dict(_delete_=True, type='DefaultSampler', shuffle=True), batch_sampler=dict(type='AspectRatioBatchSampler'), dataset=dict( _delete_=True, type='RepeatDataset', times=10, dataset=dict( type='CocoDataset', data_root=data_root, metainfo=metainfo, filter_cfg=dict(filter_empty_gt=False, min_size=32), pipeline=train_pipeline, return_classes=True, data_prefix=dict(img='leftImg8bit/train/'), ann_file='annotations/instancesonly_filtered_gtFine_train.json'))) val_dataloader = dict( dataset=dict( metainfo=metainfo, data_root=data_root, return_classes=True, ann_file='annotations/instancesonly_filtered_gtFine_val.json', data_prefix=dict(img='leftImg8bit/val/'))) test_dataloader = val_dataloader val_evaluator = dict( type='CocoMetric', ann_file=data_root + 'annotations/instancesonly_filtered_gtFine_val.json', metric='bbox', format_only=False) test_evaluator = val_evaluator optim_wrapper = dict( _delete_=True, type='OptimWrapper', optimizer=dict(type='AdamW', lr=0.0001, weight_decay=0.0001), clip_grad=dict(max_norm=0.1, norm_type=2), paramwise_cfg=dict(custom_keys={ 'absolute_pos_embed': dict(decay_mult=0.), 'backbone': dict(lr_mult=0.1) })) # learning policy max_epochs = 5 param_scheduler = [ dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[4], gamma=0.1) ] train_cfg = dict(max_epochs=max_epochs, val_interval=1) default_hooks = dict(checkpoint=dict(max_keep_ckpts=1, save_best='auto')) load_from = ''
checkpoint_config = dict(interval=1) # yapf:disable log_config = dict( interval=50, hooks=[ dict(type='TextLoggerHook'), # dict(type='TensorboardLoggerHook') ]) # yapf:enable custom_hooks = [dict(type='NumClassCheckHook')] dist_params = dict(backend='nccl') log_level = 'INFO' load_from = None resume_from = None workflow = [('train', 1)] # disable opencv multithreading to avoid system being overloaded opencv_num_threads = 0 # set multi-process start method as `fork` to speed up the training mp_start_method = 'fork'
checkpoint_config = dict(interval=1) # yapf:disable log_config = dict( interval=50, hooks=[ dict(type='TextLoggerHook'), # dict(type='TensorboardLoggerHook') ]) # yapf:enable custom_hooks = [dict(type='NumClassCheckHook')] dist_params = dict(backend='nccl') log_level = 'INFO' load_from = None resume_from = None workflow = [('train', 1)]
_base_ = './yolox_s_8xb8-300e_coco.py' # model settings model = dict( data_preprocessor=dict(batch_augments=[ dict( type='BatchSyncRandomResize', random_size_range=(320, 640), size_divisor=32, interval=10) ]), backbone=dict(deepen_factor=0.33, widen_factor=0.375), neck=dict(in_channels=[96, 192, 384], out_channels=96), bbox_head=dict(in_channels=96, feat_channels=96)) img_scale = (640, 640) # width, height train_pipeline = [ dict(type='Mosaic', img_scale=img_scale, pad_val=114.0), dict( type='RandomAffine', scaling_ratio_range=(0.5, 1.5), # img_scale is (width, height) border=(-img_scale[0] // 2, -img_scale[1] // 2)), dict(type='YOLOXHSVRandomAug'), dict(type='RandomFlip', prob=0.5), # Resize and Pad are for the last 15 epochs when Mosaic and # RandomAffine are closed by YOLOXModeSwitchHook. dict(type='Resize', scale=img_scale, keep_ratio=True), dict( type='Pad', pad_to_square=True, pad_val=dict(img=(114.0, 114.0, 114.0))), dict(type='FilterAnnotations', min_gt_bbox_wh=(1, 1), keep_empty=False), dict(type='PackDetInputs') ] test_pipeline = [ dict(type='LoadImageFromFile', backend_args={{_base_.backend_args}}), dict(type='Resize', scale=(416, 416), keep_ratio=True), dict( type='Pad', pad_to_square=True, pad_val=dict(img=(114.0, 114.0, 114.0))), dict(type='LoadAnnotations', with_bbox=True), dict( type='PackDetInputs', meta_keys=('img_id', 'img_path', 'ori_shape', 'img_shape', 'scale_factor')) ] train_dataloader = dict(dataset=dict(pipeline=train_pipeline)) val_dataloader = dict(dataset=dict(pipeline=test_pipeline)) test_dataloader = val_dataloader
_base_ = './yolox_s_8xb8-300e_coco.py' # model settings model = dict( data_preprocessor=dict(batch_augments=[ dict( type='BatchSyncRandomResize', random_size_range=(320, 640), size_divisor=32, interval=10) ]), backbone=dict(deepen_factor=0.33, widen_factor=0.375), neck=dict(in_channels=[96, 192, 384], out_channels=96), bbox_head=dict(in_channels=96, feat_channels=96)) img_scale = (640, 640) # width, height # file_client_args = dict( # backend='petrel', # path_mapping=dict({ # './data/': 's3://openmmlab/datasets/detection/', # 'data/': 's3://openmmlab/datasets/detection/' # })) file_client_args = dict(backend='disk') train_pipeline = [ dict(type='Mosaic', img_scale=img_scale, pad_val=114.0), dict( type='RandomAffine', scaling_ratio_range=(0.5, 1.5), # img_scale is (width, height) border=(-img_scale[0] // 2, -img_scale[1] // 2)), dict(type='YOLOXHSVRandomAug'), dict(type='RandomFlip', prob=0.5), # Resize and Pad are for the last 15 epochs when Mosaic and # RandomAffine are closed by YOLOXModeSwitchHook. dict(type='Resize', scale=img_scale, keep_ratio=True), dict( type='Pad', pad_to_square=True, pad_val=dict(img=(114.0, 114.0, 114.0))), dict(type='FilterAnnotations', min_gt_bbox_wh=(1, 1), keep_empty=False), dict(type='PackDetInputs') ] test_pipeline = [ dict(type='LoadImageFromFile', file_client_args=file_client_args), dict(type='Resize', scale=(416, 416), keep_ratio=True), dict( type='Pad', pad_to_square=True, pad_val=dict(img=(114.0, 114.0, 114.0))), dict(type='LoadAnnotations', with_bbox=True), dict( type='PackDetInputs', meta_keys=('img_id', 'img_path', 'ori_shape', 'img_shape', 'scale_factor')) ] train_dataloader = dict(dataset=dict(pipeline=train_pipeline)) val_dataloader = dict(dataset=dict(pipeline=test_pipeline)) test_dataloader = val_dataloader
import logging from datasets import load_dataset from sentence_transformers.sparse_encoder import ( SparseEncoder, SparseTripletEvaluator, ) logging.basicConfig(format="%(message)s", level=logging.INFO) # Load a model model = SparseEncoder("naver/splade-cocondenser-ensembledistil") # Load triplets from the AllNLI dataset # The dataset contains triplets of (anchor, positive, negative) sentences dataset = load_dataset("sentence-transformers/all-nli", "triplet", split="dev[:1000]") # Initialize the SparseTripletEvaluator evaluator = SparseTripletEvaluator( anchors=dataset[:1000]["anchor"], positives=dataset[:1000]["positive"], negatives=dataset[:1000]["negative"], name="all_nli_dev", batch_size=32, show_progress_bar=True, ) # Run the evaluation results = evaluator(model) """ TripletEvaluator: Evaluating the model on the all_nli_dev dataset: Accuracy Dot Similarity: 85.10% """ # Print the results print(f"Primary metric: {evaluator.primary_metric}") # => Primary metric: all_nli_dev_dot_accuracy print(f"Primary metric value: {results[evaluator.primary_metric]:.4f}") # => Primary metric value: 0.8510
import logging from datasets import load_dataset from sentence_transformers.sparse_encoder import ( MLMTransformer, SparseEncoder, SparseTripletEvaluator, SpladePooling, ) logging.basicConfig(format="%(asctime)s - %(message)s", datefmt="%Y-%m-%d %H:%M:%S", level=logging.INFO) # Initialize the SPLADE model model_name = "naver/splade-cocondenser-ensembledistil" model = SparseEncoder( modules=[ MLMTransformer(model_name), SpladePooling(pooling_strategy="max"), # You can also use 'sum' ], device="cuda:0", ) # Load triplets from the AllNLI dataset # The dataset contains triplets of (anchor, positive, negative) sentences dataset = load_dataset("sentence-transformers/all-nli", "triplet", split="dev[:1000]") # Initialize the SparseTripletEvaluator evaluator = SparseTripletEvaluator( anchors=dataset[:1000]["anchor"], positives=dataset[:1000]["positive"], negatives=dataset[:1000]["negative"], name="all_nli_dev", batch_size=32, show_progress_bar=True, ) # Run the evaluation results = evaluator(model) # Print the results print(f"Primary metric: {evaluator.primary_metric}") print(f"Primary metric value: {results[evaluator.primary_metric]:.4f}")
""" Visual demo for survival analysis (regression) with Accelerated Failure Time (AFT) model. ========================================================================================= This demo uses 1D toy data and visualizes how XGBoost fits a tree ensemble. The ensemble model starts out as a flat line and evolves into a step function in order to account for all ranged labels. """ import matplotlib.pyplot as plt import numpy as np import xgboost as xgb plt.rcParams.update({"font.size": 13}) # Function to visualize censored labels def plot_censored_labels( X: np.ndarray, y_lower: np.ndarray, y_upper: np.ndarray ) -> None: def replace_inf(x: np.ndarray, target_value: float) -> np.ndarray: x[np.isinf(x)] = target_value return x plt.plot(X, y_lower, "o", label="y_lower", color="blue") plt.plot(X, y_upper, "o", label="y_upper", color="fuchsia") plt.vlines( X, ymin=replace_inf(y_lower, 0.01), ymax=replace_inf(y_upper, 1000.0), label="Range for y", color="gray", ) # Toy data X = np.array([1, 2, 3, 4, 5]).reshape((-1, 1)) INF = np.inf y_lower = np.array([10, 15, -INF, 30, 100]) y_upper = np.array([INF, INF, 20, 50, INF]) # Visualize toy data plt.figure(figsize=(5, 4)) plot_censored_labels(X, y_lower, y_upper) plt.ylim((6, 200)) plt.legend(loc="lower right") plt.title("Toy data") plt.xlabel("Input feature") plt.ylabel("Label") plt.yscale("log") plt.tight_layout() plt.show(block=True) # Will be used to visualize XGBoost model grid_pts = np.linspace(0.8, 5.2, 1000).reshape((-1, 1)) # Train AFT model using XGBoost dmat = xgb.DMatrix(X) dmat.set_float_info("label_lower_bound", y_lower) dmat.set_float_info("label_upper_bound", y_upper) params = {"max_depth": 3, "objective": "survival:aft", "min_child_weight": 0} accuracy_history = [] class PlotIntermediateModel(xgb.callback.TrainingCallback): """Custom callback to plot intermediate models.""" def __init__(self) -> None: super().__init__() def after_iteration( self, model: xgb.Booster, epoch: int, evals_log: xgb.callback.TrainingCallback.EvalsLog, ) -> bool: """Run after training is finished.""" # Compute y_pred = prediction using the intermediate model, at current boosting # iteration y_pred = model.predict(dmat) # "Accuracy" = the number of data points whose ranged label (y_lower, y_upper) # includes the corresponding predicted label (y_pred) acc = np.sum( np.logical_and(y_pred >= y_lower, y_pred <= y_upper) / len(X) * 100 ) accuracy_history.append(acc) # Plot ranged labels as well as predictions by the model plt.subplot(5, 3, epoch + 1) plot_censored_labels(X, y_lower, y_upper) y_pred_grid_pts = model.predict(xgb.DMatrix(grid_pts)) plt.plot( grid_pts, y_pred_grid_pts, "r-", label="XGBoost AFT model", linewidth=4 ) plt.title("Iteration {}".format(epoch), x=0.5, y=0.8) plt.xlim((0.8, 5.2)) plt.ylim((1 if np.min(y_pred) < 6 else 6, 200)) plt.yscale("log") return False res: xgb.callback.TrainingCallback.EvalsLog = {} plt.figure(figsize=(12, 13)) bst = xgb.train( params, dmat, num_boost_round=15, evals=[(dmat, "train")], evals_result=res, callbacks=[PlotIntermediateModel()], ) plt.tight_layout() plt.legend( loc="lower center", ncol=4, bbox_to_anchor=(0.5, 0), bbox_transform=plt.gcf().transFigure, ) plt.tight_layout() # Plot negative log likelihood over boosting iterations plt.figure(figsize=(8, 3)) plt.subplot(1, 2, 1) plt.plot(res["train"]["aft-nloglik"], "b-o", label="aft-nloglik") plt.xlabel("# Boosting Iterations") plt.legend(loc="best") # Plot "accuracy" over boosting iterations # "Accuracy" = the number of data points whose ranged label (y_lower, y_upper) includes # the corresponding predicted label (y_pred) plt.subplot(1, 2, 2) plt.plot(accuracy_history, "r-o", label="Accuracy (%)") plt.xlabel("# Boosting Iterations") plt.legend(loc="best") plt.tight_layout() plt.show()
""" Visual demo for survival analysis (regression) with Accelerated Failure Time (AFT) model. ========================================================================================= This demo uses 1D toy data and visualizes how XGBoost fits a tree ensemble. The ensemble model starts out as a flat line and evolves into a step function in order to account for all ranged labels. """ import matplotlib.pyplot as plt import numpy as np import xgboost as xgb plt.rcParams.update({"font.size": 13}) # Function to visualize censored labels def plot_censored_labels( X: np.ndarray, y_lower: np.ndarray, y_upper: np.ndarray ) -> None: def replace_inf(x: np.ndarray, target_value: float) -> np.ndarray: x[np.isinf(x)] = target_value return x plt.plot(X, y_lower, "o", label="y_lower", color="blue") plt.plot(X, y_upper, "o", label="y_upper", color="fuchsia") plt.vlines( X, ymin=replace_inf(y_lower, 0.01), ymax=replace_inf(y_upper, 1000.0), label="Range for y", color="gray", ) # Toy data X = np.array([1, 2, 3, 4, 5]).reshape((-1, 1)) INF = np.inf y_lower = np.array([10, 15, -INF, 30, 100]) y_upper = np.array([INF, INF, 20, 50, INF]) # Visualize toy data plt.figure(figsize=(5, 4)) plot_censored_labels(X, y_lower, y_upper) plt.ylim((6, 200)) plt.legend(loc="lower right") plt.title("Toy data") plt.xlabel("Input feature") plt.ylabel("Label") plt.yscale("log") plt.tight_layout() plt.show(block=True) # Will be used to visualize XGBoost model grid_pts = np.linspace(0.8, 5.2, 1000).reshape((-1, 1)) # Train AFT model using XGBoost dmat = xgb.DMatrix(X) dmat.set_float_info("label_lower_bound", y_lower) dmat.set_float_info("label_upper_bound", y_upper) params = {"max_depth": 3, "objective": "survival:aft", "min_child_weight": 0} accuracy_history = [] class PlotIntermediateModel(xgb.callback.TrainingCallback): """Custom callback to plot intermediate models.""" def __init__(self) -> None: super().__init__() def after_iteration( self, model: xgb.Booster, epoch: int, evals_log: xgb.callback.TrainingCallback.EvalsLog, ) -> bool: """Run after training is finished.""" # Compute y_pred = prediction using the intermediate model, at current boosting # iteration y_pred = model.predict(dmat) # "Accuracy" = the number of data points whose ranged label (y_lower, y_upper) # includes the corresponding predicted label (y_pred) acc = np.sum( np.logical_and(y_pred >= y_lower, y_pred <= y_upper) / len(X) * 100 ) accuracy_history.append(acc) # Plot ranged labels as well as predictions by the model plt.subplot(5, 3, epoch + 1) plot_censored_labels(X, y_lower, y_upper) y_pred_grid_pts = model.predict(xgb.DMatrix(grid_pts)) plt.plot( grid_pts, y_pred_grid_pts, "r-", label="XGBoost AFT model", linewidth=4 ) plt.title("Iteration {}".format(epoch), x=0.5, y=0.8) plt.xlim((0.8, 5.2)) plt.ylim((1 if np.min(y_pred) < 6 else 6, 200)) plt.yscale("log") return False res: xgb.callback.TrainingCallback.EvalsLog = {} plt.figure(figsize=(12, 13)) bst = xgb.train( params, dmat, 15, [(dmat, "train")], evals_result=res, callbacks=[PlotIntermediateModel()], ) plt.tight_layout() plt.legend( loc="lower center", ncol=4, bbox_to_anchor=(0.5, 0), bbox_transform=plt.gcf().transFigure, ) plt.tight_layout() # Plot negative log likelihood over boosting iterations plt.figure(figsize=(8, 3)) plt.subplot(1, 2, 1) plt.plot(res["train"]["aft-nloglik"], "b-o", label="aft-nloglik") plt.xlabel("# Boosting Iterations") plt.legend(loc="best") # Plot "accuracy" over boosting iterations # "Accuracy" = the number of data points whose ranged label (y_lower, y_upper) includes # the corresponding predicted label (y_pred) plt.subplot(1, 2, 2) plt.plot(accuracy_history, "r-o", label="Accuracy (%)") plt.xlabel("# Boosting Iterations") plt.legend(loc="best") plt.tight_layout() plt.show()
from typing import Any, Literal from pydantic import SecretStr from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import ( APIKeyCredentials, CredentialsField, CredentialsMetaInput, SchemaField, ) from backend.integrations.providers import ProviderName from backend.util.request import Requests TEST_CREDENTIALS = APIKeyCredentials( id="01234567-89ab-cdef-0123-456789abcdef", provider="unreal_speech", api_key=SecretStr("mock-unreal-speech-api-key"), title="Mock Unreal Speech API key", expires_at=None, ) TEST_CREDENTIALS_INPUT = { "provider": TEST_CREDENTIALS.provider, "id": TEST_CREDENTIALS.id, "type": TEST_CREDENTIALS.type, "title": TEST_CREDENTIALS.type, } class UnrealTextToSpeechBlock(Block): class Input(BlockSchema): text: str = SchemaField( description="The text to be converted to speech", placeholder="Enter the text you want to convert to speech", ) voice_id: str = SchemaField( description="The voice ID to use for text-to-speech conversion", placeholder="Scarlett", default="Scarlett", ) credentials: CredentialsMetaInput[ Literal[ProviderName.UNREAL_SPEECH], Literal["api_key"] ] = CredentialsField( description="The Unreal Speech integration can be used with " "any API key with sufficient permissions for the blocks it is used on.", ) class Output(BlockSchema): mp3_url: str = SchemaField(description="The URL of the generated MP3 file") error: str = SchemaField(description="Error message if the API call failed") def __init__(self): super().__init__( id="4ff1ff6d-cc40-4caa-ae69-011daa20c378", description="Converts text to speech using the Unreal Speech API", categories={BlockCategory.AI, BlockCategory.TEXT, BlockCategory.MULTIMEDIA}, input_schema=UnrealTextToSpeechBlock.Input, output_schema=UnrealTextToSpeechBlock.Output, test_input={ "text": "This is a test of the text to speech API.", "voice_id": "Scarlett", "credentials": TEST_CREDENTIALS_INPUT, }, test_output=[("mp3_url", "https://example.com/test.mp3")], test_mock={ "call_unreal_speech_api": lambda *args, **kwargs: { "OutputUri": "https://example.com/test.mp3" } }, test_credentials=TEST_CREDENTIALS, ) @staticmethod async def call_unreal_speech_api( api_key: SecretStr, text: str, voice_id: str ) -> dict[str, Any]: url = "https://api.v7.unrealspeech.com/speech" headers = { "Authorization": f"Bearer {api_key.get_secret_value()}", "Content-Type": "application/json", } data = { "Text": text, "VoiceId": voice_id, "Bitrate": "192k", "Speed": "0", "Pitch": "1", "TimestampType": "sentence", } response = await Requests().post(url, headers=headers, json=data) return response.json() async def run( self, input_data: Input, *, credentials: APIKeyCredentials, **kwargs ) -> BlockOutput: api_response = await self.call_unreal_speech_api( credentials.api_key, input_data.text, input_data.voice_id, ) yield "mp3_url", api_response["OutputUri"]
from typing import Any, Literal from pydantic import SecretStr from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import ( APIKeyCredentials, CredentialsField, CredentialsMetaInput, SchemaField, ) from backend.integrations.providers import ProviderName from backend.util.request import requests TEST_CREDENTIALS = APIKeyCredentials( id="01234567-89ab-cdef-0123-456789abcdef", provider="unreal_speech", api_key=SecretStr("mock-unreal-speech-api-key"), title="Mock Unreal Speech API key", expires_at=None, ) TEST_CREDENTIALS_INPUT = { "provider": TEST_CREDENTIALS.provider, "id": TEST_CREDENTIALS.id, "type": TEST_CREDENTIALS.type, "title": TEST_CREDENTIALS.type, } class UnrealTextToSpeechBlock(Block): class Input(BlockSchema): text: str = SchemaField( description="The text to be converted to speech", placeholder="Enter the text you want to convert to speech", ) voice_id: str = SchemaField( description="The voice ID to use for text-to-speech conversion", placeholder="Scarlett", default="Scarlett", ) credentials: CredentialsMetaInput[ Literal[ProviderName.UNREAL_SPEECH], Literal["api_key"] ] = CredentialsField( description="The Unreal Speech integration can be used with " "any API key with sufficient permissions for the blocks it is used on.", ) class Output(BlockSchema): mp3_url: str = SchemaField(description="The URL of the generated MP3 file") error: str = SchemaField(description="Error message if the API call failed") def __init__(self): super().__init__( id="4ff1ff6d-cc40-4caa-ae69-011daa20c378", description="Converts text to speech using the Unreal Speech API", categories={BlockCategory.AI, BlockCategory.TEXT, BlockCategory.MULTIMEDIA}, input_schema=UnrealTextToSpeechBlock.Input, output_schema=UnrealTextToSpeechBlock.Output, test_input={ "text": "This is a test of the text to speech API.", "voice_id": "Scarlett", "credentials": TEST_CREDENTIALS_INPUT, }, test_output=[("mp3_url", "https://example.com/test.mp3")], test_mock={ "call_unreal_speech_api": lambda *args, **kwargs: { "OutputUri": "https://example.com/test.mp3" } }, test_credentials=TEST_CREDENTIALS, ) @staticmethod def call_unreal_speech_api( api_key: SecretStr, text: str, voice_id: str ) -> dict[str, Any]: url = "https://api.v7.unrealspeech.com/speech" headers = { "Authorization": f"Bearer {api_key.get_secret_value()}", "Content-Type": "application/json", } data = { "Text": text, "VoiceId": voice_id, "Bitrate": "192k", "Speed": "0", "Pitch": "1", "TimestampType": "sentence", } response = requests.post(url, headers=headers, json=data) return response.json() def run( self, input_data: Input, *, credentials: APIKeyCredentials, **kwargs ) -> BlockOutput: api_response = self.call_unreal_speech_api( credentials.api_key, input_data.text, input_data.voice_id, ) yield "mp3_url", api_response["OutputUri"]
# Owner(s): ["module: inductor"] import torch from torch._inductor import config, metrics from torch._inductor.test_case import run_tests, TestCase from torch._inductor.utils import collect_defined_kernels from torch._inductor.wrapper_benchmark import get_kernel_category_by_source_code from torch.testing._internal.common_device_type import largeTensorTest from torch.testing._internal.inductor_utils import GPU_TYPE, HAS_GPU example_kernel = """ @triton_heuristics.reduction( size_hints=[1024, 2048], reduction_hint=ReductionHint.INNER, filename=__file__, triton_meta={ 'signature': {'in_out_ptr0': '*fp32', 'in_ptr0': '*fp32', 'xnumel': 'i32', 'rnumel': 'i32'}, 'device': 0, 'device_type': 'GPU_TYPE', 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3), equal_to_1=())]}, inductor_meta={ 'autotune_hints': set(), 'kernel_name': 'triton_red_fused_add_sum_2', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'kernel_num_gb': 0.0083968 } ) @triton.jit def triton_red_fused_add_sum_2(in_out_ptr0, in_ptr0, xnumel, rnumel, XBLOCK : tl.constexpr, RBLOCK : tl.constexpr): xnumel = 1024 rnumel = 2048 xoffset = tl.program_id(0) * XBLOCK xindex = xoffset + tl.arange(0, XBLOCK)[:, None] xmask = xindex < xnumel rbase = tl.arange(0, RBLOCK)[None, :] x0 = xindex _tmp2 = tl.full([XBLOCK, RBLOCK], 0, tl.float32) for roffset in range(0, rnumel, RBLOCK): rindex = roffset + rbase rmask = rindex < rnumel r1 = rindex tmp0 = tl.load(in_ptr0 + (r1 + (2048*x0)), rmask & xmask, eviction_policy='evict_first', other=0.0) tmp1 = tl.broadcast_to(tmp0, [XBLOCK, RBLOCK]) tmp3 = _tmp2 + tmp1 _tmp2 = tl.where(rmask & xmask, tmp3, _tmp2) tmp2 = tl.sum(_tmp2, 1)[:, None] tmp4 = tl.load(in_out_ptr0 + (x0), xmask, eviction_policy='evict_last') tmp5 = tmp4 + tmp2 tl.debug_barrier() tl.store(in_out_ptr0 + (x0), tmp5, xmask) """.replace("GPU_TYPE", GPU_TYPE) class TestMetrics(TestCase): def test_parse_proper_kernel_fn_code(self): proper_kernel_fn_code = metrics._parse_proper_kernel_fn_code(example_kernel) assert proper_kernel_fn_code.startswith("def ") def test_count_args(self): proper_kernel_fn_code = metrics._parse_proper_kernel_fn_code(example_kernel) self.assertEqual(6, metrics._count_args(proper_kernel_fn_code)) def test_count_pattern(self): proper_kernel_fn_code = metrics._parse_proper_kernel_fn_code(example_kernel) self.assertEqual(2, metrics._count_pattern(proper_kernel_fn_code, "tl.load")) self.assertEqual(1, metrics._count_pattern(proper_kernel_fn_code, "tl.store")) self.assertEqual(1, metrics._count_pattern(proper_kernel_fn_code, "for ")) def test_parse_reduction_hint(self): kernel_category = get_kernel_category_by_source_code(example_kernel) self.assertEqual("reduction", kernel_category) self.assertEqual( "INNER", metrics._parse_reduction_hint(kernel_category, example_kernel) ) @config.patch("fx_graph_remote_cache", False) def test_atomic_add(self): @torch.compile def f(lhs, index, rhs): return lhs.index_put_([index], rhs, accumulate=True) lhs = torch.randn(1024, device=GPU_TYPE) index = torch.randint(0, 1024, [32], device=GPU_TYPE, dtype=torch.int32) rhs = torch.randn(32, device=GPU_TYPE) kernel_list = [] with collect_defined_kernels(kernel_list): f(lhs, index, rhs) self.assertEqual(len(kernel_list), 1) kernel_code = kernel_list[0] self.assertEqual(metrics._count_pattern(kernel_code, "tl.atomic_add"), 1) @largeTensorTest(25e7 * 2 * 4, device=GPU_TYPE, inductor=True) @config.patch("fx_graph_remote_cache", False) @config.patch("benchmark_kernel", True) def test_kernel_args_num_gb(self): @torch.compile def f(x): return x + 1 x = torch.randn(int(25e7), device=GPU_TYPE) kernel_list = [] with collect_defined_kernels(kernel_list): f(x) self.assertEqual(len(kernel_list), 1) kernel_code = kernel_list[0] self.assertEqual( metrics._parse_kernel_args_num_gb(kernel_code, "pointwise"), 2.0 ) if __name__ == "__main__": if HAS_GPU: run_tests()
# Owner(s): ["module: inductor"] import torch from torch._inductor import config, metrics from torch._inductor.test_case import run_tests, TestCase from torch._inductor.utils import collect_defined_kernels from torch._inductor.wrapper_benchmark import get_kernel_category_by_source_code from torch.testing._internal.common_device_type import largeTensorTest from torch.testing._internal.inductor_utils import GPU_TYPE, HAS_GPU example_kernel = """ @triton_heuristics.reduction( size_hints=[1024, 2048], reduction_hint=ReductionHint.INNER, filename=__file__, triton_meta={ 'signature': {'in_out_ptr0': '*fp32', 'in_ptr0': '*fp32', 'xnumel': 'i32', 'rnumel': 'i32'}, 'device': 0, 'device_type': 'GPU_TYPE', 'constants': {}, 'configs': [AttrsDescriptor(divisible_by_16=(0, 1, 2, 3), equal_to_1=())]}, inductor_meta={ 'autotune_hints': set(), 'kernel_name': 'triton_red_fused_add_sum_2', 'mutated_arg_names': ['in_out_ptr0'], 'no_x_dim': False, 'kernel_num_gb': 0.0083968 } ) @triton.jit def triton_red_fused_add_sum_2(in_out_ptr0, in_ptr0, xnumel, rnumel, XBLOCK : tl.constexpr, RBLOCK : tl.constexpr): xnumel = 1024 rnumel = 2048 xoffset = tl.program_id(0) * XBLOCK xindex = xoffset + tl.arange(0, XBLOCK)[:, None] xmask = xindex < xnumel rbase = tl.arange(0, RBLOCK)[None, :] x0 = xindex _tmp2 = tl.full([XBLOCK, RBLOCK], 0, tl.float32) for roffset in range(0, rnumel, RBLOCK): rindex = roffset + rbase rmask = rindex < rnumel r1 = rindex tmp0 = tl.load(in_ptr0 + (r1 + (2048*x0)), rmask & xmask, eviction_policy='evict_first', other=0.0) tmp1 = tl.broadcast_to(tmp0, [XBLOCK, RBLOCK]) tmp3 = _tmp2 + tmp1 _tmp2 = tl.where(rmask & xmask, tmp3, _tmp2) tmp2 = tl.sum(_tmp2, 1)[:, None] tmp4 = tl.load(in_out_ptr0 + (x0), xmask, eviction_policy='evict_last') tmp5 = tmp4 + tmp2 tl.debug_barrier() tl.store(in_out_ptr0 + (x0), tmp5, xmask) """.replace( "GPU_TYPE", GPU_TYPE ) class TestMetrics(TestCase): def test_parse_proper_kernel_fn_code(self): proper_kernel_fn_code = metrics._parse_proper_kernel_fn_code(example_kernel) assert proper_kernel_fn_code.startswith("def ") def test_count_args(self): proper_kernel_fn_code = metrics._parse_proper_kernel_fn_code(example_kernel) self.assertEqual(6, metrics._count_args(proper_kernel_fn_code)) def test_count_pattern(self): proper_kernel_fn_code = metrics._parse_proper_kernel_fn_code(example_kernel) self.assertEqual(2, metrics._count_pattern(proper_kernel_fn_code, "tl.load")) self.assertEqual(1, metrics._count_pattern(proper_kernel_fn_code, "tl.store")) self.assertEqual(1, metrics._count_pattern(proper_kernel_fn_code, "for ")) def test_parse_reduction_hint(self): kernel_category = get_kernel_category_by_source_code(example_kernel) self.assertEqual("reduction", kernel_category) self.assertEqual( "INNER", metrics._parse_reduction_hint(kernel_category, example_kernel) ) @config.patch("fx_graph_remote_cache", False) def test_atomic_add(self): @torch.compile def f(lhs, index, rhs): return lhs.index_put_([index], rhs, accumulate=True) lhs = torch.randn(1024, device=GPU_TYPE) index = torch.randint(0, 1024, [32], device=GPU_TYPE, dtype=torch.int32) rhs = torch.randn(32, device=GPU_TYPE) kernel_list = [] with collect_defined_kernels(kernel_list): f(lhs, index, rhs) self.assertEqual(len(kernel_list), 1) kernel_code = kernel_list[0] self.assertEqual(metrics._count_pattern(kernel_code, "tl.atomic_add"), 1) @largeTensorTest(25e7 * 2 * 4, device=GPU_TYPE, inductor=True) @config.patch("fx_graph_remote_cache", False) @config.patch("benchmark_kernel", True) def test_kernel_args_num_gb(self): @torch.compile def f(x): return x + 1 x = torch.randn(int(25e7), device=GPU_TYPE) kernel_list = [] with collect_defined_kernels(kernel_list): f(x) self.assertEqual(len(kernel_list), 1) kernel_code = kernel_list[0] self.assertEqual( metrics._parse_kernel_args_num_gb(kernel_code, "pointwise"), 2.0 ) if __name__ == "__main__": if HAS_GPU: run_tests()
from enum import Enum # --8<-- [start:ProviderName] class ProviderName(str, Enum): ANTHROPIC = "anthropic" DISCORD = "discord" D_ID = "d_id" E2B = "e2b" EXA = "exa" FAL = "fal" GITHUB = "github" GOOGLE = "google" GOOGLE_MAPS = "google_maps" GROQ = "groq" HUBSPOT = "hubspot" IDEOGRAM = "ideogram" JINA = "jina" MEDIUM = "medium" NOTION = "notion" OLLAMA = "ollama" OPENAI = "openai" OPENWEATHERMAP = "openweathermap" OPEN_ROUTER = "open_router" PINECONE = "pinecone" REPLICATE = "replicate" REVID = "revid" SLANT3D = "slant3d" UNREAL_SPEECH = "unreal_speech" # --8<-- [end:ProviderName]
from enum import Enum class ProviderName(str, Enum): GITHUB = "github" GOOGLE = "google" NOTION = "notion"
_base_ = './yolov3_d53_8xb8-ms-608-273e_coco.py' # fp16 settings optim_wrapper = dict(type='AmpOptimWrapper', loss_scale='dynamic')
_base_ = './yolov3_d53_mstrain-608_273e_coco.py' # fp16 settings optim_wrapper = dict(type='AmpOptimWrapper', loss_scale='dynamic')
import random import pytest from pathlib import Path from typing import Dict, Tuple, Callable import opentelemetry.sdk.metrics.export import opentelemetry.sdk.metrics.view from opentelemetry.sdk.metrics.export import ( AggregationTemporality, MetricExporter, MetricExportResult, MetricsData, PeriodicExportingMetricReader, ) class DirMetricExporter(MetricExporter): """Implementation of :class:`MetricExporter` that prints metrics to a file in a given directory. This class can be used for diagnostic or testing purposes. """ def __init__( self, metric_dir: str, preferred_temporality: Dict[type, AggregationTemporality] = None, preferred_aggregation: Dict[ type, "opentelemetry.sdk.metrics.view.Aggregation" ] = None, ): super().__init__( preferred_temporality=preferred_temporality, preferred_aggregation=preferred_aggregation, ) self.metric_filename: Path = Path(metric_dir) / str(random.randint(0, 1048575)) self.f = open(self.metric_filename, 'a') def export( self, metrics_data: MetricsData, timeout_millis: float = 10_000, **kwargs, ) -> MetricExportResult: self.f.write(metrics_data.to_json()) self.f.write('\n') self.f.flush() return MetricExportResult.SUCCESS def shutdown(self, timeout_millis: float = 30_000, **kwargs) -> None: pass def force_flush(self, timeout_millis: float = 10_000) -> bool: return True def __del__(self): self.f.close() @pytest.fixture(scope='function') def monkeypatch_metric_exporter( tmpdir_factory: pytest.TempdirFactory, ) -> Tuple[Callable, Callable]: import opentelemetry.sdk.metrics.export from pathlib import Path import time import os import json collect_path = Path(tmpdir_factory.mktemp('otel-collector')) metrics_path = collect_path / 'metrics' os.mkdir(metrics_path) tick_counter_filename = collect_path / 'tick_counter' with open(tick_counter_filename, 'w') as f: f.write('0') def collect_metrics(): with open(tick_counter_filename, 'r') as f: tick_counter = int(f.read()) with open(tick_counter_filename, 'w') as f: f.write(str(tick_counter + 1)) time.sleep(2) def _get_service_name(otel_measurement): return otel_measurement[0]['resource_metrics'][0]['resource']['attributes'][ 'service.name' ] def read_metrics(): def read_metric_file(filename): with open(filename, 'r') as f: return list(map(json.loads, f.readlines())) return { _get_service_name(i): i for i in map(read_metric_file, metrics_path.glob('*')) } class PatchedTextReader(PeriodicExportingMetricReader): def __init__(self, *args, **kwargs) -> None: self.exporter = DirMetricExporter(metrics_path) self.tick_counter = 0 super().__init__( exporter=self.exporter, export_interval_millis=500, ) def _ticker(self) -> None: interval_secs = self._export_interval_millis / 1e3 while not self._shutdown_event.wait(interval_secs): with open(tick_counter_filename, 'r') as f: tick_counter = int(f.read()) if tick_counter != self.tick_counter: self.tick_counter = tick_counter self.collect(timeout_millis=self._export_timeout_millis) self.collect(timeout_millis=self._export_interval_millis) real_reader = opentelemetry.sdk.metrics.export.PeriodicExportingMetricReader opentelemetry.sdk.metrics.export.PeriodicExportingMetricReader = PatchedTextReader yield collect_metrics, read_metrics opentelemetry.sdk.metrics.export.PeriodicExportingMetricReader = real_reader
import random import pytest from pathlib import Path from typing import Dict, Tuple, Callable import opentelemetry.sdk.metrics.export import opentelemetry.sdk.metrics.view from opentelemetry.sdk.metrics.export import ( AggregationTemporality, MetricExporter, MetricExportResult, MetricsData, ) class DirMetricExporter(MetricExporter): """Implementation of :class:`MetricExporter` that prints metrics to a file in a given directory. This class can be used for diagnostic or testing purposes. """ def __init__( self, metric_dir: str, preferred_temporality: Dict[type, AggregationTemporality] = None, preferred_aggregation: Dict[ type, "opentelemetry.sdk.metrics.view.Aggregation" ] = None, ): super().__init__( preferred_temporality=preferred_temporality, preferred_aggregation=preferred_aggregation, ) self.metric_filename: Path = Path(metric_dir) / str(random.randint(0, 1048575)) self.f = open(self.metric_filename, 'a') def export( self, metrics_data: MetricsData, timeout_millis: float = 10_000, **kwargs, ) -> MetricExportResult: self.f.write(metrics_data.to_json()) self.f.write('\n') self.f.flush() return MetricExportResult.SUCCESS def shutdown(self, timeout_millis: float = 30_000, **kwargs) -> None: pass def force_flush(self, timeout_millis: float = 10_000) -> bool: return True def __del__(self): self.f.close() @pytest.fixture(scope='session') def monkeypatch_metric_exporter( tmpdir_factory: pytest.TempdirFactory, ) -> Tuple[Callable, Callable]: import opentelemetry.sdk.metrics.export from opentelemetry.sdk.metrics.export import PeriodicExportingMetricReader from pathlib import Path import time import os import json collect_path = Path(tmpdir_factory.mktemp('otel-collector')) metrics_path = collect_path / 'metrics' os.mkdir(metrics_path) tick_counter_filename = collect_path / 'tick_counter' with open(tick_counter_filename, 'w') as f: f.write('0') def collect_metrics(): with open(tick_counter_filename, 'r') as f: tick_counter = int(f.read()) with open(tick_counter_filename, 'w') as f: f.write(str(tick_counter + 1)) time.sleep(1) def _get_service_name(otel_measurement): return otel_measurement[0]['resource_metrics'][0]['resource']['attributes'][ 'service.name' ] def read_metrics(): def read_metric_file(filename): with open(filename, 'r') as f: return list(map(json.loads, f.readlines())) return { _get_service_name(i): i for i in map(read_metric_file, metrics_path.glob('*')) } class PatchedTextReader(PeriodicExportingMetricReader): def __init__(self, *args, **kwargs) -> None: self.exporter = DirMetricExporter(metrics_path) self.tick_counter = 0 super().__init__( exporter=self.exporter, export_interval_millis=1_000, ) def _ticker(self) -> None: interval_secs = self._export_interval_millis / 1e3 while not self._shutdown_event.wait(interval_secs): with open(tick_counter_filename, 'r') as f: tick_counter = int(f.read()) if tick_counter != self.tick_counter: self.tick_counter = tick_counter self.collect(timeout_millis=self._export_timeout_millis) self.collect(timeout_millis=self._export_interval_millis) opentelemetry.sdk.metrics.export.PeriodicExportingMetricReader = PatchedTextReader return collect_metrics, read_metrics
# Copyright (c) OpenMMLab. All rights reserved. from .builder import DATASETS, PIPELINES, build_dataset from .cityscapes import CityscapesDataset from .coco import CocoDataset from .coco_panoptic import CocoPanopticDataset from .dataset_wrappers import MultiImageMixDataset from .deepfashion import DeepFashionDataset from .lvis import LVISDataset, LVISV1Dataset, LVISV05Dataset from .openimages import OpenImagesChallengeDataset, OpenImagesDataset from .samplers import (AspectRatioBatchSampler, ClassAwareSampler, GroupMultiSourceSampler, MultiSourceSampler) from .utils import get_loading_pipeline from .voc import VOCDataset from .wider_face import WIDERFaceDataset from .xml_style import XMLDataset __all__ = [ 'XMLDataset', 'CocoDataset', 'DeepFashionDataset', 'VOCDataset', 'CityscapesDataset', 'LVISDataset', 'LVISV05Dataset', 'LVISV1Dataset', 'WIDERFaceDataset', 'DATASETS', 'PIPELINES', 'build_dataset', 'get_loading_pipeline', 'CocoPanopticDataset', 'MultiImageMixDataset', 'OpenImagesDataset', 'OpenImagesChallengeDataset', 'AspectRatioBatchSampler', 'ClassAwareSampler', 'MultiSourceSampler', 'GroupMultiSourceSampler' ]
# Copyright (c) OpenMMLab. All rights reserved. from .builder import DATASETS, PIPELINES, build_dataset from .cityscapes import CityscapesDataset from .coco import CocoDataset from .coco_panoptic import CocoPanopticDataset from .dataset_wrappers import MultiImageMixDataset from .deepfashion import DeepFashionDataset from .lvis import LVISDataset, LVISV1Dataset, LVISV05Dataset from .openimages import OpenImagesChallengeDataset, OpenImagesDataset from .samplers import AspectRatioBatchSampler, ClassAwareSampler from .utils import get_loading_pipeline from .voc import VOCDataset from .wider_face import WIDERFaceDataset from .xml_style import XMLDataset __all__ = [ 'XMLDataset', 'CocoDataset', 'DeepFashionDataset', 'VOCDataset', 'CityscapesDataset', 'LVISDataset', 'LVISV05Dataset', 'LVISV1Dataset', 'WIDERFaceDataset', 'DATASETS', 'PIPELINES', 'build_dataset', 'get_loading_pipeline', 'CocoPanopticDataset', 'MultiImageMixDataset', 'OpenImagesDataset', 'OpenImagesChallengeDataset', 'AspectRatioBatchSampler', 'ClassAwareSampler' ]
"""Linkup tool spec.""" from llama_index.core.tools.tool_spec.base import BaseToolSpec class LinkupToolSpec(BaseToolSpec): """Linkup tool spec.""" spec_functions = [ "search", ] def __init__(self, api_key: str, depth: str, output_type: str) -> None: """Initialize with parameters.""" from linkup import LinkupClient self.client = LinkupClient(api_key=api_key) self.depth = depth self.output_type = output_type def search(self, query: str): """ Run query through Linkup Search and return metadata. Args: query: The query to search for. """ api_params = { "query": query, "depth": self.depth, "output_type": self.output_type, } if self.output_type == "structured": if not self.structured_output_schema: raise ValueError( "structured_output_schema must be provided when output_type is 'structured'." ) api_params["structured_output_schema"] = self.structured_output_schema return self.client.search(**api_params)
"""Linkup tool spec.""" from llama_index.core.tools.tool_spec.base import BaseToolSpec class LinkupToolSpec(BaseToolSpec): """Linkup tool spec.""" spec_functions = [ "search", ] def __init__(self, api_key: str, depth: str, output_type: str) -> None: """Initialize with parameters.""" from linkup import LinkupClient self.client = LinkupClient(api_key=api_key) self.depth = depth self.output_type = output_type def search(self, query: str): """ Run query through Linkup Search and return metadata. Args: query: The query to search for. """ api_params = { "query": query, "depth": self.depth, "output_type": self.output_type, } if self.output_type == "structured": if not self.structured_output_schema: raise ValueError( "structured_output_schema must be provided when output_type is 'structured'." ) api_params["structured_output_schema"] = self.structured_output_schema return self.client.search(**api_params)
"""LLama Kibela Reader.""" from typing import Dict, Generic, List, Optional, TypeVar from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document from llama_index.core.bridge.pydantic import BaseModel NodeType = TypeVar("NodeType") class Edge(BaseModel, Generic[NodeType]): node: Optional[NodeType] cursor: Optional[str] class PageInfo(BaseModel): startCursor: Optional[str] endCursor: Optional[str] hasNextPage: Optional[bool] class Connection(BaseModel, Generic[NodeType]): nodes: Optional[List[NodeType]] = None edges: Optional[List[Edge[NodeType]]] pageInfo: Optional[PageInfo] totalCount: Optional[int] class Note(BaseModel): content: Optional[str] id: Optional[str] title: Optional[str] url: Optional[str] class KibelaReader(BaseReader): """ Kibela reader. Reads pages from Kibela. Args: team (str): Kibela team. token (str): Kibela API token. """ def __init__(self, team: str, token: str) -> None: """Initialize with parameters.""" from gql import Client from gql.transport.aiohttp import AIOHTTPTransport self.url = f"https://{team}.kibe.la/api/v1" self.headers = {"Authorization": f"Bearer {token}"} transport = AIOHTTPTransport(url=self.url, headers=self.headers) self.client = Client(transport=transport, fetch_schema_from_transport=True) def request(self, query: str, params: dict) -> Dict: from gql import gql q = gql(query) return self.client.execute(q, variable_values=params) def load_data(self) -> List[Document]: """ Load data from Kibela. Returns: List[Document]: List of documents. """ query = """ query getNotes($after: String) { notes(first: 100, after: $after) { totalCount pageInfo { endCursor startCursor hasNextPage } edges { cursor node { id url title content } } } } """ params = {"after": ""} has_next = True documents = [] # Due to the request limit of 10 requests per second on the Kibela API, we do not process in parallel. # See https://github.com/kibela/kibela-api-v1-document#1%E7%A7%92%E3%81%82%E3%81%9F%E3%82%8A%E3%81%AE%E3%83%AA%E3%82%AF%E3%82%A8%E3%82%B9%E3%83%88%E6%95%B0 while has_next: res = self.request(query, params) note_conn = Connection[Note].model_validate(res["notes"]) for note in note_conn.edges: doc = ( f"---\nurl: {note.node.url}\ntitle:" f" {note.node.title}\n---\ncontent:\n{note.node.content}\n" ) documents.append(Document(text=doc)) has_next = note_conn.pageInfo.hasNextPage params = {"after": note_conn.pageInfo.endCursor} return documents
"""LLama Kibela Reader.""" from typing import Dict, Generic, List, Optional, TypeVar from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document from llama_index.core.bridge.pydantic import BaseModel NodeType = TypeVar("NodeType") class Edge(BaseModel, Generic[NodeType]): node: Optional[NodeType] cursor: Optional[str] class PageInfo(BaseModel): startCursor: Optional[str] endCursor: Optional[str] hasNextPage: Optional[bool] class Connection(BaseModel, Generic[NodeType]): nodes: Optional[List[NodeType]] = None edges: Optional[List[Edge[NodeType]]] pageInfo: Optional[PageInfo] totalCount: Optional[int] class Note(BaseModel): content: Optional[str] id: Optional[str] title: Optional[str] url: Optional[str] class KibelaReader(BaseReader): """Kibela reader. Reads pages from Kibela. Args: team (str): Kibela team. token (str): Kibela API token. """ def __init__(self, team: str, token: str) -> None: """Initialize with parameters.""" from gql import Client from gql.transport.aiohttp import AIOHTTPTransport self.url = f"https://{team}.kibe.la/api/v1" self.headers = {"Authorization": f"Bearer {token}"} transport = AIOHTTPTransport(url=self.url, headers=self.headers) self.client = Client(transport=transport, fetch_schema_from_transport=True) def request(self, query: str, params: dict) -> Dict: from gql import gql q = gql(query) return self.client.execute(q, variable_values=params) def load_data(self) -> List[Document]: """Load data from Kibela. Returns: List[Document]: List of documents. """ query = """ query getNotes($after: String) { notes(first: 100, after: $after) { totalCount pageInfo { endCursor startCursor hasNextPage } edges { cursor node { id url title content } } } } """ params = {"after": ""} has_next = True documents = [] # Due to the request limit of 10 requests per second on the Kibela API, we do not process in parallel. # See https://github.com/kibela/kibela-api-v1-document#1%E7%A7%92%E3%81%82%E3%81%9F%E3%82%8A%E3%81%AE%E3%83%AA%E3%82%AF%E3%82%A8%E3%82%B9%E3%83%88%E6%95%B0 while has_next: res = self.request(query, params) note_conn = Connection[Note].model_validate(res["notes"]) for note in note_conn.edges: doc = ( f"---\nurl: {note.node.url}\ntitle:" f" {note.node.title}\n---\ncontent:\n{note.node.content}\n" ) documents.append(Document(text=doc)) has_next = note_conn.pageInfo.hasNextPage params = {"after": note_conn.pageInfo.endCursor} return documents
# Configuration file for the Sphinx documentation builder. # # This file only contains a selection of the most common options. For a full # list see the documentation: # https://www.sphinx-doc.org/en/master/usage/configuration.html # -- Path setup -------------------------------------------------------------- # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. # # import os # import sys # sys.path.insert(0, os.path.abspath('.')) import datetime import os from recommonmark.transform import AutoStructify from sphinx.domains import Domain # -- Project information ----------------------------------------------------- project = "Sentence-Transformers" copyright = str(datetime.datetime.now().year) author = "Nils Reimers, Tom Aarsen" # -- General configuration --------------------------------------------------- # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. extensions = [ "sphinx.ext.napoleon", "sphinx.ext.autodoc", "recommonmark", "sphinx_markdown_tables", "sphinx.ext.intersphinx", "sphinx_tabs.tabs", ] # Add any paths that contain templates here, relative to this directory. templates_path = ["_templates"] # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. # This pattern also affects html_static_path and html_extra_path. exclude_patterns = [ "_build", "Thumbs.db", ".DS_Store", "nr_examples", "archived", "dist", "build", "output", "models", "model_card_template.md", ] intersphinx_mapping = { "datasets": ("https://huggingface.co/docs/datasets/main/en/", None), "transformers": ("https://huggingface.co/docs/transformers/main/en/", None), "torch": ("https://pytorch.org/docs/stable/", None), } # -- Options for HTML output ------------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. # html_theme = "sphinx_rtd_theme" html_theme_path = ["_themes"] html_theme_options = { "logo_only": True, "canonical_url": "https://www.sbert.net", "collapse_navigation": False, } # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ["_static"] html_css_files = [ "css/custom.css", ] html_js_files = [ "js/custom.js", ] html_show_sourcelink = False html_context = { "display_github": True, "github_user": "UKPLab", "github_repo": "sentence-transformers", "github_version": "master/", } html_logo = "img/logo.png" html_favicon = "img/favicon.ico" autoclass_content = "both" class GithubURLDomain(Domain): """ Resolve .py links to their respective Github URL """ name = "githuburl" ROOT = "https://github.com/UKPLab/sentence-transformers/tree/master" def resolve_any_xref(self, env, fromdocname, builder, target, node, contnode): if (target.endswith(".py") or target.endswith(".ipynb")) and not target.startswith("http"): from_folder = os.path.dirname(fromdocname) contnode["refuri"] = "/".join([self.ROOT, from_folder, target]) return [("githuburl:any", contnode)] return [] def setup(app): app.add_domain(GithubURLDomain) app.add_config_value( "recommonmark_config", { #'url_resolver': lambda url: github_doc_root + url, "auto_toc_tree_section": "Contents", }, True, ) app.add_transform(AutoStructify)
# Configuration file for the Sphinx documentation builder. # # This file only contains a selection of the most common options. For a full # list see the documentation: # https://www.sphinx-doc.org/en/master/usage/configuration.html # -- Path setup -------------------------------------------------------------- # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. # # import os # import sys # sys.path.insert(0, os.path.abspath('.')) from recommonmark.transform import AutoStructify import os from sphinx.domains import Domain import datetime # -- Project information ----------------------------------------------------- project = "Sentence-Transformers" copyright = str(datetime.datetime.now().year) author = "Nils Reimers, Tom Aarsen" # -- General configuration --------------------------------------------------- # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. extensions = [ "sphinx.ext.napoleon", "sphinx.ext.autodoc", "recommonmark", "sphinx_markdown_tables", "sphinx.ext.intersphinx", "sphinx_tabs.tabs", ] # Add any paths that contain templates here, relative to this directory. templates_path = ["_templates"] # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. # This pattern also affects html_static_path and html_extra_path. exclude_patterns = [ "_build", "Thumbs.db", ".DS_Store", "nr_examples", "archived", "dist", "build", "output", "models", "model_card_template.md", ] intersphinx_mapping = { "datasets": ("https://huggingface.co/docs/datasets/main/en/", None), "transformers": ("https://huggingface.co/docs/transformers/main/en/", None), "torch": ("https://pytorch.org/docs/stable/", None), } # -- Options for HTML output ------------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. # html_theme = "sphinx_rtd_theme" html_theme_path = ["_themes"] html_theme_options = { "logo_only": True, "canonical_url": "https://www.sbert.net", "collapse_navigation": False, } # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ["_static"] html_css_files = [ "css/custom.css", ] html_js_files = [ "js/custom.js", ] html_show_sourcelink = False html_context = { "display_github": True, "github_user": "UKPLab", "github_repo": "sentence-transformers", "github_version": "master/", } html_logo = "img/logo.png" html_favicon = "img/favicon.ico" autoclass_content = "both" class GithubURLDomain(Domain): """ Resolve .py links to their respective Github URL """ name = "githuburl" ROOT = "https://github.com/UKPLab/sentence-transformers/tree/master" def resolve_any_xref(self, env, fromdocname, builder, target, node, contnode): if (target.endswith(".py") or target.endswith(".ipynb")) and not target.startswith("http"): from_folder = os.path.dirname(fromdocname) contnode["refuri"] = "/".join([self.ROOT, from_folder, target]) return [("githuburl:any", contnode)] return [] def setup(app): app.add_domain(GithubURLDomain) app.add_config_value( "recommonmark_config", { #'url_resolver': lambda url: github_doc_root + url, "auto_toc_tree_section": "Contents", }, True, ) app.add_transform(AutoStructify)
# Copyright (c) OpenMMLab. All rights reserved. from .det_inferencer import DetInferencer from .inference import (async_inference_detector, inference_detector, init_detector) __all__ = [ 'init_detector', 'async_inference_detector', 'inference_detector', 'DetInferencer' ]
# Copyright (c) OpenMMLab. All rights reserved. from .inference import (async_inference_detector, inference_detector, init_detector) __all__ = [ 'init_detector', 'async_inference_detector', 'inference_detector', ]
import pytest from llama_index.voice_agents.openai.types import ( ConversationDeltaEvent, ConversationDoneEvent, ConversationSession, ConversationSessionUpdate, ) @pytest.fixture() def session_json() -> dict: return { "modalities": ["text", "audio"], "instructions": "You are a helpful assistant.", "voice": "sage", "input_audio_format": "pcm16", "output_audio_format": "pcm16", "input_audio_transcription": {"model": "whisper-1"}, "turn_detection": { "type": "server_vad", "threshold": 0.5, "prefix_padding_ms": 300, "silence_duration_ms": 500, "create_response": True, }, "tools": [], "tool_choice": "auto", "temperature": 0.8, "max_response_output_tokens": "inf", "speed": 1.1, "tracing": "auto", } def test_serialization(session_json: dict) -> None: start_event = ConversationSessionUpdate( type_t="session.update", session=ConversationSession(), ) assert start_event.model_dump(by_alias=True) == { "type": "session.update", "session": session_json, } def test_deserialization() -> None: message = {"type_t": "response.text.delta", "delta": "Hello", "item_id": "msg_001"} assert ConversationDeltaEvent.model_validate(message) == ConversationDeltaEvent( type_t="response.text.delta", delta="Hello", item_id="msg_001" ) message1 = { "type_t": "response.text.done", "text": "Hello world, this is a test!", "item_id": "msg_001", } assert ConversationDoneEvent.model_validate(message1) == ConversationDoneEvent( type_t="response.text.done", text="Hello world, this is a test!", item_id="msg_001", ) message2 = { "type_t": "response.audio_transcript.done", "transcript": "Hello world, this is a test!", "item_id": "msg_001", } assert ConversationDoneEvent.model_validate(message2) == ConversationDoneEvent( type_t="response.audio_transcript.done", transcript="Hello world, this is a test!", item_id="msg_001", )
import pytest from llama_index.voice_agents.openai.types import ( ConversationDeltaEvent, ConversationDoneEvent, ConversationSession, ConversationSessionUpdate, ) @pytest.fixture() def session_json() -> dict: return { "modalities": ["text", "audio"], "instructions": "You are a helpful assistant", "voice": "sage", "input_audio_format": "pcm16", "output_audio_format": "pcm16", "input_audio_transcription": {"model": "whisper-1"}, "turn_detection": { "type": "server_vad", "threshold": 0.5, "prefix_padding_ms": 300, "silence_duration_ms": 500, "create_response": True, }, "tools": [], "tool_choice": "auto", "temperature": 0.5, "max_response_output_tokens": "inf", "speed": 1.1, "tracing": "auto", } def test_serialization(session_json: dict) -> None: start_event = ConversationSessionUpdate( type_t="session.update", session=ConversationSession(), ) assert start_event.model_dump(by_alias=True) == { "type": "session.update", "session": session_json, } def test_deserialization() -> None: message = {"type_t": "response.text.delta", "delta": "Hello", "item_id": "msg_001"} assert ConversationDeltaEvent.model_validate(message) == ConversationDeltaEvent( type_t="response.text.delta", delta="Hello", item_id="msg_001" ) message1 = { "type_t": "response.text.done", "text": "Hello world, this is a test!", "item_id": "msg_001", } assert ConversationDoneEvent.model_validate(message1) == ConversationDoneEvent( type_t="response.text.done", text="Hello world, this is a test!", item_id="msg_001", ) message2 = { "type_t": "response.audio_transcript.done", "transcript": "Hello world, this is a test!", "item_id": "msg_001", } assert ConversationDoneEvent.model_validate(message2) == ConversationDoneEvent( type_t="response.audio_transcript.done", transcript="Hello world, this is a test!", item_id="msg_001", )
_base_ = [ '../_base_/models/ssd300.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] # dataset settings input_size = 300 train_pipeline = [ dict(type='LoadImageFromFile', backend_args={{_base_.backend_args}}), dict(type='LoadAnnotations', with_bbox=True), dict( type='Expand', mean={{_base_.model.data_preprocessor.mean}}, to_rgb={{_base_.model.data_preprocessor.bgr_to_rgb}}, ratio_range=(1, 4)), dict( type='MinIoURandomCrop', min_ious=(0.1, 0.3, 0.5, 0.7, 0.9), min_crop_size=0.3), dict(type='Resize', scale=(input_size, input_size), keep_ratio=False), dict(type='RandomFlip', prob=0.5), dict( type='PhotoMetricDistortion', brightness_delta=32, contrast_range=(0.5, 1.5), saturation_range=(0.5, 1.5), hue_delta=18), dict(type='PackDetInputs') ] test_pipeline = [ dict(type='LoadImageFromFile', backend_args={{_base_.backend_args}}), dict(type='Resize', scale=(input_size, input_size), keep_ratio=False), dict(type='LoadAnnotations', with_bbox=True), dict( type='PackDetInputs', meta_keys=('img_id', 'img_path', 'ori_shape', 'img_shape', 'scale_factor')) ] train_dataloader = dict( batch_size=8, num_workers=2, batch_sampler=None, dataset=dict( _delete_=True, type='RepeatDataset', times=5, dataset=dict( type={{_base_.dataset_type}}, data_root={{_base_.data_root}}, ann_file='annotations/instances_train2017.json', data_prefix=dict(img='train2017/'), filter_cfg=dict(filter_empty_gt=True, min_size=32), pipeline=train_pipeline, backend_args={{_base_.backend_args}}))) val_dataloader = dict(batch_size=8, dataset=dict(pipeline=test_pipeline)) test_dataloader = val_dataloader # optimizer optim_wrapper = dict( type='OptimWrapper', optimizer=dict(type='SGD', lr=2e-3, momentum=0.9, weight_decay=5e-4)) custom_hooks = [ dict(type='NumClassCheckHook'), dict(type='CheckInvalidLossHook', interval=50, priority='VERY_LOW') ] # NOTE: `auto_scale_lr` is for automatically scaling LR, # USER SHOULD NOT CHANGE ITS VALUES. # base_batch_size = (8 GPUs) x (8 samples per GPU) auto_scale_lr = dict(base_batch_size=64)
_base_ = [ '../_base_/models/ssd300.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] # dataset settings input_size = 300 train_pipeline = [ dict(type='LoadImageFromFile'), dict(type='LoadAnnotations', with_bbox=True), dict( type='Expand', mean={{_base_.model.data_preprocessor.mean}}, to_rgb={{_base_.model.data_preprocessor.bgr_to_rgb}}, ratio_range=(1, 4)), dict( type='MinIoURandomCrop', min_ious=(0.1, 0.3, 0.5, 0.7, 0.9), min_crop_size=0.3), dict(type='Resize', scale=(input_size, input_size), keep_ratio=False), dict(type='RandomFlip', prob=0.5), dict( type='PhotoMetricDistortion', brightness_delta=32, contrast_range=(0.5, 1.5), saturation_range=(0.5, 1.5), hue_delta=18), dict(type='PackDetInputs') ] test_pipeline = [ dict(type='LoadImageFromFile'), dict(type='Resize', scale=(input_size, input_size), keep_ratio=False), dict(type='LoadAnnotations', with_bbox=True), dict( type='PackDetInputs', meta_keys=('img_id', 'img_path', 'ori_shape', 'img_shape', 'scale_factor')) ] train_dataloader = dict( batch_size=8, num_workers=2, batch_sampler=None, dataset=dict( _delete_=True, type='RepeatDataset', times=5, dataset=dict( type={{_base_.dataset_type}}, data_root={{_base_.data_root}}, ann_file='annotations/instances_train2017.json', data_prefix=dict(img='train2017/'), filter_cfg=dict(filter_empty_gt=True, min_size=32), pipeline=train_pipeline))) val_dataloader = dict(batch_size=8, dataset=dict(pipeline=test_pipeline)) test_dataloader = val_dataloader # optimizer optim_wrapper = dict( type='OptimWrapper', optimizer=dict(type='SGD', lr=2e-3, momentum=0.9, weight_decay=5e-4)) custom_hooks = [ dict(type='NumClassCheckHook'), dict(type='CheckInvalidLossHook', interval=50, priority='VERY_LOW') ] # NOTE: `auto_scale_lr` is for automatically scaling LR, # USER SHOULD NOT CHANGE ITS VALUES. # base_batch_size = (8 GPUs) x (8 samples per GPU) auto_scale_lr = dict(base_batch_size=64)
_base_ = './ms-rcnn_x101-64x4d_fpn_1x_coco.py' # learning policy max_epochs = 24 train_cfg = dict( type='EpochBasedTrainLoop', max_epochs=max_epochs, val_interval=1) param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[16, 22], gamma=0.1) ]
_base_ = './ms_rcnn_x101_64x4d_fpn_1x_coco.py' # learning policy max_epochs = 24 train_cfg = dict( type='EpochBasedTrainLoop', max_epochs=max_epochs, val_interval=1) param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[16, 22], gamma=0.1) ]
from pathlib import Path from typing import Dict, Tuple, Union import torchaudio from torch import Tensor from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import extract_archive _URL = "https://datashare.ed.ac.uk/bitstream/handle/10283/3038/DR-VCTK.zip" _CHECKSUM = "781f12f4406ed36ed27ae3bce55da47ba176e2d8bae67319e389e07b2c9bd769" _SUPPORTED_SUBSETS = {"train", "test"} class DR_VCTK(Dataset): """*Device Recorded VCTK (Small subset version)* :cite:`Sarfjoo2018DeviceRV` dataset. Args: root (str or Path): Root directory where the dataset's top level directory is found. subset (str): The subset to use. Can be one of ``"train"`` and ``"test"``. (default: ``"train"``). download (bool): Whether to download the dataset if it is not found at root path. (default: ``False``). url (str): The URL to download the dataset from. (default: ``"https://datashare.ed.ac.uk/bitstream/handle/10283/3038/DR-VCTK.zip"``) """ def __init__( self, root: Union[str, Path], subset: str = "train", *, download: bool = False, url: str = _URL, ) -> None: if subset not in _SUPPORTED_SUBSETS: raise RuntimeError( f"The subset '{subset}' does not match any of the supported subsets: {_SUPPORTED_SUBSETS}" ) root = Path(root).expanduser() archive = root / "DR-VCTK.zip" self._subset = subset self._path = root / "DR-VCTK" / "DR-VCTK" self._clean_audio_dir = self._path / f"clean_{self._subset}set_wav_16k" self._noisy_audio_dir = self._path / f"device-recorded_{self._subset}set_wav_16k" self._config_filepath = self._path / "configurations" / f"{self._subset}_ch_log.txt" if not self._path.is_dir(): if not archive.is_file(): if not download: raise RuntimeError("Dataset not found. Please use `download=True` to download it.") download_url_to_file(url, archive, hash_prefix=_CHECKSUM) extract_archive(archive, root) self._config = self._load_config(self._config_filepath) self._filename_list = sorted(self._config) def _load_config(self, filepath: str) -> Dict[str, Tuple[str, int]]: # Skip header skip_rows = 2 if self._subset == "train" else 1 config = {} with open(filepath) as f: for i, line in enumerate(f): if i < skip_rows or not line: continue filename, source, channel_id = line.strip().split("\t") config[filename] = (source, int(channel_id)) return config def _load_dr_vctk_item(self, filename: str) -> Tuple[Tensor, int, Tensor, int, str, str, str, int]: speaker_id, utterance_id = filename.split(".")[0].split("_") source, channel_id = self._config[filename] file_clean_audio = self._clean_audio_dir / filename file_noisy_audio = self._noisy_audio_dir / filename waveform_clean, sample_rate_clean = torchaudio.load(file_clean_audio) waveform_noisy, sample_rate_noisy = torchaudio.load(file_noisy_audio) return ( waveform_clean, sample_rate_clean, waveform_noisy, sample_rate_noisy, speaker_id, utterance_id, source, channel_id, ) def __getitem__(self, n: int) -> Tuple[Tensor, int, Tensor, int, str, str, str, int]: """Load the n-th sample from the dataset. Args: n (int): The index of the sample to be loaded Returns: Tuple of the following items; Tensor: Clean waveform int: Sample rate of the clean waveform Tensor: Noisy waveform int: Sample rate of the noisy waveform str: Speaker ID str: Utterance ID str: Source int: Channel ID """ filename = self._filename_list[n] return self._load_dr_vctk_item(filename) def __len__(self) -> int: return len(self._filename_list)
from pathlib import Path from typing import Dict, Tuple, Union import torchaudio from torch import Tensor from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import extract_archive _URL = "https://datashare.ed.ac.uk/bitstream/handle/10283/3038/DR-VCTK.zip" _CHECKSUM = "781f12f4406ed36ed27ae3bce55da47ba176e2d8bae67319e389e07b2c9bd769" _SUPPORTED_SUBSETS = {"train", "test"} class DR_VCTK(Dataset): """Create a dataset for *Device Recorded VCTK (Small subset version)* :cite:`Sarfjoo2018DeviceRV`. Args: root (str or Path): Root directory where the dataset's top level directory is found. subset (str): The subset to use. Can be one of ``"train"`` and ``"test"``. (default: ``"train"``). download (bool): Whether to download the dataset if it is not found at root path. (default: ``False``). url (str): The URL to download the dataset from. (default: ``"https://datashare.ed.ac.uk/bitstream/handle/10283/3038/DR-VCTK.zip"``) """ def __init__( self, root: Union[str, Path], subset: str = "train", *, download: bool = False, url: str = _URL, ) -> None: if subset not in _SUPPORTED_SUBSETS: raise RuntimeError( f"The subset '{subset}' does not match any of the supported subsets: {_SUPPORTED_SUBSETS}" ) root = Path(root).expanduser() archive = root / "DR-VCTK.zip" self._subset = subset self._path = root / "DR-VCTK" / "DR-VCTK" self._clean_audio_dir = self._path / f"clean_{self._subset}set_wav_16k" self._noisy_audio_dir = self._path / f"device-recorded_{self._subset}set_wav_16k" self._config_filepath = self._path / "configurations" / f"{self._subset}_ch_log.txt" if not self._path.is_dir(): if not archive.is_file(): if not download: raise RuntimeError("Dataset not found. Please use `download=True` to download it.") download_url_to_file(url, archive, hash_prefix=_CHECKSUM) extract_archive(archive, root) self._config = self._load_config(self._config_filepath) self._filename_list = sorted(self._config) def _load_config(self, filepath: str) -> Dict[str, Tuple[str, int]]: # Skip header skip_rows = 2 if self._subset == "train" else 1 config = {} with open(filepath) as f: for i, line in enumerate(f): if i < skip_rows or not line: continue filename, source, channel_id = line.strip().split("\t") config[filename] = (source, int(channel_id)) return config def _load_dr_vctk_item(self, filename: str) -> Tuple[Tensor, int, Tensor, int, str, str, str, int]: speaker_id, utterance_id = filename.split(".")[0].split("_") source, channel_id = self._config[filename] file_clean_audio = self._clean_audio_dir / filename file_noisy_audio = self._noisy_audio_dir / filename waveform_clean, sample_rate_clean = torchaudio.load(file_clean_audio) waveform_noisy, sample_rate_noisy = torchaudio.load(file_noisy_audio) return ( waveform_clean, sample_rate_clean, waveform_noisy, sample_rate_noisy, speaker_id, utterance_id, source, channel_id, ) def __getitem__(self, n: int) -> Tuple[Tensor, int, Tensor, int, str, str, str, int]: """Load the n-th sample from the dataset. Args: n (int): The index of the sample to be loaded Returns: (Tensor, int, Tensor, int, str, str, str, int): ``(waveform_clean, sample_rate_clean, waveform_noisy, sample_rate_noisy, speaker_id,\ utterance_id, source, channel_id)`` """ filename = self._filename_list[n] return self._load_dr_vctk_item(filename) def __len__(self) -> int: return len(self._filename_list)
_base_ = './fovea_r50_fpn_4xb4-1x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
_base_ = './fovea_r50_fpn_4x4_1x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
from typing import Dict, List import numpy as np import pytest from docarray import DocList from docarray.base_doc import AnyDoc, BaseDoc from docarray.typing import NdArray def test_any_doc(): class InnerDocument(BaseDoc): text: str tensor: NdArray class CustomDoc(BaseDoc): inner: InnerDocument text: str doc = CustomDoc( text='bye', inner=InnerDocument(text='hello', tensor=np.zeros((3, 224, 224))) ) any_doc = AnyDoc(**doc.__dict__) assert any_doc.text == doc.text assert any_doc.inner.text == doc.inner.text assert (any_doc.inner.tensor == doc.inner.tensor).all() @pytest.mark.parametrize('protocol', ['proto', 'json']) def test_any_document_from_to(protocol): class InnerDoc(BaseDoc): text: str t: Dict[str, str] class DocTest(BaseDoc): text: str tags: Dict[str, int] l_: List[int] d: InnerDoc ld: DocList[InnerDoc] inner_doc = InnerDoc(text='I am inner', t={'a': 'b'}) da = DocList[DocTest]( [ DocTest( text='type1', tags={'type': 1}, l_=[1, 2], d=inner_doc, ld=DocList[InnerDoc]([inner_doc]), ), DocTest( text='type2', tags={'type': 2}, l_=[1, 2], d=inner_doc, ld=DocList[InnerDoc]([inner_doc]), ), ] ) from docarray.base_doc import AnyDoc if protocol == 'proto': aux = DocList[AnyDoc].from_protobuf(da.to_protobuf()) else: aux = DocList[AnyDoc].from_json(da.to_json()) assert len(aux) == 2 assert len(aux.id) == 2 for i, d in enumerate(aux): assert d.tags['type'] == i + 1 assert d.text == f'type{i + 1}' assert d.l_ == [1, 2] if protocol == 'proto': assert isinstance(d.d, AnyDoc) assert d.d.text == 'I am inner' # inner Document is a Dict assert d.d.t == {'a': 'b'} else: assert isinstance(d.d, dict) assert d.d['text'] == 'I am inner' # inner Document is a Dict assert d.d['t'] == {'a': 'b'} assert len(d.ld) == 1 if protocol == 'proto': assert isinstance(d.ld[0], AnyDoc) assert d.ld[0].text == 'I am inner' assert d.ld[0].t == {'a': 'b'} else: assert isinstance(d.ld[0], dict) assert d.ld[0]['text'] == 'I am inner' assert d.ld[0]['t'] == {'a': 'b'}
from typing import Dict, List import numpy as np import pytest from orjson import orjson from docarray import DocList from docarray.base_doc import AnyDoc, BaseDoc from docarray.base_doc.io.json import orjson_dumps_and_decode from docarray.typing import NdArray from docarray.typing.tensor.abstract_tensor import AbstractTensor from docarray.utils._internal.pydantic import is_pydantic_v2 def test_any_doc(): class InnerDocument(BaseDoc): text: str tensor: NdArray class CustomDoc(BaseDoc): inner: InnerDocument text: str doc = CustomDoc( text='bye', inner=InnerDocument(text='hello', tensor=np.zeros((3, 224, 224))) ) any_doc = AnyDoc(**doc.__dict__) assert any_doc.text == doc.text assert any_doc.inner.text == doc.inner.text assert (any_doc.inner.tensor == doc.inner.tensor).all() @pytest.mark.parametrize('protocol', ['proto', 'json']) def test_any_document_from_to(protocol): class InnerDoc(BaseDoc): text: str t: Dict[str, str] class DocTest(BaseDoc): text: str tags: Dict[str, int] l_: List[int] d: InnerDoc ld: DocList[InnerDoc] inner_doc = InnerDoc(text='I am inner', t={'a': 'b'}) da = DocList[DocTest]( [ DocTest( text='type1', tags={'type': 1}, l_=[1, 2], d=inner_doc, ld=DocList[InnerDoc]([inner_doc]), ), DocTest( text='type2', tags={'type': 2}, l_=[1, 2], d=inner_doc, ld=DocList[InnerDoc]([inner_doc]), ), ] ) from docarray.base_doc import AnyDoc if protocol == 'proto': aux = DocList[AnyDoc].from_protobuf(da.to_protobuf()) else: aux = DocList[AnyDoc].from_json(da.to_json()) assert len(aux) == 2 assert len(aux.id) == 2 for i, d in enumerate(aux): assert d.tags['type'] == i + 1 assert d.text == f'type{i + 1}' assert d.l_ == [1, 2] if protocol == 'proto': assert isinstance(d.d, AnyDoc) assert d.d.text == 'I am inner' # inner Document is a Dict assert d.d.t == {'a': 'b'} else: assert isinstance(d.d, dict) assert d.d['text'] == 'I am inner' # inner Document is a Dict assert d.d['t'] == {'a': 'b'} assert len(d.ld) == 1 if protocol == 'proto': assert isinstance(d.ld[0], AnyDoc) assert d.ld[0].text == 'I am inner' assert d.ld[0].t == {'a': 'b'} else: assert isinstance(d.ld[0], dict) assert d.ld[0]['text'] == 'I am inner' assert d.ld[0]['t'] == {'a': 'b'} @pytest.mark.skipif(is_pydantic_v2, reason="Not working with pydantic v2 for now") def test_subclass_config(): class MyDoc(BaseDoc): x: str class Config(BaseDoc.Config): arbitrary_types_allowed = True # just an example setting assert MyDoc.Config.json_loads == orjson.loads assert MyDoc.Config.json_dumps == orjson_dumps_and_decode assert ( MyDoc.Config.json_encoders[AbstractTensor](3) == 3 ) # dirty check that it is identity assert MyDoc.Config.validate_assignment assert not MyDoc.Config._load_extra_fields_from_protobuf assert MyDoc.Config.arbitrary_types_allowed
_base_ = [ '../_base_/models/cascade-mask-rcnn_r50_fpn.py', '../_base_/datasets/coco_instance.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')), roi_head=dict( bbox_head=[ dict( type='Shared2FCBBoxHead', in_channels=256, fc_out_channels=1024, roi_feat_size=7, num_classes=1203, bbox_coder=dict( type='DeltaXYWHBBoxCoder', target_means=[0., 0., 0., 0.], target_stds=[0.1, 0.1, 0.2, 0.2]), reg_class_agnostic=True, cls_predictor_cfg=dict(type='NormedLinear', tempearture=20), loss_cls=dict( type='SeesawLoss', p=0.8, q=2.0, num_classes=1203, loss_weight=1.0), loss_bbox=dict(type='SmoothL1Loss', beta=1.0, loss_weight=1.0)), dict( type='Shared2FCBBoxHead', in_channels=256, fc_out_channels=1024, roi_feat_size=7, num_classes=1203, bbox_coder=dict( type='DeltaXYWHBBoxCoder', target_means=[0., 0., 0., 0.], target_stds=[0.05, 0.05, 0.1, 0.1]), reg_class_agnostic=True, cls_predictor_cfg=dict(type='NormedLinear', tempearture=20), loss_cls=dict( type='SeesawLoss', p=0.8, q=2.0, num_classes=1203, loss_weight=1.0), loss_bbox=dict(type='SmoothL1Loss', beta=1.0, loss_weight=1.0)), dict( type='Shared2FCBBoxHead', in_channels=256, fc_out_channels=1024, roi_feat_size=7, num_classes=1203, bbox_coder=dict( type='DeltaXYWHBBoxCoder', target_means=[0., 0., 0., 0.], target_stds=[0.033, 0.033, 0.067, 0.067]), reg_class_agnostic=True, cls_predictor_cfg=dict(type='NormedLinear', tempearture=20), loss_cls=dict( type='SeesawLoss', p=0.8, q=2.0, num_classes=1203, loss_weight=1.0), loss_bbox=dict(type='SmoothL1Loss', beta=1.0, loss_weight=1.0)) ], mask_head=dict(num_classes=1203)), test_cfg=dict( rcnn=dict( score_thr=0.0001, # LVIS allows up to 300 max_per_img=300))) # dataset settings train_pipeline = [ dict(type='LoadImageFromFile', backend_args={{_base_.backend_args}}), dict(type='LoadAnnotations', with_bbox=True, with_mask=True), dict( type='RandomChoiceResize', scales=[(1333, 640), (1333, 672), (1333, 704), (1333, 736), (1333, 768), (1333, 800)], keep_ratio=True), dict(type='RandomFlip', prob=0.5), dict(type='PackDetInputs') ] dataset_type = 'LVISV1Dataset' data_root = 'data/lvis_v1/' train_dataloader = dict( dataset=dict( type=dataset_type, data_root=data_root, ann_file='annotations/lvis_v1_train.json', data_prefix=dict(img=''), pipeline=train_pipeline)) val_dataloader = dict( dataset=dict( type=dataset_type, data_root=data_root, ann_file='annotations/lvis_v1_val.json', data_prefix=dict(img=''))) test_dataloader = val_dataloader val_evaluator = dict( type='LVISMetric', ann_file=data_root + 'annotations/lvis_v1_val.json', metric=['bbox', 'segm']) test_evaluator = val_evaluator train_cfg = dict(val_interval=24)
_base_ = [ '../_base_/models/cascade-mask-rcnn_r50_fpn.py', '../_base_/datasets/coco_instance.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')), roi_head=dict( bbox_head=[ dict( type='Shared2FCBBoxHead', in_channels=256, fc_out_channels=1024, roi_feat_size=7, num_classes=1203, bbox_coder=dict( type='DeltaXYWHBBoxCoder', target_means=[0., 0., 0., 0.], target_stds=[0.1, 0.1, 0.2, 0.2]), reg_class_agnostic=True, cls_predictor_cfg=dict(type='NormedLinear', tempearture=20), loss_cls=dict( type='SeesawLoss', p=0.8, q=2.0, num_classes=1203, loss_weight=1.0), loss_bbox=dict(type='SmoothL1Loss', beta=1.0, loss_weight=1.0)), dict( type='Shared2FCBBoxHead', in_channels=256, fc_out_channels=1024, roi_feat_size=7, num_classes=1203, bbox_coder=dict( type='DeltaXYWHBBoxCoder', target_means=[0., 0., 0., 0.], target_stds=[0.05, 0.05, 0.1, 0.1]), reg_class_agnostic=True, cls_predictor_cfg=dict(type='NormedLinear', tempearture=20), loss_cls=dict( type='SeesawLoss', p=0.8, q=2.0, num_classes=1203, loss_weight=1.0), loss_bbox=dict(type='SmoothL1Loss', beta=1.0, loss_weight=1.0)), dict( type='Shared2FCBBoxHead', in_channels=256, fc_out_channels=1024, roi_feat_size=7, num_classes=1203, bbox_coder=dict( type='DeltaXYWHBBoxCoder', target_means=[0., 0., 0., 0.], target_stds=[0.033, 0.033, 0.067, 0.067]), reg_class_agnostic=True, cls_predictor_cfg=dict(type='NormedLinear', tempearture=20), loss_cls=dict( type='SeesawLoss', p=0.8, q=2.0, num_classes=1203, loss_weight=1.0), loss_bbox=dict(type='SmoothL1Loss', beta=1.0, loss_weight=1.0)) ], mask_head=dict(num_classes=1203)), test_cfg=dict( rcnn=dict( score_thr=0.0001, # LVIS allows up to 300 max_per_img=300))) # dataset settings train_pipeline = [ dict( type='LoadImageFromFile', file_client_args={{_base_.file_client_args}}), dict(type='LoadAnnotations', with_bbox=True, with_mask=True), dict( type='RandomChoiceResize', scales=[(1333, 640), (1333, 672), (1333, 704), (1333, 736), (1333, 768), (1333, 800)], keep_ratio=True), dict(type='RandomFlip', prob=0.5), dict(type='PackDetInputs') ] dataset_type = 'LVISV1Dataset' data_root = 'data/lvis_v1/' train_dataloader = dict( dataset=dict( type=dataset_type, data_root=data_root, ann_file='annotations/lvis_v1_train.json', data_prefix=dict(img=''), pipeline=train_pipeline)) val_dataloader = dict( dataset=dict( type=dataset_type, data_root=data_root, ann_file='annotations/lvis_v1_val.json', data_prefix=dict(img=''))) test_dataloader = val_dataloader val_evaluator = dict( type='LVISMetric', ann_file=data_root + 'annotations/lvis_v1_val.json', metric=['bbox', 'segm']) test_evaluator = val_evaluator train_cfg = dict(val_interval=24)
import importlib.util import warnings from functools import wraps from typing import Optional def is_module_available(*modules: str) -> bool: r"""Returns if a top-level module with :attr:`name` exists *without** importing it. This is generally safer than try-catch block around a `import X`. It avoids third party libraries breaking assumptions of some of our tests, e.g., setting multiprocessing start method when imported (see librosa/#747, torchvision/#544). """ return all(importlib.util.find_spec(m) is not None for m in modules) def requires_module(*modules: str): """Decorate function to give error message if invoked without required optional modules. This decorator is to give better error message to users rather than raising ``NameError: name 'module' is not defined`` at random places. """ missing = [m for m in modules if not is_module_available(m)] if not missing: # fall through. If all the modules are available, no need to decorate def decorator(func): return func else: req = f"module: {missing[0]}" if len(missing) == 1 else f"modules: {missing}" def decorator(func): @wraps(func) def wrapped(*args, **kwargs): raise RuntimeError(f"{func.__module__}.{func.__name__} requires {req}") return wrapped return decorator def deprecated(direction: str, version: Optional[str] = None): """Decorator to add deprecation message Args: direction (str): Migration steps to be given to users. version (str or int): The version when the object will be removed """ def decorator(func): @wraps(func) def wrapped(*args, **kwargs): message = ( f"{func.__module__}.{func.__name__} has been deprecated " f'and will be removed from {"future" if version is None else version} release. ' f"{direction}" ) warnings.warn(message, stacklevel=2) return func(*args, **kwargs) return wrapped return decorator def is_kaldi_available(): try: import torchaudio.lib._torchaudio return torchaudio.lib._torchaudio.is_kaldi_available() except Exception: return False def requires_kaldi(): if is_kaldi_available(): def decorator(func): return func else: def decorator(func): @wraps(func) def wrapped(*args, **kwargs): raise RuntimeError(f"{func.__module__}.{func.__name__} requires kaldi") return wrapped return decorator def _check_soundfile_importable(): if not is_module_available("soundfile"): return False try: import soundfile # noqa: F401 return True except Exception: warnings.warn("Failed to import soundfile. 'soundfile' backend is not available.") return False _is_soundfile_importable = _check_soundfile_importable() def is_soundfile_available(): return _is_soundfile_importable def requires_soundfile(): if is_soundfile_available(): def decorator(func): return func else: def decorator(func): @wraps(func) def wrapped(*args, **kwargs): raise RuntimeError(f"{func.__module__}.{func.__name__} requires soundfile") return wrapped return decorator def is_sox_available(): return is_module_available("torchaudio.lib._torchaudio_sox") def requires_sox(): if is_sox_available(): def decorator(func): return func else: def decorator(func): @wraps(func) def wrapped(*args, **kwargs): raise RuntimeError(f"{func.__module__}.{func.__name__} requires sox") return wrapped return decorator
import importlib.util import warnings from functools import wraps from typing import Optional import torch def is_module_available(*modules: str) -> bool: r"""Returns if a top-level module with :attr:`name` exists *without** importing it. This is generally safer than try-catch block around a `import X`. It avoids third party libraries breaking assumptions of some of our tests, e.g., setting multiprocessing start method when imported (see librosa/#747, torchvision/#544). """ return all(importlib.util.find_spec(m) is not None for m in modules) def requires_module(*modules: str): """Decorate function to give error message if invoked without required optional modules. This decorator is to give better error message to users rather than raising ``NameError: name 'module' is not defined`` at random places. """ missing = [m for m in modules if not is_module_available(m)] if not missing: # fall through. If all the modules are available, no need to decorate def decorator(func): return func else: req = f"module: {missing[0]}" if len(missing) == 1 else f"modules: {missing}" def decorator(func): @wraps(func) def wrapped(*args, **kwargs): raise RuntimeError(f"{func.__module__}.{func.__name__} requires {req}") return wrapped return decorator def deprecated(direction: str, version: Optional[str] = None): """Decorator to add deprecation message Args: direction (str): Migration steps to be given to users. version (str or int): The version when the object will be removed """ def decorator(func): @wraps(func) def wrapped(*args, **kwargs): message = ( f"{func.__module__}.{func.__name__} has been deprecated " f'and will be removed from {"future" if version is None else version} release. ' f"{direction}" ) warnings.warn(message, stacklevel=2) return func(*args, **kwargs) return wrapped return decorator def is_kaldi_available(): try: import torchaudio.lib._torchaudio return torchaudio.lib._torchaudio.is_kaldi_available() except Exception: return False def requires_kaldi(): if is_kaldi_available(): def decorator(func): return func else: def decorator(func): @wraps(func) def wrapped(*args, **kwargs): raise RuntimeError(f"{func.__module__}.{func.__name__} requires kaldi") return wrapped return decorator def _check_soundfile_importable(): if not is_module_available("soundfile"): return False try: import soundfile # noqa: F401 return True except Exception: warnings.warn("Failed to import soundfile. 'soundfile' backend is not available.") return False _is_soundfile_importable = _check_soundfile_importable() def is_soundfile_available(): return _is_soundfile_importable def requires_soundfile(): if is_soundfile_available(): def decorator(func): return func else: def decorator(func): @wraps(func) def wrapped(*args, **kwargs): raise RuntimeError(f"{func.__module__}.{func.__name__} requires soundfile") return wrapped return decorator def is_sox_available(): return is_module_available("torchaudio.lib._torchaudio_sox") def requires_sox(): if is_sox_available(): def decorator(func): return func else: def decorator(func): @wraps(func) def wrapped(*args, **kwargs): raise RuntimeError(f"{func.__module__}.{func.__name__} requires sox") return wrapped return decorator
try: from llama_index.readers.imdb_review.scraper import main_scraper except ImportError: from scraper import main_scraper from typing import List from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class IMDBReviews(BaseReader): def __init__( self, movie_name_year: str, webdriver_engine: str = "google", generate_csv: bool = False, multithreading: bool = False, max_workers: int = 0, reviews_folder: str = "movie_reviews", ): """ Get the IMDB reviews of a movie. Args: movie_name_year (str): movie name alongwith year webdriver_engine (str, optional): webdriver engine to use. Defaults to "google". generate_csv (bool, optional): whether to generate csv. Defaults to False. multithreading (bool, optional): whether to use multithreading. Defaults to False. max_workers (int, optional): number of workers if you are using multithreading. Defaults to 0. """ assert webdriver_engine in [ "google", "edge", "firefox", ], "The webdriver should be in ['google','edge','firefox']" self.movie_name_year = movie_name_year self.webdriver_engine = webdriver_engine self.generate_csv = generate_csv self.multithreading = multithreading self.max_workers = max_workers self.reviews_folder = reviews_folder def load_data(self) -> List[Document]: """ Scrapes the data from the IMDB website movie reviews. Returns: List[Document]: document object in llama index with date and rating as extra information """ ( reviews_date, reviews_title, reviews_comment, reviews_rating, reviews_link, review_helpful, review_total_votes, review_if_spoiler, ) = main_scraper( self.movie_name_year, self.webdriver_engine, self.generate_csv, self.multithreading, self.max_workers, self.reviews_folder, ) all_docs = [] for i in range(len(reviews_date)): all_docs.append( Document( text=reviews_title[i] + " " + reviews_comment[i], extra_info={ "date": reviews_date[i], "rating": reviews_rating[i], "link": reviews_link[i], "found_helpful_votes": review_helpful[i], "total_votes": review_total_votes[i], "spolier": review_if_spoiler[i], }, ) ) return all_docs
try: from llama_index.readers.imdb_review.scraper import main_scraper except ImportError: from scraper import main_scraper from typing import List from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class IMDBReviews(BaseReader): def __init__( self, movie_name_year: str, webdriver_engine: str = "google", generate_csv: bool = False, multithreading: bool = False, max_workers: int = 0, reviews_folder: str = "movie_reviews", ): """Get the IMDB reviews of a movie. Args: movie_name_year (str): movie name alongwith year webdriver_engine (str, optional): webdriver engine to use. Defaults to "google". generate_csv (bool, optional): whether to generate csv. Defaults to False. multithreading (bool, optional): whether to use multithreading. Defaults to False. max_workers (int, optional): number of workers if you are using multithreading. Defaults to 0. """ assert webdriver_engine in [ "google", "edge", "firefox", ], "The webdriver should be in ['google','edge','firefox']" self.movie_name_year = movie_name_year self.webdriver_engine = webdriver_engine self.generate_csv = generate_csv self.multithreading = multithreading self.max_workers = max_workers self.reviews_folder = reviews_folder def load_data(self) -> List[Document]: """Scrapes the data from the IMDB website movie reviews. Returns: List[Document]: document object in llama index with date and rating as extra information """ ( reviews_date, reviews_title, reviews_comment, reviews_rating, reviews_link, review_helpful, review_total_votes, review_if_spoiler, ) = main_scraper( self.movie_name_year, self.webdriver_engine, self.generate_csv, self.multithreading, self.max_workers, self.reviews_folder, ) all_docs = [] for i in range(len(reviews_date)): all_docs.append( Document( text=reviews_title[i] + " " + reviews_comment[i], extra_info={ "date": reviews_date[i], "rating": reviews_rating[i], "link": reviews_link[i], "found_helpful_votes": review_helpful[i], "total_votes": review_total_votes[i], "spolier": review_if_spoiler[i], }, ) ) return all_docs
from __future__ import annotations from collections import Counter import pytest from sentence_transformers.sampler import GroupByLabelBatchSampler from sentence_transformers.util import is_datasets_available if is_datasets_available(): from datasets import Dataset else: pytest.skip( reason='Sentence Transformers was not installed with the `["train"]` extra.', allow_module_level=True, ) @pytest.fixture def dummy_dataset(): """ Dummy dataset for testing purposes. The dataset looks as follows: { "data": [0, 1, 2, ..., 99], "label_a": [0, 1, 0, 1, ..., 0, 1], "label_b": [0, 1, 2, 3, 4, 0, ..., 4] } """ data = {"data": list(range(100)), "label_a": [i % 2 for i in range(100)], "label_b": [i % 5 for i in range(100)]} return Dataset.from_dict(data) @pytest.fixture def dummy_uneven_dataset(): """ Dummy dataset for testing purposes. The dataset looks as follows: { "data": ["a"] * 51, "label": [0] * 17 + [1] * 17 + [2] * 17, } """ data = {"data": ["a"] * 51, "label": [0] * 17 + [1] * 17 + [2] * 17} return Dataset.from_dict(data) def test_group_by_label_batch_sampler_label_a(dummy_dataset: Dataset) -> None: batch_size = 10 sampler = GroupByLabelBatchSampler( dataset=dummy_dataset, batch_size=batch_size, drop_last=False, valid_label_columns=["label_a", "label_b"] ) batches = list(iter(sampler)) assert all(len(batch) == batch_size for batch in batches) # Check if all labels within each batch are identical # In this case, label_a has 50 0's and 50 1's, so with a batch size of 10 we expect each batch to # have only 0's or only 1's. for batch in batches: labels = [dummy_dataset[int(idx)]["label_a"] for idx in batch] assert len(set(labels)) == 1, f"Batch {batch} does not have identical labels: {labels}" def test_group_by_label_batch_sampler_label_b(dummy_dataset: Dataset) -> None: batch_size = 8 sampler = GroupByLabelBatchSampler( dataset=dummy_dataset, batch_size=batch_size, drop_last=True, valid_label_columns=["label_b"] ) # drop_last=True, so each batch should be the same length and the last batch is dropped. batches = list(iter(sampler)) assert all( len(batch) == batch_size for batch in batches ), "Not all batches are the same size, while drop_last was True." # Assert that we have the expected number of total samples in the batches. assert sum(len(batch) for batch in batches) == 100 // batch_size * batch_size # Since we have 20 occurrences each of label_b values 0, 1, 2, 3 and 4 and a batch_size of 8, we expect each batch # to have either 4 or 8 samples with the same label. (The first two batches are 16 samples of the first label, # leaving 4 for the third batch. There 4 of the next label are added, leaving 16 for the next two batches, and so on.) for batch in batches: labels = [dummy_dataset[int(idx)]["label_b"] for idx in batch] counts = list(Counter(labels).values()) assert counts == [8] or counts == [4, 4] def test_group_by_label_batch_sampler_uneven_dataset(dummy_uneven_dataset: Dataset) -> None: batch_size = 8 sampler = GroupByLabelBatchSampler( dataset=dummy_uneven_dataset, batch_size=batch_size, drop_last=False, valid_label_columns=["label"] ) # With a batch_size of 8 and 17 samples per label; verify that every label in a batch occurs at least twice. # We accept some tiny data loss (1 sample per label) due to the uneven number of samples per label. batches = list(iter(sampler)) for batch in batches: labels = [dummy_uneven_dataset[int(idx)]["label"] for idx in batch] counts = list(Counter(labels).values()) assert [count > 1 for count in counts]
from __future__ import annotations from collections import Counter import pytest from datasets import Dataset from sentence_transformers.sampler import GroupByLabelBatchSampler @pytest.fixture def dummy_dataset(): """ Dummy dataset for testing purposes. The dataset looks as follows: { "data": [0, 1, 2, ..., 99], "label_a": [0, 1, 0, 1, ..., 0, 1], "label_b": [0, 1, 2, 3, 4, 0, ..., 4] } """ data = {"data": list(range(100)), "label_a": [i % 2 for i in range(100)], "label_b": [i % 5 for i in range(100)]} return Dataset.from_dict(data) @pytest.fixture def dummy_uneven_dataset(): """ Dummy dataset for testing purposes. The dataset looks as follows: { "data": ["a"] * 51, "label": [0] * 17 + [1] * 17 + [2] * 17, } """ data = {"data": ["a"] * 51, "label": [0] * 17 + [1] * 17 + [2] * 17} return Dataset.from_dict(data) def test_group_by_label_batch_sampler_label_a(dummy_dataset: Dataset) -> None: batch_size = 10 sampler = GroupByLabelBatchSampler( dataset=dummy_dataset, batch_size=batch_size, drop_last=False, valid_label_columns=["label_a", "label_b"] ) batches = list(iter(sampler)) assert all(len(batch) == batch_size for batch in batches) # Check if all labels within each batch are identical # In this case, label_a has 50 0's and 50 1's, so with a batch size of 10 we expect each batch to # have only 0's or only 1's. for batch in batches: labels = [dummy_dataset[int(idx)]["label_a"] for idx in batch] assert len(set(labels)) == 1, f"Batch {batch} does not have identical labels: {labels}" def test_group_by_label_batch_sampler_label_b(dummy_dataset: Dataset) -> None: batch_size = 8 sampler = GroupByLabelBatchSampler( dataset=dummy_dataset, batch_size=batch_size, drop_last=True, valid_label_columns=["label_b"] ) # drop_last=True, so each batch should be the same length and the last batch is dropped. batches = list(iter(sampler)) assert all( len(batch) == batch_size for batch in batches ), "Not all batches are the same size, while drop_last was True." # Assert that we have the expected number of total samples in the batches. assert sum(len(batch) for batch in batches) == 100 // batch_size * batch_size # Since we have 20 occurrences each of label_b values 0, 1, 2, 3 and 4 and a batch_size of 8, we expect each batch # to have either 4 or 8 samples with the same label. (The first two batches are 16 samples of the first label, # leaving 4 for the third batch. There 4 of the next label are added, leaving 16 for the next two batches, and so on.) for batch in batches: labels = [dummy_dataset[int(idx)]["label_b"] for idx in batch] counts = list(Counter(labels).values()) assert counts == [8] or counts == [4, 4] def test_group_by_label_batch_sampler_uneven_dataset(dummy_uneven_dataset: Dataset) -> None: batch_size = 8 sampler = GroupByLabelBatchSampler( dataset=dummy_uneven_dataset, batch_size=batch_size, drop_last=False, valid_label_columns=["label"] ) # With a batch_size of 8 and 17 samples per label; verify that every label in a batch occurs at least twice. # We accept some tiny data loss (1 sample per label) due to the uneven number of samples per label. batches = list(iter(sampler)) for batch in batches: labels = [dummy_uneven_dataset[int(idx)]["label"] for idx in batch] counts = list(Counter(labels).values()) assert [count > 1 for count in counts]
from typing import Any, Dict, Iterator import torch from ..utils import _log_api_usage_once try: from ._load_gpu_decoder import _HAS_GPU_VIDEO_DECODER except ModuleNotFoundError: _HAS_GPU_VIDEO_DECODER = False from ._video_opt import ( _HAS_CPU_VIDEO_DECODER, _HAS_VIDEO_OPT, _probe_video_from_file, _probe_video_from_memory, _read_video_from_file, _read_video_from_memory, _read_video_timestamps_from_file, _read_video_timestamps_from_memory, Timebase, VideoMetaData, ) from .image import ( decode_gif, decode_image, decode_jpeg, decode_png, decode_webp, encode_jpeg, encode_png, ImageReadMode, read_file, read_image, write_file, write_jpeg, write_png, ) from .video import read_video, read_video_timestamps, write_video from .video_reader import VideoReader __all__ = [ "write_video", "read_video", "read_video_timestamps", "_read_video_from_file", "_read_video_timestamps_from_file", "_probe_video_from_file", "_read_video_from_memory", "_read_video_timestamps_from_memory", "_probe_video_from_memory", "_HAS_CPU_VIDEO_DECODER", "_HAS_VIDEO_OPT", "_HAS_GPU_VIDEO_DECODER", "_read_video_clip_from_memory", "_read_video_meta_data", "VideoMetaData", "Timebase", "ImageReadMode", "decode_image", "decode_jpeg", "decode_png", "decode_heic", "decode_webp", "decode_gif", "encode_jpeg", "encode_png", "read_file", "read_image", "write_file", "write_jpeg", "write_png", "Video", "VideoReader", ]
from typing import Any, Dict, Iterator import torch from ..utils import _log_api_usage_once try: from ._load_gpu_decoder import _HAS_GPU_VIDEO_DECODER except ModuleNotFoundError: _HAS_GPU_VIDEO_DECODER = False from ._video_opt import ( _HAS_CPU_VIDEO_DECODER, _HAS_VIDEO_OPT, _probe_video_from_file, _probe_video_from_memory, _read_video_from_file, _read_video_from_memory, _read_video_timestamps_from_file, _read_video_timestamps_from_memory, Timebase, VideoMetaData, ) from .image import ( decode_gif, decode_image, decode_jpeg, decode_png, decode_webp, encode_jpeg, encode_png, ImageReadMode, read_file, read_image, write_file, write_jpeg, write_png, ) from .video import read_video, read_video_timestamps, write_video from .video_reader import VideoReader __all__ = [ "write_video", "read_video", "read_video_timestamps", "_read_video_from_file", "_read_video_timestamps_from_file", "_probe_video_from_file", "_read_video_from_memory", "_read_video_timestamps_from_memory", "_probe_video_from_memory", "_HAS_CPU_VIDEO_DECODER", "_HAS_VIDEO_OPT", "_HAS_GPU_VIDEO_DECODER", "_read_video_clip_from_memory", "_read_video_meta_data", "VideoMetaData", "Timebase", "ImageReadMode", "decode_image", "decode_jpeg", "decode_png", "decode_webp", "decode_gif", "encode_jpeg", "encode_png", "read_file", "read_image", "write_file", "write_jpeg", "write_png", "Video", "VideoReader", ]
"""Wrapper around in-memory storage.""" from __future__ import annotations from typing import Any, Dict, List, Literal, Optional from langchain_core.embeddings import Embeddings from langchain_community.vectorstores.docarray.base import ( DocArrayIndex, _check_docarray_import, ) class DocArrayInMemorySearch(DocArrayIndex): """In-memory `DocArray` storage for exact search. To use it, you should have the ``docarray`` package with version >=0.32.0 installed. You can install it with `pip install docarray`. """ @classmethod def from_params( cls, embedding: Embeddings, metric: Literal[ "cosine_sim", "euclidian_dist", "sgeuclidean_dist" ] = "cosine_sim", **kwargs: Any, ) -> DocArrayInMemorySearch: """Initialize DocArrayInMemorySearch store. Args: embedding (Embeddings): Embedding function. metric (str): metric for exact nearest-neighbor search. Can be one of: "cosine_sim", "euclidean_dist" and "sqeuclidean_dist". Defaults to "cosine_sim". **kwargs: Other keyword arguments to be passed to the get_doc_cls method. """ _check_docarray_import() from docarray.index import InMemoryExactNNIndex doc_cls = cls._get_doc_cls(space=metric, **kwargs) doc_index = InMemoryExactNNIndex[doc_cls]() return cls(doc_index, embedding) @classmethod def from_texts( cls, texts: List[str], embedding: Embeddings, metadatas: Optional[List[Dict[Any, Any]]] = None, **kwargs: Any, ) -> DocArrayInMemorySearch: """Create an DocArrayInMemorySearch store and insert data. Args: texts (List[str]): Text data. embedding (Embeddings): Embedding function. metadatas (Optional[List[Dict[Any, Any]]]): Metadata for each text if it exists. Defaults to None. **kwargs: Other keyword arguments to be passed to the from_params method. Returns: DocArrayInMemorySearch Vector Store """ store = cls.from_params(embedding, **kwargs) store.add_texts(texts=texts, metadatas=metadatas) return store
"""Wrapper around in-memory storage.""" from __future__ import annotations from typing import Any, Dict, List, Literal, Optional from langchain_core.embeddings import Embeddings from langchain_community.vectorstores.docarray.base import ( DocArrayIndex, _check_docarray_import, ) class DocArrayInMemorySearch(DocArrayIndex): """In-memory `DocArray` storage for exact search. To use it, you should have the ``docarray`` package with version >=0.32.0 installed. You can install it with `pip install docarray`. """ @classmethod def from_params( cls, embedding: Embeddings, metric: Literal[ "cosine_sim", "euclidian_dist", "sgeuclidean_dist" ] = "cosine_sim", **kwargs: Any, ) -> DocArrayInMemorySearch: """Initialize DocArrayInMemorySearch store. Args: embedding (Embeddings): Embedding function. metric (str): metric for exact nearest-neighbor search. Can be one of: "cosine_sim", "euclidean_dist" and "sqeuclidean_dist". Defaults to "cosine_sim". **kwargs: Other keyword arguments to be passed to the get_doc_cls method. """ _check_docarray_import() from docarray.index import InMemoryExactNNIndex doc_cls = cls._get_doc_cls(space=metric, **kwargs) doc_index = InMemoryExactNNIndex[doc_cls]() # type: ignore return cls(doc_index, embedding) @classmethod def from_texts( cls, texts: List[str], embedding: Embeddings, metadatas: Optional[List[Dict[Any, Any]]] = None, **kwargs: Any, ) -> DocArrayInMemorySearch: """Create an DocArrayInMemorySearch store and insert data. Args: texts (List[str]): Text data. embedding (Embeddings): Embedding function. metadatas (Optional[List[Dict[Any, Any]]]): Metadata for each text if it exists. Defaults to None. **kwargs: Other keyword arguments to be passed to the from_params method. Returns: DocArrayInMemorySearch Vector Store """ store = cls.from_params(embedding, **kwargs) store.add_texts(texts=texts, metadatas=metadatas) return store
# Copyright 2025 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from packaging import version from .. import __version__ from .constants import ( CONFIG_NAME, DEPRECATED_REVISION_ARGS, DIFFUSERS_DYNAMIC_MODULE_NAME, FLAX_WEIGHTS_NAME, GGUF_FILE_EXTENSION, HF_MODULES_CACHE, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MIN_PEFT_VERSION, ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFETENSORS_FILE_EXTENSION, SAFETENSORS_WEIGHTS_NAME, USE_PEFT_BACKEND, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .deprecation_utils import deprecate from .doc_utils import replace_example_docstring from .dynamic_modules_utils import get_class_from_dynamic_module from .export_utils import export_to_gif, export_to_obj, export_to_ply, export_to_video from .hub_utils import ( PushToHubMixin, _add_variant, _get_checkpoint_shard_files, _get_model_file, extract_commit_hash, http_user_agent, ) from .import_utils import ( BACKENDS_MAPPING, DIFFUSERS_SLOW_IMPORT, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, USE_JAX, USE_TF, USE_TORCH, DummyObject, OptionalDependencyNotAvailable, _LazyModule, get_objects_from_module, is_accelerate_available, is_accelerate_version, is_bitsandbytes_available, is_bitsandbytes_version, is_bs4_available, is_flax_available, is_ftfy_available, is_gguf_available, is_gguf_version, is_google_colab, is_hf_hub_version, is_inflect_available, is_invisible_watermark_available, is_k_diffusion_available, is_k_diffusion_version, is_librosa_available, is_matplotlib_available, is_note_seq_available, is_onnx_available, is_opencv_available, is_optimum_quanto_available, is_optimum_quanto_version, is_peft_available, is_peft_version, is_safetensors_available, is_scipy_available, is_sentencepiece_available, is_tensorboard_available, is_timm_available, is_torch_available, is_torch_npu_available, is_torch_version, is_torch_xla_available, is_torch_xla_version, is_torchao_available, is_torchao_version, is_torchsde_available, is_torchvision_available, is_transformers_available, is_transformers_version, is_unidecode_available, is_wandb_available, is_xformers_available, requires_backends, ) from .loading_utils import get_module_from_name, get_submodule_by_name, load_image, load_video from .logging import get_logger from .outputs import BaseOutput from .peft_utils import ( check_peft_version, delete_adapter_layers, get_adapter_name, get_peft_kwargs, recurse_remove_peft_layers, scale_lora_layers, set_adapter_layers, set_weights_and_activate_adapters, unscale_lora_layers, ) from .pil_utils import PIL_INTERPOLATION, make_image_grid, numpy_to_pil, pt_to_pil from .remote_utils import remote_decode from .state_dict_utils import ( convert_all_state_dict_to_peft, convert_state_dict_to_diffusers, convert_state_dict_to_kohya, convert_state_dict_to_peft, convert_unet_state_dict_to_peft, state_dict_all_zero, ) from .typing_utils import _get_detailed_type, _is_valid_type logger = get_logger(__name__) def check_min_version(min_version): if version.parse(__version__) < version.parse(min_version): if "dev" in min_version: error_message = ( "This example requires a source install from HuggingFace diffusers (see " "`https://huggingface.co/docs/diffusers/installation#install-from-source`)," ) else: error_message = f"This example requires a minimum version of {min_version}," error_message += f" but the version found is {__version__}.\n" raise ImportError(error_message)
# Copyright 2025 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from packaging import version from .. import __version__ from .constants import ( CONFIG_NAME, DEPRECATED_REVISION_ARGS, DIFFUSERS_DYNAMIC_MODULE_NAME, FLAX_WEIGHTS_NAME, GGUF_FILE_EXTENSION, HF_MODULES_CACHE, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MIN_PEFT_VERSION, ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFETENSORS_FILE_EXTENSION, SAFETENSORS_WEIGHTS_NAME, USE_PEFT_BACKEND, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .deprecation_utils import deprecate from .doc_utils import replace_example_docstring from .dynamic_modules_utils import get_class_from_dynamic_module from .export_utils import export_to_gif, export_to_obj, export_to_ply, export_to_video from .hub_utils import ( PushToHubMixin, _add_variant, _get_checkpoint_shard_files, _get_model_file, extract_commit_hash, http_user_agent, ) from .import_utils import ( BACKENDS_MAPPING, DIFFUSERS_SLOW_IMPORT, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, USE_JAX, USE_TF, USE_TORCH, DummyObject, OptionalDependencyNotAvailable, _LazyModule, get_objects_from_module, is_accelerate_available, is_accelerate_version, is_bitsandbytes_available, is_bitsandbytes_version, is_bs4_available, is_flax_available, is_ftfy_available, is_gguf_available, is_gguf_version, is_google_colab, is_hf_hub_version, is_inflect_available, is_invisible_watermark_available, is_k_diffusion_available, is_k_diffusion_version, is_librosa_available, is_matplotlib_available, is_note_seq_available, is_onnx_available, is_optimum_quanto_available, is_optimum_quanto_version, is_peft_available, is_peft_version, is_safetensors_available, is_scipy_available, is_sentencepiece_available, is_tensorboard_available, is_timm_available, is_torch_available, is_torch_npu_available, is_torch_version, is_torch_xla_available, is_torch_xla_version, is_torchao_available, is_torchao_version, is_torchsde_available, is_torchvision_available, is_transformers_available, is_transformers_version, is_unidecode_available, is_wandb_available, is_xformers_available, requires_backends, ) from .loading_utils import get_module_from_name, get_submodule_by_name, load_image, load_video from .logging import get_logger from .outputs import BaseOutput from .peft_utils import ( check_peft_version, delete_adapter_layers, get_adapter_name, get_peft_kwargs, recurse_remove_peft_layers, scale_lora_layers, set_adapter_layers, set_weights_and_activate_adapters, unscale_lora_layers, ) from .pil_utils import PIL_INTERPOLATION, make_image_grid, numpy_to_pil, pt_to_pil from .remote_utils import remote_decode from .state_dict_utils import ( convert_all_state_dict_to_peft, convert_state_dict_to_diffusers, convert_state_dict_to_kohya, convert_state_dict_to_peft, convert_unet_state_dict_to_peft, state_dict_all_zero, ) from .typing_utils import _get_detailed_type, _is_valid_type logger = get_logger(__name__) def check_min_version(min_version): if version.parse(__version__) < version.parse(min_version): if "dev" in min_version: error_message = ( "This example requires a source install from HuggingFace diffusers (see " "`https://huggingface.co/docs/diffusers/installation#install-from-source`)," ) else: error_message = f"This example requires a minimum version of {min_version}," error_message += f" but the version found is {__version__}.\n" raise ImportError(error_message)
""" This example trains a SparseEncoder for the Natural Questions (NQ) dataset. The training script fine-tunes a SparseEncoder using the Splade loss function for retrieval. It loads a subset of the Natural Questions dataset, splits it into training and evaluation subsets, and trains the model as a retriever. After training, the model is evaluated and saved locally, with an optional step to push the trained model to the Hugging Face Hub. Usage: python train_splade_nq.py """ import logging import traceback from datasets import load_dataset from sentence_transformers import ( SparseEncoder, SparseEncoderModelCardData, SparseEncoderTrainer, SparseEncoderTrainingArguments, ) from sentence_transformers.sparse_encoder import evaluation, losses from sentence_transformers.training_args import BatchSamplers # Set the log level to INFO to get more information logging.basicConfig(format="%(asctime)s - %(message)s", datefmt="%Y-%m-%d %H:%M:%S", level=logging.INFO) def main(): model_name = "distilbert/distilbert-base-uncased" train_batch_size = 12 num_epochs = 1 # 1a. Load a model to finetune with 1b. (Optional) model card data model = SparseEncoder( model_name, model_card_data=SparseEncoderModelCardData( language="en", license="apache-2.0", model_name="splade-distilbert-base-uncased trained on Natural Questions", ), ) model.max_seq_length = 256 # Set the max sequence length to 256 for the training print("Model max length:", model.max_seq_length) # 2. Load the NQ dataset: https://huggingface.co/datasets/sentence-transformers/natural-questions logging.info("Read the Natural Questions training dataset") full_dataset = load_dataset("sentence-transformers/natural-questions", split="train").select(range(100_000)) dataset_dict = full_dataset.train_test_split(test_size=1_000, seed=12) train_dataset = dataset_dict["train"] eval_dataset = dataset_dict["test"] logging.info(train_dataset) logging.info(eval_dataset) # 3. Define our training loss. lambda_query = 5e-5 lambda_corpus = 3e-5 loss = losses.SpladeLoss( model=model, loss=losses.SparseMultipleNegativesRankingLoss(model=model), lambda_query=lambda_query, # Weight for query loss lambda_corpus=lambda_corpus, # Weight for document loss ) # 4. Define evaluator. We use the SparseNanoBEIREvaluator, which is a light-weight evaluator evaluator = evaluation.SparseNanoBEIREvaluator(show_progress_bar=True, batch_size=train_batch_size) # 5. Define the training arguments short_model_name = model_name if "/" not in model_name else model_name.split("/")[-1] run_name = f"splade-{short_model_name}-nq" training_args = SparseEncoderTrainingArguments( # Required parameter: output_dir=f"models/{run_name}", # Optional training parameters: num_train_epochs=num_epochs, per_device_train_batch_size=train_batch_size, per_device_eval_batch_size=train_batch_size, learning_rate=2e-5, fp16=False, # Set to False if you get an error that your GPU can't run on FP16 bf16=True, # Set to True if you have a GPU that supports BF16 batch_sampler=BatchSamplers.NO_DUPLICATES, # MultipleNegativesRankingLoss benefits from no duplicate samples in a batch load_best_model_at_end=True, metric_for_best_model="eval_NanoBEIR_mean_dot_ndcg@10", # Optional tracking/debugging parameters: eval_strategy="steps", eval_steps=1650, save_strategy="steps", save_steps=1650, save_total_limit=2, logging_steps=200, run_name=run_name, # Will be used in W&B if `wandb` is installed seed=42, ) # 6. Create the trainer & start training trainer = SparseEncoderTrainer( model=model, args=training_args, train_dataset=train_dataset, eval_dataset=eval_dataset, loss=loss, evaluator=evaluator, ) trainer.train() # 7. Evaluate the final model, useful to include these in the model card evaluator(model) # 8. Save the final model final_output_dir = f"models/{run_name}/final" model.save_pretrained(final_output_dir) # 9. (Optional) save the model to the Hugging Face Hub! # It is recommended to run `huggingface-cli login` to log into your Hugging Face account first try: model.push_to_hub(run_name) except Exception: logging.error( f"Error uploading model to the Hugging Face Hub:\n{traceback.format_exc()}To upload it manually, you can run " f"`huggingface-cli login`, followed by loading the model using `model = SparseEncoder({final_output_dir!r})` " f"and saving it using `model.push_to_hub('{run_name}')`." ) if __name__ == "__main__": main()
""" This example trains a SparseEncoder for the Natural Questions (NQ) dataset. The training script fine-tunes a SparseEncoder using the Splade loss function for retrieval. It loads a subset of the Natural Questions dataset, splits it into training and evaluation subsets, and trains the model as a retriever. After training, the model is evaluated and saved locally, with an optional step to push the trained model to the Hugging Face Hub. Usage: python train_splade_nq.py """ import logging import traceback from datasets import load_dataset from sentence_transformers import ( SparseEncoder, SparseEncoderModelCardData, SparseEncoderTrainer, SparseEncoderTrainingArguments, ) from sentence_transformers.sparse_encoder import evaluation, losses from sentence_transformers.training_args import BatchSamplers # Set the log level to INFO to get more information logging.basicConfig(format="%(asctime)s - %(message)s", datefmt="%Y-%m-%d %H:%M:%S", level=logging.INFO) def main(): model_name = "distilbert/distilbert-base-uncased" train_batch_size = 12 num_epochs = 1 # 1a. Load a model to finetune with 1b. (Optional) model card data model = SparseEncoder( model_name, model_card_data=SparseEncoderModelCardData( language="en", license="apache-2.0", model_name="splade-distilbert-base-uncased trained on Natural Questions", ), ) model.max_seq_length = 256 # Set the max sequence length to 256 for the training print("Model max length:", model.max_seq_length) # 2. Load the NQ dataset: https://huggingface.co/datasets/sentence-transformers/natural-questions logging.info("Read the Natural Questions training dataset") full_dataset = load_dataset("sentence-transformers/natural-questions", split="train").select(range(100_000)) dataset_dict = full_dataset.train_test_split(test_size=1_000, seed=12) train_dataset = dataset_dict["train"] eval_dataset = dataset_dict["test"] logging.info(train_dataset) logging.info(eval_dataset) # 3. Define our training loss. lambda_query = 5e-5 lambda_corpus = 3e-5 loss = losses.SpladeLoss( model=model, loss=losses.SparseMultipleNegativesRankingLoss(model=model), lambda_query=lambda_query, # Weight for query loss lambda_corpus=lambda_corpus, # Weight for document loss ) # 4. Define evaluator. We use the SparseNanoBEIREvaluator, which is a light-weight evaluator evaluator = evaluation.SparseNanoBEIREvaluator(show_progress_bar=True, batch_size=train_batch_size) # 5. Define the training arguments short_model_name = model_name if "/" not in model_name else model_name.split("/")[-1] run_name = f"splade-{short_model_name}-nq" training_args = SparseEncoderTrainingArguments( # Required parameter: output_dir=f"models/{run_name}", # Optional training parameters: num_train_epochs=num_epochs, per_device_train_batch_size=train_batch_size, per_device_eval_batch_size=train_batch_size, learning_rate=2e-5, fp16=False, # Set to False if you get an error that your GPU can't run on FP16 bf16=True, # Set to True if you have a GPU that supports BF16 batch_sampler=BatchSamplers.NO_DUPLICATES, # MultipleNegativesRankingLoss benefits from no duplicate samples in a batch load_best_model_at_end=True, metric_for_best_model="eval_NanoBEIR_mean_dot_ndcg@10", # Optional tracking/debugging parameters: eval_strategy="steps", eval_steps=1650, save_strategy="steps", save_steps=1650, save_total_limit=2, logging_steps=200, run_name=run_name, # Will be used in W&B if `wandb` is installed seed=42, ) # 6. Create the trainer & start training trainer = SparseEncoderTrainer( model=model, args=training_args, train_dataset=train_dataset, eval_dataset=eval_dataset, loss=loss, evaluator=evaluator, ) trainer.train() # 7. Evaluate the final model, useful to include these in the model card evaluator(model) # 8. Save the final model final_output_dir = f"models/{run_name}/final" model.save_pretrained(final_output_dir) # 9. (Optional) save the model to the Hugging Face Hub! # It is recommended to run `huggingface-cli login` to log into your Hugging Face account first try: model.push_to_hub(run_name) except Exception: logging.error( f"Error uploading model to the Hugging Face Hub:\n{traceback.format_exc()}To upload it manually, you can run " f"`huggingface-cli login`, followed by loading the model using `model = CrossEncoder({final_output_dir!r})` " f"and saving it using `model.push_to_hub('{run_name}')`." ) if __name__ == "__main__": main()
import json from enum import Enum from typing import Any from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField from backend.util.request import requests class HttpMethod(Enum): GET = "GET" POST = "POST" PUT = "PUT" DELETE = "DELETE" PATCH = "PATCH" OPTIONS = "OPTIONS" HEAD = "HEAD" class SendWebRequestBlock(Block): class Input(BlockSchema): url: str = SchemaField( description="The URL to send the request to", placeholder="https://api.example.com", ) method: HttpMethod = SchemaField( description="The HTTP method to use for the request", default=HttpMethod.POST, ) headers: dict[str, str] = SchemaField( description="The headers to include in the request", default={}, ) json_format: bool = SchemaField( title="JSON format", description="Whether to send and receive body as JSON", default=True, ) body: Any = SchemaField( description="The body of the request", default=None, ) class Output(BlockSchema): response: object = SchemaField(description="The response from the server") client_error: object = SchemaField(description="The error on 4xx status codes") server_error: object = SchemaField(description="The error on 5xx status codes") def __init__(self): super().__init__( id="6595ae1f-b924-42cb-9a41-551a0611c4b4", description="This block makes an HTTP request to the given URL.", categories={BlockCategory.OUTPUT}, input_schema=SendWebRequestBlock.Input, output_schema=SendWebRequestBlock.Output, ) def run(self, input_data: Input, **kwargs) -> BlockOutput: if isinstance(input_data.body, str): input_data.body = json.loads(input_data.body) response = requests.request( input_data.method.value, input_data.url, headers=input_data.headers, json=input_data.body if input_data.json_format else None, data=input_data.body if not input_data.json_format else None, ) result = response.json() if input_data.json_format else response.text if response.status_code // 100 == 2: yield "response", result elif response.status_code // 100 == 4: yield "client_error", result elif response.status_code // 100 == 5: yield "server_error", result else: raise ValueError(f"Unexpected status code: {response.status_code}")
import json from enum import Enum import requests from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField class HttpMethod(Enum): GET = "GET" POST = "POST" PUT = "PUT" DELETE = "DELETE" PATCH = "PATCH" OPTIONS = "OPTIONS" HEAD = "HEAD" class SendWebRequestBlock(Block): class Input(BlockSchema): url: str = SchemaField( description="The URL to send the request to", placeholder="https://api.example.com", ) method: HttpMethod = SchemaField( description="The HTTP method to use for the request", default=HttpMethod.POST, ) headers: dict[str, str] = SchemaField( description="The headers to include in the request", default={}, ) body: object = SchemaField( description="The body of the request", default={}, ) class Output(BlockSchema): response: object = SchemaField(description="The response from the server") client_error: object = SchemaField(description="The error on 4xx status codes") server_error: object = SchemaField(description="The error on 5xx status codes") def __init__(self): super().__init__( id="6595ae1f-b924-42cb-9a41-551a0611c4b4", description="This block makes an HTTP request to the given URL.", categories={BlockCategory.OUTPUT}, input_schema=SendWebRequestBlock.Input, output_schema=SendWebRequestBlock.Output, ) def run(self, input_data: Input, **kwargs) -> BlockOutput: if isinstance(input_data.body, str): input_data.body = json.loads(input_data.body) response = requests.request( input_data.method.value, input_data.url, headers=input_data.headers, json=input_data.body, ) if response.status_code // 100 == 2: yield "response", response.json() elif response.status_code // 100 == 4: yield "client_error", response.json() elif response.status_code // 100 == 5: yield "server_error", response.json() else: raise ValueError(f"Unexpected status code: {response.status_code}")
_base_ = './mask-rcnn_r50-caffe_fpn_ms-poly-1x_coco.py' train_cfg = dict(max_epochs=36) # learning rate param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=24, by_epoch=True, milestones=[28, 34], gamma=0.1) ]
_base_ = './mask_rcnn_r50_caffe_fpn_mstrain-poly_1x_coco.py' train_cfg = dict(max_epochs=36) # learning rate param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=24, by_epoch=True, milestones=[28, 34], gamma=0.1) ]
"""Base classes for chain routing.""" from __future__ import annotations from abc import ABC from collections.abc import Mapping from typing import Any, NamedTuple, Optional from langchain_core.callbacks import ( AsyncCallbackManagerForChainRun, CallbackManagerForChainRun, Callbacks, ) from pydantic import ConfigDict from langchain.chains.base import Chain class Route(NamedTuple): destination: Optional[str] next_inputs: dict[str, Any] class RouterChain(Chain, ABC): """Chain that outputs the name of a destination chain and the inputs to it.""" @property def output_keys(self) -> list[str]: return ["destination", "next_inputs"] def route(self, inputs: dict[str, Any], callbacks: Callbacks = None) -> Route: """ Route inputs to a destination chain. Args: inputs: inputs to the chain callbacks: callbacks to use for the chain Returns: a Route object """ result = self(inputs, callbacks=callbacks) return Route(result["destination"], result["next_inputs"]) async def aroute( self, inputs: dict[str, Any], callbacks: Callbacks = None ) -> Route: result = await self.acall(inputs, callbacks=callbacks) return Route(result["destination"], result["next_inputs"]) class MultiRouteChain(Chain): """Use a single chain to route an input to one of multiple candidate chains.""" router_chain: RouterChain """Chain that routes inputs to destination chains.""" destination_chains: Mapping[str, Chain] """Chains that return final answer to inputs.""" default_chain: Chain """Default chain to use when none of the destination chains are suitable.""" silent_errors: bool = False """If True, use default_chain when an invalid destination name is provided. Defaults to False.""" model_config = ConfigDict( arbitrary_types_allowed=True, extra="forbid", ) @property def input_keys(self) -> list[str]: """Will be whatever keys the router chain prompt expects. :meta private: """ return self.router_chain.input_keys @property def output_keys(self) -> list[str]: """Will always return text key. :meta private: """ return [] def _call( self, inputs: dict[str, Any], run_manager: Optional[CallbackManagerForChainRun] = None, ) -> dict[str, Any]: _run_manager = run_manager or CallbackManagerForChainRun.get_noop_manager() callbacks = _run_manager.get_child() route = self.router_chain.route(inputs, callbacks=callbacks) _run_manager.on_text( str(route.destination) + ": " + str(route.next_inputs), verbose=self.verbose ) if not route.destination: return self.default_chain(route.next_inputs, callbacks=callbacks) elif route.destination in self.destination_chains: return self.destination_chains[route.destination]( route.next_inputs, callbacks=callbacks ) elif self.silent_errors: return self.default_chain(route.next_inputs, callbacks=callbacks) else: raise ValueError( f"Received invalid destination chain name '{route.destination}'" ) async def _acall( self, inputs: dict[str, Any], run_manager: Optional[AsyncCallbackManagerForChainRun] = None, ) -> dict[str, Any]: _run_manager = run_manager or AsyncCallbackManagerForChainRun.get_noop_manager() callbacks = _run_manager.get_child() route = await self.router_chain.aroute(inputs, callbacks=callbacks) await _run_manager.on_text( str(route.destination) + ": " + str(route.next_inputs), verbose=self.verbose ) if not route.destination: return await self.default_chain.acall( route.next_inputs, callbacks=callbacks ) elif route.destination in self.destination_chains: return await self.destination_chains[route.destination].acall( route.next_inputs, callbacks=callbacks ) elif self.silent_errors: return await self.default_chain.acall( route.next_inputs, callbacks=callbacks ) else: raise ValueError( f"Received invalid destination chain name '{route.destination}'" )
"""Base classes for chain routing.""" from __future__ import annotations from abc import ABC from typing import Any, Dict, List, Mapping, NamedTuple, Optional from langchain_core.callbacks import ( AsyncCallbackManagerForChainRun, CallbackManagerForChainRun, Callbacks, ) from pydantic import ConfigDict from langchain.chains.base import Chain class Route(NamedTuple): destination: Optional[str] next_inputs: Dict[str, Any] class RouterChain(Chain, ABC): """Chain that outputs the name of a destination chain and the inputs to it.""" @property def output_keys(self) -> List[str]: return ["destination", "next_inputs"] def route(self, inputs: Dict[str, Any], callbacks: Callbacks = None) -> Route: """ Route inputs to a destination chain. Args: inputs: inputs to the chain callbacks: callbacks to use for the chain Returns: a Route object """ result = self(inputs, callbacks=callbacks) return Route(result["destination"], result["next_inputs"]) async def aroute( self, inputs: Dict[str, Any], callbacks: Callbacks = None ) -> Route: result = await self.acall(inputs, callbacks=callbacks) return Route(result["destination"], result["next_inputs"]) class MultiRouteChain(Chain): """Use a single chain to route an input to one of multiple candidate chains.""" router_chain: RouterChain """Chain that routes inputs to destination chains.""" destination_chains: Mapping[str, Chain] """Chains that return final answer to inputs.""" default_chain: Chain """Default chain to use when none of the destination chains are suitable.""" silent_errors: bool = False """If True, use default_chain when an invalid destination name is provided. Defaults to False.""" model_config = ConfigDict( arbitrary_types_allowed=True, extra="forbid", ) @property def input_keys(self) -> List[str]: """Will be whatever keys the router chain prompt expects. :meta private: """ return self.router_chain.input_keys @property def output_keys(self) -> List[str]: """Will always return text key. :meta private: """ return [] def _call( self, inputs: Dict[str, Any], run_manager: Optional[CallbackManagerForChainRun] = None, ) -> Dict[str, Any]: _run_manager = run_manager or CallbackManagerForChainRun.get_noop_manager() callbacks = _run_manager.get_child() route = self.router_chain.route(inputs, callbacks=callbacks) _run_manager.on_text( str(route.destination) + ": " + str(route.next_inputs), verbose=self.verbose ) if not route.destination: return self.default_chain(route.next_inputs, callbacks=callbacks) elif route.destination in self.destination_chains: return self.destination_chains[route.destination]( route.next_inputs, callbacks=callbacks ) elif self.silent_errors: return self.default_chain(route.next_inputs, callbacks=callbacks) else: raise ValueError( f"Received invalid destination chain name '{route.destination}'" ) async def _acall( self, inputs: Dict[str, Any], run_manager: Optional[AsyncCallbackManagerForChainRun] = None, ) -> Dict[str, Any]: _run_manager = run_manager or AsyncCallbackManagerForChainRun.get_noop_manager() callbacks = _run_manager.get_child() route = await self.router_chain.aroute(inputs, callbacks=callbacks) await _run_manager.on_text( str(route.destination) + ": " + str(route.next_inputs), verbose=self.verbose ) if not route.destination: return await self.default_chain.acall( route.next_inputs, callbacks=callbacks ) elif route.destination in self.destination_chains: return await self.destination_chains[route.destination].acall( route.next_inputs, callbacks=callbacks ) elif self.silent_errors: return await self.default_chain.acall( route.next_inputs, callbacks=callbacks ) else: raise ValueError( f"Received invalid destination chain name '{route.destination}'" )
# Copyright (c) OpenMMLab. All rights reserved. import torch.nn as nn from mmcv.runner import BaseModule, auto_fp16 from mmdet.models.builder import HEADS @HEADS.register_module() class FeatureRelayHead(BaseModule): """Feature Relay Head used in `SCNet <https://arxiv.org/abs/2012.10150>`_. Args: in_channels (int, optional): number of input channels. Default: 256. conv_out_channels (int, optional): number of output channels before classification layer. Default: 256. roi_feat_size (int, optional): roi feat size at box head. Default: 7. scale_factor (int, optional): scale factor to match roi feat size at mask head. Default: 2. init_cfg (dict or list[dict], optional): Initialization config dict. """ def __init__(self, in_channels=1024, out_conv_channels=256, roi_feat_size=7, scale_factor=2, init_cfg=dict(type='Kaiming', layer='Linear')): super(FeatureRelayHead, self).__init__(init_cfg) assert isinstance(roi_feat_size, int) self.in_channels = in_channels self.out_conv_channels = out_conv_channels self.roi_feat_size = roi_feat_size self.out_channels = (roi_feat_size**2) * out_conv_channels self.scale_factor = scale_factor self.fp16_enabled = False self.fc = nn.Linear(self.in_channels, self.out_channels) self.upsample = nn.Upsample( scale_factor=scale_factor, mode='bilinear', align_corners=True) @auto_fp16() def forward(self, x): """Forward function.""" N, in_C = x.shape if N > 0: out_C = self.out_conv_channels out_HW = self.roi_feat_size x = self.fc(x) x = x.reshape(N, out_C, out_HW, out_HW) x = self.upsample(x) return x return None
import torch.nn as nn from mmcv.runner import BaseModule, auto_fp16 from mmdet.models.builder import HEADS @HEADS.register_module() class FeatureRelayHead(BaseModule): """Feature Relay Head used in `SCNet <https://arxiv.org/abs/2012.10150>`_. Args: in_channels (int, optional): number of input channels. Default: 256. conv_out_channels (int, optional): number of output channels before classification layer. Default: 256. roi_feat_size (int, optional): roi feat size at box head. Default: 7. scale_factor (int, optional): scale factor to match roi feat size at mask head. Default: 2. init_cfg (dict or list[dict], optional): Initialization config dict. """ def __init__(self, in_channels=1024, out_conv_channels=256, roi_feat_size=7, scale_factor=2, init_cfg=dict(type='Kaiming', layer='Linear')): super(FeatureRelayHead, self).__init__(init_cfg) assert isinstance(roi_feat_size, int) self.in_channels = in_channels self.out_conv_channels = out_conv_channels self.roi_feat_size = roi_feat_size self.out_channels = (roi_feat_size**2) * out_conv_channels self.scale_factor = scale_factor self.fp16_enabled = False self.fc = nn.Linear(self.in_channels, self.out_channels) self.upsample = nn.Upsample( scale_factor=scale_factor, mode='bilinear', align_corners=True) @auto_fp16() def forward(self, x): """Forward function.""" N, in_C = x.shape if N > 0: out_C = self.out_conv_channels out_HW = self.roi_feat_size x = self.fc(x) x = x.reshape(N, out_C, out_HW, out_HW) x = self.upsample(x) return x return None
from typing import Union import numpy as np Matrix = Union[list[list[float]], list[np.ndarray], np.ndarray] def maximal_marginal_relevance( query_embedding: np.ndarray, embedding_list: list, lambda_mult: float = 0.5, k: int = 4, ) -> list[int]: """Calculate maximal marginal relevance.""" if min(k, len(embedding_list)) <= 0: return [] if query_embedding.ndim == 1: query_embedding = np.expand_dims(query_embedding, axis=0) similarity_to_query = cosine_similarity(query_embedding, embedding_list)[0] most_similar = int(np.argmax(similarity_to_query)) idxs = [most_similar] selected = np.array([embedding_list[most_similar]]) while len(idxs) < min(k, len(embedding_list)): best_score = -np.inf idx_to_add = -1 similarity_to_selected = cosine_similarity(embedding_list, selected) for i, query_score in enumerate(similarity_to_query): if i in idxs: continue redundant_score = max(similarity_to_selected[i]) equation_score = ( lambda_mult * query_score - (1 - lambda_mult) * redundant_score ) if equation_score > best_score: best_score = equation_score idx_to_add = i idxs.append(idx_to_add) selected = np.append(selected, [embedding_list[idx_to_add]], axis=0) return idxs def cosine_similarity(X: Matrix, Y: Matrix) -> np.ndarray: """Row-wise cosine similarity between two equal-width matrices.""" if len(X) == 0 or len(Y) == 0: return np.array([]) X = np.array(X) Y = np.array(Y) if X.shape[1] != Y.shape[1]: raise ValueError( f"Number of columns in X and Y must be the same. X has shape {X.shape} " f"and Y has shape {Y.shape}." ) try: import simsimd as simd X = np.array(X, dtype=np.float32) Y = np.array(Y, dtype=np.float32) Z = 1 - np.array(simd.cdist(X, Y, metric="cosine")) return Z except ImportError: X_norm = np.linalg.norm(X, axis=1) Y_norm = np.linalg.norm(Y, axis=1) # Ignore divide by zero errors run time warnings as those are handled below. with np.errstate(divide="ignore", invalid="ignore"): similarity = np.dot(X, Y.T) / np.outer(X_norm, Y_norm) similarity[np.isnan(similarity) | np.isinf(similarity)] = 0.0 return similarity
from typing import List, Union import numpy as np Matrix = Union[List[List[float]], List[np.ndarray], np.ndarray] def maximal_marginal_relevance( query_embedding: np.ndarray, embedding_list: list, lambda_mult: float = 0.5, k: int = 4, ) -> List[int]: """Calculate maximal marginal relevance.""" if min(k, len(embedding_list)) <= 0: return [] if query_embedding.ndim == 1: query_embedding = np.expand_dims(query_embedding, axis=0) similarity_to_query = cosine_similarity(query_embedding, embedding_list)[0] most_similar = int(np.argmax(similarity_to_query)) idxs = [most_similar] selected = np.array([embedding_list[most_similar]]) while len(idxs) < min(k, len(embedding_list)): best_score = -np.inf idx_to_add = -1 similarity_to_selected = cosine_similarity(embedding_list, selected) for i, query_score in enumerate(similarity_to_query): if i in idxs: continue redundant_score = max(similarity_to_selected[i]) equation_score = ( lambda_mult * query_score - (1 - lambda_mult) * redundant_score ) if equation_score > best_score: best_score = equation_score idx_to_add = i idxs.append(idx_to_add) selected = np.append(selected, [embedding_list[idx_to_add]], axis=0) return idxs def cosine_similarity(X: Matrix, Y: Matrix) -> np.ndarray: """Row-wise cosine similarity between two equal-width matrices.""" if len(X) == 0 or len(Y) == 0: return np.array([]) X = np.array(X) Y = np.array(Y) if X.shape[1] != Y.shape[1]: raise ValueError( f"Number of columns in X and Y must be the same. X has shape {X.shape} " f"and Y has shape {Y.shape}." ) try: import simsimd as simd X = np.array(X, dtype=np.float32) Y = np.array(Y, dtype=np.float32) Z = 1 - np.array(simd.cdist(X, Y, metric="cosine")) return Z except ImportError: X_norm = np.linalg.norm(X, axis=1) Y_norm = np.linalg.norm(Y, axis=1) # Ignore divide by zero errors run time warnings as those are handled below. with np.errstate(divide="ignore", invalid="ignore"): similarity = np.dot(X, Y.T) / np.outer(X_norm, Y_norm) similarity[np.isnan(similarity) | np.isinf(similarity)] = 0.0 return similarity
# Copyright (c) OpenMMLab. All rights reserved. import unittest import torch from mmengine.config import Config from mmdet.models.seg_heads.panoptic_fusion_heads import MaskFormerFusionHead from mmdet.structures import DetDataSample class TestMaskFormerFusionHead(unittest.TestCase): def test_loss(self): head = MaskFormerFusionHead(num_things_classes=2, num_stuff_classes=2) result = head.loss() self.assertTrue(not head.with_loss) self.assertDictEqual(result, dict()) def test_predict(self): mask_cls_results = torch.rand((2, 10, 5)) mask_pred_results = torch.rand((2, 10, 32, 32)) batch_data_samples = [ DetDataSample( metainfo={ 'batch_input_shape': (32, 32), 'img_shape': (32, 30), 'ori_shape': (30, 30) }), DetDataSample( metainfo={ 'batch_input_shape': (32, 32), 'img_shape': (32, 30), 'ori_shape': (29, 30) }) ] # get panoptic and instance segmentation results test_cfg = Config( dict( panoptic_on=True, semantic_on=False, instance_on=True, max_per_image=10, object_mask_thr=0.3, iou_thr=0.3, filter_low_score=False)) head = MaskFormerFusionHead( num_things_classes=2, num_stuff_classes=2, test_cfg=test_cfg) results = head.predict( mask_cls_results, mask_pred_results, batch_data_samples, rescale=False) for i in range(len(results)): self.assertEqual(results[i]['pan_results'].sem_seg.shape[-2:], batch_data_samples[i].img_shape) self.assertEqual(results[i]['ins_results'].masks.shape[-2:], batch_data_samples[i].img_shape) results = head.predict( mask_cls_results, mask_pred_results, batch_data_samples, rescale=True) for i in range(len(results)): self.assertEqual(results[i]['pan_results'].sem_seg.shape[-2:], batch_data_samples[i].ori_shape) self.assertEqual(results[i]['ins_results'].masks.shape[-2:], batch_data_samples[i].ori_shape) # get empty results test_cfg = Config( dict( panoptic_on=False, semantic_on=False, instance_on=False, max_per_image=10, object_mask_thr=0.3, iou_thr=0.3, filter_low_score=False)) head = MaskFormerFusionHead( num_things_classes=2, num_stuff_classes=2, test_cfg=test_cfg) results = head.predict( mask_cls_results, mask_pred_results, batch_data_samples, rescale=True) for i in range(len(results)): self.assertEqual(results[i], dict()) # semantic segmentation is not supported test_cfg = Config( dict( panoptic_on=False, semantic_on=True, instance_on=False, max_per_image=10, object_mask_thr=0.3, iou_thr=0.3, filter_low_score=False)) head = MaskFormerFusionHead( num_things_classes=2, num_stuff_classes=2, test_cfg=test_cfg) with self.assertRaises(AssertionError): results = head.predict( mask_cls_results, mask_pred_results, batch_data_samples, rescale=True)
# Copyright (c) OpenMMLab. All rights reserved. import unittest import torch from mmengine.config import Config from mmdet.data_elements import DetDataSample from mmdet.models.seg_heads.panoptic_fusion_heads import MaskFormerFusionHead class TestMaskFormerFusionHead(unittest.TestCase): def test_loss(self): head = MaskFormerFusionHead(num_things_classes=2, num_stuff_classes=2) result = head.loss() self.assertTrue(not head.with_loss) self.assertDictEqual(result, dict()) def test_predict(self): mask_cls_results = torch.rand((2, 10, 5)) mask_pred_results = torch.rand((2, 10, 32, 32)) batch_data_samples = [ DetDataSample( metainfo={ 'batch_input_shape': (32, 32), 'img_shape': (32, 30), 'ori_shape': (30, 30) }), DetDataSample( metainfo={ 'batch_input_shape': (32, 32), 'img_shape': (32, 30), 'ori_shape': (29, 30) }) ] # get panoptic and instance segmentation results test_cfg = Config( dict( panoptic_on=True, semantic_on=False, instance_on=True, max_per_image=10, object_mask_thr=0.3, iou_thr=0.3, filter_low_score=False)) head = MaskFormerFusionHead( num_things_classes=2, num_stuff_classes=2, test_cfg=test_cfg) results = head.predict( mask_cls_results, mask_pred_results, batch_data_samples, rescale=False) for i in range(len(results)): self.assertEqual(results[i]['pan_results'].sem_seg.shape[-2:], batch_data_samples[i].img_shape) self.assertEqual(results[i]['ins_results'].masks.shape[-2:], batch_data_samples[i].img_shape) results = head.predict( mask_cls_results, mask_pred_results, batch_data_samples, rescale=True) for i in range(len(results)): self.assertEqual(results[i]['pan_results'].sem_seg.shape[-2:], batch_data_samples[i].ori_shape) self.assertEqual(results[i]['ins_results'].masks.shape[-2:], batch_data_samples[i].ori_shape) # get empty results test_cfg = Config( dict( panoptic_on=False, semantic_on=False, instance_on=False, max_per_image=10, object_mask_thr=0.3, iou_thr=0.3, filter_low_score=False)) head = MaskFormerFusionHead( num_things_classes=2, num_stuff_classes=2, test_cfg=test_cfg) results = head.predict( mask_cls_results, mask_pred_results, batch_data_samples, rescale=True) for i in range(len(results)): self.assertEqual(results[i], dict()) # semantic segmentation is not supported test_cfg = Config( dict( panoptic_on=False, semantic_on=True, instance_on=False, max_per_image=10, object_mask_thr=0.3, iou_thr=0.3, filter_low_score=False)) head = MaskFormerFusionHead( num_things_classes=2, num_stuff_classes=2, test_cfg=test_cfg) with self.assertRaises(AssertionError): results = head.predict( mask_cls_results, mask_pred_results, batch_data_samples, rescale=True)
import torch from parameterized import parameterized from torchaudio.prototype.models import squim_objective_base, squim_subjective_base from torchaudio_unittest.common_utils import skipIfNoCuda, torch_script, TorchaudioTestCase class TestSquimObjective(TorchaudioTestCase): def _smoke_test_objective(self, model, device, dtype): model = model.to(device=device, dtype=dtype) model = model.eval() batch_size, num_frames = 3, 16000 waveforms = torch.randn(batch_size, num_frames, device=device, dtype=dtype) model(waveforms) @parameterized.expand([(torch.float32,), (torch.float64,)]) def test_cpu_smoke_test(self, dtype): model = squim_objective_base() self._smoke_test_objective(model, torch.device("cpu"), dtype) @parameterized.expand([(torch.float32,), (torch.float64,)]) @skipIfNoCuda def test_cuda_smoke_test(self, dtype): model = squim_objective_base() self._smoke_test_objective(model, torch.device("cuda"), dtype) def test_batch_consistency(self): model = squim_objective_base() model.eval() batch_size, num_frames = 3, 16000 waveforms = torch.randn(batch_size, num_frames) ref_scores = model(waveforms) hyp_scores = [torch.zeros(batch_size), torch.zeros(batch_size), torch.zeros(batch_size)] for i in range(batch_size): scores = model(waveforms[i : i + 1]) for j in range(3): hyp_scores[j][i] = scores[j] self.assertEqual(len(hyp_scores), len(ref_scores)) for i in range(len(ref_scores)): self.assertEqual(hyp_scores[i], ref_scores[i]) def test_torchscript_consistency(self): model = squim_objective_base() model.eval() batch_size, num_frames = 3, 16000 waveforms = torch.randn(batch_size, num_frames) ref_scores = model(waveforms) scripted = torch_script(model) hyp_scores = scripted(waveforms) self.assertEqual(len(hyp_scores), len(ref_scores)) for i in range(len(ref_scores)): self.assertEqual(hyp_scores[i], ref_scores[i]) class TestSquimSubjective(TorchaudioTestCase): def _smoke_test_subjective(self, model, device, dtype): model = model.to(device=device, dtype=dtype) model = model.eval() batch_size, num_frames = 3, 16000 waveforms = torch.randn(batch_size, num_frames, device=device, dtype=dtype) reference = torch.randn(batch_size, num_frames, device=device, dtype=dtype) model(waveforms, reference) @parameterized.expand([(torch.float32,), (torch.float64,)]) def test_cpu_smoke_test(self, dtype): model = squim_subjective_base() self._smoke_test_subjective(model, torch.device("cpu"), dtype) @parameterized.expand([(torch.float32,), (torch.float64,)]) @skipIfNoCuda def test_cuda_smoke_test(self, dtype): model = squim_subjective_base() self._smoke_test_subjective(model, torch.device("cuda"), dtype) def test_batch_consistency(self): model = squim_subjective_base() model.eval() batch_size, num_frames = 3, 16000 waveforms = torch.randn(batch_size, num_frames) reference = torch.randn(batch_size, num_frames) ref_scores = model(waveforms, reference) hyp_scores = [] for i in range(batch_size): scores = model(waveforms[i : i + 1], reference[i : i + 1]) hyp_scores.append(scores) hyp_scores = torch.tensor(hyp_scores) self.assertEqual(hyp_scores, ref_scores) def test_torchscript_consistency(self): model = squim_subjective_base() model.eval() batch_size, num_frames = 3, 16000 waveforms = torch.randn(batch_size, num_frames) reference = torch.randn(batch_size, num_frames) ref_scores = model(waveforms, reference) scripted = torch_script(model) hyp_scores = scripted(waveforms, reference) self.assertEqual(hyp_scores, ref_scores)
import torch from parameterized import parameterized from torchaudio.prototype.models import squim_objective_base from torchaudio_unittest.common_utils import skipIfNoCuda, torch_script, TorchaudioTestCase class TestSQUIM(TorchaudioTestCase): def _smoke_test_objective(self, model, device, dtype): model = model.to(device=device, dtype=dtype) model = model.eval() batch_size, num_frames = 3, 16000 waveforms = torch.randn(batch_size, num_frames, device=device, dtype=dtype) model(waveforms) @parameterized.expand([(torch.float32,), (torch.float64,)]) def test_cpu_smoke_test(self, dtype): model = squim_objective_base() self._smoke_test_objective(model, torch.device("cpu"), dtype) @parameterized.expand([(torch.float32,), (torch.float64,)]) @skipIfNoCuda def test_cuda_smoke_test(self, dtype): model = squim_objective_base() self._smoke_test_objective(model, torch.device("cuda"), dtype) def test_batch_consistency(self): model = squim_objective_base() model.eval() batch_size, num_frames = 3, 16000 waveforms = torch.randn(batch_size, num_frames) ref_scores = model(waveforms) hyp_scores = [torch.zeros(batch_size), torch.zeros(batch_size), torch.zeros(batch_size)] for i in range(batch_size): scores = model(waveforms[i : i + 1]) for j in range(3): hyp_scores[j][i] = scores[j] self.assertEqual(len(hyp_scores), len(ref_scores)) for i in range(len(ref_scores)): self.assertEqual(hyp_scores[i], ref_scores[i]) def test_torchscript_consistency(self): model = squim_objective_base() model.eval() batch_size, num_frames = 3, 16000 waveforms = torch.randn(batch_size, num_frames) ref_scores = model(waveforms) scripted = torch_script(model) hyp_scores = scripted(waveforms) self.assertEqual(len(hyp_scores), len(ref_scores)) for i in range(len(ref_scores)): self.assertEqual(hyp_scores[i], ref_scores[i])
from __future__ import annotations from .CrossEncoder import CrossEncoder from .model_card import CrossEncoderModelCardData from .trainer import CrossEncoderTrainer from .training_args import CrossEncoderTrainingArguments __all__ = [ "CrossEncoder", "CrossEncoderTrainer", "CrossEncoderTrainingArguments", "CrossEncoderModelCardData", ]
from __future__ import annotations from .CrossEncoder import CrossEncoder __all__ = ["CrossEncoder"]
_base_ = '../mask_rcnn/mask-rcnn_r101_fpn_1x_coco.py' model = dict( backbone=dict(plugins=[ dict( cfg=dict(type='ContextBlock', ratio=1. / 16), stages=(False, True, True, True), position='after_conv3') ]))
_base_ = '../mask_rcnn/mask_rcnn_r101_fpn_1x_coco.py' model = dict( backbone=dict(plugins=[ dict( cfg=dict(type='ContextBlock', ratio=1. / 16), stages=(False, True, True, True), position='after_conv3') ]))
from typing import ( Union, TYPE_CHECKING, TypeVar, Sequence, Optional, List, Dict, Generator, Iterable, Tuple, ForwardRef, ) if TYPE_CHECKING: # pragma: no cover import scipy.sparse import tensorflow import torch import numpy as np from PIL.Image import Image as PILImage from docarray import Document ArrayType = TypeVar( 'ArrayType', np.ndarray, scipy.sparse.spmatrix, tensorflow.SparseTensor, tensorflow.Tensor, torch.Tensor, Sequence[float], ) DocumentContentType = Union[bytes, str, ArrayType] ProtoValueType = Optional[Union[str, bool, float]] StructValueType = Union[ ProtoValueType, List[ProtoValueType], Dict[str, ProtoValueType] ] DocumentArraySourceType = Union[ Sequence[Document], Document, Generator[Document], Iterable[Document] ] T = TypeVar('T') AnyDNN = TypeVar( 'AnyDNN' ) #: The type of any implementation of a Deep Neural Network object DocumentArraySingletonIndexType = Union[int, str] DocumentArrayMultipleIndexType = Union[ slice, Sequence[int], Sequence[str], Sequence[bool], Ellipsis ] DocumentArraySingleAttributeType = Tuple[ Union[DocumentArraySingletonIndexType, DocumentArrayMultipleIndexType], str ] DocumentArrayMultipleAttributeType = Tuple[ Union[DocumentArraySingletonIndexType, DocumentArrayMultipleIndexType], Sequence[str], ] DocumentArrayIndexType = Union[ DocumentArraySingletonIndexType, DocumentArrayMultipleIndexType, DocumentArraySingleAttributeType, DocumentArrayMultipleAttributeType, ] Image = TypeVar( 'Image', str, ForwardRef('np.ndarray'), ForwardRef('PILImage'), ) Text = TypeVar('Text', bound=str) URI = TypeVar('URI', bound=str) Audio = TypeVar('Audio', str, ForwardRef('np.ndarray')) Video = TypeVar('Video', str, ForwardRef('np.ndarray')) Mesh = TypeVar('Mesh', str, ForwardRef('np.ndarray')) Tabular = TypeVar('Tabular', bound=str) Blob = TypeVar('Blob', str, bytes) JSON = TypeVar('JSON', str, dict)
from typing import ( Union, TYPE_CHECKING, TypeVar, Sequence, Optional, List, Dict, Generator, Iterable, Tuple, ForwardRef, ) if TYPE_CHECKING: import scipy.sparse import tensorflow import torch import numpy as np from PIL.Image import Image as PILImage from docarray import Document ArrayType = TypeVar( 'ArrayType', np.ndarray, scipy.sparse.spmatrix, tensorflow.SparseTensor, tensorflow.Tensor, torch.Tensor, Sequence[float], ) DocumentContentType = Union[bytes, str, ArrayType] ProtoValueType = Optional[Union[str, bool, float]] StructValueType = Union[ ProtoValueType, List[ProtoValueType], Dict[str, ProtoValueType] ] DocumentArraySourceType = Union[ Sequence[Document], Document, Generator[Document], Iterable[Document] ] T = TypeVar('T') AnyDNN = TypeVar( 'AnyDNN' ) #: The type of any implementation of a Deep Neural Network object DocumentArraySingletonIndexType = Union[int, str] DocumentArrayMultipleIndexType = Union[ slice, Sequence[int], Sequence[str], Sequence[bool], Ellipsis ] DocumentArraySingleAttributeType = Tuple[ Union[DocumentArraySingletonIndexType, DocumentArrayMultipleIndexType], str ] DocumentArrayMultipleAttributeType = Tuple[ Union[DocumentArraySingletonIndexType, DocumentArrayMultipleIndexType], Sequence[str], ] DocumentArrayIndexType = Union[ DocumentArraySingletonIndexType, DocumentArrayMultipleIndexType, DocumentArraySingleAttributeType, DocumentArrayMultipleAttributeType, ] Image = TypeVar( 'Image', str, ForwardRef('np.ndarray'), ForwardRef('PILImage'), ) Text = TypeVar('Text', bound=str) URI = TypeVar('URI', bound=str) Audio = TypeVar('Audio', str, ForwardRef('np.ndarray')) Video = TypeVar('Video', str, ForwardRef('np.ndarray')) Mesh = TypeVar('Mesh', str, ForwardRef('np.ndarray')) Tabular = TypeVar('Tabular', bound=str) Blob = TypeVar('Blob', str, bytes) JSON = TypeVar('JSON', str, dict)
import tracemalloc from functools import wraps from docarray import DocList from docarray.documents import TextDoc def get_test_da(n: int): return DocList[TextDoc](gen_text_docs(n)) def gen_text_docs(n: int): for i in range(n): yield TextDoc(text=f'text {i}') def profile_memory(func): """Decorator to profile memory usage of a function. Returns: original function return value, (current memory usage, peak memory usage) """ @wraps(func) def _inner(*args, **kwargs): tracemalloc.start() ret = func(*args, **kwargs) current, peak = tracemalloc.get_traced_memory() tracemalloc.stop() return ret, (current, peak) return _inner
import tracemalloc from functools import wraps from docarray import DocArray from docarray.documents import TextDoc def get_test_da(n: int): return DocArray[TextDoc](gen_text_docs(n)) def gen_text_docs(n: int): for i in range(n): yield TextDoc(text=f'text {i}') def profile_memory(func): """Decorator to profile memory usage of a function. Returns: original function return value, (current memory usage, peak memory usage) """ @wraps(func) def _inner(*args, **kwargs): tracemalloc.start() ret = func(*args, **kwargs) current, peak = tracemalloc.get_traced_memory() tracemalloc.stop() return ret, (current, peak) return _inner
"""Init composability.""" from llama_index.core.composability.base import ComposableGraph from llama_index.core.composability.joint_qa_summary import ( QASummaryQueryEngineBuilder, ) __all__ = ["ComposableGraph", "QASummaryQueryEngineBuilder"]
"""Init composability.""" from llama_index.core.composability.base import ComposableGraph from llama_index.core.composability.joint_qa_summary import ( QASummaryQueryEngineBuilder, ) __all__ = ["ComposableGraph", "QASummaryQueryEngineBuilder"]
# Copyright (c) OpenMMLab. All rights reserved. import torch.nn as nn from mmcv.cnn import ConvModule, is_norm from mmengine.model import caffe2_xavier_init, constant_init, normal_init from torch.nn import BatchNorm2d from mmdet.registry import MODELS class Bottleneck(nn.Module): """Bottleneck block for DilatedEncoder used in `YOLOF. <https://arxiv.org/abs/2103.09460>`. The Bottleneck contains three ConvLayers and one residual connection. Args: in_channels (int): The number of input channels. mid_channels (int): The number of middle output channels. dilation (int): Dilation rate. norm_cfg (dict): Dictionary to construct and config norm layer. """ def __init__(self, in_channels, mid_channels, dilation, norm_cfg=dict(type='BN', requires_grad=True)): super(Bottleneck, self).__init__() self.conv1 = ConvModule( in_channels, mid_channels, 1, norm_cfg=norm_cfg) self.conv2 = ConvModule( mid_channels, mid_channels, 3, padding=dilation, dilation=dilation, norm_cfg=norm_cfg) self.conv3 = ConvModule( mid_channels, in_channels, 1, norm_cfg=norm_cfg) def forward(self, x): identity = x out = self.conv1(x) out = self.conv2(out) out = self.conv3(out) out = out + identity return out @MODELS.register_module() class DilatedEncoder(nn.Module): """Dilated Encoder for YOLOF <https://arxiv.org/abs/2103.09460>`. This module contains two types of components: - the original FPN lateral convolution layer and fpn convolution layer, which are 1x1 conv + 3x3 conv - the dilated residual block Args: in_channels (int): The number of input channels. out_channels (int): The number of output channels. block_mid_channels (int): The number of middle block output channels num_residual_blocks (int): The number of residual blocks. block_dilations (list): The list of residual blocks dilation. """ def __init__(self, in_channels, out_channels, block_mid_channels, num_residual_blocks, block_dilations): super(DilatedEncoder, self).__init__() self.in_channels = in_channels self.out_channels = out_channels self.block_mid_channels = block_mid_channels self.num_residual_blocks = num_residual_blocks self.block_dilations = block_dilations self._init_layers() def _init_layers(self): self.lateral_conv = nn.Conv2d( self.in_channels, self.out_channels, kernel_size=1) self.lateral_norm = BatchNorm2d(self.out_channels) self.fpn_conv = nn.Conv2d( self.out_channels, self.out_channels, kernel_size=3, padding=1) self.fpn_norm = BatchNorm2d(self.out_channels) encoder_blocks = [] for i in range(self.num_residual_blocks): dilation = self.block_dilations[i] encoder_blocks.append( Bottleneck( self.out_channels, self.block_mid_channels, dilation=dilation)) self.dilated_encoder_blocks = nn.Sequential(*encoder_blocks) def init_weights(self): caffe2_xavier_init(self.lateral_conv) caffe2_xavier_init(self.fpn_conv) for m in [self.lateral_norm, self.fpn_norm]: constant_init(m, 1) for m in self.dilated_encoder_blocks.modules(): if isinstance(m, nn.Conv2d): normal_init(m, mean=0, std=0.01) if is_norm(m): constant_init(m, 1) def forward(self, feature): out = self.lateral_norm(self.lateral_conv(feature[-1])) out = self.fpn_norm(self.fpn_conv(out)) return self.dilated_encoder_blocks(out),
# Copyright (c) OpenMMLab. All rights reserved. import torch.nn as nn from mmcv.cnn import ConvModule, is_norm from mmengine.model.utils import caffe2_xavier_init, constant_init, normal_init from torch.nn import BatchNorm2d from mmdet.registry import MODELS class Bottleneck(nn.Module): """Bottleneck block for DilatedEncoder used in `YOLOF. <https://arxiv.org/abs/2103.09460>`. The Bottleneck contains three ConvLayers and one residual connection. Args: in_channels (int): The number of input channels. mid_channels (int): The number of middle output channels. dilation (int): Dilation rate. norm_cfg (dict): Dictionary to construct and config norm layer. """ def __init__(self, in_channels, mid_channels, dilation, norm_cfg=dict(type='BN', requires_grad=True)): super(Bottleneck, self).__init__() self.conv1 = ConvModule( in_channels, mid_channels, 1, norm_cfg=norm_cfg) self.conv2 = ConvModule( mid_channels, mid_channels, 3, padding=dilation, dilation=dilation, norm_cfg=norm_cfg) self.conv3 = ConvModule( mid_channels, in_channels, 1, norm_cfg=norm_cfg) def forward(self, x): identity = x out = self.conv1(x) out = self.conv2(out) out = self.conv3(out) out = out + identity return out @MODELS.register_module() class DilatedEncoder(nn.Module): """Dilated Encoder for YOLOF <https://arxiv.org/abs/2103.09460>`. This module contains two types of components: - the original FPN lateral convolution layer and fpn convolution layer, which are 1x1 conv + 3x3 conv - the dilated residual block Args: in_channels (int): The number of input channels. out_channels (int): The number of output channels. block_mid_channels (int): The number of middle block output channels num_residual_blocks (int): The number of residual blocks. block_dilations (list): The list of residual blocks dilation. """ def __init__(self, in_channels, out_channels, block_mid_channels, num_residual_blocks, block_dilations): super(DilatedEncoder, self).__init__() self.in_channels = in_channels self.out_channels = out_channels self.block_mid_channels = block_mid_channels self.num_residual_blocks = num_residual_blocks self.block_dilations = block_dilations self._init_layers() def _init_layers(self): self.lateral_conv = nn.Conv2d( self.in_channels, self.out_channels, kernel_size=1) self.lateral_norm = BatchNorm2d(self.out_channels) self.fpn_conv = nn.Conv2d( self.out_channels, self.out_channels, kernel_size=3, padding=1) self.fpn_norm = BatchNorm2d(self.out_channels) encoder_blocks = [] for i in range(self.num_residual_blocks): dilation = self.block_dilations[i] encoder_blocks.append( Bottleneck( self.out_channels, self.block_mid_channels, dilation=dilation)) self.dilated_encoder_blocks = nn.Sequential(*encoder_blocks) def init_weights(self): caffe2_xavier_init(self.lateral_conv) caffe2_xavier_init(self.fpn_conv) for m in [self.lateral_norm, self.fpn_norm]: constant_init(m, 1) for m in self.dilated_encoder_blocks.modules(): if isinstance(m, nn.Conv2d): normal_init(m, mean=0, std=0.01) if is_norm(m): constant_init(m, 1) def forward(self, feature): out = self.lateral_norm(self.lateral_conv(feature[-1])) out = self.fpn_norm(self.fpn_conv(out)) return self.dilated_encoder_blocks(out),