input
stringlengths
33
5k
output
stringlengths
32
5k
import asyncio import copy from typing import Any, List, Optional from jina.serve.gateway import BaseGateway class CompositeGateway(BaseGateway): """GRPC Gateway implementation""" def __init__( self, **kwargs, ): """Initialize the gateway :param kwargs: keyword args ...
import asyncio import copy from typing import Any, List, Optional from jina.serve.gateway import BaseGateway class CompositeGateway(BaseGateway): """GRPC Gateway implementation""" def __init__( self, **kwargs, ): """Initialize the gateway :param kwargs: keyword args ...
_base_ = './retinanet_r50-caffe_fpn_1x_coco.py' train_pipeline = [ dict(type='LoadImageFromFile'), dict(type='LoadAnnotations', with_bbox=True), dict( type='RandomResize', scale=[(1333, 640), (1333, 672), (1333, 704), (1333, 736), (1333, 768), (1333, 800)], keep_ratio...
_base_ = './retinanet_r50_caffe_fpn_1x_coco.py' train_pipeline = [ dict(type='LoadImageFromFile'), dict(type='LoadAnnotations', with_bbox=True), dict( type='RandomResize', scale=[(1333, 640), (1333, 672), (1333, 704), (1333, 736), (1333, 768), (1333, 800)], keep_ratio...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.core import ConfigType, OptConfigType, OptMultiConfig from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class PAA(SingleStageDetector): """Implementation of `PAA <https://arxiv.org/pdf/2007.08103.pdf...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.core import ConfigType, OptConfigType, OptMultiConfig from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class PAA(SingleStageDetector): """Implementation of `PAA <https://arxiv.org/pdf/2007.08103.pdf...
import re from typing import Union from langchain_core.agents import AgentAction, AgentFinish from langchain_core.exceptions import OutputParserException from langchain.agents.agent import AgentOutputParser from langchain.agents.mrkl.prompt import FORMAT_INSTRUCTIONS FINAL_ANSWER_ACTION = "Final Answer:" MISSING_ACT...
import re from typing import Union from langchain_core.agents import AgentAction, AgentFinish from langchain_core.exceptions import OutputParserException from langchain.agents.agent import AgentOutputParser from langchain.agents.mrkl.prompt import FORMAT_INSTRUCTIONS FINAL_ANSWER_ACTION = "Final Answer:" MISSING_ACT...
from docarray.typing.tensor.video.video_ndarray import VideoNdArray __all__ = ['VideoNdArray'] from docarray.utils._internal.misc import is_tf_available, is_torch_available torch_available = is_torch_available() if torch_available: from docarray.typing.tensor.video.video_torch_tensor import VideoTorchTensor # n...
from docarray.typing.tensor.video.video_ndarray import VideoNdArray __all__ = ['VideoNdArray'] from docarray.utils.misc import is_tf_available, is_torch_available torch_available = is_torch_available() if torch_available: from docarray.typing.tensor.video.video_torch_tensor import VideoTorchTensor # noqa _...
from typing import Any, Dict, Optional, Union import PIL.Image import torch from torchvision.prototype import features from torchvision.prototype.transforms import functional as F, Transform class ConvertBoundingBoxFormat(Transform): _transformed_types = (features.BoundingBox,) def __init__(self, format: U...
from typing import Any, Dict, Optional, Union import PIL.Image import torch from torchvision.prototype import features from torchvision.prototype.transforms import functional as F, Transform class ConvertBoundingBoxFormat(Transform): _transformed_types = (features.BoundingBox,) def __init__(self, format: U...
from langchain.chains.structured_output.base import ( create_openai_fn_runnable, create_structured_output_runnable, ) __all__ = ["create_openai_fn_runnable", "create_structured_output_runnable"]
from langchain.chains.structured_output.base import ( create_openai_fn_runnable, create_structured_output_runnable, ) __all__ = ["create_structured_output_runnable", "create_openai_fn_runnable"]
import asyncio import copy from typing import Any, List from jina.serve.runtimes.servers import BaseServer class CompositeServer(BaseServer): """Composite Server implementation""" def __init__( self, **kwargs, ): """Initialize the gateway :param kwargs: keyword ar...
import asyncio import copy from typing import Any, List from jina.serve.runtimes.servers import BaseServer class CompositeServer(BaseServer): """Composite Server implementation""" def __init__( self, **kwargs, ): """Initialize the gateway :param kwargs: keyword ar...
from enum import Enum from fsspec import AbstractFileSystem from pathlib import Path from typing import Any, Dict, Iterable, Optional, Protocol, runtime_checkable import json import uuid from docling.document_converter import DocumentConverter from docling_core.types import DoclingDocument as DLDocument from llama_ind...
from enum import Enum from fsspec import AbstractFileSystem from pathlib import Path from typing import Any, Dict, Iterable, Optional, Protocol, runtime_checkable import json import uuid from docling.document_converter import DocumentConverter from docling_core.types import DoclingDocument as DLDocument from llama_ind...
from typing import Optional, Union import torch from torch import nn, Tensor def _cat(tensors: list[Tensor], dim: int = 0) -> Tensor: """ Efficient version of torch.cat that avoids a copy if there is only a single element in a list """ # TODO add back the assert # assert isinstance(tensors, (list...
from typing import List, Optional, Tuple, Union import torch from torch import nn, Tensor def _cat(tensors: List[Tensor], dim: int = 0) -> Tensor: """ Efficient version of torch.cat that avoids a copy if there is only a single element in a list """ # TODO add back the assert # assert isinstance(t...
# Copyright (c) OpenMMLab. All rights reserved. import random from typing import Sequence import numpy as np import torch DATA_BATCH = Sequence[dict] def worker_init_fn(worker_id: int, num_workers: int, rank: int, seed: int) -> None: """This function will be called on each worker subprocess a...
# Copyright (c) OpenMMLab. All rights reserved. import random from typing import Any, Sequence, Tuple import numpy as np import torch from .base_data_element import BaseDataElement DATA_BATCH = Sequence[Tuple[Any, BaseDataElement]] def worker_init_fn(worker_id: int, num_workers: int, rank: int, ...
#!/usr/bin/env python """Script to sync libraries from various repositories into the main langchain repository.""" import os import shutil import yaml from pathlib import Path from typing import Dict, Any def load_packages_yaml() -> Dict[str, Any]: """Load and parse the packages.yml file.""" with open("langc...
#!/usr/bin/env python """Script to sync libraries from various repositories into the main langchain repository.""" import os import shutil import yaml from pathlib import Path from typing import Dict, Any def load_packages_yaml() -> Dict[str, Any]: """Load and parse the packages.yml file.""" with open("langc...
from functools import partial from torchaudio.models import emformer_rnnt_base from torchaudio.pipelines import RNNTBundle EMFORMER_RNNT_BASE_MUSTC = RNNTBundle( _rnnt_path="models/emformer_rnnt_base_mustc.pt", _rnnt_factory_func=partial(emformer_rnnt_base, num_symbols=501), _global_stats_path="pipeline-...
from functools import partial from torchaudio.models import emformer_rnnt_base from torchaudio.pipelines import RNNTBundle EMFORMER_RNNT_BASE_MUSTC = RNNTBundle( _rnnt_path="emformer_rnnt_base_mustc.pt", _rnnt_factory_func=partial(emformer_rnnt_base, num_symbols=501), _global_stats_path="global_stats_rnn...
""" Demo for using and defining callback functions ============================================== .. versionadded:: 1.3.0 """ import argparse import os import tempfile import numpy as np from matplotlib import pyplot as plt from sklearn.datasets import load_breast_cancer from sklearn.model_selection import train_...
""" Demo for using and defining callback functions ============================================== .. versionadded:: 1.3.0 """ import argparse import os import tempfile import numpy as np from matplotlib import pyplot as plt from sklearn.datasets import load_breast_cancer from sklearn.model_selection import train_...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.tree.tree_api import assert_same_paths from keras.src.tree.tree_api import assert_same_structure from keras.src.tree.tree_api import flatten from keras.src.tree.tree_api import flatte...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.tree.tree_api import assert_same_structure from keras.src.tree.tree_api import flatten from keras.src.tree.tree_api import is_nested from keras.src.tree.tree_api import lists_to_tuple...
from tempfile import NamedTemporaryFile import pytest import requests from datasets.utils.file_utils import fsspec_get, fsspec_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline, require_not_windows @pytest.mark.integration @require_not_windows # fsspec get keeps a file hand...
from tempfile import NamedTemporaryFile import huggingface_hub import pytest import requests from packaging import version from datasets.utils.file_utils import fsspec_get, fsspec_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline, require_not_windows @pytest.mark.integration...
from typing import List, Sequence from llama_index.core.agent.workflow.base_agent import BaseWorkflowAgent from llama_index.core.agent.workflow.single_agent_workflow import SingleAgentRunnerMixin from llama_index.core.agent.workflow.workflow_events import ( AgentInput, AgentOutput, AgentStream, ToolCal...
from typing import List, Sequence from llama_index.core.agent.workflow.base_agent import BaseWorkflowAgent from llama_index.core.agent.workflow.single_agent_workflow import SingleAgentRunnerMixin from llama_index.core.agent.workflow.workflow_events import ( AgentInput, AgentOutput, AgentStream, ToolCal...
from pathlib import Path import numpy as np import pytest from jina import Document, DocumentArray, Executor from ...custom_image_torch_encoder import CustomImageTorchEncoder @pytest.fixture def encoder(): model_dir = Path(__file__).parents[1] / 'model' return CustomImageTorchEncoder( model_definiti...
import os import numpy as np import pytest from jina import Document, DocumentArray from ...custom_image_torch_encoder import CustomImageTorchEncoder cur_dir = os.path.dirname(os.path.abspath(__file__)) @pytest.fixture def encoder(tmpdir): model_state_dict_path = os.path.join(cur_dir, '../model/model_state_dic...
# Copyright (c) OpenMMLab. All rights reserved. from .coco_metric import CocoMetric __all__ = ['CocoMetric']
# Copyright (c) OpenMMLab. All rights reserved.
"""Various utilities to help with development.""" # Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from ..exceptions import DataConversionWarning from . import metadata_routing from ._bunch import Bunch from ._chunking import gen_batches, gen_even_slices from ._estimator_html_repr import...
"""Various utilities to help with development.""" # Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from ..exceptions import DataConversionWarning from . import metadata_routing from ._bunch import Bunch from ._chunking import gen_batches, gen_even_slices from ._estimator_html_repr import...
__version__ = '0.34.1' import logging from docarray.array import DocList, DocVec from docarray.base_doc.doc import BaseDoc from docarray.utils._internal.misc import _get_path_from_docarray_root_level __all__ = ['BaseDoc', 'DocList', 'DocVec'] logger = logging.getLogger('docarray') handler = logging.StreamHandler()...
__version__ = '0.34.0' import logging from docarray.array import DocList, DocVec from docarray.base_doc.doc import BaseDoc from docarray.utils._internal.misc import _get_path_from_docarray_root_level __all__ = ['BaseDoc', 'DocList', 'DocVec'] logger = logging.getLogger('docarray') handler = logging.StreamHandler()...
import sys from dataclasses import dataclass from typing import TYPE_CHECKING, Dict, List, Optional, Tuple, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast if TYPE_CHECKING: im...
import sys from dataclasses import dataclass from typing import TYPE_CHECKING, Dict, List, Optional, Tuple, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast if TYPE_CHECKING: im...
__copyright__ = 'Copyright (c) 2020-2021 Jina AI Limited. All rights reserved.' __license__ = 'Apache-2.0' from typing import Any, Iterable, Optional import librosa as lr import numpy as np import torch from jina import DocumentArray, Executor, requests from jina.excepts import BadDocType from .audio_clip.model impo...
__copyright__ = 'Copyright (c) 2020-2021 Jina AI Limited. All rights reserved.' __license__ = 'Apache-2.0' from typing import Any, Iterable, Optional import librosa as lr import numpy as np import torch from jina import DocumentArray, Executor, requests from jina.excepts import BadDocType from .audio_clip.model impo...
import json import os import pytest from hubble.executor import HubExecutor from hubble.executor.hubio import HubIO from jina import __version__ from jina.orchestrate.deployments.config.helper import ( get_base_executor_version, get_image_name, to_compatible_name, ) @pytest.mark.parametrize('is_master',...
import json import os import pytest from hubble.executor import HubExecutor from hubble.executor.hubio import HubIO from jina import __version__ from jina.orchestrate.deployments.config.helper import ( get_base_executor_version, get_image_name, to_compatible_name, ) @pytest.mark.parametrize('is_master',...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import warnings from mmcv import Config, DictAction from mmdet.utils import update_data_root def parse_args(): parser = argparse.ArgumentParser(description='Print the whole config') parser.add_argument('config', help='config file path') par...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import warnings from mmcv import Config, DictAction def parse_args(): parser = argparse.ArgumentParser(description='Print the whole config') parser.add_argument('config', help='config file path') parser.add_argument( '--options', ...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/coco_panoptic.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] img_norm_cfg = dict( mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True) model = dict( type='PanopticFPN', img_n...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/coco_panoptic.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( type='PanopticFPN', semantic_head=dict( type='PanopticFPNHead', num_things_classes=80, num_stuff_cla...
import os import random import time from typing import Dict, OrderedDict import numpy as np import pytest from jina import Document, DocumentArray, Executor, Flow, requests from jina_commons.indexers.dump import dump_docs from jinahub.indexers.compound.FaissLMDBSearcher.faiss_lmdb import FaissLMDBSearcher from jinahu...
import os import random import time from typing import Dict, OrderedDict import numpy as np import pytest from jina import Document, DocumentArray, Executor, Flow, requests from jina_commons.indexers.dump import dump_docs from jinahub.indexers.compound.FaissLMDBSearcher.faiss_lmdb import FaissLMDBSearcher from jinahu...
"""Configuration for unit tests.""" from collections.abc import Sequence from importlib import util import pytest from pytest import Config, Function, Parser def pytest_addoption(parser: Parser) -> None: """Add custom command line options to pytest.""" parser.addoption( "--only-extended", ac...
"""Configuration for unit tests.""" from collections.abc import Sequence from importlib import util import pytest from pytest import Config, Function, Parser def pytest_addoption(parser: Parser) -> None: """Add custom command line options to pytest.""" parser.addoption( "--only-extended", ac...
# dataset settings dataset_type = 'CityscapesDataset' data_root = 'data/cityscapes/' train_pipeline = [ dict(type='LoadImageFromFile'), dict(type='LoadAnnotations', with_bbox=True, with_mask=True), dict( type='RandomResize', scale=[(2048, 800), (2048, 1024)], keep_ratio=True), d...
# dataset settings dataset_type = 'CityscapesDataset' data_root = 'data/cityscapes/' train_pipeline = [ dict(type='LoadImageFromFile'), dict(type='LoadAnnotations', with_bbox=True, with_mask=True), dict(type='RandomResize', scale=[(2048, 800), (2048, 1024)]), dict(type='RandomFlip', prob=0.5), dict...
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
_base_ = './retinanet_r50_fpn_ghm-1x_coco.py' model = dict( backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), style='pytorch...
_base_ = './retinanet_ghm_r50_fpn_1x_coco.py' model = dict( backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), style='pytorch...
"""Test EdenAi's image moderation Tool . In order to run this test, you need to have an EdenAI api key. You can get it by registering for free at https://app.edenai.run/user/register. A test key can be found at https://app.edenai.run/admin/account/settings by clicking on the 'sandbox' toggle. (calls will be free, and ...
"""Test EdenAi's image moderation Tool . In order to run this test, you need to have an EdenAI api key. You can get it by registering for free at https://app.edenai.run/user/register. A test key can be found at https://app.edenai.run/admin/account/settings by clicking on the 'sandbox' toggle. (calls will be free, and ...
# Copyright (c) OpenMMLab. All rights reserved. import asyncio from argparse import ArgumentParser from mmdet.apis import (async_inference_detector, inference_detector, init_detector, show_result_pyplot) def parse_args(): parser = ArgumentParser() parser.add_argument('img', help='Imag...
# Copyright (c) OpenMMLab. All rights reserved. import asyncio from argparse import ArgumentParser from mmdet.apis import (async_inference_detector, inference_detector, init_detector, show_result_pyplot) def parse_args(): parser = ArgumentParser() parser.add_argument('img', help='Imag...
from functools import partial from torchaudio.models import emformer_rnnt_base from torchaudio.pipelines import RNNTBundle EMFORMER_RNNT_BASE_MUSTC = RNNTBundle( _rnnt_path="models/emformer_rnnt_base_mustc.pt", _rnnt_factory_func=partial(emformer_rnnt_base, num_symbols=501), _global_stats_path="pipeline-...
from functools import partial from torchaudio.models import emformer_rnnt_base from torchaudio.pipelines import RNNTBundle EMFORMER_RNNT_BASE_MUSTC = RNNTBundle( _rnnt_path="models/emformer_rnnt_base_mustc.pt", _rnnt_factory_func=partial(emformer_rnnt_base, num_symbols=501), _global_stats_path="pipeline-...
from __future__ import annotations from collections.abc import Iterable from torch import Tensor from sentence_transformers import util from sentence_transformers.losses.MultipleNegativesRankingLoss import MultipleNegativesRankingLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder cla...
from __future__ import annotations from collections.abc import Iterable from torch import Tensor from sentence_transformers import util from sentence_transformers.losses.MultipleNegativesRankingLoss import MultipleNegativesRankingLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder cla...
from .filtering import ( allpass_biquad, band_biquad, bandpass_biquad, bandreject_biquad, bass_biquad, biquad, contrast, dcshift, deemph_biquad, dither, equalizer_biquad, filtfilt, flanger, gain, highpass_biquad, lfilter, lowpass_biquad, overdrive,...
from .filtering import ( allpass_biquad, band_biquad, bandpass_biquad, bandreject_biquad, bass_biquad, biquad, contrast, dcshift, deemph_biquad, dither, equalizer_biquad, filtfilt, flanger, gain, highpass_biquad, lfilter, lowpass_biquad, overdrive,...
# Copyright (c) OpenMMLab. All rights reserved. import random import warnings import torch from mmcv.runner import get_dist_info from mmcv.runner.hooks import HOOKS, Hook from torch import distributed as dist @HOOKS.register_module() class SyncRandomSizeHook(Hook): """Change and synchronize the random image size...
# Copyright (c) OpenMMLab. All rights reserved. import random import warnings import torch from mmcv.runner import get_dist_info from mmcv.runner.hooks import HOOKS, Hook from torch import distributed as dist @HOOKS.register_module() class SyncRandomSizeHook(Hook): """Change and synchronize the random image size...
""" This file contains some utilities functions used to find parallel sentences in two monolingual corpora. Code in this file has been adapted from the LASER repository: https://github.com/facebookresearch/LASER """ import gzip import lzma import time import faiss import numpy as np ######## Functions to find and...
""" This file contains some utilities functions used to find parallel sentences in two monolingual corpora. Code in this file has been adapted from the LASER repository: https://github.com/facebookresearch/LASER """ import faiss import numpy as np import time import gzip import lzma ######## Functions to find and ...
import numpy as np import torch from docarray import BaseDocument from docarray.document import AnyDocument from docarray.typing import ( AnyEmbedding, AnyUrl, ImageUrl, Mesh3DUrl, NdArray, PointCloud3DUrl, TextUrl, TorchTensor, ) def test_proto_all_types(): class Mymmdoc(BaseDocu...
import numpy as np import torch from docarray import BaseDocument from docarray.document import AnyDocument from docarray.typing import ( AnyUrl, Embedding, ImageUrl, Mesh3DUrl, NdArray, PointCloud3DUrl, TextUrl, TorchTensor, ) def test_proto_all_types(): class Mymmdoc(BaseDocumen...
# Copyright (c) OpenMMLab. All rights reserved. """Collecting some commonly used type hint in mmdetection.""" from typing import Dict, List, Optional, Tuple, Union import torch from mmengine.config import ConfigDict from mmengine.data import InstanceData from ..bbox.samplers import SamplingResult from ..data_structur...
# Copyright (c) OpenMMLab. All rights reserved. """Collecting some commonly used type hint in mmdetection.""" from typing import List, Optional, Union from mmengine.config import ConfigDict from mmengine.data import InstanceData from ..bbox.samplers import SamplingResult from ..data_structures import DetDataSample #...
#!/usr/bin/env python3 import logging import pathlib from argparse import ArgumentParser from common import MODEL_TYPE_LIBRISPEECH, MODEL_TYPE_MUSTC, MODEL_TYPE_TEDLIUM3 from librispeech.lightning import LibriSpeechRNNTModule from mustc.lightning import MuSTCRNNTModule from pytorch_lightning import Trainer from pytorc...
#!/usr/bin/env python3 import logging import pathlib from argparse import ArgumentParser from common import MODEL_TYPE_LIBRISPEECH, MODEL_TYPE_TEDLIUM3, MODEL_TYPE_MUSTC from librispeech.lightning import LibriSpeechRNNTModule from mustc.lightning import MuSTCRNNTModule from pytorch_lightning import Trainer from pytorc...
import os import time import uuid from contextlib import contextmanager from typing import Optional import pytest import requests from huggingface_hub.hf_api import HfApi, RepositoryNotFoundError CI_HUB_USER = "DSUser" CI_HUB_USER_FULL_NAME = "Dummy Datasets User" CI_HUB_USER_TOKEN = "hf_iiTdXZFWohTKHEfuQWoEmmmaEVCF...
import os import time import uuid from contextlib import contextmanager from typing import Optional import pytest import requests from huggingface_hub.hf_api import HfApi, RepositoryNotFoundError CI_HUB_USER = "__DUMMY_TRANSFORMERS_USER__" CI_HUB_USER_FULL_NAME = "Dummy User" CI_HUB_USER_TOKEN = "hf_hZEmnoOEYISjraJt...
from typing import Dict, List, Tuple import pytest from opentelemetry.metrics import Meter from opentelemetry.sdk.metrics import MeterProvider from opentelemetry.sdk.metrics.export import ( HistogramDataPoint, InMemoryMetricReader, Metric, ) from jina.serve.networking import _NetworkingHistograms @pytes...
import pytest from typing import Tuple, List, Dict from opentelemetry.sdk.metrics import MeterProvider from opentelemetry.sdk.metrics.export import ( InMemoryMetricReader, Metric, HistogramDataPoint, ) from opentelemetry.metrics import Meter from jina.serve.networking import _NetworkingHistograms @pytest...
import numpy as np from docarray import BaseDoc from docarray.array import DocArrayStacked from docarray.array.stacked.column_storage import ColumnStorageView from docarray.typing import AnyTensor def test_column_storage_init(): class InnerDoc(BaseDoc): price: int class MyDoc(BaseDoc): tenso...
import numpy as np from docarray import BaseDocument from docarray.array import DocumentArrayStacked from docarray.array.stacked.column_storage import ColumnStorageView from docarray.typing import AnyTensor def test_column_storage_init(): class InnerDoc(BaseDocument): price: int class MyDoc(BaseDocu...
from __future__ import annotations from typing_extensions import deprecated from sentence_transformers import InputExample from sentence_transformers.cross_encoder.evaluation.CEClassificationEvaluator import CEClassificationEvaluator @deprecated( "This evaluator has been deprecated in favor of the more general ...
from __future__ import annotations import csv import logging import os import numpy as np from sentence_transformers import InputExample logger = logging.getLogger(__name__) class CEBinaryAccuracyEvaluator: """ This evaluator can be used with the CrossEncoder class. It is designed for CrossEncoders w...
import ast from langchain_community.utilities.steam import SteamWebAPIWrapper def test_get_game_details() -> None: """Test for getting game details on Steam""" steam = SteamWebAPIWrapper() output = steam.run("get_game_details", "Terraria") assert "id" in output assert "link" in output assert ...
import ast from langchain_community.utilities.steam import SteamWebAPIWrapper def test_get_game_details() -> None: """Test for getting game details on Steam""" steam = SteamWebAPIWrapper() # type: ignore[call-arg] output = steam.run("get_game_details", "Terraria") assert "id" in output assert "l...
# -*- coding: utf-8 -*- # Generated by the protocol buffer compiler. DO NOT EDIT! # source: docarray.proto """Generated protocol buffer code.""" from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool from google.protobuf import symbol_database as _symbol_d...
# -*- coding: utf-8 -*- # Generated by the protocol buffer compiler. DO NOT EDIT! # source: docarray.proto """Generated protocol buffer code.""" from google.protobuf.internal import builder as _builder from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool...
"""Tool for the Google Trends""" from typing import Optional from langchain_core.callbacks import CallbackManagerForToolRun from langchain_core.tools import BaseTool from langchain_community.utilities.google_jobs import GoogleJobsAPIWrapper class GoogleJobsQueryRun(BaseTool): """Tool that queries the Google Jo...
"""Tool for the Google Trends""" from typing import Optional from langchain_core.callbacks import CallbackManagerForToolRun from langchain_core.tools import BaseTool from langchain_community.utilities.google_jobs import GoogleJobsAPIWrapper class GoogleJobsQueryRun(BaseTool): # type: ignore[override] """Tool ...
# Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
import json import os from typing import Dict import torch from torch import Tensor, nn from sentence_transformers.util import fullname, import_from_string class Dense(nn.Module): """ Feed-forward function with activiation function. This layer takes a fixed-sized sentence embedding and passes it throu...
import torch from torch import Tensor from torch import nn from typing import Dict import os import json from ..util import fullname, import_from_string class Dense(nn.Module): """Feed-forward function with activiation function. This layer takes a fixed-sized sentence embedding and passes it through a feed-...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
from llama_index.core.base.llms.types import ( ChatMessage, ChatResponse, ChatResponseAsyncGen, ChatResponseGen, CompletionResponse, CompletionResponseAsyncGen, CompletionResponseGen, ImageBlock, LLMMetadata, MessageRole, TextBlock, AudioBlock, DocumentBlock, Cach...
from llama_index.core.base.llms.types import ( ChatMessage, ChatResponse, ChatResponseAsyncGen, ChatResponseGen, CompletionResponse, CompletionResponseAsyncGen, CompletionResponseGen, ImageBlock, LLMMetadata, MessageRole, TextBlock, AudioBlock, DocumentBlock, ) from l...
# Copyright 2019 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
# Copyright 2019 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
"""Conftest.""" from typing import List import pytest from llama_index.core.schema import Document @pytest.fixture() def documents() -> List[Document]: """Get documents.""" # NOTE: one document for now doc_text = ( "Hello world.\nThis is a test.\nThis is another test.\nThis is a test v2." ) ...
"""Conftest.""" from typing import List import pytest from llama_index.core.schema import Document @pytest.fixture() def documents() -> List[Document]: """Get documents.""" # NOTE: one document for now doc_text = ( "Hello world.\n" "This is a test.\n" "This is another test.\n" ...
"""Chroma Auto-retrieval Pack.""" from typing import Any, Dict, List, Optional from llama_index.core.indices.vector_store import VectorStoreIndex from llama_index.core.indices.vector_store.retrievers import ( VectorIndexAutoRetriever, ) from llama_index.core.llama_pack.base import BaseLlamaPack from llama_index.c...
"""Chroma Auto-retrieval Pack.""" from typing import Any, Dict, List, Optional from llama_index.core.indices.vector_store import VectorStoreIndex from llama_index.core.indices.vector_store.retrievers import ( VectorIndexAutoRetriever, ) from llama_index.core.llama_pack.base import BaseLlamaPack from llama_index....
# Copyright 2025 Stability AI, The HuggingFace Team and The InstantX Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-...
# Copyright 2024 Stability AI, The HuggingFace Team and The InstantX Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.callbacks.arize_callback import ArizeCallbackHandler # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handling op...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.callbacks.arize_callback import ArizeCallbackHandler # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handling op...
from pathlib import Path from typing import Any, Optional, TypedDict from tomlkit import load def get_package_root(cwd: Optional[Path] = None) -> Path: # traverse path for routes to host (any directory holding a pyproject.toml file) package_root = Path.cwd() if cwd is None else cwd visited: set[Path] = s...
from pathlib import Path from typing import Any, Dict, Optional, Set, TypedDict from tomlkit import load def get_package_root(cwd: Optional[Path] = None) -> Path: # traverse path for routes to host (any directory holding a pyproject.toml file) package_root = Path.cwd() if cwd is None else cwd visited: Se...
# dataset settings dataset_type = 'CityscapesDataset' data_root = 'data/cityscapes/' # Example to use different file client # Method 1: simply set the data root and let the file I/O module # automatically infer from prefix (not support LMDB and Memcache yet) # data_root = 's3://openmmlab/datasets/segmentation/citysca...
# dataset settings dataset_type = 'CityscapesDataset' data_root = 'data/cityscapes/' train_pipeline = [ dict(type='LoadImageFromFile'), dict(type='LoadAnnotations', with_bbox=True, with_mask=True), dict( type='RandomResize', scale=[(2048, 800), (2048, 1024)], keep_ratio=True), d...
""" This file contains deprecated code that can only be used with the old `model.fit`-style Sentence Transformers v2.X training. It exists for backwards compatibility with the `model.old_fit` method, but will be removed in a future version. Nowadays, with Sentence Transformers v3+, it is recommended to use the `Senten...
""" This file contains deprecated code that can only be used with the old `model.fit`-style Sentence Transformers v2.X training. It exists for backwards compatibility with the `model.old_fit` method, but will be removed in a future version. Nowadays, with Sentence Transformers v3+, it is recommended to use the `Senten...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.visualization.draw_bounding_boxes import ( draw_bounding_boxes as draw_bounding_boxes, ) from keras.src.visualization.draw_segmentation_masks import ( draw_segmentation_masks ...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.visualization.draw_bounding_boxes import draw_bounding_boxes from keras.src.visualization.draw_segmentation_masks import ( draw_segmentation_masks, ) from keras.src.visualization....
_base_ = './mask-rcnn_r50_fpn_gn-all_2x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://detectron/resnet101_gn')))
_base_ = './mask_rcnn_r50_fpn_gn-all_2x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://detectron/resnet101_gn')))
""" This examples trains a CrossEncoder for the STSbenchmark task. A CrossEncoder takes a sentence pair as input and outputs a label. Here, it output a continuous labels 0...1 to indicate the similarity between the input pair. It does NOT produce a sentence embedding and does NOT work for individual sentences. Usage:...
""" This examples trains a CrossEncoder for the STSbenchmark task. A CrossEncoder takes a sentence pair as input and outputs a label. Here, it output a continuous labels 0...1 to indicate the similarity between the input pair. It does NOT produce a sentence embedding and does NOT work for individual sentences. Usage:...
from typing import TYPE_CHECKING, Any, Generic, Type, TypeVar, Union import numpy as np from docarray.typing.tensor.abstract_tensor import AbstractTensor from docarray.typing.tensor.ndarray import NdArray from docarray.utils._internal.misc import is_tf_available, is_torch_available # noqa torch_available = is_torch...
from typing import TYPE_CHECKING, Any, Generic, Type, TypeVar, Union import numpy as np from docarray.typing.tensor.abstract_tensor import AbstractTensor from docarray.typing.tensor.ndarray import NdArray from docarray.utils._internal.misc import is_tf_available, is_torch_available # noqa torch_available = is_torch...
from typing import overload from urllib.parse import urlparse from backend.blocks.github._auth import ( GithubCredentials, GithubFineGrainedAPICredentials, ) from backend.util.request import URL, Requests @overload def _convert_to_api_url(url: str) -> str: ... @overload def _convert_to_api_url(url: URL) ->...
from urllib.parse import urlparse from backend.blocks.github._auth import ( GithubCredentials, GithubFineGrainedAPICredentials, ) from backend.util.request import Requests def _convert_to_api_url(url: str) -> str: """ Converts a standard GitHub URL to the corresponding GitHub API URL. Handles rep...
from typing import TYPE_CHECKING from ...utils import ( DIFFUSERS_SLOW_IMPORT, OptionalDependencyNotAvailable, _LazyModule, get_objects_from_module, is_torch_available, is_transformers_available, ) _dummy_objects = {} _import_structure = {} try: if not (is_transformers_available() and i...
from typing import TYPE_CHECKING from ...utils import ( DIFFUSERS_SLOW_IMPORT, OptionalDependencyNotAvailable, _LazyModule, get_objects_from_module, is_torch_available, is_transformers_available, ) _dummy_objects = {} _import_structure = {} try: if not (is_transformers_available() and i...
from pathlib import Path import click from rich.console import Console from rich.theme import Theme from .pkg import pkg from .test import test LLAMA_DEV_THEME = Theme( { "repr.path": "", "repr.filename": "", "repr.str": "", "traceback.note": "cyan", "info": "dim cyan", ...
from pathlib import Path import click from rich.console import Console from rich.theme import Theme from .pkg import pkg from .test import test LLAMA_DEV_THEME = Theme( { "repr.path": "", "repr.filename": "", "repr.str": "", "traceback.note": "cyan", "info": "dim cyan", ...
import random from typing import Optional, TYPE_CHECKING if TYPE_CHECKING: from docarray.array.document import DocumentArray class SampleMixin: """A mixin that provides search functionality to DocumentArrays""" def sample(self, k: int, seed: Optional[int] = None) -> 'DocumentArray': """random sa...
import random from typing import Optional, TYPE_CHECKING if TYPE_CHECKING: from ..document import DocumentArray class SampleMixin: """ A mixin that provides search functionality to DocumentArrays""" def sample(self, k: int, seed: Optional[int] = None) -> 'DocumentArray': """random sample k eleme...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../common/lsj_100e_coco_instance.py' ] image_size = (1024, 1024) batch_augments = [dict(type='BatchFixedSizePad', size=image_size)] norm_cfg = dict(type='SyncBN', requires_grad=True) # Use MMSyncBN that handles empty tensor in head. It can be changed to # Sy...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../common/lsj_100e_coco_instance.py' ] norm_cfg = dict(type='SyncBN', requires_grad=True) # Use MMSyncBN that handles empty tensor in head. It can be changed to # SyncBN after https://github.com/pytorch/pytorch/issues/36530 is fixed # Requires MMCV-full afte...
from backend.blocks.linear._api import LinearAPIException, LinearClient from backend.blocks.linear._auth import ( LINEAR_OAUTH_IS_CONFIGURED, TEST_CREDENTIALS_INPUT_OAUTH, TEST_CREDENTIALS_OAUTH, LinearCredentials, LinearCredentialsField, LinearCredentialsInput, LinearScope, ) from backend.b...
from backend.blocks.linear._api import LinearAPIException, LinearClient from backend.blocks.linear._auth import ( LINEAR_OAUTH_IS_CONFIGURED, TEST_CREDENTIALS_INPUT_OAUTH, TEST_CREDENTIALS_OAUTH, LinearCredentials, LinearCredentialsField, LinearCredentialsInput, LinearScope, ) from backend.b...
from __future__ import annotations import functools import operator from typing import Any, TYPE_CHECKING import torch # NOTE: other files rely on the imports below from torch._dynamo import callback as compilation_callback # noqa: F401 from torch._inductor.runtime.cache_dir_utils import ( # noqa: F401 cache_d...
from __future__ import annotations import functools import operator from typing import Any, TYPE_CHECKING import torch # NOTE: other files rely on the imports below from torch._dynamo import callback as compilation_callback # noqa: F401 from torch._inductor.runtime.cache_dir_utils import ( # noqa: F401 cache_d...
import os.path import numpy as np from whisper.audio import SAMPLE_RATE, load_audio, log_mel_spectrogram def test_audio(): audio_path = os.path.join(os.path.dirname(__file__), "jfk.flac") audio = load_audio(audio_path) assert audio.ndim == 1 assert SAMPLE_RATE * 10 < audio.shape[0] < SAMPLE_RATE * 1...
import os.path import numpy as np from whisper.audio import load_audio, log_mel_spectrogram, SAMPLE_RATE def test_audio(): audio_path = os.path.join(os.path.dirname(__file__), "jfk.flac") audio = load_audio(audio_path) assert audio.ndim == 1 assert SAMPLE_RATE * 10 < audio.shape[0] < SAMPLE_RATE * 1...
import tweepy from backend.blocks.twitter._auth import ( TEST_CREDENTIALS, TEST_CREDENTIALS_INPUT, TWITTER_OAUTH_IS_CONFIGURED, TwitterCredentials, TwitterCredentialsField, TwitterCredentialsInput, ) from backend.blocks.twitter.tweepy_exceptions import handle_tweepy_exception from backend.data....
import tweepy from backend.blocks.twitter._auth import ( TEST_CREDENTIALS, TEST_CREDENTIALS_INPUT, TWITTER_OAUTH_IS_CONFIGURED, TwitterCredentials, TwitterCredentialsField, TwitterCredentialsInput, ) from backend.blocks.twitter.tweepy_exceptions import handle_tweepy_exception from backend.data....
from docarray.document.any_document import AnyDocument from docarray.document.document import BaseDocument
from .any_document import AnyDocument from .document import BaseDocument
_base_ = './mask_rcnn_r50_caffe_fpn_syncbn-all_rpn-2conv_lsj_100e_coco.py' # Use RepeatDataset to speed up training # change repeat time from 4 (for 100 epochs) to 16 (for 400 epochs) train_dataloader = dict(dataset=dict(times=4 * 4)) param_scheduler = [ dict( type='LinearLR', start_factor=0.067, ...
_base_ = './mask_rcnn_r50_caffe_fpn_syncbn-all_rpn-2conv_lsj_100e_coco.py' # Use RepeatDataset to speed up training # change repeat time from 4 (for 100 epochs) to 16 (for 400 epochs) data = dict(train=dict(times=4 * 4)) lr_config = dict(warmup_iters=500 * 4)
from docarray.typing.tensor.image.image_ndarray import ImageNdArray from docarray.typing.tensor.image.image_tensor import ImageTensor __all__ = ['ImageNdArray', 'ImageTensor'] from docarray.utils._internal.misc import is_tf_available, is_torch_available torch_available = is_torch_available() if torch_available: ...
from docarray.typing.tensor.image.image_ndarray import ImageNdArray from docarray.typing.tensor.image.image_tensor import ImageTensor __all__ = ['ImageNdArray', 'ImageTensor'] from docarray.utils.misc import is_tf_available, is_torch_available torch_available = is_torch_available() if torch_available: from docar...
from typing import Optional from llama_index.core.storage.docstore.keyval_docstore import KVDocumentStore from llama_index.core.storage.docstore.types import DEFAULT_BATCH_SIZE from llama_index.storage.kvstore.firestore import FirestoreKVStore class FirestoreDocumentStore(KVDocumentStore): """ Firestore Docu...
from typing import Optional from llama_index.core.storage.docstore.keyval_docstore import KVDocumentStore from llama_index.core.storage.docstore.types import DEFAULT_BATCH_SIZE from llama_index.storage.kvstore.firestore import FirestoreKVStore class FirestoreDocumentStore(KVDocumentStore): """Firestore Document ...
from typing import TYPE_CHECKING from docarray.math.ndarray import get_array_type if TYPE_CHECKING: from docarray.typing import ArrayType import numpy as np def pdist( x_mat: 'ArrayType', metric: str, ) -> 'np.ndarray': """Computes Pairwise distances between observations in n-dimensional space. ...
from typing import TYPE_CHECKING from docarray.math.ndarray import get_array_type if TYPE_CHECKING: from docarray.typing import ArrayType import numpy as np def pdist( x_mat: 'ArrayType', metric: str, ) -> 'np.ndarray': """Computes Pairwise distances between observations in n-dimensional space. ...
from typing import Any, Dict, List, Tuple, Type, cast from docarray import BaseDoc, DocList from docarray.index.abstract import BaseDocIndex from docarray.utils.filter import filter_docs from docarray.utils.find import FindResult def _collect_query_args(method_name: str): # TODO: use partialmethod instead def i...
from typing import Any, Dict, List, Tuple, Type, cast from docarray import BaseDoc, DocList from docarray.index.abstract import BaseDocIndex from docarray.utils.filter import filter_docs from docarray.utils.find import FindResult def _collect_query_args(method_name: str): # TODO: use partialmethod instead def i...
""" This examples trains a CrossEncoder for the STSbenchmark task. A CrossEncoder takes a sentence pair as input and outputs a label. Here, it output a continuous labels 0...1 to indicate the similarity between the input pair. It does NOT produce a sentence embedding and does NOT work for individual sentences. Usage:...
""" This examples trains a CrossEncoder for the STSbenchmark task. A CrossEncoder takes a sentence pair as input and outputs a label. Here, it output a continuous labels 0...1 to indicate the similarity between the input pair. It does NOT produce a sentence embedding and does NOT work for individual sentences. Usage:...
from keras.src import backend from keras.src.api_export import keras_export from keras.src.layers.preprocessing.image_preprocessing.base_image_preprocessing_layer import ( # noqa: E501 BaseImagePreprocessingLayer, ) @keras_export("keras.layers.RandomGrayscale") class RandomGrayscale(BaseImagePreprocessingLayer):...
from keras.src import backend from keras.src.api_export import keras_export from keras.src.layers.preprocessing.image_preprocessing.base_image_preprocessing_layer import ( # noqa: E501 BaseImagePreprocessingLayer, ) @keras_export("keras.layers.RandomGrayscale") class RandomGrayscale(BaseImagePreprocessingLayer):...
_base_ = './retinanet_r50_fpn_1x_coco_v1.py' model = dict( data_preprocessor=dict( type='DetDataPreprocessor', # use caffe img_norm mean=[102.9801, 115.9465, 122.7717], std=[1.0, 1.0, 1.0], bgr_to_rgb=False, pad_size_divisor=32), backbone=dict( norm_cfg=di...
_base_ = './retinanet_r50_fpn_1x_coco_v1.py' model = dict( backbone=dict( norm_cfg=dict(requires_grad=False), norm_eval=True, style='caffe', init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://detectron/resnet50_caffe'))) # use caffe img_norm img_norm_c...
_base_ = [ '../common/mstrain_3x_coco_instance.py', '../_base_/models/cascade_mask_rcnn_r50_fpn.py' ] preprocess_cfg = dict( mean=[103.530, 116.280, 123.675], std=[1.0, 1.0, 1.0], to_rgb=False, pad_size_divisor=32) model = dict( # use caffe img_norm preprocess_cfg=preprocess_cfg, ba...
_base_ = ['./cascade_mask_rcnn_r50_fpn_mstrain_3x_coco.py'] model = dict( backbone=dict( norm_cfg=dict(requires_grad=False), norm_eval=True, style='caffe', init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://detectron2/resnet50_caffe'))) # use caffe im...
from typing import Dict, Optional, Union import pytest from docarray.typing import NdArray, TorchTensor from docarray.typing.tensor.abstract_tensor import AbstractTensor from docarray.utils._typing import is_tensor_union, is_type_tensor try: from docarray.typing import TensorFlowTensor except (ImportError, TypeE...
from typing import Dict, Optional, Union import pytest from docarray.typing import NdArray, TorchTensor from docarray.typing.tensor.abstract_tensor import AbstractTensor from docarray.utils._typing import is_tensor_union, is_type_tensor @pytest.mark.parametrize( 'type_, is_tensor', [ (int, False), ...
import pytest from backend.data import db from backend.executor import Scheduler from backend.server.model import CreateGraph from backend.usecases.sample import create_test_graph, create_test_user from backend.util.service import get_service_client from backend.util.test import SpinTestServer @pytest.mark.asyncio(l...
import pytest from backend.data import db from backend.executor import Scheduler from backend.server.model import CreateGraph from backend.usecases.sample import create_test_graph, create_test_user from backend.util.service import get_service_client from backend.util.test import SpinTestServer @pytest.mark.asyncio(s...
__version__ = "2.8.0.dev0" __MODEL_HUB_ORGANIZATION__ = "sentence-transformers" from .datasets import SentencesDataset, ParallelSentencesDataset from .LoggingHandler import LoggingHandler from .SentenceTransformer import SentenceTransformer from .readers import InputExample from .cross_encoder.CrossEncoder import Cross...
__version__ = "2.7.0.dev0" __MODEL_HUB_ORGANIZATION__ = "sentence-transformers" from .datasets import SentencesDataset, ParallelSentencesDataset from .LoggingHandler import LoggingHandler from .SentenceTransformer import SentenceTransformer from .readers import InputExample from .cross_encoder.CrossEncoder import Cross...
""" This tool allows agents to interact with the python-gitlab library and operate on a GitLab repository. To use this tool, you must first set as environment variables: GITLAB_PRIVATE_ACCESS_TOKEN GITLAB_REPOSITORY -> format: {owner}/{repo} """ from typing import Optional from langchain_core.callbacks impo...
""" This tool allows agents to interact with the python-gitlab library and operate on a GitLab repository. To use this tool, you must first set as environment variables: GITLAB_PRIVATE_ACCESS_TOKEN GITLAB_REPOSITORY -> format: {owner}/{repo} """ from typing import Optional from langchain_core.callbacks impo...
_base_ = '../cascade_rcnn/cascade-rcnn_r50_fpn_20e_coco.py' model = dict( backbone=dict( type='Res2Net', depth=101, scales=4, base_width=26, init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://res2net101_v1d_26w_4s')))
_base_ = '../cascade_rcnn/cascade_rcnn_r50_fpn_20e_coco.py' model = dict( backbone=dict( type='Res2Net', depth=101, scales=4, base_width=26, init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://res2net101_v1d_26w_4s')))
from langchain_core.language_models import BaseLanguageModel from langchain_core.output_parsers import StrOutputParser from langchain_core.prompts.few_shot import FewShotPromptTemplate from langchain_core.prompts.prompt import PromptTemplate TEST_GEN_TEMPLATE_SUFFIX = "Add another example." def generate_example( ...
from typing import List from langchain_core.language_models import BaseLanguageModel from langchain_core.output_parsers import StrOutputParser from langchain_core.prompts.few_shot import FewShotPromptTemplate from langchain_core.prompts.prompt import PromptTemplate TEST_GEN_TEMPLATE_SUFFIX = "Add another example." ...
# Copyright (c) OpenMMLab. All rights reserved. import time from typing import Any, Optional, Sequence, Tuple, Union from mmengine.data import BaseDataSample from mmengine.registry import HOOKS from .hook import Hook DATA_BATCH = Optional[Sequence[Tuple[Any, BaseDataSample]]] @HOOKS.register_module() class IterTime...
# Copyright (c) OpenMMLab. All rights reserved. import time from typing import Any, Optional, Sequence, Tuple, Union from mmengine.data import BaseDataSample from mmengine.registry import HOOKS from .hook import Hook DATA_BATCH = Optional[Sequence[Tuple[Any, BaseDataSample]]] @HOOKS.register_module() class IterTime...
# Copyright (c) OpenMMLab. All rights reserved. from .collect_env import collect_env from .compat_config import compat_cfg from .dist_utils import (DistOptimizerHook, all_reduce_dict, allreduce_grads, reduce_mean, sync_random_seed) from .logger import get_caller_name, get_root_logger, log_img_s...
# Copyright (c) OpenMMLab. All rights reserved. from .collect_env import collect_env from .compat_config import compat_cfg from .logger import get_caller_name, get_root_logger, log_img_scale from .memory import AvoidCUDAOOM, AvoidOOM from .misc import find_latest_checkpoint, update_data_root from .parallel import MMDat...
import os from typing import Optional import numpy as np import pytest import torch from pydantic.tools import parse_obj_as, schema_json_of from docarray import BaseDoc from docarray.base_doc.io.json import orjson_dumps from docarray.typing import ( AudioNdArray, NdArray, VideoBytes, VideoNdArray, ...
from typing import Optional import numpy as np import pytest import torch from pydantic.tools import parse_obj_as, schema_json_of from docarray import BaseDoc from docarray.base_doc.io.json import orjson_dumps from docarray.typing import ( AudioNdArray, NdArray, VideoBytes, VideoNdArray, VideoTorc...
# Copyright (c) OpenMMLab. All rights reserved. from argparse import ArgumentParser, Namespace from pathlib import Path from tempfile import TemporaryDirectory from mmengine.config import Config from mmengine.utils import mkdir_or_exist try: from model_archiver.model_packaging import package_model from model_...
# Copyright (c) OpenMMLab. All rights reserved. from argparse import ArgumentParser, Namespace from pathlib import Path from tempfile import TemporaryDirectory from mmengine.config import Config from mmengine.utils import mkdir_or_exist try: from model_archiver.model_packaging import package_model from model_...
"""Evaluation metrics for cluster analysis results. - Supervised evaluation uses a ground truth class values for each sample. - Unsupervised evaluation does not use ground truths and measures the "quality" of the model itself. """ # Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from ...
"""Evaluation metrics for cluster analysis results. - Supervised evaluation uses a ground truth class values for each sample. - Unsupervised evaluation does not use ground truths and measures the "quality" of the model itself. """ # Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from ...
# Copyright (c) OpenMMLab. All rights reserved. from .collect_env import collect_env from .compat_config import compat_cfg from .logger import get_caller_name, get_root_logger, log_img_scale from .misc import find_latest_checkpoint, update_data_root from .setup_env import setup_multi_processes from .split_batch import ...
# Copyright (c) OpenMMLab. All rights reserved. from .collect_env import collect_env from .logger import get_caller_name, get_root_logger, log_img_scale from .misc import find_latest_checkpoint, update_data_root from .setup_env import setup_multi_processes __all__ = [ 'get_root_logger', 'collect_env', 'find_latest...
from sentence_transformers.similarity_functions import SimilarityFunction __all__ = ["SimilarityFunction"]
from enum import Enum class SimilarityFunction(Enum): COSINE = 0 EUCLIDEAN = 1 MANHATTAN = 2 DOT_PRODUCT = 3
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
from typing import TYPE_CHECKING, Any, Type, TypeVar, Union, cast import numpy as np from docarray.typing.tensor.tensor import AnyTensor from docarray.typing.tensor.video.video_ndarray import VideoNdArray from docarray.typing.tensor.video.video_tensor_mixin import VideoTensorMixin from docarray.utils._internal.misc i...
from typing import TYPE_CHECKING, Any, Type, TypeVar, Union, cast import numpy as np from docarray.typing.tensor.tensor import AnyTensor from docarray.typing.tensor.video.video_ndarray import VideoNdArray from docarray.typing.tensor.video.video_tensor_mixin import VideoTensorMixin from docarray.utils._internal.misc i...
# Copyright (c) OpenMMLab. All rights reserved. import numpy as np import pytest import torch from mmdet.core import BboxOverlaps2D, bbox_overlaps def test_bbox_overlaps_2d(eps=1e-7): def _construct_bbox(num_bbox=None): img_h = int(np.random.randint(3, 1000)) img_w = int(np.random.randint(3, 100...
import numpy as np import pytest import torch from mmdet.core import BboxOverlaps2D, bbox_overlaps def test_bbox_overlaps_2d(eps=1e-7): def _construct_bbox(num_bbox=None): img_h = int(np.random.randint(3, 1000)) img_w = int(np.random.randint(3, 1000)) if num_bbox is None: num...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( type='YOLOF', backbone=dict( type='ResNet', depth=50, num_stages=4, out_indices=(3, ), frozen_stages=1, norm_cfg=dict(ty...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( type='YOLOF', backbone=dict( type='ResNet', depth=50, num_stages=4, out_indices=(3, ), frozen_stages=1, norm_cfg=dict(ty...