input
stringlengths
33
5k
output
stringlengths
32
5k
# Copyright (c) OpenMMLab. All rights reserved. from .evaluator import * # noqa: F401,F403 from .functional import * # noqa: F401,F403 from .metrics import * # noqa: F401,F403
# Copyright (c) OpenMMLab. All rights reserved. from .functional import * # noqa: F401,F403 from .metrics import * # noqa: F401,F403
import logging from functools import wraps from typing import Any, Callable from packaging import version MIN_ADS_VERSION = "2.12.9" logger = logging.getLogger(__name__) class UnsupportedOracleAdsVersionError(Exception): """ Custom exception for unsupported `oracle-ads` versions. Attributes: c...
import logging from functools import wraps from typing import Any, Callable from packaging import version MIN_ADS_VERSION = "2.12.9" logger = logging.getLogger(__name__) class UnsupportedOracleAdsVersionError(Exception): """Custom exception for unsupported `oracle-ads` versions. Attributes: curren...
import warnings from typing import Any, List, Union import PIL.Image import torch from torchvision.prototype import datapoints from torchvision.transforms import functional as _F from ._utils import is_simple_tensor @torch.jit.unused def to_grayscale(inpt: PIL.Image.Image, num_output_channels: int = 1) -> PIL.Imag...
import warnings from typing import Any, List, Union import PIL.Image import torch from torchvision.prototype import datapoints from torchvision.transforms import functional as _F @torch.jit.unused def to_grayscale(inpt: PIL.Image.Image, num_output_channels: int = 1) -> PIL.Image.Image: call = ", num_output_chan...
import numpy as np import orjson import pytest from pydantic.tools import parse_obj_as, schema_json_of from docarray.base_document.io.json import orjson_dumps from docarray.typing import NdArray from docarray.typing.tensor import NdArrayEmbedding def test_proto_tensor(): tensor = parse_obj_as(NdArray, np.zeros(...
import numpy as np import orjson import pytest from pydantic.tools import parse_obj_as, schema_json_of from docarray.base_document.io.json import orjson_dumps from docarray.typing import NdArray from docarray.typing.tensor import NdArrayEmbedding def test_proto_tensor(): tensor = parse_obj_as(NdArray, np.zeros(...
from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.audio.abstract_audio_tensor import AbstractAudioTensor from docarray.typing.tensor.torch_tensor import TorchTensor, metaTorchAndNode @_register_proto(proto_type_name='audio_torch_tensor') class AudioTorchTensor(AbstractAudioTensor,...
from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.audio.abstract_audio_tensor import AbstractAudioTensor from docarray.typing.tensor.torch_tensor import TorchTensor, metaTorchAndNode @_register_proto(proto_type_name='audio_torch_tensor') class AudioTorchTensor(AbstractAudioTensor,...
# Copyright (c) OpenMMLab. All rights reserved. from .augment_wrappers import AutoAugment, RandAugment from .colorspace import (AutoContrast, Brightness, Color, ColorTransform, Contrast, Equalize, Invert, Posterize, Sharpness, Solarize, SolarizeAdd) from .formatting imp...
# Copyright (c) OpenMMLab. All rights reserved. from .augment_wrappers import AutoAugment, RandAugment from .colorspace import (AutoContrast, Brightness, Color, ColorTransform, Contrast, Equalize, Invert, Posterize, Sharpness, Solarize, SolarizeAdd) from .formatting imp...
"""**Tools** are classes that an Agent uses to interact with the world. Each tool has a **description**. Agent uses the description to choose the right tool for the job. **Class hierarchy:** .. code-block:: RunnableSerializable --> BaseTool --> <name>Tool # Examples: AIPluginTool, BaseGraphQLTool ...
"""**Tools** are classes that an Agent uses to interact with the world. Each tool has a **description**. Agent uses the description to choose the right tool for the job. **Class hierarchy:** .. code-block:: RunnableSerializable --> BaseTool --> <name>Tool # Examples: AIPluginTool, BaseGraphQLTool ...
from keras.src import activations from keras.src.api_export import keras_export from keras.src.layers.layer import Layer @keras_export("keras.layers.Activation") class Activation(Layer): """Applies an activation function to an output. Args: activation: Activation function. It could be a callable, or ...
from keras.src import activations from keras.src.api_export import keras_export from keras.src.layers.layer import Layer @keras_export("keras.layers.Activation") class Activation(Layer): """Applies an activation function to an output. Args: activation: Activation function. It could be a callable, or ...
from pathlib import Path from typing import Dict, Tuple, Union import torchaudio from torch import Tensor from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import ( extract_archive, ) _URL = "https://datashare.ed.ac.uk/bitstream/handle/10283/3038/DR-VC...
from pathlib import Path from typing import Dict, Tuple, Union import torchaudio from torch import Tensor from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import ( extract_archive, ) _URL = "https://datashare.ed.ac.uk/bitstream/handle/10283/3038/DR-VC...
# Copyright (c) OpenMMLab. All rights reserved. from .inference import (async_inference_detector, inference_detector, init_detector, show_result_pyplot) from .test import multi_gpu_test, single_gpu_test from .train import get_root_logger, set_random_seed, train_detector __all__ = [ 'get_roo...
from .inference import (async_inference_detector, inference_detector, init_detector, show_result_pyplot) from .test import multi_gpu_test, single_gpu_test from .train import get_root_logger, set_random_seed, train_detector __all__ = [ 'get_root_logger', 'set_random_seed', 'train_detector', ...
__version__ = '0.36.1' import logging from docarray.array import DocList, DocVec from docarray.base_doc.doc import BaseDoc from docarray.utils._internal.misc import _get_path_from_docarray_root_level __all__ = ['BaseDoc', 'DocList', 'DocVec'] logger = logging.getLogger('docarray') handler = logging.StreamHandler()...
__version__ = '0.36.0' import logging from docarray.array import DocList, DocVec from docarray.base_doc.doc import BaseDoc from docarray.utils._internal.misc import _get_path_from_docarray_root_level __all__ = ['BaseDoc', 'DocList', 'DocVec'] logger = logging.getLogger('docarray') handler = logging.StreamHandler()...
# Copyright (c) OpenMMLab. All rights reserved. import numpy as np from mmengine.testing import assert_allclose from mmdet.structures.bbox import BaseBoxes, HorizontalBoxes from mmdet.structures.mask import BitmapMasks, PolygonMasks def create_random_bboxes(num_bboxes, img_w, img_h): bboxes_left_top = np.random....
# Copyright (c) OpenMMLab. All rights reserved. import numpy as np from mmengine.testing import assert_allclose from mmdet.structures.bbox import BaseBoxes, HorizontalBoxes from mmdet.structures.mask import BitmapMasks, PolygonMasks def create_random_bboxes(num_bboxes, img_w, img_h): bboxes_left_top = np.random....
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import numpy as np import pytest from jina import Document, DocumentArray from ...transformer_tf_text_encode import TransformerTFTextEncoder target_dim = 768 @pytest.fixture() def docs_generator(): return ...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import numpy as np import pytest from jina import Document, DocumentArray from jinahub.encoder.transformer_tf_text_encode import TransformerTFTextEncoder target_dim = 768 @pytest.fixture() def docs_generator()...
from typing import Union from docarray.typing.tensor.ndarray import NdArray try: import torch # noqa: F401 from docarray.typing.tensor.torch_tensor import TorchTensor # noqa: F401 is_torch_available = True except ImportError: is_torch_available = False try: import tensorflow as tf # type: ig...
from typing import Union from docarray.typing.tensor.ndarray import NdArray try: import torch # noqa: F401 except ImportError: AnyTensor = Union[NdArray] # type: ignore else: from docarray.typing.tensor.torch_tensor import TorchTensor # noqa: F401 AnyTensor = Union[NdArray, TorchTensor] # type: ...
import logging from datasets import load_dataset from sentence_transformers import SparseEncoder from sentence_transformers.sparse_encoder.evaluation import SparseEmbeddingSimilarityEvaluator logging.basicConfig(format="%(message)s", level=logging.INFO) # Load a model model = SparseEncoder("naver/splade-cocondenser...
import logging from datasets import load_dataset from sentence_transformers import SparseEncoder from sentence_transformers.sparse_encoder.evaluation import SparseEmbeddingSimilarityEvaluator logging.basicConfig(format="%(message)s", level=logging.INFO) # Load a model model = SparseEncoder("naver/splade-cocondenser...
# Copyright (c) OpenMMLab. All rights reserved. import copy from typing import Dict, List, Optional import numpy as np from mmcv.transforms import BaseTransform, Compose from mmcv.transforms.utils import cache_randomness from mmdet.registry import TRANSFORMS @TRANSFORMS.register_module() class MultiBranch(BaseTrans...
# Copyright (c) OpenMMLab. All rights reserved. import copy from typing import List, Optional from mmcv.transforms import BaseTransform, Compose from mmdet.registry import TRANSFORMS @TRANSFORMS.register_module() class MultiBranch(BaseTransform): r"""Multiple branch pipeline wrapper. Generate multiple data...
import unittest import torch from mmengine.config import Config from mmengine.structures import InstanceData from mmengine.testing import assert_allclose from mmdet.evaluation import INSTANCE_OFFSET from mmdet.models.seg_heads.panoptic_fusion_heads import HeuristicFusionHead class TestHeuristicFusionHead(unittest.T...
import unittest import torch from mmengine.config import Config from mmengine.data import InstanceData from mmengine.testing import assert_allclose from mmdet.evaluation import INSTANCE_OFFSET from mmdet.models.seg_heads.panoptic_fusion_heads import HeuristicFusionHead class TestHeuristicFusionHead(unittest.TestCas...
"""Integration test for DallE API Wrapper.""" from langchain_community.utilities.dalle_image_generator import DallEAPIWrapper def test_call() -> None: """Test that call returns a URL in the output.""" search = DallEAPIWrapper() output = search.run("volcano island") assert "https://oaidalleapi" in out...
"""Integration test for DallE API Wrapper.""" from langchain_community.utilities.dalle_image_generator import DallEAPIWrapper def test_call() -> None: """Test that call returns a URL in the output.""" search = DallEAPIWrapper() # type: ignore[call-arg] output = search.run("volcano island") assert "h...
"""Test cohere embeddings.""" from langchain_community.embeddings.cohere import CohereEmbeddings def test_cohere_embedding_documents() -> None: """Test cohere embeddings.""" documents = ["foo bar"] embedding = CohereEmbeddings() output = embedding.embed_documents(documents) assert len(output) == ...
"""Test cohere embeddings.""" from langchain_community.embeddings.cohere import CohereEmbeddings def test_cohere_embedding_documents() -> None: """Test cohere embeddings.""" documents = ["foo bar"] embedding = CohereEmbeddings() # type: ignore[call-arg] output = embedding.embed_documents(documents) ...
from typing import TYPE_CHECKING, Any, Dict, Type, TypeVar from docarray.base_document.abstract_document import AbstractDocument from docarray.base_document.base_node import BaseNode from docarray.typing.proto_register import _PROTO_TYPE_NAME_TO_CLASS if TYPE_CHECKING: from docarray.proto import DocumentProto, No...
from typing import TYPE_CHECKING, Any, Dict, Type, TypeVar from docarray.base_document.abstract_document import AbstractDocument from docarray.base_document.base_node import BaseNode from docarray.typing.proto_register import _PROTO_TYPE_NAME_TO_CLASS if TYPE_CHECKING: from docarray.proto import DocumentProto, No...
_base_ = [ '../_base_/models/rpn_r50-caffe-c4.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] val_evaluator = dict(metric='proposal_fast') test_evaluator = val_evaluator
_base_ = [ '../_base_/models/rpn_r50_caffe_c4.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] val_evaluator = dict(metric='proposal_fast') test_evaluator = val_evaluator
import os from typing import Any, Optional from llama_index.llms.openai_like import OpenAILike from llama_index.llms.deepseek.utils import get_context_window, FUNCTION_CALLING_MODELS class DeepSeek(OpenAILike): """ DeepSeek LLM. Examples: `pip install llama-index-llms-deepseek` ```pytho...
import os from typing import Any, Optional from llama_index.llms.openai_like import OpenAILike from llama_index.llms.deepseek.utils import get_context_window class DeepSeek(OpenAILike): """ DeepSeek LLM. Examples: `pip install llama-index-llms-deepseek` ```python from llama_inde...
# Licensed to the LF AI & Data foundation under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the "License"); # you may not use this fil...
import os import time import pytest cur_dir = os.path.dirname(os.path.abspath(__file__)) epsilla_yml = os.path.abspath(os.path.join(cur_dir, 'docker-compose.yml')) @pytest.fixture(scope='session', autouse=True) def start_storage(): os.system(f"docker compose -f {epsilla_yml} up -d --remove-orphans") time.sl...
"""Utilities to init Vertex AI.""" from importlib import metadata from typing import Optional from google.api_core.gapic_v1.client_info import ClientInfo def get_user_agent(module: Optional[str] = None) -> str: r""" Returns a custom user agent header. Args: module (Optional[str]): Op...
"""Utilities to init Vertex AI.""" from importlib import metadata from typing import Optional from google.api_core.gapic_v1.client_info import ClientInfo def get_user_agent(module: Optional[str] = None) -> str: r"""Returns a custom user agent header. Args: module (Optional[str]): Optiona...
"""Internal utilities for the in memory implementation of VectorStore. These are part of a private API, and users should not use them directly as they can change without notice. """ from __future__ import annotations import logging from typing import TYPE_CHECKING, Union if TYPE_CHECKING: import numpy as np ...
"""Internal utilities for the in memory implementation of VectorStore. These are part of a private API, and users should not use them directly as they can change without notice. """ from __future__ import annotations import logging from typing import TYPE_CHECKING, Union if TYPE_CHECKING: import numpy as np ...
import logging from datasets import load_dataset from sentence_transformers import SparseEncoder from sentence_transformers.sparse_encoder.evaluation import SparseBinaryClassificationEvaluator logging.basicConfig(format="%(message)s", level=logging.INFO) # Initialize the SPLADE model model = SparseEncoder("naver/sp...
import logging from datasets import load_dataset from sentence_transformers.sparse_encoder import ( SparseBinaryClassificationEvaluator, SparseEncoder, ) logging.basicConfig(format="%(message)s", level=logging.INFO) # Initialize the SPLADE model model = SparseEncoder("naver/splade-cocondenser-ensembledistil...
from langchain_community.utilities.openweathermap import OpenWeatherMapAPIWrapper def test_openweathermap_api_wrapper() -> None: """Test that OpenWeatherMapAPIWrapper returns correct data for London, GB.""" weather = OpenWeatherMapAPIWrapper() weather_data = weather.run("London,GB") assert weather_d...
from langchain_community.utilities.openweathermap import OpenWeatherMapAPIWrapper def test_openweathermap_api_wrapper() -> None: """Test that OpenWeatherMapAPIWrapper returns correct data for London, GB.""" weather = OpenWeatherMapAPIWrapper() # type: ignore[call-arg] weather_data = weather.run("London,...
"""Question-answering with sources over a vector database.""" import warnings from typing import Any from langchain_core.callbacks import ( AsyncCallbackManagerForChainRun, CallbackManagerForChainRun, ) from langchain_core.documents import Document from langchain_core.vectorstores import VectorStore from pyda...
"""Question-answering with sources over a vector database.""" import warnings from typing import Any from langchain_core.callbacks import ( AsyncCallbackManagerForChainRun, CallbackManagerForChainRun, ) from langchain_core.documents import Document from langchain_core.vectorstores import VectorStore from pyda...
from .base import OutlookEmailReader __all__ = ["OutlookEmailReader"]
from llama_index.readers.outlook_emails.base import OutlookEmailReader __all__ = ["OutlookEmailReader"]
import inspect import re from typing import Dict, List from huggingface_hub.utils import insecure_hashlib from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql imp...
import inspect import re from typing import Dict, List from huggingface_hub.utils import insecure_hashlib from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql imp...
_base_ = './cascade_mask_rcnn_convnext-t_p4_w7_fpn_giou_4conv1f_fp16_ms-crop_3x_coco.py' # noqa # please install mmcls>=1.0 # import mmcls.models to trigger register_module in mmcls custom_imports = dict(imports=['mmcls.models'], allow_failed_imports=False) checkpoint_file = 'https://download.openmmlab.com/mmclassifi...
_base_ = './cascade_mask_rcnn_convnext-t_p4_w7_fpn_giou_4conv1f_fp16_ms-crop_3x_coco.py' # noqa # please install mmcls>=0.22.0 # import mmcls.models to trigger register_module in mmcls custom_imports = dict(imports=['mmcls.models'], allow_failed_imports=False) checkpoint_file = 'https://download.openmmlab.com/mmclass...
"""Chat loaders.""" from abc import ABC, abstractmethod from collections.abc import Iterator from langchain_core.chat_sessions import ChatSession class BaseChatLoader(ABC): """Base class for chat loaders.""" @abstractmethod def lazy_load(self) -> Iterator[ChatSession]: """Lazy load the chat ses...
from abc import ABC, abstractmethod from collections.abc import Iterator from langchain_core.chat_sessions import ChatSession class BaseChatLoader(ABC): """Base class for chat loaders.""" @abstractmethod def lazy_load(self) -> Iterator[ChatSession]: """Lazy load the chat sessions. Retur...
from typing import List from llama_index.core.instrumentation.events.base import BaseEvent from llama_index.core.schema import QueryType, NodeWithScore class RetrievalStartEvent(BaseEvent): """ RetrievalStartEvent. Args: str_or_query_bundle (QueryType): Query bundle. """ str_or_query_bu...
from typing import List from llama_index.core.instrumentation.events.base import BaseEvent from llama_index.core.schema import QueryType, NodeWithScore class RetrievalStartEvent(BaseEvent): """RetrievalStartEvent. Args: str_or_query_bundle (QueryType): Query bundle. """ str_or_query_bundle: ...
# mypy: allow-untyped-defs import torch._C._lazy def reset(): """Resets all metric counters.""" torch._C._lazy._reset_metrics() def counter_names(): """Retrieves all the currently active counter names.""" return torch._C._lazy._counter_names() def counter_value(name: str): """Return the value ...
# mypy: allow-untyped-defs import torch._C._lazy def reset(): """Resets all metric counters.""" torch._C._lazy._reset_metrics() def counter_names(): """Retrieves all the currently active counter names.""" return torch._C._lazy._counter_names() def counter_value(name: str): """Return the value ...
import logging import os import zlib from contextlib import asynccontextmanager from urllib.parse import parse_qsl, urlencode, urlparse, urlunparse from uuid import uuid4 from dotenv import load_dotenv from prisma import Prisma from pydantic import BaseModel, Field, field_validator from backend.util.retry import conn...
import logging import os import zlib from contextlib import asynccontextmanager from urllib.parse import parse_qsl, urlencode, urlparse, urlunparse from uuid import uuid4 from dotenv import load_dotenv from prisma import Prisma from pydantic import BaseModel, Field, field_validator from backend.util.retry import conn...
# THIS FILE HAS BEEN AUTOGENERATED. To update: # 1. modify the `_deps` dict in setup.py # 2. run `make deps_table_update`` deps = { "Pillow": "Pillow>=10.0.1,<=15.0", "accelerate": "accelerate>=0.26.0", "av": "av", "beautifulsoup4": "beautifulsoup4", "blobfile": "blobfile", "codecarbon": "codeca...
# THIS FILE HAS BEEN AUTOGENERATED. To update: # 1. modify the `_deps` dict in setup.py # 2. run `make deps_table_update`` deps = { "Pillow": "Pillow>=10.0.1,<=15.0", "accelerate": "accelerate>=0.26.0", "av": "av", "beautifulsoup4": "beautifulsoup4", "blobfile": "blobfile", "codecarbon": "codeca...
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_vision_available from .base import GenericTensor, Pipeline, build_pipeline_init_args if is_vision_available(): from PIL import Image from ..image_utils import load_image @add_end_docstrings( build_pipeline_init_args(h...
from typing import Dict from ..utils import add_end_docstrings, is_vision_available from .base import GenericTensor, Pipeline, build_pipeline_init_args if is_vision_available(): from ..image_utils import load_image @add_end_docstrings( build_pipeline_init_args(has_image_processor=True), """ ima...
"""Gmail tools.""" from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import ( GmailCreateDraft, GmailGetMessage, GmailGetThread, GmailSearch, GmailSendMessage, ) # Create a way to dynamica...
"""Gmail tools.""" from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import ( GmailCreateDraft, GmailGetMessage, GmailGetThread, GmailSearch, GmailSendMessage, ) # Create a way to dynamica...
import importlib import pytest from dirty_equals import IsDict from fastapi.testclient import TestClient from ...utils import needs_py310 @pytest.fixture( name="client", params=[ "tutorial001", pytest.param("tutorial001_py310", marks=needs_py310), "tutorial001_an", pytest.par...
import pytest from dirty_equals import IsDict from fastapi.testclient import TestClient from docs_src.header_params.tutorial001 import app client = TestClient(app) @pytest.mark.parametrize( "path,headers,expected_status,expected_response", [ ("/items", None, 200, {"User-Agent": "testclient"}), ...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
import os from pathlib import Path import pytest from jina import Flow from jina.excepts import RuntimeFailToStart from jina.orchestrate.deployments import Deployment from jina.parsers import set_deployment_parser from jina.serve.executors import BaseExecutor cur_dir = os.path.dirname(os.path.abspath(__file__)) @py...
import os from pathlib import Path import pytest from jina import Flow from jina.excepts import RuntimeFailToStart from jina.orchestrate.deployments import Deployment from jina.parsers import set_deployment_parser from jina.serve.executors import BaseExecutor cur_dir = os.path.dirname(os.path.abspath(__file__)) de...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.activations import deserialize from keras.src.activations import get from keras.src.activations import serialize from keras.src.activations.activations import celu from keras.src.acti...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.activations import deserialize from keras.src.activations import get from keras.src.activations import serialize from keras.src.activations.activations import celu from keras.src.acti...
from keras.src import backend from keras.src import ops class DropoutRNNCell: """Object that holds dropout-related functionality for RNN cells. This class is not a standalone RNN cell. It suppose to be used with a RNN cell by multiple inheritance. Any cell that mix with class should have following fi...
from keras.src import backend from keras.src import ops class DropoutRNNCell: """Object that holds dropout-related functionality for RNN cells. This class is not a standalone RNN cell. It suppose to be used with a RNN cell by multiple inheritance. Any cell that mix with class should have following fi...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/coco_instance.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] train_pipeline = [ dict(type='LoadImageFromFile'), dict( type='LoadAnnotations', with_bbox=True, with_mask=True, ...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/coco_instance.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] img_norm_cfg = dict( mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True) train_pipeline = [ dict(type='LoadImageFro...
from argparse import ArgumentParser from pathlib import Path import mir_eval import torch from lightning_train import _get_dataloader, _get_model, sisdri_metric def _eval(model, data_loader, device): results = torch.zeros(4) with torch.no_grad(): for _, batch in enumerate(data_loader): mi...
from argparse import ArgumentParser from pathlib import Path import mir_eval import torch from lightning_train import _get_model, _get_dataloader, sisdri_metric def _eval(model, data_loader, device): results = torch.zeros(4) with torch.no_grad(): for _, batch in enumerate(data_loader): mi...
from keras.src import activations from keras.src.api_export import keras_export from keras.src.layers.layer import Layer @keras_export("keras.layers.ReLU") class ReLU(Layer): """Rectified Linear Unit activation function layer. Formula: ``` python f(x) = max(x,0) f(x) = max_value if x >= max_value...
from keras.src import activations from keras.src.api_export import keras_export from keras.src.layers.layer import Layer @keras_export("keras.layers.ReLU") class ReLU(Layer): """Rectified Linear Unit activation function layer. Formula: ``` python f(x) = max(x,0) f(x) = max_value if x >= max_value...
# Copyright (c) OpenMMLab. All rights reserved. from .checkloss_hook import CheckInvalidLossHook from .ema import ExpMomentumEMAHook, LinearMomentumEMAHook from .set_epoch_info_hook import SetEpochInfoHook from .sync_norm_hook import SyncNormHook from .sync_random_size_hook import SyncRandomSizeHook from .yolox_lrupdat...
# Copyright (c) OpenMMLab. All rights reserved. from .checkloss_hook import CheckInvalidLossHook from .ema import ExpMomentumEMAHook, LinearMomentumEMAHook from .sync_norm_hook import SyncNormHook from .sync_random_size_hook import SyncRandomSizeHook from .yolox_lrupdater_hook import YOLOXLrUpdaterHook from .yolox_mode...
__version__ = '0.13.27' import os from .document import Document from .array import DocumentArray from .dataclasses import dataclass, field if 'DA_RICH_HANDLER' in os.environ: from rich.traceback import install install()
__version__ = '0.13.26' import os from .document import Document from .array import DocumentArray from .dataclasses import dataclass, field if 'DA_RICH_HANDLER' in os.environ: from rich.traceback import install install()
_base_ = './retinanet_r50_fpn_ghm-1x_coco.py' model = dict( backbone=dict( type='ResNeXt', depth=101, groups=32, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), style='pytorch...
_base_ = './retinanet_ghm_r50_fpn_1x_coco.py' model = dict( backbone=dict( type='ResNeXt', depth=101, groups=32, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), style='pytorch...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.ops.image import affine_transform from keras.src.ops.image import crop_images from keras.src.ops.image import elastic_transform from keras.src.ops.image import extract_patches from ke...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.ops.image import affine_transform from keras.src.ops.image import crop_images from keras.src.ops.image import extract_patches from keras.src.ops.image import gaussian_blur from keras....
_base_ = './retinanet_r50_fpn_ghm-1x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
_base_ = './retinanet_ghm_r50_fpn_1x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
from __future__ import annotations from typing import Any, Literal, Optional, Union from exa_py import Exa # type: ignore[untyped-import] from exa_py.api import ( HighlightsContentsOptions, # type: ignore[untyped-import] TextContentsOptions, # type: ignore[untyped-import] ) from langchain_core.callbacks im...
from typing import Any, Literal, Optional, Union from exa_py import Exa # type: ignore[untyped-import] from exa_py.api import ( HighlightsContentsOptions, # type: ignore[untyped-import] TextContentsOptions, # type: ignore[untyped-import] ) from langchain_core.callbacks import CallbackManagerForRetrieverRun ...
from typing import Optional import os from llama_index.core.tools.tool_spec.base import BaseToolSpec from llama_index.tools.agentql.const import ( DEFAULT_API_TIMEOUT_SECONDS, DEFAULT_IS_STEALTH_MODE_ENABLED, DEFAULT_WAIT_FOR_PAGE_LOAD_SECONDS, DEFAULT_IS_SCROLL_TO_BOTTOM_ENABLED, DEFAULT_RESPONSE...
from typing import Optional import os from llama_index.core.tools.tool_spec.base import BaseToolSpec from llama_index.tools.agentql.const import ( DEFAULT_API_TIMEOUT_SECONDS, DEFAULT_IS_STEALTH_MODE_ENABLED, DEFAULT_WAIT_FOR_PAGE_LOAD_SECONDS, DEFAULT_IS_SCROLL_TO_BOTTOM_ENABLED, DEFAULT_RESPONSE...
from jina import DocumentArray, Executor, Flow, requests def test_gateway_metric_labels(monkeypatch_metric_exporter): collect_metrics, read_metrics = monkeypatch_metric_exporter class FirstExec(Executor): @requests() def meow(self, docs, **kwargs): return DocumentArray.empty(3) ...
from jina import Executor, Flow, requests, DocumentArray def test_gateway_metric_labels(monkeypatch_metric_exporter): collect_metrics, read_metrics = monkeypatch_metric_exporter class FirstExec(Executor): @requests() def meow(self, docs, **kwargs): return DocumentArray.empty(3) ...
_base_ = './mask_rcnn_hrnetv2p_w18_1x_coco.py' # learning policy max_epochs = 24 train_cfg = dict(max_epochs=max_epochs) param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_epochs, b...
_base_ = './mask_rcnn_hrnetv2p_w18_1x_coco.py' # learning policy lr_config = dict(step=[16, 22]) runner = dict(type='EpochBasedRunner', max_epochs=24)
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import os import numpy as np import pytest from jina import Document, DocumentArray, Flow from jina.executors.metas import get_default_metas from jina_commons.indexers.dump import import_vectors from .. import Hnswl...
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import os import numpy as np import pytest from jina import Document, DocumentArray, Flow from jina.executors.metas import get_default_metas from jina_commons.indexers.dump import import_vectors from .. import Hnswl...
from __future__ import annotations from sentence_transformers.training_args import SentenceTransformerTrainingArguments class CrossEncoderTrainingArguments(SentenceTransformerTrainingArguments): r""" CrossEncoderTrainingArguments extends :class:`~transformers.TrainingArguments` with additional arguments ...
from __future__ import annotations from sentence_transformers.training_args import SentenceTransformerTrainingArguments class CrossEncoderTrainingArguments(SentenceTransformerTrainingArguments): r""" CrossEncoderTrainingArguments extends :class:`~transformers.TrainingArguments` with additional arguments ...
#!/usr/bin/env python3 """Run smoke tests""" import argparse import logging def base_smoke_test(): import torchaudio # noqa: F401 import torchaudio.compliance.kaldi # noqa: F401 import torchaudio.datasets # noqa: F401 import torchaudio.functional # noqa: F401 import torchaudio.models # noqa: ...
"""Run smoke tests""" import argparse import logging def base_smoke_test(): import torchaudio # noqa: F401 import torchaudio.compliance.kaldi # noqa: F401 import torchaudio.datasets # noqa: F401 import torchaudio.functional # noqa: F401 import torchaudio.models # noqa: F401 import torchau...
from __future__ import annotations __version__ = "4.2.0.dev0" __MODEL_HUB_ORGANIZATION__ = "sentence-transformers" import importlib import os import warnings from sentence_transformers.backend import ( export_dynamic_quantized_onnx_model, export_optimized_onnx_model, export_static_quantized_openvino_mode...
from __future__ import annotations __version__ = "4.2.0.dev0" __MODEL_HUB_ORGANIZATION__ = "sentence-transformers" import importlib import os import warnings from sentence_transformers.backend import ( export_dynamic_quantized_onnx_model, export_optimized_onnx_model, export_static_quantized_openvino_mode...
from __future__ import annotations from collections.abc import Iterable from torch import Tensor from sentence_transformers.losses.TripletLoss import TripletDistanceMetric, TripletLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseTripletLoss(TripletLoss): def __init_...
from __future__ import annotations from sentence_transformers.losses.TripletLoss import TripletDistanceMetric, TripletLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseTripletLoss(TripletLoss): def __init__( self, model: SparseEncoder, distance_metric=TripletDi...
_base_ = '../faster_rcnn/faster-rcnn_r101_fpn_1x_coco.py' model = dict( backbone=dict( dcn=dict(type='DCN', deform_groups=1, fallback_on_stride=False), stage_with_dcn=(False, True, True, True)))
_base_ = '../faster_rcnn/faster_rcnn_r101_fpn_1x_coco.py' model = dict( backbone=dict( dcn=dict(type='DCN', deform_groups=1, fallback_on_stride=False), stage_with_dcn=(False, True, True, True)))
_base_ = ['faster_rcnn_r50_fpn_32x2_1x_openimages.py'] model = dict( roi_head=dict(bbox_head=dict(num_classes=500)), test_cfg=dict(rcnn=dict(score_thr=0.01))) # dataset settings dataset_type = 'OpenImagesChallengeDataset' data_root = 'data/OpenImages/' data = dict( train=dict( type=dataset_type, ...
_base_ = ['faster_rcnn_r50_fpn_32x2_1x_openimages.py'] model = dict( roi_head=dict(bbox_head=dict(num_classes=500)), test_cfg=dict(rcnn=dict(score_thr=0.01))) # dataset settings dataset_type = 'OpenImagesChallengeDataset' data_root = 'data/OpenImages/' data = dict( train=dict( type=dataset_type, ...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from parameterized import parameterized from mmdet.registry import MODELS from mmdet.testing import demo_mm_inputs, demo_mm_proposals, get_roi_head_cfg from mmdet.utils import register_all_modules class TestDy...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from parameterized import parameterized from mmdet.registry import MODELS from mmdet.testing import demo_mm_inputs, demo_mm_proposals, get_roi_head_cfg from mmdet.utils import register_all_modules class TestDy...
"""Fake LLM wrapper for testing purposes.""" from collections.abc import Mapping from typing import Any, Optional, cast from langchain_core.callbacks.manager import CallbackManagerForLLMRun from langchain_core.language_models.llms import LLM from pydantic import model_validator class FakeLLM(LLM): """Fake LLM w...
"""Fake LLM wrapper for testing purposes.""" from collections.abc import Mapping from typing import Any, Optional, cast from langchain_core.callbacks.manager import CallbackManagerForLLMRun from langchain_core.language_models.llms import LLM from pydantic import model_validator class FakeLLM(LLM): """Fake LLM w...
# flake8: noqa """Tools for working with JSON specs.""" from __future__ import annotations import json import re from pathlib import Path from typing import Dict, List, Optional, Union from pydantic import BaseModel from langchain_core.callbacks import ( AsyncCallbackManagerForToolRun, CallbackManagerForToo...
# flake8: noqa """Tools for working with JSON specs.""" from __future__ import annotations import json import re from pathlib import Path from typing import Dict, List, Optional, Union from pydantic import BaseModel from langchain_core.callbacks import ( AsyncCallbackManagerForToolRun, CallbackManagerForToo...
"""Tools for interacting with vectorstores.""" import json from typing import Any, Dict, Optional from langchain_core.callbacks import ( AsyncCallbackManagerForToolRun, CallbackManagerForToolRun, ) from langchain_core.language_models import BaseLanguageModel from langchain_core.tools import BaseTool from lang...
"""Tools for interacting with vectorstores.""" import json from typing import Any, Dict, Optional from langchain_core.callbacks import ( AsyncCallbackManagerForToolRun, CallbackManagerForToolRun, ) from langchain_core.language_models import BaseLanguageModel from langchain_core.tools import BaseTool from lang...
from typing import Any, Optional, Type, TypeVar, Union import numpy as np from docarray.base_doc import BaseDoc from docarray.documents import AudioDoc from docarray.typing import AnyEmbedding, AnyTensor from docarray.typing.tensor.abstract_tensor import AbstractTensor from docarray.typing.tensor.video.video_tensor i...
from typing import Any, Optional, Type, TypeVar, Union import numpy as np from docarray.base_doc import BaseDoc from docarray.documents import AudioDoc from docarray.typing import AnyEmbedding, AnyTensor from docarray.typing.tensor.abstract_tensor import AbstractTensor from docarray.typing.tensor.video.video_tensor i...
from typing import Any, Union from torchvision import tv_tensors from torchvision.transforms.v2 import functional as F, Transform class ConvertBoundingBoxFormat(Transform): """Convert bounding box coordinates to the given ``format``, eg from "CXCYWH" to "XYXY". Args: format (str or tv_tensors.Boundi...
from typing import Any, Union from torchvision import tv_tensors from torchvision.transforms.v2 import functional as F, Transform class ConvertBoundingBoxFormat(Transform): """Convert bounding box coordinates to the given ``format``, eg from "CXCYWH" to "XYXY". Args: format (str or tv_tensors.Boundi...
"""**Prompt** is the input to the model. Prompt is often constructed from multiple components and prompt values. Prompt classes and functions make constructing and working with prompts easy. **Class hierarchy:** .. code-block:: BasePromptTemplate --> PipelinePromptTemplate StringProm...
"""**Prompt** is the input to the model. Prompt is often constructed from multiple components and prompt values. Prompt classes and functions make constructing and working with prompts easy. **Class hierarchy:** .. code-block:: BasePromptTemplate --> PipelinePromptTemplate StringProm...
import copy import clip import numpy as np import pytest import torch from jina import Document, DocumentArray from ...clip_text import CLIPTextEncoder @pytest.fixture(scope="module") def encoder() -> CLIPTextEncoder: return CLIPTextEncoder() def test_no_documents(encoder: CLIPTextEncoder): docs = Document...
import copy import clip import numpy as np import pytest import torch from jina import Document, DocumentArray from ...clip_text import CLIPTextEncoder @pytest.fixture(scope="module") def encoder() -> CLIPTextEncoder: return CLIPTextEncoder() def test_no_documents(encoder: CLIPTextEncoder): ...
import os import numpy as np import pytest from jina import Document, DocumentArray from .. import NumpySearcher TOP_K = 5 cur_dir = os.path.dirname(os.path.abspath(__file__)) def test_query_vector(tmpdir): runtime = { 'workspace': str(tmpdir), 'name': 'searcher', 'pea_id': 0, '...
import numpy as np from jina import Document, DocumentArray from .. import NumpySearcher def test_query_vector(tmpdir): runtime = { 'workspace': str(tmpdir), 'name': 'searcher', 'pea_id': 0, 'replica_id': 0, } indexer = NumpySearcher(dump_path='tests/dump1', runtime_args=r...
"""Test NLPCloud API wrapper.""" from pathlib import Path from typing import cast from pydantic import SecretStr from pytest import CaptureFixture, MonkeyPatch from langchain_community.llms.loading import load_llm from langchain_community.llms.nlpcloud import NLPCloud from tests.integration_tests.llms.utils import a...
"""Test NLPCloud API wrapper.""" from pathlib import Path from typing import cast from pydantic import SecretStr from pytest import CaptureFixture, MonkeyPatch from langchain_community.llms.loading import load_llm from langchain_community.llms.nlpcloud import NLPCloud from tests.integration_tests.llms.utils import a...
"""monday.com reader.""" from typing import Dict, List import requests from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class MondayReader(BaseReader): """ monday.com reader. Reads board's data by a GraphQL query. Args: api_key (str): monday.com A...
"""monday.com reader.""" from typing import Dict, List import requests from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class MondayReader(BaseReader): """monday.com reader. Reads board's data by a GraphQL query. Args: api_key (str): monday.com API ke...
_base_ = [ '../common/ms-poly_3x_coco-instance.py', '../_base_/models/mask-rcnn_r50_fpn.py' ] model = dict( backbone=dict( _delete_=True, type='RegNet', arch='regnetx_800mf', out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=Tr...
_base_ = [ '../common/mstrain-poly_3x_coco_instance.py', '../_base_/models/mask_rcnn_r50_fpn.py' ] model = dict( backbone=dict( _delete_=True, type='RegNet', arch='regnetx_800mf', out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_gr...
import os import re from pathlib import Path from typing import Tuple, Union, Optional import torch import torchaudio from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import extract_archive URL = "https://speech.fit.vutbr.cz/files/quesst14Database.tgz" _C...
import os import re from pathlib import Path from typing import Tuple, Union, Optional import torch import torchaudio from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import extract_archive URL = "https://speech.fit.vutbr.cz/files/quesst14Database.tgz" _C...
import copy from dataclasses import dataclass from pathlib import Path from typing import Dict, Optional, Union @dataclass class DownloadConfig: """Configuration for our cached path manager. Attributes: cache_dir (`str` or `Path`, *optional*): Specify a cache directory to save the file to...
import copy from dataclasses import dataclass from pathlib import Path from typing import Dict, Optional, Union @dataclass class DownloadConfig: """Configuration for our cached path manager. Attributes: cache_dir (`str` or `Path`, *optional*): Specify a cache directory to save the file to...
"""Argparser module for WorkerRuntime""" from jina import __default_host__, helper from jina.enums import PollingType from jina.parsers.helper import KVAppendAction, add_arg_group from jina.parsers.orchestrate.runtimes.runtime import mixin_base_runtime_parser def mixin_worker_runtime_parser(parser): """Mixing in ...
"""Argparser module for WorkerRuntime""" from jina import __default_host__, helper from jina.parsers.helper import KVAppendAction, add_arg_group from jina.parsers.orchestrate.runtimes.runtime import mixin_base_runtime_parser def mixin_worker_runtime_parser(parser): """Mixing in arguments required by :class:`Worke...
import os import shutil from pathlib import Path from typing import Tuple import numpy as np import pytest from big_transfer import BigTransferEncoder from jina import Document, DocumentArray, Executor from PIL import Image directory = os.path.dirname(os.path.realpath(__file__)) _INPUT_DIM = 512 _EMBEDDING_DIM = 20...
import os import shutil from pathlib import Path import numpy as np import PIL.Image as Image import pytest from big_transfer import BigTransferEncoder from jina import Document, DocumentArray, Executor directory = os.path.dirname(os.path.realpath(__file__)) def test_config(): ex = Executor.load_config(str(Path...
_base_ = [ '../_base_/models/faster_rcnn_r50_fpn.py', '../_base_/datasets/voc0712.py', '../_base_/default_runtime.py' ] model = dict(roi_head=dict(bbox_head=dict(num_classes=20))) METAINFO = { 'CLASSES': ('aeroplane', 'bicycle', 'bird', 'boat', 'bottle', 'bus', 'car', 'cat', 'chair', 'cow', 'dinin...
_base_ = [ '../_base_/models/faster_rcnn_r50_fpn.py', '../_base_/datasets/voc0712.py', '../_base_/default_runtime.py' ] model = dict(roi_head=dict(bbox_head=dict(num_classes=20))) CLASSES = ('aeroplane', 'bicycle', 'bird', 'boat', 'bottle', 'bus', 'car', 'cat', 'chair', 'cow', 'diningtable', 'dog', ...
import textwrap import pyarrow as pa import pytest from datasets import Features, Value from datasets.packaged_modules.json.json import Json @pytest.fixture def jsonl_file(tmp_path): filename = tmp_path / "file.jsonl" data = textwrap.dedent( """\ {"col_1": -1} {"col_1": 1, "col_2": 2...
import textwrap import pyarrow as pa import pytest from datasets import Features, Value from datasets.packaged_modules.json.json import Json @pytest.fixture def jsonl_file(tmp_path): filename = tmp_path / "file.jsonl" data = textwrap.dedent( """\ {"col_1": -1} {"col_1": 1, "col_2": 2...
""" This script contains an example how to perform semantic search with Seismic. For more information, please refer to the documentation: https://github.com/TusKANNy/seismic/blob/main/docs/Guidelines.md All you need is installing the `pyseismic-lsr` package: ``` pip install pyseismic-lsr ``` """ import time from dat...
""" This script contains an example how to perform semantic search with Seismic. For more information, please refer to the documentation: https://github.com/TusKANNy/seismic/blob/main/docs/Guidelines.md All you need is installing the `pyseismic-lsr` package: ``` pip install pyseismic-lsr ``` """ import time from dat...
import os import re from pathlib import Path from typing import Optional, Tuple, Union import torch import torchaudio from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import extract_archive URL = "https://speech.fit.vutbr.cz/files/quesst14Database.tgz" _C...
import os import re from pathlib import Path from typing import Optional, Tuple, Union import torch import torchaudio from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import extract_archive URL = "https://speech.fit.vutbr.cz/files/quesst14Database.tgz" _C...
import random from collections import defaultdict from typing import Dict, Any, TYPE_CHECKING, Generator, List import numpy as np from docarray.helper import dunder_get if TYPE_CHECKING: from docarray import DocumentArray class GroupMixin: """These helpers yield groups of :class:`DocumentArray` from a ...
import random from collections import defaultdict from typing import Dict, Any, TYPE_CHECKING, Generator, List import numpy as np from docarray.helper import dunder_get if TYPE_CHECKING: from docarray import DocumentArray class GroupMixin: """These helpers yield groups of :class:`DocumentArray` from a ...
# Copyright (c) OpenMMLab. All rights reserved. from .misc import (check_prerequisites, concat_list, deprecated_api_warning, has_method, import_modules_from_strings, is_list_of, is_method_overridden, is_seq_of, is_str, is_tuple_of, iter_cast, list_cast, requires_...
# Copyright (c) OpenMMLab. All rights reserved. from .fileio import (FileClient, dict_from_file, dump, list_from_file, load, register_handler) from .misc import (check_prerequisites, concat_list, deprecated_api_warning, has_method, import_modules_from_strings, is_list_of, ...
# Copyright 2022 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writ...
# Copyright 2022 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writ...
import copy from dataclasses import dataclass, field from pathlib import Path from typing import Any, Dict, Optional, Union from .. import config @dataclass class DownloadConfig: """Configuration for our cached path manager. Attributes: cache_dir (`str` or `Path`, *optional*): Specify a ...
import copy from dataclasses import dataclass, field from pathlib import Path from typing import Any, Dict, Optional, Union from .. import config @dataclass class DownloadConfig: """Configuration for our cached path manager. Attributes: cache_dir (`str` or `Path`, *optional*): Specify a ...
_base_ = './cascade-mask-rcnn_s50_fpn_syncbn-backbone+head_ms-1x_coco.py' model = dict( backbone=dict( stem_channels=128, depth=101, init_cfg=dict(type='Pretrained', checkpoint='open-mmlab://resnest101')))
_base_ = './cascade_mask_rcnn_s50_fpn_syncbn-backbone+head_mstrain_1x_coco.py' model = dict( backbone=dict( stem_channels=128, depth=101, init_cfg=dict(type='Pretrained', checkpoint='open-mmlab://resnest101')))
"""Criteria or rubric based evaluators. These evaluators are useful for evaluating the output of a language model or chain against specified criteria or rubric. Classes ------- CriteriaEvalChain : Evaluates the output of a language model or chain against specified criteria. Examples -------- Using a predefined crite...
"""Criteria or rubric based evaluators. These evaluators are useful for evaluating the output of a language model or chain against specified criteria or rubric. Classes ------- CriteriaEvalChain : Evaluates the output of a language model or chain against specified criteria. Examples -------- Using a predefined crite...
from setuptools import find_packages, setup with open("README.md", mode="r", encoding="utf-8") as readme_file: readme = readme_file.read() setup( name="sentence-transformers", version="3.1.0.dev0", author="Nils Reimers, Tom Aarsen", author_email="info@nils-reimers.de", description="Multilingu...
from setuptools import find_packages, setup with open("README.md", mode="r", encoding="utf-8") as readme_file: readme = readme_file.read() setup( name="sentence-transformers", version="3.0.0.dev0", author="Nils Reimers", author_email="info@nils-reimers.de", description="Multilingual text embe...
from langchain_core.prompts.loading import ( _load_examples, _load_few_shot_prompt, _load_output_parser, _load_prompt, _load_prompt_from_file, _load_template, load_prompt, load_prompt_from_config, ) from langchain_core.utils.loading import try_load_from_hub __all__ = [ "_load_exampl...
from langchain_core.prompts.loading import ( _load_examples, _load_few_shot_prompt, _load_output_parser, _load_prompt, _load_prompt_from_file, _load_template, load_prompt, load_prompt_from_config, ) from langchain_core.utils.loading import try_load_from_hub __all__ = [ "load_prompt_...
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
_base_ = [ '../_base_/models/retinanet_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/default_runtime.py' ] cudnn_benchmark = True norm_cfg = dict(type='BN', requires_grad=True) checkpoint = 'https://download.openmmlab.com/mmclassification/v0/efficientnet/efficientnet-b3_3rdparty_8xb32-aa_in1k...
_base_ = [ '../_base_/models/retinanet_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/default_runtime.py' ] cudnn_benchmark = True norm_cfg = dict(type='BN', requires_grad=True) checkpoint = 'https://download.openmmlab.com/mmclassification/v0/efficientnet/efficientnet-b3_3rdparty_8xb32-aa_in1k...
import json import os import pytest from jina import __version__ from jina.hubble import HubExecutor from jina.hubble.hubio import HubIO from jina.orchestrate.deployments.config.helper import ( get_base_executor_version, get_image_name, to_compatible_name, ) @pytest.mark.parametrize('is_master', (True, ...
import os import pytest from jina import __version__ from jina.hubble import HubExecutor from jina.hubble.hubio import HubIO from jina.orchestrate.deployments.config.helper import ( get_base_executor_version, get_image_name, to_compatible_name, ) @pytest.mark.parametrize('is_master', (True, False)) def ...
"""Tools for interacting with an Apache Cassandra database.""" from typing import List from llama_index.core.bridge.pydantic import Field from llama_index.core.schema import Document from llama_index.core.tools.tool_spec.base import BaseToolSpec from llama_index.tools.cassandra.cassandra_database_wrapper import ( ...
"""Tools for interacting with an Apache Cassandra database.""" from typing import List from llama_index.core.bridge.pydantic import Field from llama_index.core.schema import Document from llama_index.core.tools.tool_spec.base import BaseToolSpec from llama_index.tools.cassandra.cassandra_database_wrapper import ( ...
import torch from torchaudio._internal.module_utils import is_module_available if is_module_available("PIL"): from PIL import Image def save_image(path, data, mode=None): """Save image. The input image is expected to be CHW order """ if torch.is_tensor(data): data = data.numpy() if m...
import torch from torchaudio._internal.module_utils import is_module_available if is_module_available("PIL"): from PIL import Image def save_image(path, data, mode=None): """Save image. The input image is expected to be CHW order """ if torch.is_tensor(data): data = data.numpy() if m...
import tempfile import os import time import pytest from elasticsearch import Elasticsearch cur_dir = os.path.dirname(os.path.abspath(__file__)) compose_yml = os.path.abspath( os.path.join(cur_dir, 'unit', 'array', 'docker-compose.yml') ) @pytest.fixture(autouse=True) def tmpfile(tmpdir): tmpfile = f'docarr...
import tempfile import pytest @pytest.fixture(autouse=True) def tmpfile(tmpdir): tmpfile = f'docarray_test_{next(tempfile._get_candidate_names())}.db' return tmpdir / tmpfile
_base_ = '../_base_/default_runtime.py' # dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' # file_client_args = dict( # backend='petrel', # path_mapping=dict({ # './data/': 's3://openmmlab/datasets/detection/', # 'data/': 's3://openmmlab/datasets/detection/' # })) file...
_base_ = '../_base_/default_runtime.py' # dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' # file_client_args = dict( # backend='petrel', # path_mapping=dict({ # './data/': 's3://openmmlab/datasets/detection/', # 'data/': 's3://openmmlab/datasets/detection/' # })) file...
_base_ = ['./yolov3_mobilenetv2_8xb24-ms-416-300e_coco.py'] # yapf:disable model = dict( bbox_head=dict( anchor_generator=dict( base_sizes=[[(220, 125), (128, 222), (264, 266)], [(35, 87), (102, 96), (60, 170)], [(10, 15), (24, 36), (72, 42)]]))) ...
_base_ = ['./yolov3_mobilenetv2_8xb24-ms-416-300e_coco.py'] # yapf:disable model = dict( bbox_head=dict( anchor_generator=dict( base_sizes=[[(220, 125), (128, 222), (264, 266)], [(35, 87), (102, 96), (60, 170)], [(10, 15), (24, 36), (72, 42)]]))) ...
from docarray.document.mixins.attribute import GetAttributesMixin from docarray.document.mixins.audio import AudioDataMixin from docarray.document.mixins.blob import BlobDataMixin from docarray.document.mixins.content import ContentPropertyMixin from docarray.document.mixins.convert import ConvertMixin from docarray.do...
from .attribute import GetAttributesMixin from .audio import AudioDataMixin from .blob import BlobDataMixin from .content import ContentPropertyMixin from .convert import ConvertMixin from .dump import UriFileMixin from .featurehash import FeatureHashMixin from .image import ImageDataMixin from .mesh import MeshDataMix...
# Copyright (c) OpenMMLab. All rights reserved. from mmengine.config import ConfigDict from mmdet.core.utils import OptConfigType, OptMultiConfig from mmdet.registry import MODELS from .two_stage import TwoStageDetector @MODELS.register_module() class MaskRCNN(TwoStageDetector): """Implementation of `Mask R-CNN ...
# Copyright (c) OpenMMLab. All rights reserved. from mmengine.config import ConfigDict from mmdet.core.utils import OptConfigType, OptMultiConfig from mmdet.registry import MODELS from .two_stage import TwoStageDetector @MODELS.register_module() class MaskRCNN(TwoStageDetector): """Implementation of `Mask R-CNN ...