input
stringlengths
33
5k
output
stringlengths
32
5k
import os import urllib import pytest from pydantic import parse_obj_as, schema_json_of from docarray.base_document.io.json import orjson_dumps from docarray.typing import TextUrl from tests import TOYDATA_DIR REMOTE_TEXT_FILE = 'https://de.wikipedia.org/wiki/Brixen' CUR_DIR = os.path.dirname(os.path.abspath(__file...
import os import urllib import pytest from pydantic import parse_obj_as, schema_json_of from docarray.base_document.io.json import orjson_dumps from docarray.typing import TextUrl REMOTE_TXT = 'https://de.wikipedia.org/wiki/Brixen' CUR_DIR = os.path.dirname(os.path.abspath(__file__)) LOCAL_TXT = os.path.join(CUR_DIR...
from docutils import nodes from docutils.parsers.rst import Directive class BetaStatus(Directive): has_content = True text = "The {api_name} is in Beta stage, and backward compatibility is not guaranteed." node = nodes.warning def run(self): text = self.text.format(api_name=" ".join(self.cont...
from docutils import nodes from docutils.parsers.rst import Directive class BetaStatus(Directive): has_content = True text = "The {api_name} is in Beta stage, and backward compatibility is not guaranteed." node = nodes.warning def run(self): text = self.text.format(api_name=" ".join(self.cont...
from fastapi.testclient import TestClient from docs_src.configure_swagger_ui.tutorial002 import app client = TestClient(app) def test_swagger_ui(): response = client.get("/docs") assert response.status_code == 200, response.text assert ( '"syntaxHighlight": false' not in response.text ), "no...
from fastapi.testclient import TestClient from docs_src.configure_swagger_ui.tutorial002 import app client = TestClient(app) def test_swagger_ui(): response = client.get("/docs") assert response.status_code == 200, response.text assert ( '"syntaxHighlight": false' not in response.text ), "no...
import numpy as np from pydantic.tools import parse_obj_as, schema_json_of from docarray.base_document.io.json import orjson_dumps from docarray.typing import AnyEmbedding def test_proto_embedding(): embedding = parse_obj_as(AnyEmbedding, np.zeros((3, 224, 224))) embedding._to_node_protobuf() def test_js...
import numpy as np from pydantic.tools import parse_obj_as, schema_json_of from docarray.document.io.json import orjson_dumps from docarray.typing import AnyEmbedding def test_proto_embedding(): embedding = parse_obj_as(AnyEmbedding, np.zeros((3, 224, 224))) embedding._to_node_protobuf() def test_json_sc...
import os import re import subprocess from keras.src import backend # For torch, use index url to avoid installing nvidia drivers for the test. BACKEND_REQ = { "tensorflow": ("tensorflow-cpu", ""), "torch": ( "torch", "--extra-index-url https://download.pytorch.org/whl/cpu ", ), "jax":...
import os import re import subprocess from keras.src import backend # For torch, use index url to avoid installing nvidia drivers for the test. BACKEND_REQ = { "tensorflow": ("tensorflow-cpu", ""), "torch": ( "torch torchvision", "--extra-index-url https://download.pytorch.org/whl/cpu ", )...
# Copyright (c) OpenMMLab. All rights reserved. from .optimizer import (OPTIM_WRAPPER_CONSTRUCTORS, OPTIMIZERS, AmpOptimWrapper, DefaultOptimWrapperConstructor, OptimWrapper, OptimWrapperDict, build_optim_wrapper) from .scheduler import (ConstantLR, ConstantMomentum, Cons...
# Copyright (c) OpenMMLab. All rights reserved. from .optimizer import (OPTIMIZER_CONSTRUCTORS, OPTIMIZERS, DefaultOptimizerConstructor, build_optimizer) from .scheduler import (ConstantLR, ConstantMomentum, ConstantParamScheduler, CosineAnnealingLR, CosineAnnealingMoment...
from .autograd_utils import use_deterministic_algorithms from .backend_utils import set_audio_backend from .case_utils import ( HttpServerMixin, is_ffmpeg_available, PytorchTestCase, skipIfCudaSmallMemory, skipIfNoAudioDevice, skipIfNoCtcDecoder, skipIfNoCuda, skipIfNoExec, skipIfNoF...
from .backend_utils import set_audio_backend from .case_utils import ( HttpServerMixin, is_ffmpeg_available, PytorchTestCase, skipIfCudaSmallMemory, skipIfNoAudioDevice, skipIfNoCtcDecoder, skipIfNoCuda, skipIfNoExec, skipIfNoFFmpeg, skipIfNoKaldi, skipIfNoMacOS, skipIfNo...
from langchain_core.documents import Document from langchain_core.language_models import FakeListChatModel from langchain.retrievers.document_compressors import LLMChainFilter def test_llm_chain_filter() -> None: documents = [ Document( page_content="Candlepin bowling is popular in New Englan...
from langchain_core.documents import Document from langchain_core.language_models import FakeListChatModel from langchain.retrievers.document_compressors import LLMChainFilter def test_llm_chain_filter() -> None: documents = [ Document( page_content="Candlepin bowling is popular in New Englan...
_base_ = './grid-rcnn_r50_fpn_gn-head_2x_coco.py' model = dict( backbone=dict( type='ResNeXt', depth=101, groups=32, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, style='pytorch', init_cfg=dict( type='Pretra...
_base_ = './grid_rcnn_r50_fpn_gn-head_2x_coco.py' model = dict( backbone=dict( type='ResNeXt', depth=101, groups=32, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, style='pytorch', init_cfg=dict( type='Pretra...
import pytest from docarray import BaseDocument, DocumentArray from docarray.documents import ImageDoc from docarray.typing import NdArray class MyDoc(BaseDocument): embedding: NdArray text: str image: ImageDoc @pytest.mark.parametrize( 'protocol', ['pickle-array', 'protobuf-array', 'protobuf', 'pi...
import pytest from docarray import BaseDocument from docarray.typing import NdArray from docarray.documents import Image from docarray import DocumentArray class MyDoc(BaseDocument): embedding: NdArray text: str image: Image @pytest.mark.parametrize( 'protocol', ['pickle-array', 'protobuf-array', '...
import sys from os import path from setuptools import find_packages from setuptools import setup if sys.version_info < (3, 7, 0): raise OSError(f'DocArray requires Python >=3.7, but yours is {sys.version}') try: pkg_name = 'docarray' libinfo_py = path.join(pkg_name, '__init__.py') libinfo_content = o...
import sys from os import path from setuptools import find_packages from setuptools import setup if sys.version_info < (3, 7, 0): raise OSError(f'DocArray requires Python >=3.7, but yours is {sys.version}') try: pkg_name = 'docarray' libinfo_py = path.join(pkg_name, '__init__.py') libinfo_content = o...
# Copyright (c) OpenMMLab. All rights reserved. from mmengine.utils.dl_utils import TORCH_VERSION from mmengine.utils.version_utils import digit_version from .averaged_model import (BaseAveragedModel, ExponentialMovingAverage, MomentumAnnealingEMA, StochasticWeightAverage) from .base_model ...
# Copyright (c) OpenMMLab. All rights reserved. from mmengine.utils.dl_utils import TORCH_VERSION from mmengine.utils.version_utils import digit_version from .averaged_model import (BaseAveragedModel, ExponentialMovingAverage, MomentumAnnealingEMA, StochasticWeightAverage) from .base_model ...
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .single_stage_instance_seg import SingleStageInstanceSegmentor @DETECTORS.register_module() class SOLO(SingleStageInstanceSegmentor): """`SOLO: Segmenting Objects by Locations <https://arxiv.org/abs/1912.04488>`_ """ ...
from ..builder import DETECTORS from .single_stage_instance_seg import SingleStageInstanceSegmentor @DETECTORS.register_module() class SOLO(SingleStageInstanceSegmentor): """`SOLO: Segmenting Objects by Locations <https://arxiv.org/abs/1912.04488>`_ """ def __init__(self, backbone, ...
__version__ = '0.35.0' import logging from docarray.array import DocList, DocVec from docarray.base_doc.doc import BaseDoc from docarray.utils._internal.misc import _get_path_from_docarray_root_level __all__ = ['BaseDoc', 'DocList', 'DocVec'] logger = logging.getLogger('docarray') handler = logging.StreamHandler()...
__version__ = '0.34.1' import logging from docarray.array import DocList, DocVec from docarray.base_doc.doc import BaseDoc from docarray.utils._internal.misc import _get_path_from_docarray_root_level __all__ = ['BaseDoc', 'DocList', 'DocVec'] logger = logging.getLogger('docarray') handler = logging.StreamHandler()...
_base_ = './mask-rcnn_swin-t-p4-w7_fpn_ms-crop-3x_coco.py' # Enable automatic-mixed-precision training with AmpOptimWrapper. optim_wrapper = dict(type='AmpOptimWrapper')
_base_ = './mask_rcnn_swin-t-p4-w7_fpn_ms-crop-3x_coco.py' # Enable automatic-mixed-precision training with AmpOptimWrapper. optim_wrapper = dict(type='AmpOptimWrapper')
import numpy as np import pytest import torch from pydantic import parse_obj_as from docarray import BaseDoc from docarray.documents import VideoDoc from docarray.typing import AudioNdArray, NdArray, VideoNdArray from docarray.utils._internal.misc import is_tf_available from docarray.utils._internal.pydantic import is...
import numpy as np import pytest import torch from pydantic import parse_obj_as from docarray import BaseDoc from docarray.documents import VideoDoc from docarray.typing import AudioNdArray, NdArray, VideoNdArray from docarray.utils._internal.misc import is_tf_available from tests import TOYDATA_DIR tf_available = is...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '3.0.0rc4' short_version = __version__ def parse_version_info(version_str): """Parse a version string into a tuple. Args: version_str (str): The version string. Returns: tuple[int | str]: The version info, e.g., "1.3.0" is par...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '3.0.0rc3' short_version = __version__ def parse_version_info(version_str): """Parse a version string into a tuple. Args: version_str (str): The version string. Returns: tuple[int | str]: The version info, e.g., "1.3.0" is par...
_base_ = './fcos_hrnetv2p-w32-gn-head_ms-640-800-4xb4-2x_coco.py' model = dict( backbone=dict( type='HRNet', extra=dict( stage2=dict(num_channels=(40, 80)), stage3=dict(num_channels=(40, 80, 160)), stage4=dict(num_channels=(40, 80, 160, 320))), init_cfg=di...
_base_ = './fcos_hrnetv2p_w32_gn-head_mstrain_640-800_4x4_2x_coco.py' model = dict( backbone=dict( type='HRNet', extra=dict( stage2=dict(num_channels=(40, 80)), stage3=dict(num_channels=(40, 80, 160)), stage4=dict(num_channels=(40, 80, 160, 320))), init_cf...
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from jina import Flow from simpleranker import SimpleRanker def test_integration(documents_chunk): with Flow().add(uses=SimpleRanker, uses_with={'metric': 'cosine'}) as flow: resp = flow.post(on='/search...
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from jina import Flow from ...simpleranker import SimpleRanker def test_integration(documents_chunk): with Flow().add(uses=SimpleRanker, uses_with={'metric': 'cosine'}) as flow: resp = flow.post(on='/se...
from backend.data.credit import get_user_credit_model from backend.data.execution import ( ExecutionResult, NodeExecutionEntry, RedisExecutionEventBus, create_graph_execution, get_execution_results, get_incomplete_executions, get_latest_execution, update_execution_status, update_grap...
from functools import wraps from typing import Any, Callable, Concatenate, Coroutine, ParamSpec, TypeVar, cast from backend.data.credit import get_user_credit_model from backend.data.execution import ( ExecutionResult, NodeExecutionEntry, RedisExecutionEventBus, create_graph_execution, get_executio...
from keras.src import backend from keras.src.utils.module_utils import tensorflow as tf def get_tensor_spec(t, dynamic_batch=False, name=None): """Returns a `TensorSpec` given a single `Tensor` or `TensorSpec`.""" if isinstance(t, tf.TypeSpec): spec = t elif isinstance(t, tf.__internal__.Composite...
from keras.src.utils.module_utils import tensorflow as tf def get_tensor_spec(t, dynamic_batch=False, name=None): """Returns a `TensorSpec` given a single `Tensor` or `TensorSpec`.""" if isinstance(t, tf.TypeSpec): spec = t elif isinstance(t, tf.__internal__.CompositeTensor): # Check for E...
import typing import pydantic class LibraryAgent(pydantic.BaseModel): id: str # Changed from agent_id to match GraphMeta version: int # Changed from agent_version to match GraphMeta is_active: bool # Added to match GraphMeta name: str description: str isCreatedByUser: bool # Made inpu...
import datetime import json import typing import prisma.models import pydantic import backend.data.block import backend.data.graph import backend.server.model class LibraryAgent(pydantic.BaseModel): id: str # Changed from agent_id to match GraphMeta agent_id: str agent_version: int # Changed from age...
from langchain_core.load import dumpd, dumps, load, loads from langchain_openai import ChatOpenAI, OpenAI def test_loads_openai_llm() -> None: llm = OpenAI(model="davinci", temperature=0.5, openai_api_key="hello", top_p=0.8) # type: ignore[call-arg] llm_string = dumps(llm) llm2 = loads(llm_string, secre...
from langchain_core.load.dump import dumpd, dumps from langchain_core.load.load import load, loads from langchain_openai import ChatOpenAI, OpenAI def test_loads_openai_llm() -> None: llm = OpenAI(model="davinci", temperature=0.5, openai_api_key="hello", top_p=0.8) # type: ignore[call-arg] llm_string = dump...
import os from typing import Any, Callable, List, Optional, Tuple import torch import torch.utils.data as data from ..utils import _log_api_usage_once class VisionDataset(data.Dataset): """ Base Class For making datasets which are compatible with torchvision. It is necessary to override the ``__getitem_...
import os from typing import Any, Callable, List, Optional, Tuple import torch import torch.utils.data as data from ..utils import _log_api_usage_once class VisionDataset(data.Dataset): """ Base Class For making datasets which are compatible with torchvision. It is necessary to override the ``__getitem_...
""" This examples loads a pre-trained model and evaluates it on the STSbenchmark dataset Usage: python evaluation_stsbenchmark.py OR python evaluation_stsbenchmark.py model_name """ from sentence_transformers import SentenceTransformer from sentence_transformers.evaluation import EmbeddingSimilarityEvaluator from dat...
""" This examples loads a pre-trained model and evaluates it on the STSbenchmark dataset Usage: python evaluation_stsbenchmark.py OR python evaluation_stsbenchmark.py model_name """ from sentence_transformers import SentenceTransformer, util, LoggingHandler, InputExample from sentence_transformers.evaluation import E...
_base_ = '../cascade_rcnn/cascade-mask-rcnn_x101-32x4d_fpn_1x_coco.py' model = dict( backbone=dict( norm_cfg=dict(type='SyncBN', requires_grad=True), norm_eval=False, plugins=[ dict( cfg=dict(type='ContextBlock', ratio=1. / 16), stages=(False, True...
_base_ = '../cascade_rcnn/cascade_mask_rcnn_x101_32x4d_fpn_1x_coco.py' model = dict( backbone=dict( norm_cfg=dict(type='SyncBN', requires_grad=True), norm_eval=False, plugins=[ dict( cfg=dict(type='ContextBlock', ratio=1. / 16), stages=(False, True...
_base_ = [ '../_base_/default_runtime.py', '../_base_/datasets/coco_detection.py' ] data_preprocessor = dict( type='DetDataPreprocessor', mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], bgr_to_rgb=True) # model settings model = dict( type='CornerNet', data_preprocessor=data_pr...
_base_ = [ '../_base_/default_runtime.py', '../_base_/datasets/coco_detection.py' ] data_preprocessor = dict( type='DetDataPreprocessor', mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], bgr_to_rgb=True) # model settings model = dict( type='CornerNet', data_preprocessor=data_pr...
import asyncio import pytest from llama_index.core.workflow.context import Context from llama_index.core.workflow.decorators import step from llama_index.core.workflow.errors import WorkflowRuntimeError, WorkflowTimeoutError from llama_index.core.workflow.events import Event, StartEvent, StopEvent from llama_index.cor...
import asyncio import pytest from llama_index.core.workflow.context import Context from llama_index.core.workflow.decorators import step from llama_index.core.workflow.errors import WorkflowRuntimeError, WorkflowTimeoutError from llama_index.core.workflow.events import Event, StartEvent, StopEvent from llama_index.cor...
from __future__ import annotations from .CSRSparsity import CSRSparsity from .TopKActivation import TopKActivation __all__ = ["CSRSparsity", "TopKActivation"] # TODO : Add in models the possibility to have the MLM head(for splade)
from __future__ import annotations from .CSRSparsity import CSRSparsity __all__ = ["CSRSparsity"]
# Copyright (c) OpenMMLab. All rights reserved. from abc import ABCMeta, abstractmethod from typing import Dict, Union from torch.utils.data import DataLoader class BaseLoop(metaclass=ABCMeta): """Base loop class. All subclasses inherited from ``BaseLoop`` should overwrite the :meth:`run` method. A...
# Copyright (c) OpenMMLab. All rights reserved. from abc import ABCMeta, abstractmethod from typing import Dict, Union from torch.utils.data import DataLoader class BaseLoop(metaclass=ABCMeta): """Base loop class. All subclasses inherited from ``BaseLoop`` should overwrite the :meth:`run` method. A...
import asyncio import logging import os from jina import __default_host__ from jina.importer import ImportExtensions from jina.serve.runtimes.gateway import GatewayRuntime from jina.serve.runtimes.gateway.http.app import get_fastapi_app __all__ = ['HTTPGatewayRuntime'] class HTTPGatewayRuntime(GatewayRuntime): ...
import asyncio import logging import os from jina import __default_host__ from jina.importer import ImportExtensions from jina.serve.runtimes.gateway import GatewayRuntime from jina.serve.runtimes.gateway.http.app import get_fastapi_app __all__ = ['HTTPGatewayRuntime'] class HTTPGatewayRuntime(GatewayRuntime): ...
import types from typing import TYPE_CHECKING from docarray.index.backends.in_memory import InMemoryExactNNIndex from docarray.utils._internal.misc import ( _get_path_from_docarray_root_level, import_library, ) if TYPE_CHECKING: from docarray.index.backends.elastic import ElasticDocIndex # noqa: F401 ...
import types from typing import TYPE_CHECKING from docarray.index.backends.in_memory import InMemoryExactNNIndex from docarray.utils._internal.misc import ( _get_path_from_docarray_root_level, import_library, ) if TYPE_CHECKING: from docarray.index.backends.elastic import ElasticDocIndex # noqa: F401 ...
import sys import numpy as np import pytest from hypothesis import given, settings, strategies import xgboost as xgb from xgboost import testing as tm from xgboost.testing import no_cupy from xgboost.testing.updater import check_extmem_qdm, check_quantile_loss_extmem sys.path.append("tests/python") from test_data_it...
import sys import numpy as np import pytest from hypothesis import given, settings, strategies import xgboost as xgb from xgboost import testing as tm from xgboost.testing import no_cupy from xgboost.testing.updater import check_extmem_qdm, check_quantile_loss_extmem sys.path.append("tests/python") from test_data_it...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '0.10.0' def parse_version_info(version_str): """Parse the version information. Args: version_str (str): version string like '0.1.0'. Returns: tuple: version information contains major, minor, micro version. """ versi...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '0.9.1' def parse_version_info(version_str): """Parse the version information. Args: version_str (str): version string like '0.1.0'. Returns: tuple: version information contains major, minor, micro version. """ versio...
from abc import abstractmethod from typing import List, Sequence from llama_index.core.bridge.pydantic import BaseModel from llama_index.core.instrumentation import DispatcherSpanMixin from llama_index.core.prompts.mixin import PromptMixin, PromptMixinType from llama_index.core.schema import QueryBundle from llama_ind...
from abc import abstractmethod from typing import List, Sequence from llama_index.core.bridge.pydantic import BaseModel from llama_index.core.instrumentation import DispatcherSpanMixin from llama_index.core.prompts.mixin import PromptMixin, PromptMixinType from llama_index.core.schema import QueryBundle from llama_ind...
import pytest from jina import Flow from jina.enums import GatewayProtocolType from tests import random_docs @pytest.mark.slow @pytest.mark.parametrize('protocol', ['http', 'websocket', 'grpc']) @pytest.mark.parametrize('changeto_protocol', ['grpc', 'http', 'websocket']) def test_change_gateway(protocol, changeto_pr...
import pytest from jina import Flow from jina.enums import GatewayProtocolType from tests import random_docs @pytest.mark.slow @pytest.mark.parametrize('protocol', ['http', 'websocket', 'grpc']) @pytest.mark.parametrize('changeto_protocol', ['grpc', 'http', 'websocket']) def test_change_gateway(protocol, changeto_pr...
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
from keras.src.api_export import keras_export from keras.src.layers.pooling.base_pooling import BasePooling @keras_export(["keras.layers.MaxPooling1D", "keras.layers.MaxPool1D"]) class MaxPooling1D(BasePooling): """Max pooling operation for 1D temporal data. Downsamples the input representation by taking the...
from keras.src.api_export import keras_export from keras.src.layers.pooling.base_pooling import BasePooling @keras_export(["keras.layers.MaxPooling1D", "keras.layers.MaxPool1D"]) class MaxPooling1D(BasePooling): """Max pooling operation for 1D temporal data. Downsamples the input representation by taking the...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
import asyncio from typing import Any, AsyncGenerator, List, Optional from llama_index.core.workflow.context import Context from llama_index.core.workflow.errors import WorkflowDone from llama_index.core.workflow.events import Event, StopEvent from .types import RunResultT from .utils import BUSY_WAIT_DELAY class W...
import asyncio from typing import Any, AsyncGenerator, List, Optional from llama_index.core.workflow.context import Context from llama_index.core.workflow.errors import WorkflowDone from llama_index.core.workflow.events import Event, StopEvent from .types import RunResultT from .utils import BUSY_WAIT_DELAY class W...
# Copyright 2025 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
# Copyright 2025 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
# Copyright (c) OpenMMLab. All rights reserved. from .base_dataset import BaseDataset, Compose, force_full_init from .dataset_wrapper import ClassBalancedDataset, ConcatDataset, RepeatDataset from .sampler import DefaultSampler, InfiniteSampler from .utils import (COLLATE_FUNCTIONS, default_collate, pseudo_collate, ...
# Copyright (c) OpenMMLab. All rights reserved. from .base_dataset import BaseDataset, Compose, force_full_init from .dataset_wrapper import ClassBalancedDataset, ConcatDataset, RepeatDataset from .sampler import DefaultSampler, InfiniteSampler from .utils import pseudo_collate, worker_init_fn __all__ = [ 'BaseDat...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.utilities.dataforseo_api_search import DataForSeoAPIWrapper # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # hand...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.utilities.dataforseo_api_search import DataForSeoAPIWrapper # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # hand...
# model settings norm_cfg = dict(type='BN', requires_grad=False) model = dict( type='MaskRCNN', data_preprocessor=dict( type='DetDataPreprocessor', mean=[103.530, 116.280, 123.675], std=[1.0, 1.0, 1.0], bgr_to_rgb=False, pad_size_divisor=32), backbone=dict( ty...
# model settings preprocess_cfg = dict( mean=[103.530, 116.280, 123.675], std=[1.0, 1.0, 1.0], to_rgb=False, pad_size_divisor=32) norm_cfg = dict(type='BN', requires_grad=False) model = dict( type='MaskRCNN', preprocess_cfg=preprocess_cfg, backbone=dict( type='ResNet', depth=...
"""Test ChatYuan2 wrapper.""" import pytest from langchain_core.messages import ( AIMessage, HumanMessage, SystemMessage, ) from langchain_community.chat_models.yuan2 import ( ChatYuan2, _convert_dict_to_message, _convert_message_to_dict, ) @pytest.mark.requires("openai") def test_yuan2_mode...
"""Test ChatYuan2 wrapper.""" import pytest from langchain_core.messages import ( AIMessage, HumanMessage, SystemMessage, ) from langchain_community.chat_models.yuan2 import ( ChatYuan2, _convert_dict_to_message, _convert_message_to_dict, ) @pytest.mark.requires("openai") def test_yuan2_mode...
import logging import os from typing import Optional from jina import __default_host__ from jina.importer import ImportExtensions from jina.serve.gateway import BaseGateway from jina.serve.runtimes.gateway.websocket.app import get_fastapi_app class WebSocketGateway(BaseGateway): """WebSocket Gateway implementati...
import logging import os from typing import Optional from jina import __default_host__ from jina.importer import ImportExtensions from jina.serve.gateway import BaseGateway from jina.serve.runtimes.gateway.websocket.app import get_fastapi_app class WebSocketGateway(BaseGateway): """WebSocket Gateway implementati...
from .dpr_text import DPRTextEncoder
from .dpr_text import DPRTextEncoder
from langchain_core.prompt_values import ChatPromptValue, ChatPromptValueConcrete from langchain_core.prompts.chat import ( AIMessagePromptTemplate, BaseChatPromptTemplate, BaseStringMessagePromptTemplate, ChatMessagePromptTemplate, ChatPromptTemplate, HumanMessagePromptTemplate, MessageLike...
from langchain_core.prompt_values import ChatPromptValue, ChatPromptValueConcrete from langchain_core.prompts.chat import ( AIMessagePromptTemplate, BaseChatPromptTemplate, BaseStringMessagePromptTemplate, ChatMessagePromptTemplate, ChatPromptTemplate, HumanMessagePromptTemplate, MessageLike...
# Copyright (c) OpenMMLab. All rights reserved. from .augment_wrappers import AutoAugment, RandAugment from .colorspace import (AutoContrast, Brightness, Color, ColorTransform, Contrast, Equalize, Invert, Posterize, Sharpness, Solarize, SolarizeAdd) from .formatting imp...
# Copyright (c) OpenMMLab. All rights reserved. from .augment_wrappers import AutoAugment, RandAugment from .colorspace import (AutoContrast, Brightness, Color, ColorTransform, Contrast, Equalize, Invert, Posterize, Sharpness, Solarize, SolarizeAdd) from .formatting imp...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import CopyFileTool from langchain_community.tools.file_management.copy import FileCopyInput # Create a way to dynamically look up deprecated imports. # Used to consolidate logic ...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import CopyFileTool from langchain_community.tools.file_management.copy import FileCopyInput # Create a way to dynamically look up deprecated imports. # Used to consolidate logic ...
_base_ = '../htc/htc_r50_fpn_20e_coco.py' model = dict( backbone=dict( type='Res2Net', depth=101, scales=4, base_width=26, init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://res2net101_v1d_26w_4s')))
_base_ = '../htc/htc_r50_fpn_1x_coco.py' model = dict( backbone=dict( type='Res2Net', depth=101, scales=4, base_width=26, init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://res2net101_v1d_26w_4s'))) # learning policy lr_config = dict(step=[16, ...
from __future__ import annotations import logging from typing import TYPE_CHECKING, Any, Callable from sentence_transformers.evaluation import InformationRetrievalEvaluator if TYPE_CHECKING: import numpy as np from torch import Tensor from sentence_transformers.similarity_functions import SimilarityFunc...
from __future__ import annotations import logging from typing import TYPE_CHECKING, Any, Callable from sentence_transformers.evaluation import InformationRetrievalEvaluator if TYPE_CHECKING: import numpy as np from torch import Tensor from sentence_transformers.similarity_functions import SimilarityFunc...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] # model settings preprocess_cfg = dict( mean=[103.530, 116.280, 123.675], std=[1.0, 1.0, 1.0], to_rgb=False, pad_size_divisor=32) model = dict( type='NASFCOS', prepr...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] # model settings preprocess_cfg = dict( mean=[103.530, 116.280, 123.675], std=[1.0, 1.0, 1.0], to_rgb=False, pad_size_divisor=32) model = dict( type='NASFCOS', prepr...
import posixpath from pathlib import Path from unittest.mock import patch import fsspec import pytest from fsspec.implementations.local import AbstractFileSystem, LocalFileSystem, stringify_path class MockFileSystem(AbstractFileSystem): protocol = "mock" def __init__(self, *args, local_root_dir, **kwargs): ...
import posixpath from pathlib import Path import fsspec import pytest from fsspec.implementations.local import AbstractFileSystem, LocalFileSystem, stringify_path class MockFileSystem(AbstractFileSystem): protocol = "mock" def __init__(self, *args, local_root_dir, **kwargs): super().__init__() ...
__all__ = ['reduce', 'reduce_all'] from typing import Dict, List, Optional from docarray import DocList def reduce( left: DocList, right: DocList, left_id_map: Optional[Dict] = None ) -> 'DocList': """ Reduces left and right DocList into one DocList in-place. Changes are applied to the left DocList....
__all__ = ['reduce', 'reduce_all'] from typing import Dict, List, Optional from docarray import DocArray def reduce( left: DocArray, right: DocArray, left_id_map: Optional[Dict] = None ) -> 'DocArray': """ Reduces left and right DocArray into one DocArray in-place. Changes are applied to the left Do...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '0.10.7' def parse_version_info(version_str): """Parse the version information. Args: version_str (str): version string like '0.1.0'. Returns: tuple: version information contains major, minor, micro version. """ versi...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '0.10.6' def parse_version_info(version_str): """Parse the version information. Args: version_str (str): version string like '0.1.0'. Returns: tuple: version information contains major, minor, micro version. """ versi...
import warnings from keras.src import activations from keras.src.api_export import keras_export from keras.src.layers.layer import Layer @keras_export("keras.layers.LeakyReLU") class LeakyReLU(Layer): """Leaky version of a Rectified Linear Unit activation layer. This layer allows a small gradient when the u...
import warnings from keras.src import activations from keras.src.api_export import keras_export from keras.src.layers.layer import Layer @keras_export("keras.layers.LeakyReLU") class LeakyReLU(Layer): """Leaky version of a Rectified Linear Unit activation layer. This layer allows a small gradient when the u...
""" This basic example loads a pre-trained model from the web and uses it to generate sentence embeddings for a given list of sentences. """ import logging import numpy as np from sentence_transformers import LoggingHandler, SentenceTransformer #### Just some code to print debug information to stdout np.set_printop...
""" This basic example loads a pre-trained model from the web and uses it to generate sentence embeddings for a given list of sentences. """ from sentence_transformers import SentenceTransformer, LoggingHandler import numpy as np import logging #### Just some code to print debug information to stdout np.set_printopti...
_base_ = '../_base_/default_runtime.py' # model settings model = dict( type='YOLOV3', backbone=dict( type='Darknet', depth=53, out_indices=(3, 4, 5), init_cfg=dict(type='Pretrained', checkpoint='open-mmlab://darknet53')), neck=dict( type='YOLOV3Neck', num_scal...
_base_ = '../_base_/default_runtime.py' # model settings model = dict( type='YOLOV3', backbone=dict( type='Darknet', depth=53, out_indices=(3, 4, 5), init_cfg=dict(type='Pretrained', checkpoint='open-mmlab://darknet53')), neck=dict( type='YOLOV3Neck', num_scal...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import MODELS from mmdet.utils import ConfigType, OptConfigType, OptMultiConfig from .two_stage import TwoStageDetector @MODELS.register_module() class SparseRCNN(TwoStageDetector): r"""Implementation of `Sparse R-CNN: End-to-End Object Detection...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import MODELS from .two_stage import TwoStageDetector @MODELS.register_module() class SparseRCNN(TwoStageDetector): r"""Implementation of `Sparse R-CNN: End-to-End Object Detection with Learnable Proposals <https://arxiv.org/abs/2011.12450>`_...
__version__ = '0.19.0' import os from docarray.document import Document from docarray.array import DocumentArray from docarray.dataclasses import dataclass, field from docarray.helper import login, logout if 'DA_RICH_HANDLER' in os.environ: from rich.traceback import install install()
__version__ = '0.18.2' import os from docarray.document import Document from docarray.array import DocumentArray from docarray.dataclasses import dataclass, field from docarray.helper import login, logout if 'DA_RICH_HANDLER' in os.environ: from rich.traceback import install install()
# Copyright (c) OpenMMLab. All rights reserved. import mmcv import numpy as np import pycocotools.mask as mask_util import torch def split_combined_polys(polys, poly_lens, polys_per_mask): """Split the combined 1-D polys into masks. A mask is represented as a list of polys, and a poly is represented as a...
# Copyright (c) OpenMMLab. All rights reserved. import mmcv import numpy as np import pycocotools.mask as mask_util def split_combined_polys(polys, poly_lens, polys_per_mask): """Split the combined 1-D polys into masks. A mask is represented as a list of polys, and a poly is represented as a 1-D array. I...
# coding=utf-8 # Copyright 2025 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
# coding=utf-8 # Copyright 2025 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
# Copyright (c) OpenMMLab. All rights reserved. """Collecting some commonly used type hint in mmdetection.""" from typing import List, Optional, Union from mmengine.config import ConfigDict from mmengine.data import InstanceData from ..data_structures import DetDataSample # Type hint of config data ConfigType = Uni...
# Copyright (c) OpenMMLab. All rights reserved. """Collecting some commonly used type hint in mmdetection.""" from typing import List, Optional, Union from mmengine.config import ConfigDict from mmengine.data import InstanceData from mmdet.core import DetDataSample # Type hint of config data ConfigType = Union[Conf...
class MediaUploadError(Exception): """Base exception for media upload errors""" pass class InvalidFileTypeError(MediaUploadError): """Raised when file type is not supported""" pass class FileSizeTooLargeError(MediaUploadError): """Raised when file size exceeds maximum limit""" pass clas...
class MediaUploadError(Exception): """Base exception for media upload errors""" pass class InvalidFileTypeError(MediaUploadError): """Raised when file type is not supported""" pass class FileSizeTooLargeError(MediaUploadError): """Raised when file size exceeds maximum limit""" pass clas...
# Copyright (c) OpenMMLab. All rights reserved. import logging import random from typing import List, Optional, Tuple import numpy as np import torch from mmengine.dist import get_rank, sync_random_seed from mmengine.logging import print_log from mmengine.utils import digit_version, is_list_of from mmengine.utils.dl_...
# Copyright (c) OpenMMLab. All rights reserved. import logging import random from typing import List, Optional, Tuple import numpy as np import torch from mmengine.dist import get_rank, sync_random_seed from mmengine.logging import print_log from mmengine.utils import digit_version, is_list_of from mmengine.utils.dl_...
_base_ = [ './yolox_x_8xb4-80e_crowdhuman-mot17halftrain_test-mot17halfval.py', # noqa: E501 ] dataset_type = 'MOTChallengeDataset' detector = _base_.model detector.pop('data_preprocessor') del _base_.model model = dict( type='StrongSORT', data_preprocessor=dict( type='TrackDataPreprocessor', ...
_base_ = [ './yolox_x_8xb4-80e_crowdhuman-mot17halftrain_test-mot17halfval.py', # noqa: E501 ] dataset_type = 'MOTChallengeDataset' detector = _base_.model detector.pop('data_preprocessor') del _base_.model model = dict( type='StrongSORT', data_preprocessor=dict( type='TrackDataPreprocessor', ...
# Copyright (c) OpenMMLab. All rights reserved. from .bbox_overlaps import bbox_overlaps from .cityscapes_utils import evaluateImgLists from .class_names import (cityscapes_classes, coco_classes, coco_panoptic_classes, dataset_aliases, get_classes, imagenet_det_classe...
# Copyright (c) OpenMMLab. All rights reserved. from .bbox_overlaps import bbox_overlaps from .cityscapes_utils import evaluateImgLists from .class_names import (cityscapes_classes, coco_classes, coco_panoptic_classes, dataset_aliases, get_classes, imagenet_det_classe...
_base_ = './atss_r50_fpn_8xb8-amp-lsj-200e_coco.py' model = dict( backbone=dict( depth=18, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet18')), neck=dict(in_channels=[64, 128, 256, 512]))
_base_ = './atss_r50_fpn_lsj_200e_8x8_fp16_coco.py' model = dict( backbone=dict( depth=18, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet18')), neck=dict(in_channels=[64, 128, 256, 512]))
from .gateway import HTTPGateway
import asyncio import os from jina import __default_host__ from jina.serve.runtimes.gateway import GatewayRuntime from jina.serve.runtimes.gateway.http.app import get_fastapi_app __all__ = ['HTTPGatewayRuntime'] from jina.serve.runtimes.gateway.http.gateway import HTTPGateway class HTTPGatewayRuntime(GatewayRuntim...
# Copyright (c) OpenMMLab. All rights reserved. from .averaged_model import (ExponentialMovingAverage, MomentumAnnealingEMA, StochasticWeightAverage) from .wrappers import (MMDataParallel, MMDistributedDataParallel, is_model_wrapper) __all__ = [ 'MMDistributedDat...
# Copyright (c) OpenMMLab. All rights reserved. from .wrappers import (MMDataParallel, MMDistributedDataParallel, is_model_wrapper) __all__ = ['MMDistributedDataParallel', 'MMDataParallel', 'is_model_wrapper']
from langchain_core.embeddings import DeterministicFakeEmbedding, Embeddings from langchain_tests.integration_tests import EmbeddingsIntegrationTests from langchain_tests.unit_tests import EmbeddingsUnitTests class TestFakeEmbeddingsUnit(EmbeddingsUnitTests): @property def embeddings_class(self) -> type[Embe...
from typing import Type from langchain_core.embeddings import DeterministicFakeEmbedding, Embeddings from langchain_tests.integration_tests import EmbeddingsIntegrationTests from langchain_tests.unit_tests import EmbeddingsUnitTests class TestFakeEmbeddingsUnit(EmbeddingsUnitTests): @property def embeddings...
"""Hive data reader.""" from typing import List, Optional from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class HiveReader(BaseReader): """ Read documents from a Hive. These documents can then be used in a downstream Llama Index data structure. Arg...
"""Hive data reader.""" from typing import List, Optional from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class HiveReader(BaseReader): """ Read documents from a Hive. These documents can then be used in a downstream Llama Index data structure. Arg...
from __future__ import annotations from sentence_transformers.training_args import SentenceTransformerTrainingArguments class CrossEncoderTrainingArguments(SentenceTransformerTrainingArguments): r""" CrossEncoderTrainingArguments extends :class:`~transformers.TrainingArguments` with additional arguments ...
from __future__ import annotations from sentence_transformers.training_args import SentenceTransformerTrainingArguments class CrossEncoderTrainingArguments(SentenceTransformerTrainingArguments): """ CrossEncoderTrainingArguments extends :class:`~transformers.TrainingArguments` with additional arguments s...
_base_ = [ '../_base_/models/mask-rcnn_r50_fpn.py', '../_base_/datasets/lvis_v1_instance.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] model = dict( roi_head=dict( bbox_head=dict( num_classes=1203, cls_predictor_cfg=dict(type='NormedLinear', ...
_base_ = [ '../_base_/models/mask-rcnn_r50_fpn.py', '../_base_/datasets/lvis_v1_instance.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] model = dict( roi_head=dict( bbox_head=dict( num_classes=1203, cls_predictor_cfg=dict(type='NormedLinear', ...
# coding=utf-8 # Copyright 2024 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
# coding=utf-8 # Copyright 2024 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
from typing import Optional import pytest import torch from docarray import BaseDocument, DocumentArray from docarray.array.abstract_array import AnyDocumentArray from docarray.documents import Text from docarray.typing import TorchTensor num_docs = 5 num_sub_docs = 2 num_sub_sub_docs = 3 @pytest.fixture def multi...
from typing import Optional import pytest import torch from docarray import BaseDocument, DocumentArray, Text from docarray.array.abstract_array import AnyDocumentArray from docarray.typing import TorchTensor num_docs = 5 num_sub_docs = 2 num_sub_sub_docs = 3 @pytest.fixture def multi_model_docs(): class SubSu...
import os import torch import torchaudio.prototype.transforms as T import torchaudio.transforms as transforms from torchaudio_unittest.common_utils import nested_params, TorchaudioTestCase class BatchConsistencyTest(TorchaudioTestCase): def assert_batch_consistency(self, transform, batch, *args, atol=1e-8, rtol=...
import os import torch import torchaudio.prototype.transforms as T import torchaudio.transforms as transforms from torchaudio_unittest.common_utils import nested_params, TorchaudioTestCase class BatchConsistencyTest(TorchaudioTestCase): def assert_batch_consistency(self, transform, batch, *args, atol=1e-8, rtol=...
from __future__ import annotations import sys from .BoW import BoW from .CLIPModel import CLIPModel from .CNN import CNN from .Dense import Dense from .Dropout import Dropout from .InputModule import InputModule from .LayerNorm import LayerNorm from .LSTM import LSTM from .Module import Module from .Normalize import ...
from __future__ import annotations from .Asym import Asym from .BoW import BoW from .CLIPModel import CLIPModel from .CNN import CNN from .Dense import Dense from .Dropout import Dropout from .InputModule import InputModule from .LayerNorm import LayerNorm from .LSTM import LSTM from .Module import Module from .Normal...
import logging import random from datasets import load_dataset from sentence_transformers import SparseEncoder from sentence_transformers.sparse_encoder.evaluation import SparseInformationRetrievalEvaluator logging.basicConfig(format="%(message)s", level=logging.INFO) # Load a model model = SparseEncoder("naver/spl...
import logging import random from datasets import load_dataset from sentence_transformers import SparseEncoder from sentence_transformers.sparse_encoder.evaluation import SparseInformationRetrievalEvaluator logging.basicConfig(format="%(message)s", level=logging.INFO) # Load a model model = SparseEncoder("naver/spl...
from abc import abstractmethod from typing import TYPE_CHECKING, Generic, List, Sequence, Type, TypeVar, Union from docarray.document import BaseDocument from docarray.typing.abstract_type import AbstractType if TYPE_CHECKING: from docarray.proto import DocumentArrayProto, NodeProto from docarray.typing impor...
from abc import abstractmethod from typing import TYPE_CHECKING, Dict, Iterable, List, Optional, Sequence, Type, Union from docarray.document import BaseDocument if TYPE_CHECKING: from docarray.typing import NdArray, TorchTensor class AbstractDocumentArray(Sequence): document_type: Type[BaseDocument] _...
_base_ = 'faster-rcnn_r50-caffe_fpn_ms-1x_coco.py' max_iter = 90000 param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_iter, by_epoch=False, milestones=[60000, 80000], ...
_base_ = 'faster-rcnn_r50-caffe_fpn_ms-1x_coco.py' # learning policy lr_config = dict( policy='step', warmup='linear', warmup_iters=500, warmup_ratio=0.001, step=[60000, 80000]) # Runner type runner = dict(_delete_=True, type='IterBasedRunner', max_iters=90000) checkpoint_config = dict(interval=1...
from ...utils import is_torch_available if is_torch_available(): from .auraflow_transformer_2d import AuraFlowTransformer2DModel from .cogvideox_transformer_3d import CogVideoXTransformer3DModel from .dit_transformer_2d import DiTTransformer2DModel from .dual_transformer_2d import DualTransformer2DMod...
from ...utils import is_torch_available if is_torch_available(): from .auraflow_transformer_2d import AuraFlowTransformer2DModel from .cogvideox_transformer_3d import CogVideoXTransformer3DModel from .dit_transformer_2d import DiTTransformer2DModel from .dual_transformer_2d import DualTransformer2DMod...
from backend.integrations.providers import ProviderName from backend.util.settings import Config app_config = Config() # TODO: add test to assert this matches the actual API route def webhook_ingress_url(provider_name: ProviderName, webhook_id: str) -> str: return ( f"{app_config.platform_base_url}/api/i...
from backend.integrations.providers import ProviderName from backend.util.settings import Config app_config = Config() # TODO: add test to assert this matches the actual API route def webhook_ingress_url(provider_name: ProviderName, webhook_id: str) -> str: return ( f"{app_config.platform_base_url}/api/i...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from mmengine.config import ConfigDict from mmengine.data import InstanceData from parameterized import parameterized from mmdet.data_elements.mask import mask_target from mmdet.models.roi_heads.mask_heads impor...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from mmengine.config import ConfigDict from mmengine.data import InstanceData from parameterized import parameterized from mmdet.core import mask_target from mmdet.models.roi_heads.mask_heads import MaskIoUHead ...
"""LLM Compiler Output Parser.""" import re from typing import Any, Dict, List, Sequence from llama_index.core.tools import BaseTool from llama_index.core.types import BaseOutputParser from .schema import JoinerOutput, LLMCompilerParseResult from .utils import get_graph_dict THOUGHT_PATTERN = r"Thought: ([^\n]*)" A...
"""LLM Compiler Output Parser.""" import re from typing import Any, Dict, List, Sequence from llama_index.core.tools import BaseTool from llama_index.core.types import BaseOutputParser from .schema import JoinerOutput, LLMCompilerParseResult from .utils import get_graph_dict THOUGHT_PATTERN = r"Thought: ([^\n]*)" A...
from typing import Union, Iterable from docarray.array.storage.base.seqlike import BaseSequenceLikeMixin from docarray.array.memory import DocumentArrayInMemory from docarray import Document class SequenceLikeMixin(BaseSequenceLikeMixin): """Implement sequence-like methods""" def _extend(self, values: Itera...
from typing import Union, Iterable from docarray.array.storage.base.seqlike import BaseSequenceLikeMixin from docarray.array.memory import DocumentArrayInMemory from docarray import Document class SequenceLikeMixin(BaseSequenceLikeMixin): """Implement sequence-like methods""" def extend(self, values: Iterab...
from prisma.models import User from backend.blocks.basic import AgentInputBlock, PrintToConsoleBlock from backend.blocks.text import FillTextTemplateBlock from backend.data import graph from backend.data.graph import create_graph from backend.data.user import get_or_create_user from backend.util.test import SpinTestSe...
from prisma.models import User from backend.blocks.basic import AgentInputBlock, PrintToConsoleBlock from backend.blocks.text import FillTextTemplateBlock from backend.data import graph from backend.data.graph import create_graph from backend.data.user import get_or_create_user from backend.util.test import SpinTestSe...
__version__ = '0.15.2' import os from docarray.document import Document from docarray.array import DocumentArray from docarray.dataclasses import dataclass, field if 'DA_RICH_HANDLER' in os.environ: from rich.traceback import install install()
__version__ = '0.15.1' import os from docarray.document import Document from docarray.array import DocumentArray from docarray.dataclasses import dataclass, field if 'DA_RICH_HANDLER' in os.environ: from rich.traceback import install install()
from docarray.proto.pb2.docarray_pb2 import DocumentArrayProto, DocumentProto, NdArrayProto, NodeProto
from .pb2.docarray_pb2 import DocumentArrayProto, DocumentProto, NdArrayProto, NodeProto
_base_ = './ga-rpn_r50-caffe_fpn_1x_coco.py' # model settings model = dict( backbone=dict( depth=101, init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://detectron2/resnet101_caffe')))
_base_ = './ga_rpn_r50_caffe_fpn_1x_coco.py' # model settings model = dict( backbone=dict( depth=101, init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://detectron2/resnet101_caffe')))
from urllib.parse import quote from backend.blocks.jina._auth import ( JinaCredentials, JinaCredentialsField, JinaCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField from backend.util.request import requests class Fa...
from urllib.parse import quote import requests from backend.blocks.jina._auth import ( JinaCredentials, JinaCredentialsField, JinaCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField class FactCheckerBlock(Block): ...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import cv2 import mmcv import numpy as np import torch import torch.nn as nn from mmcv.transforms import Compose from mmengine.utils import track_iter_progress from mmdet.apis import init_detector from mmdet.registry import VISUALIZERS from mmdet.structu...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import cv2 import mmcv import numpy as np import torch import torch.nn as nn from mmcv.transforms import Compose from mmdet.apis import init_detector from mmdet.registry import VISUALIZERS from mmdet.structures import DetDataSample from mmdet.utils impor...
from keras.src import activations from keras.src import applications from keras.src import backend from keras.src import constraints from keras.src import datasets from keras.src import initializers from keras.src import layers from keras.src import models from keras.src import ops from keras.src import optimizers from...
from keras.src import activations from keras.src import applications from keras.src import backend from keras.src import constraints from keras.src import datasets from keras.src import initializers from keras.src import layers from keras.src import models from keras.src import ops from keras.src import optimizers from...
# Copyright (c) OpenMMLab. All rights reserved. from abc import ABCMeta, abstractmethod import torch import torch.nn as nn from mmcv import ops from mmcv.runner import BaseModule class BaseRoIExtractor(BaseModule, metaclass=ABCMeta): """Base class for RoI extractor. Args: roi_layer (dict): Specify R...
from abc import ABCMeta, abstractmethod import torch import torch.nn as nn from mmcv import ops from mmcv.runner import BaseModule class BaseRoIExtractor(BaseModule, metaclass=ABCMeta): """Base class for RoI extractor. Args: roi_layer (dict): Specify RoI layer type and arguments. out_channel...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Dict, Optional, Sequence from ..registry import HOOKS from ..utils import get_git_hash from .hook import Hook DATA_BATCH = Optional[Sequence[dict]] @HOOKS.register_module() class RuntimeInfoHook(Hook): """A hook that updates runtime information ...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Dict, Optional, Sequence from ..registry import HOOKS from ..utils import get_git_hash from .hook import Hook DATA_BATCH = Optional[Sequence[dict]] @HOOKS.register_module() class RuntimeInfoHook(Hook): """A hook that updates runtime information ...
from ._bounding_box import BoundingBox, BoundingBoxFormat from ._datapoint import FillType, FillTypeJIT, InputType, InputTypeJIT from ._image import Image, ImageType, ImageTypeJIT, TensorImageType, TensorImageTypeJIT from ._label import Label, OneHotLabel from ._mask import Mask from ._video import TensorVideoType, Ten...
from ._bounding_box import BoundingBox, BoundingBoxFormat from ._datapoint import FillType, FillTypeJIT, InputType, InputTypeJIT from ._image import ColorSpace, Image, ImageType, ImageTypeJIT, TensorImageType, TensorImageTypeJIT from ._label import Label, OneHotLabel from ._mask import Mask from ._video import TensorVi...
# Copyright (c) OpenMMLab. All rights reserved. from .compare import (assert_allclose, assert_attrs_equal, assert_dict_contains_subset, assert_dict_has_keys, assert_is_norm_layer, assert_keys_equal, assert_params_all_zeros, check_python_script) __all__ ...
# Copyright (c) OpenMMLab. All rights reserved. from .compare import assert_allclose __all__ = ['assert_allclose']
# pylint: disable=invalid-name,unused-import """For compatibility and optional dependencies.""" import importlib.util import logging import sys import types from typing import Any, Sequence, cast import numpy as np from ._typing import _T assert sys.version_info[0] == 3, "Python 2 is no longer supported." def py_s...
# pylint: disable=invalid-name,unused-import """For compatibility and optional dependencies.""" import importlib.util import logging import sys import types from typing import Any, Sequence, cast import numpy as np from ._typing import _T assert sys.version_info[0] == 3, "Python 2 is no longer supported." def py_s...
import logging import aiohttp from fastapi import APIRouter from backend.util.settings import Settings from .models import TurnstileVerifyRequest, TurnstileVerifyResponse logger = logging.getLogger(__name__) router = APIRouter() settings = Settings() @router.post( "/verify", response_model=TurnstileVerifyRes...
import logging import aiohttp from fastapi import APIRouter from backend.util.settings import Settings from .models import TurnstileVerifyRequest, TurnstileVerifyResponse logger = logging.getLogger(__name__) router = APIRouter() settings = Settings() @router.post("/verify", response_model=TurnstileVerifyResponse...