input
stringlengths
33
5k
output
stringlengths
32
5k
"""Schemas for tracers.""" from __future__ import annotations import warnings from datetime import datetime, timezone from typing import Any, Optional from uuid import UUID from langsmith import RunTree from langsmith.schemas import RunTypeEnum as RunTypeEnumDep from pydantic import PydanticDeprecationWarning from p...
"""Schemas for tracers.""" from __future__ import annotations import datetime import warnings from typing import Any, Optional from uuid import UUID from langsmith import RunTree from langsmith.schemas import RunTypeEnum as RunTypeEnumDep from pydantic import PydanticDeprecationWarning from pydantic.v1 import BaseMo...
from abc import abstractmethod from typing import TYPE_CHECKING, Dict, Iterable, Type from pydantic.fields import ModelField if TYPE_CHECKING: from docarray.document.mixins.proto import ProtoMixin class AbstractDocument(Iterable): __fields__: Dict[str, ModelField] @classmethod @abstractmethod d...
from typing import Dict, Iterable from pydantic.fields import ModelField class AbstractDocument(Iterable): __fields__: Dict[str, ModelField]
import json from json import JSONDecodeError from typing import Union from langchain_core.agents import AgentAction, AgentActionMessageLog, AgentFinish from langchain_core.exceptions import OutputParserException from langchain_core.messages import ( AIMessage, BaseMessage, ToolCall, ) from langchain_core.o...
import json from json import JSONDecodeError from typing import Union from langchain_core.agents import AgentAction, AgentActionMessageLog, AgentFinish from langchain_core.exceptions import OutputParserException from langchain_core.messages import ( AIMessage, BaseMessage, ToolCall, ) from langchain_core.o...
import numpy as np import pytest from numpy.testing import assert_allclose from sklearn.datasets import load_iris from sklearn.linear_model import LogisticRegression from sklearn.metrics import DetCurveDisplay, det_curve @pytest.mark.parametrize("constructor_name", ["from_estimator", "from_predictions"]) @pytest.mar...
import numpy as np import pytest from numpy.testing import assert_allclose from sklearn.datasets import load_iris from sklearn.linear_model import LogisticRegression from sklearn.metrics import DetCurveDisplay, det_curve @pytest.mark.parametrize("constructor_name", ["from_estimator", "from_predictions"]) @pytest.mar...
from __future__ import annotations import os from copy import deepcopy import numpy as np import pytest from tokenizers import Tokenizer from sentence_transformers import SentenceTransformer from sentence_transformers.models import Pooling, StaticEmbedding, Transformer from sentence_transformers.util import is_datas...
from __future__ import annotations import os from copy import deepcopy import numpy as np import pytest from tokenizers import Tokenizer from sentence_transformers import SentenceTransformer from sentence_transformers.models import Pooling, StaticEmbedding, Transformer from sentence_transformers.util import is_datas...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '2.21.0' short_version = __version__ def parse_version_info(version_str): version_info = [] for x in version_str.split('.'): if x.isdigit(): version_info.append(int(x)) elif x.find('rc') != -1: patch_version...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '2.20.0' short_version = __version__ def parse_version_info(version_str): version_info = [] for x in version_str.split('.'): if x.isdigit(): version_info.append(int(x)) elif x.find('rc') != -1: patch_version...
from ._vggish import VGGISH, VGGishBundle from .hifigan_pipeline import HIFIGAN_VOCODER_V3_LJSPEECH as _HIFIGAN_VOCODER_V3_LJSPEECH, HiFiGANVocoderBundle from .rnnt_pipeline import ( EMFORMER_RNNT_BASE_MUSTC as _EMFORMER_RNNT_BASE_MUSTC, EMFORMER_RNNT_BASE_TEDLIUM3 as _EMFORMER_RNNT_BASE_TEDLIUM3 ) from torchau...
from ._vggish import VGGISH, VGGishBundle from .hifigan_pipeline import HIFIGAN_VOCODER_V3_LJSPEECH, HiFiGANVocoderBundle from .rnnt_pipeline import EMFORMER_RNNT_BASE_MUSTC, EMFORMER_RNNT_BASE_TEDLIUM3 __all__ = [ "EMFORMER_RNNT_BASE_MUSTC", "EMFORMER_RNNT_BASE_TEDLIUM3", "HIFIGAN_VOCODER_V3_LJSPEECH", ...
_base_ = './cascade-rcnn_r50_fpn_8xb8-amp-lsj-200e_coco.py' model = dict( backbone=dict( depth=18, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet18')), neck=dict(in_channels=[64, 128, 256, 512]))
_base_ = './cascade_rcnn_r50_fpn_lsj_200e_8x8_fp16_coco.py' model = dict( backbone=dict( depth=18, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet18')), neck=dict(in_channels=[64, 128, 256, 512]))
import numpy as np from docarray.proto import DocumentProto, NdArrayProto, NodeProto from docarray.typing import Tensor def test_nested_item_proto(): NodeProto(text='hello') NodeProto(nested=DocumentProto()) def test_nested_optional_item_proto(): NodeProto() def test_ndarray(): nd_proto = NdArray...
import numpy as np from docarray.proto import DocumentProto, NdArrayProto, NodeProto from docarray.proto.io import flush_ndarray, read_ndarray def test_nested_item_proto(): NodeProto(text='hello') NodeProto(nested=DocumentProto()) def test_nested_optional_item_proto(): NodeProto() def test_ndarray():...
from typing import Any, Optional from llama_index.core.base.base_query_engine import BaseQueryEngine from llama_index.core.evaluation import ( AnswerRelevancyEvaluator, BaseEvaluator, EvaluationResult, ) from llama_index.core.tools import QueryEngineTool from llama_index.core.tools.types import ToolMetadat...
from typing import Any, Optional from llama_index.core.base.base_query_engine import BaseQueryEngine from llama_index.core.evaluation import ( AnswerRelevancyEvaluator, BaseEvaluator, EvaluationResult, ) from llama_index.core.tools import QueryEngineTool from llama_index.core.tools.types import ToolMetadat...
_base_ = '../_base_/default_runtime.py' # dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' # Example to use different file client # Method 1: simply set the data root and let the file I/O module # automatically infer from prefix (not support LMDB and Memcache yet) # data_root = 's3://openmmlab/...
_base_ = '../_base_/default_runtime.py' # dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' # file_client_args = dict( # backend='petrel', # path_mapping=dict({ # './data/': 's3://openmmlab/datasets/detection/', # 'data/': 's3://openmmlab/datasets/detection/' # })) fil...
from typing import TYPE_CHECKING, Any, Dict, Optional, TypeVar import numpy as np from pydantic import parse_obj_as from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.ndarray import NdArray from docarray.typing.url.url_3d.url_3d import Url3D if TYPE_CHECKING: from docarray.doc...
from typing import TYPE_CHECKING, Any, Dict, Optional, TypeVar import numpy as np from pydantic import parse_obj_as from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.ndarray import NdArray from docarray.typing.url.url_3d.url_3d import Url3D if TYPE_CHECKING: from docarray.doc...
from .base import ElevenLabsVoiceAgent, ElevenLabsVoiceAgentInterface __all__ = ["ElevenLabsVoiceAgent", "ElevenLabsVoiceAgentInterface"]
from .base import ElevenLabsConversation __all__ = ["ElevenLabsConversation"]
import mimetypes from typing import TYPE_CHECKING, Optional from docarray.document.mixins._property import _PropertyMixin if TYPE_CHECKING: from docarray.typing import DocumentContentType, ArrayType from docarray import DocumentArray _all_mime_types = set(mimetypes.types_map.values()) class PropertyMixin(_...
import mimetypes from typing import TYPE_CHECKING, Optional from ._property import _PropertyMixin if TYPE_CHECKING: from ...typing import DocumentContentType, ArrayType from ... import DocumentArray _all_mime_types = set(mimetypes.types_map.values()) class PropertyMixin(_PropertyMixin): def _clear_cont...
# Copyright (c) OpenMMLab. All rights reserved. import os import os.path as osp from pathlib import Path from .misc import is_str def is_filepath(x): return is_str(x) or isinstance(x, Path) def fopen(filepath, *args, **kwargs): if is_str(filepath): return open(filepath, *args, **kwargs) elif is...
# Copyright (c) OpenMMLab. All rights reserved. import os import os.path as osp from pathlib import Path from .misc import is_str def is_filepath(x): return is_str(x) or isinstance(x, Path) def fopen(filepath, *args, **kwargs): if is_str(filepath): return open(filepath, *args, **kwargs) elif is...
from docarray.typing.id import ID from docarray.typing.tensor.audio import AudioNdArray from docarray.typing.tensor.embedding.embedding import AnyEmbedding from docarray.typing.tensor.ndarray import NdArray from docarray.typing.tensor.tensor import AnyTensor from docarray.typing.tensor.video import VideoNdArray from do...
from docarray.typing.id import ID from docarray.typing.tensor.audio import AudioNdArray from docarray.typing.tensor.embedding.embedding import AnyEmbedding from docarray.typing.tensor.ndarray import NdArray from docarray.typing.tensor.tensor import AnyTensor from docarray.typing.url import ( AnyUrl, AudioUrl, ...
# Copyright (c) OpenMMLab. All rights reserved. from .checkloss_hook import CheckInvalidLossHook from .ema import ExpMomentumEMAHook, LinearMomentumEMAHook from .set_epoch_info_hook import SetEpochInfoHook from .sync_norm_hook import SyncNormHook from .sync_random_size_hook import SyncRandomSizeHook from .yolox_lrupdat...
# Copyright (c) OpenMMLab. All rights reserved. from .checkloss_hook import CheckInvalidLossHook from .ema import ExpMomentumEMAHook, LinearMomentumEMAHook from .sync_norm_hook import SyncNormHook from .sync_random_size_hook import SyncRandomSizeHook from .yolox_lrupdater_hook import YOLOXLrUpdaterHook from .yolox_mode...
""" Implements the Generalized R-CNN framework """ import warnings from collections import OrderedDict from typing import Optional, Union import torch from torch import nn from ...utils import _log_api_usage_once class GeneralizedRCNN(nn.Module): """ Main class for Generalized R-CNN. Args: bac...
""" Implements the Generalized R-CNN framework """ import warnings from collections import OrderedDict from typing import Dict, List, Optional, Tuple, Union import torch from torch import nn, Tensor from ...utils import _log_api_usage_once class GeneralizedRCNN(nn.Module): """ Main class for Generalized R-...
# pants requires this import to recognize the dep import pytest_asyncio # noqa: F401 import pytest import os from llama_index.embeddings.nvidia import NVIDIAEmbedding as Interface from llama_index.embeddings.nvidia.base import DEFAULT_MODEL from typing import Generator # this fixture is used to mask the NVIDIA_AP...
# pants requires this import to recognize the dep import pytest_asyncio # noqa: F401 import pytest import os from llama_index.embeddings.nvidia import NVIDIAEmbedding as Interface from llama_index.embeddings.nvidia.base import DEFAULT_MODEL from typing import Generator # this fixture is used to mask the NVIDIA_AP...
_base_ = ['co_dino_5scale_r50_8xb2_1x_coco.py'] pretrained = 'https://github.com/SwinTransformer/storage/releases/download/v1.0.0/swin_large_patch4_window12_384_22k.pth' # noqa load_from = 'https://download.openmmlab.com/mmdetection/v3.0/codetr/co_dino_5scale_swin_large_16e_o365tococo-614254c9.pth' # noqa # model s...
_base_ = ['co_dino_5scale_r50_8xb2_1x_coco.py'] pretrained = 'https://github.com/SwinTransformer/storage/releases/download/v1.0.0/swin_large_patch4_window12_384_22k.pth' # noqa load_from = 'https://download.openmmlab.com/mmdetection/v3.0/codetr/co_dino_5scale_swin_large_22e_o365-0a33e247.pth' # noqa # model setting...
# Copyright (c) Meta Platforms, Inc. and affiliates # Owner(s): ["oncall: distributed"] from model_registry import MLPModule, ModelWithParamAlias import torch from torch.distributed.pipelining import pipe_split, pipeline from torch.testing._internal.common_utils import ( instantiate_parametrized_tests, paramet...
# Copyright (c) Meta Platforms, Inc. and affiliates # Owner(s): ["oncall: distributed"] from model_registry import MLPModule, ModelWithParamAlias import torch from torch.distributed.pipelining import pipe_split, pipeline from torch.testing._internal.common_utils import ( instantiate_parametrized_tests, paramet...
"""Init file.""" from llama_index.readers.web.agentql_web.base import ( AgentQLWebReader, ) from llama_index.readers.web.async_web.base import ( AsyncWebPageReader, ) from llama_index.readers.web.beautiful_soup_web.base import ( BeautifulSoupWebReader, ) from llama_index.readers.web.browserbase_web.base im...
"""Init file.""" from llama_index.readers.web.agentql_web.base import ( AgentQLWebReader, ) from llama_index.readers.web.async_web.base import ( AsyncWebPageReader, ) from llama_index.readers.web.beautiful_soup_web.base import ( BeautifulSoupWebReader, ) from llama_index.readers.web.browserbase_web.base imp...
import argparse import urllib from abc import ABC from http import HTTPStatus from typing import TYPE_CHECKING, Optional, Union from jina.serve.runtimes.asyncio import AsyncNewLoopRuntime if TYPE_CHECKING: import asyncio import multiprocessing import threading class GatewayRuntime(AsyncNewLoopRuntime, A...
import argparse from abc import ABC from typing import TYPE_CHECKING, Optional, Union from jina.serve.runtimes.asyncio import AsyncNewLoopRuntime if TYPE_CHECKING: import asyncio import multiprocessing import threading class GatewayRuntime(AsyncNewLoopRuntime, ABC): """ The Runtime from which th...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.quantizers import deserialize as deserialize from keras.src.quantizers import get as get from keras.src.quantizers import serialize as serialize from keras.src.quantizers.quantizers i...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.quantizers import deserialize as deserialize from keras.src.quantizers import get as get from keras.src.quantizers import serialize as serialize from keras.src.quantizers.quantizers i...
from backend.blocks.jina._auth import ( JinaCredentials, JinaCredentialsField, JinaCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField from backend.util.request import requests class JinaChunkingBlock(Block): clas...
import requests from backend.blocks.jina._auth import ( JinaCredentials, JinaCredentialsField, JinaCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField class JinaChunkingBlock(Block): class Input(BlockSchema): ...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.core import ConfigType, OptConfigType, OptMultiConfig from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class FSAF(SingleStageDetector): """Implementation of `FSAF <https://arxiv.org/abs/1903.00621>`...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.core import ConfigType, OptConfigType, OptMultiConfig from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class FSAF(SingleStageDetector): """Implementation of `FSAF <https://arxiv.org/abs/1903.00621>`...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.document_loaders import WhatsAppChatLoader from langchain_community.document_loaders.whatsapp_chat import concatenate_rows # Create a way to dynamically look up deprecated imports. # Us...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.document_loaders import WhatsAppChatLoader from langchain_community.document_loaders.whatsapp_chat import concatenate_rows # Create a way to dynamically look up deprecated imports. # Us...
import logging import os import json from typing import Any, List, Optional, cast from llama_index.core.base.base_retriever import BaseRetriever from llama_index.core.constants import DEFAULT_SIMILARITY_TOP_K from llama_index.core.schema import NodeWithScore, QueryBundle, TextNode from llama_index.core.vector_stores.u...
import logging import os import json from typing import Any, List, Optional, cast from llama_index.core.base.base_retriever import BaseRetriever from llama_index.core.constants import DEFAULT_SIMILARITY_TOP_K from llama_index.core.schema import NodeWithScore, QueryBundle, TextNode from llama_index.core.vector_stores.u...
# ReAct agent formatter import logging from abc import abstractmethod from typing import List, Optional, Sequence from llama_index.core.agent.react.prompts import ( CONTEXT_REACT_CHAT_SYSTEM_HEADER, REACT_CHAT_SYSTEM_HEADER, ) from llama_index.core.agent.react.types import ( BaseReasoningStep, Observa...
# ReAct agent formatter import logging from abc import abstractmethod from typing import List, Optional, Sequence from llama_index.core.agent.react.prompts import ( CONTEXT_REACT_CHAT_SYSTEM_HEADER, REACT_CHAT_SYSTEM_HEADER, ) from llama_index.core.agent.react.types import ( BaseReasoningStep, Observa...
from typing import TYPE_CHECKING, Optional if TYPE_CHECKING: from docarray.proto import NodeProto from docarray.typing.url.any_url import AnyUrl from docarray.typing.url.helper import _uri_to_blob class TextUrl(AnyUrl): """ URL to a text file. Cane be remote (web) URL, or a local file path. """ ...
from typing import TYPE_CHECKING, Optional if TYPE_CHECKING: from docarray.proto import NodeProto from docarray.typing.url.any_url import AnyUrl from docarray.typing.url.helper import _uri_to_blob class TextUrl(AnyUrl): """ URL to a text file. Cane be remote (web) URL, or a local file path. """ ...
# Initialize extension and backend first from . import ( # noqa # usort: skip _extension, _backend, ) from . import ( # noqa: F401 backend, # For BC compliance, datasets, functional, io, kaldi_io, models, pipelines, sox_effects, transforms, utils, ) from ._backend...
from . import ( # noqa: F401 _extension, compliance, datasets, functional, io, kaldi_io, models, pipelines, sox_effects, transforms, utils, ) from ._backend.common import AudioMetaData # noqa try: from .version import __version__, git_version # noqa: F401 except Impor...
# Licensed to the LF AI & Data foundation under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the "License"); # you may not use this fil...
from abc import abstractmethod from typing import TYPE_CHECKING, Any, Type, TypeVar from pydantic import parse_obj_as from docarray.typing.abstract_type import AbstractType from docarray.utils._internal.pydantic import bytes_validator, is_pydantic_v2 if is_pydantic_v2: from pydantic_core import core_schema if T...
""" ================================== Getting started with transforms v2 ================================== Most computer vision tasks are not supported out of the box by ``torchvision.transforms`` v1, since it only supports images. ``torchvision.transforms.v2`` enables jointly transforming images, videos, bounding b...
""" ================================== Getting started with transforms v2 ================================== Most computer vision tasks are not supported out of the box by ``torchvision.transforms`` v1, since it only supports images. ``torchvision.transforms.v2`` enables jointly transforming images, videos, bounding b...
# Copyright 2024 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
# Copyright 2024 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
from __future__ import annotations import os from typing import Any, Callable, Optional, Tuple import PIL.Image from .utils import download_and_extract_archive, verify_str_arg from .vision import VisionDataset class FGVCAircraft(VisionDataset): """`FGVC Aircraft <https://www.robots.ox.ac.uk/~vgg/data/fgvc-airc...
from __future__ import annotations import os from typing import Any, Callable, Optional, Tuple import PIL.Image from .utils import download_and_extract_archive, verify_str_arg from .vision import VisionDataset class FGVCAircraft(VisionDataset): """`FGVC Aircraft <https://www.robots.ox.ac.uk/~vgg/data/fgvc-airc...
import pathlib from typing import Any, Dict, List, Optional, Tuple, Union from torchdata.datapipes.iter import CSVDictParser, Demultiplexer, Filter, IterDataPipe, Mapper, Zipper from torchvision.prototype.datapoints import BoundingBox, Label from torchvision.prototype.datasets.utils import Dataset, EncodedImage, HttpR...
import pathlib from typing import Any, Dict, List, Optional, Tuple, Union from torchdata.datapipes.iter import CSVDictParser, Demultiplexer, Filter, IterDataPipe, Mapper, Zipper from torchvision.prototype.datasets.utils import Dataset, EncodedImage, HttpResource, OnlineResource from torchvision.prototype.datasets.util...
# ruff: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICE...
# ruff: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICE...
import pytest from jina import Document, Flow @pytest.mark.parametrize('endpoint', ['foo', 'bar']) @pytest.mark.parametrize( 'uses', ['jinaai+sandbox://jina-ai/Hello'] ) def test_sandbox(endpoint, uses): with Flow().add(uses=uses) as f: da = f.post( endpoint, [ ...
import pytest from jina import Document, Flow @pytest.mark.parametrize('endpoint', ['foo', 'bar']) def test_sandbox(endpoint): with Flow().add(uses='jinahub+sandbox://Hello') as f: da = f.post( endpoint, [ Document(text="dog world"), Document(text="...
import numpy as np import pytest import scipy.ndimage import torch from whisper.timing import dtw_cpu, dtw_cuda, median_filter sizes = [ (10, 20), (32, 16), (123, 1500), (234, 189), ] shapes = [ (10,), (1, 15), (4, 5, 345), (6, 12, 240, 512), ] @pytest.mark.parametrize("N, M", sizes)...
import pytest import numpy as np import scipy.ndimage import torch from whisper.timing import dtw_cpu, dtw_cuda, median_filter sizes = [ (10, 20), (32, 16), (123, 1500), (234, 189), ] shapes = [ (10,), (1, 15), (4, 5, 345), (6, 12, 240, 512), ] @pytest.mark.parametrize("N, M", sizes) def test_dtw(N: int, ...
from keras.src import backend from keras.src import ops from keras.src.api_export import keras_export from keras.src.layers.layer import Layer from keras.src.saving.serialization_lib import deserialize_keras_object @keras_export("keras.layers.Masking") class Masking(Layer): """Masks a sequence by using a mask val...
from keras.src import backend from keras.src import ops from keras.src.api_export import keras_export from keras.src.layers.layer import Layer @keras_export("keras.layers.Masking") class Masking(Layer): """Masks a sequence by using a mask value to skip timesteps. For each timestep in the input tensor (dimens...
"""LLMResult class.""" from __future__ import annotations from copy import deepcopy from typing import Literal, Optional, Union from pydantic import BaseModel from langchain_core.outputs.chat_generation import ChatGeneration, ChatGenerationChunk from langchain_core.outputs.generation import Generation, GenerationCh...
"""LLMResult class.""" from __future__ import annotations from copy import deepcopy from typing import Literal, Optional, Union from pydantic import BaseModel from langchain_core.outputs.chat_generation import ChatGeneration, ChatGenerationChunk from langchain_core.outputs.generation import Generation, GenerationCh...
from __future__ import annotations from collections.abc import Sequence from typing import Any, Optional from langchain_core.callbacks import ( AsyncCallbackManagerForChainRun, CallbackManagerForChainRun, ) from langchain_core.documents import Document from langchain_core.embeddings import Embeddings from lan...
from __future__ import annotations from collections.abc import Sequence from typing import Any, Optional from langchain_core.callbacks import ( AsyncCallbackManagerForChainRun, CallbackManagerForChainRun, ) from langchain_core.documents import Document from langchain_core.embeddings import Embeddings from lan...
# Copyright (c) OpenMMLab. All rights reserved. import warnings from typing import Tuple import torch.nn as nn import torch.nn.functional as F from mmcv.cnn import ConvModule from mmengine.config import ConfigDict from mmengine.model import BaseModule from torch import Tensor from mmdet.registry import MODELS from mm...
# Copyright (c) OpenMMLab. All rights reserved. import warnings from typing import Tuple import torch.nn as nn import torch.nn.functional as F from mmcv.cnn import ConvModule from mmengine.config import ConfigDict from mmengine.model import BaseModule from torch import Tensor from mmdet.registry import MODELS from mm...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Sequence from torch.utils.data import BatchSampler, Sampler from mmdet.datasets.samplers.track_img_sampler import TrackImgSampler from mmdet.registry import DATA_SAMPLERS # TODO: maybe replace with a data_loader wrapper @DATA_SAMPLERS.register_modul...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Sequence from torch.utils.data import BatchSampler, Sampler from mmdet.registry import DATA_SAMPLERS # TODO: maybe replace with a data_loader wrapper @DATA_SAMPLERS.register_module() class AspectRatioBatchSampler(BatchSampler): """A sampler wrap...
"""Benchmarks of Lasso regularization path computation using Lars and CD The input data is mostly low rank but is a fat infinite tail. """ import gc import sys from collections import defaultdict from time import time import numpy as np from sklearn.datasets import make_regression from sklearn.linear_model import l...
"""Benchmarks of Lasso regularization path computation using Lars and CD The input data is mostly low rank but is a fat infinite tail. """ import gc import sys from collections import defaultdict from time import time import numpy as np from sklearn.datasets import make_regression from sklearn.linear_model import l...
""" =================================================== Recursive feature elimination with cross-validation =================================================== A Recursive Feature Elimination (RFE) example with automatic tuning of the number of features selected with cross-validation. """ # Authors: The scikit-learn...
""" =================================================== Recursive feature elimination with cross-validation =================================================== A Recursive Feature Elimination (RFE) example with automatic tuning of the number of features selected with cross-validation. """ # Authors: The scikit-learn...
_base_ = './fcos_hrnetv2p_w32_gn-head_4x4_1x_coco.py' # learning policy max_epochs = 24 train_cfg = dict(max_epochs=max_epochs) param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_epochs, ...
_base_ = './fcos_hrnetv2p_w32_gn-head_4x4_1x_coco.py' # learning policy lr_config = dict(step=[16, 22]) runner = dict(type='EpochBasedRunner', max_epochs=24)
"""Integration test for Sms.""" from langchain_community.utilities.twilio import TwilioAPIWrapper def test_call() -> None: """Test that call runs.""" twilio = TwilioAPIWrapper() output = twilio.run("Message", "+16162904619") assert output
"""Integration test for Sms.""" from langchain_community.utilities.twilio import TwilioAPIWrapper def test_call() -> None: """Test that call runs.""" twilio = TwilioAPIWrapper() # type: ignore[call-arg] output = twilio.run("Message", "+16162904619") assert output
from keras.src import testing from keras.src import tree from keras.src.backend import KerasTensor from keras.src.ops.symbolic_arguments import SymbolicArguments class SymbolicArgumentsTest(testing.TestCase): # Testing multiple args and empty kwargs def test_args(self): shape = (2, 3, 4) a = K...
from keras.src import testing from keras.src import tree from keras.src.backend import KerasTensor from keras.src.ops.symbolic_arguments import SymbolicArguments class SymbolicArgumentsTest(testing.TestCase): # Testing multiple args and empty kwargs def test_args(self): shape = (2, 3, 4) a = K...
import tantivy # noqa from llama_index.core.vector_stores.types import BasePydanticVectorStore from llama_index.vector_stores.lancedb import LanceDBVectorStore from llama_index.core import VectorStoreIndex import pytest import lance # noqa: F401 import pytest import pytest_asyncio from llama_index.core import VectorS...
import tantivy # noqa from llama_index.core.vector_stores.types import BasePydanticVectorStore from llama_index.vector_stores.lancedb import LanceDBVectorStore from llama_index.core import VectorStoreIndex def test_class(): names_of_base_classes = [b.__name__ for b in LanceDBVectorStore.__mro__] assert BaseP...
_base_ = [ '../common/ms-poly_3x_coco-instance.py', '../_base_/models/mask-rcnn_r50_fpn.py' ] model = dict( backbone=dict( _delete_=True, type='RegNet', arch='regnetx_400mf', out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=Tr...
_base_ = [ '../common/mstrain-poly_3x_coco_instance.py', '../_base_/models/mask_rcnn_r50_fpn.py' ] model = dict( backbone=dict( _delete_=True, type='RegNet', arch='regnetx_400mf', out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_gr...
# dataset settings dataset_type = 'MOTChallengeDataset' data_root = 'data/MOT17/' img_scale = (1088, 1088) backend_args = None # data pipeline train_pipeline = [ dict( type='UniformRefFrameSample', num_ref_imgs=1, frame_range=10, filter_key_img=True), dict( type='Transfo...
# dataset settings dataset_type = 'MOTChallengeDataset' data_root = 'data/MOT17/' img_scale = (1088, 1088) # data pipeline train_pipeline = [ dict( type='UniformRefFrameSample', num_ref_imgs=1, frame_range=10, filter_key_img=True), dict( type='TransformBroadcaster', ...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase import torch from mmengine.data import InstanceData from mmdet.models.dense_heads import AutoAssignHead class TestAutoAssignHead(TestCase): def test_autoassign_head_loss(self): """Tests autoassign head loss when truth is empt...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase import torch from mmengine.data import InstanceData from mmdet.models.dense_heads import AutoAssignHead class TestAutoAssignHead(TestCase): def test_autoassign_head_loss(self): """Tests autoassign head loss when truth is empt...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import ( BaseRequestsTool, RequestsDeleteTool, RequestsGetTool, RequestsPatchTool, RequestsPostTool, RequestsPutTool, ) # Create a ...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import ( BaseRequestsTool, RequestsDeleteTool, RequestsGetTool, RequestsPatchTool, RequestsPostTool, RequestsPutTool, ) # Create a ...
import warnings from typing import Any, Dict, List, Union import numpy as np import PIL.Image import torch from torchvision.prototype import datapoints from torchvision.prototype.transforms import Transform from torchvision.transforms import functional as _F from typing_extensions import Literal from ._transform imp...
import warnings from typing import Any, Dict, List, Union import numpy as np import PIL.Image import torch from torchvision.prototype import datapoints from torchvision.prototype.transforms import Transform from torchvision.transforms import functional as _F from typing_extensions import Literal from ._transform imp...
import os import pytest import time import uuid import pinecone.db_data from pinecone import Pinecone, ServerlessSpec from typing import List from llama_index.core import StorageContext, VectorStoreIndex from llama_index.core.embeddings import MockEmbedding from llama_index.core.schema import TextNode from llama_inde...
from llama_index.core.vector_stores.types import BasePydanticVectorStore from llama_index.vector_stores.pinecone import PineconeVectorStore def test_class(): names_of_base_classes = [b.__name__ for b in PineconeVectorStore.__mro__] assert BasePydanticVectorStore.__name__ in names_of_base_classes
from typing import Any, Optional, Type, TypeVar, Union import numpy as np from docarray.base_document import BaseDocument from docarray.typing import AnyEmbedding, AudioUrl from docarray.typing.tensor.abstract_tensor import AbstractTensor from docarray.typing.tensor.audio.audio_tensor import AudioTensor try: imp...
from typing import Optional, TypeVar from docarray.base_document import BaseDocument from docarray.typing import AnyEmbedding, AudioUrl from docarray.typing.tensor.audio.audio_tensor import AudioTensor T = TypeVar('T', bound='Audio') class Audio(BaseDocument): """ Document for handling audios. The Audi...
import itertools import torch from parameterized import parameterized from torchaudio_unittest.common_utils import get_asset_path, skipIfNoCtcDecoder, TempDirMixin, TorchaudioTestCase NUM_TOKENS = 8 @skipIfNoCtcDecoder class CTCDecoderTest(TempDirMixin, TorchaudioTestCase): def _get_decoder(self, tokens=None, u...
import itertools import torch from parameterized import parameterized from torchaudio_unittest.common_utils import ( get_asset_path, skipIfNoCtcDecoder, TempDirMixin, TorchaudioTestCase, ) NUM_TOKENS = 8 @skipIfNoCtcDecoder class CTCDecoderTest(TempDirMixin, TorchaudioTestCase): def _get_decoder...
# THIS FILE HAS BEEN AUTOGENERATED. To update: # 1. modify the `_deps` dict in setup.py # 2. run `make deps_table_update`` deps = { "Pillow": "Pillow>=10.0.1,<=15.0", "accelerate": "accelerate>=0.26.0", "av": "av", "beautifulsoup4": "beautifulsoup4", "blobfile": "blobfile", "codecarbon": "codeca...
# THIS FILE HAS BEEN AUTOGENERATED. To update: # 1. modify the `_deps` dict in setup.py # 2. run `make deps_table_update`` deps = { "Pillow": "Pillow>=10.0.1,<=15.0", "accelerate": "accelerate>=0.26.0", "av": "av", "beautifulsoup4": "beautifulsoup4", "blobfile": "blobfile", "codecarbon": "codeca...
""" This script contains an example how to perform semantic search with OpenSearch. You need OpenSearch up and running locally: https://docs.opensearch.org/docs/latest/getting-started/quickstart/ Further, you need the Python OpenSearch Client installed: https://docs.opensearch.org/docs/latest/clients/python-low-level...
""" This script contains an example how to perform semantic search with OpenSearch. You need OpenSearch up and running locally: https://docs.opensearch.org/docs/latest/getting-started/quickstart/ Further, you need the Python OpenSearch Client installed: https://docs.opensearch.org/docs/latest/clients/python-low-level...
import math import random class NoDuplicatesDataLoader: def __init__(self, train_examples, batch_size): """ A special data loader to be used with MultipleNegativesRankingLoss. The data loader ensures that there are no duplicate sentences within the same batch """ self.batch...
import random import math class NoDuplicatesDataLoader: def __init__(self, train_examples, batch_size): """ A special data loader to be used with MultipleNegativesRankingLoss. The data loader ensures that there are no duplicate sentences within the same batch """ self.batch...
from __future__ import annotations try: from typing import Self except ImportError: from typing_extensions import Self import torch from torch import nn from sentence_transformers.models.Module import Module class CNN(Module): """CNN-layer with multiple kernel-sizes over the word embeddings""" con...
from __future__ import annotations import json import os import torch from safetensors.torch import load_model as load_safetensors_model from safetensors.torch import save_model as save_safetensors_model from torch import nn class CNN(nn.Module): """CNN-layer with multiple kernel-sizes over the word embeddings"...
import re from typing import Any, Optional from langchain_text_splitters import RecursiveCharacterTextSplitter class JSFrameworkTextSplitter(RecursiveCharacterTextSplitter): """Text splitter that handles React (JSX), Vue, and Svelte code. This splitter extends RecursiveCharacterTextSplitter to handle Re...
import re from typing import Any, List, Optional from langchain_text_splitters import RecursiveCharacterTextSplitter class JSFrameworkTextSplitter(RecursiveCharacterTextSplitter): """Text splitter that handles React (JSX), Vue, and Svelte code. This splitter extends RecursiveCharacterTextSplitter to handle ...
# Copyright (c) OpenMMLab. All rights reserved. import glob import os import os.path as osp import urllib import warnings from typing import Union import torch from mmengine.config import Config, ConfigDict from mmengine.logging import print_log from mmengine.utils import scandir IMG_EXTENSIONS = ('.jpg', '.jpeg', '....
# Copyright (c) OpenMMLab. All rights reserved. import glob import os import os.path as osp import warnings from typing import Union from mmengine.config import Config, ConfigDict from mmengine.logging import print_log def find_latest_checkpoint(path, suffix='pth'): """Find the latest checkpoint from the working...
# Copyright (c) OpenMMLab. All rights reserved. from .base_data_element import BaseDataElement from .instance_data import InstanceData from .label_data import LabelData from .pixel_data import PixelData from .sampler import DefaultSampler, InfiniteSampler from .utils import pseudo_collate, worker_init_fn __all__ = [ ...
# Copyright (c) OpenMMLab. All rights reserved. from .base_data_element import BaseDataElement from .instance_data import InstanceData from .sampler import DefaultSampler, InfiniteSampler from .utils import pseudo_collate, worker_init_fn __all__ = [ 'BaseDataElement', 'DefaultSampler', 'InfiniteSampler', 'worker_i...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.activations import deserialize from keras.src.activations import get from keras.src.activations import serialize from keras.src.activations.activations import celu from keras.src.acti...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.activations import deserialize from keras.src.activations import get from keras.src.activations import serialize from keras.src.activations.activations import celu from keras.src.acti...
from __future__ import annotations from collections.abc import Iterable from torch import Tensor from sentence_transformers import util from sentence_transformers.losses.MultipleNegativesRankingLoss import MultipleNegativesRankingLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder cla...
from __future__ import annotations from collections.abc import Iterable from torch import Tensor from sentence_transformers import util from sentence_transformers.losses.MultipleNegativesRankingLoss import MultipleNegativesRankingLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder cla...
import os import warnings from modulefinder import Module import torch from torchvision import datasets, io, models, ops, transforms, utils from .extension import _HAS_OPS, _load_library try: from .version import __version__ # noqa: F401 except ImportError: pass try: _load_library("Decoder") _HAS_G...
import os import warnings import torch from torchvision import datasets, io, models, ops, transforms, utils from .extension import _HAS_OPS try: from .version import __version__ # noqa: F401 except ImportError: pass # Check if torchvision is being imported within the root folder if not _HAS_OPS and os.path...
import pytest from absl.testing import parameterized from keras.src import backend from keras.src import layers from keras.src import testing from keras.src.backend.common.keras_tensor import KerasTensor class ReshapeTest(testing.TestCase): @parameterized.named_parameters( [ {"testcase_name":...
import pytest from absl.testing import parameterized from keras.src import backend from keras.src import layers from keras.src import testing from keras.src.backend.common.keras_tensor import KerasTensor class ReshapeTest(testing.TestCase, parameterized.TestCase): @parameterized.named_parameters( [ ...
import enum from typing import Any, List, Optional, Union import pydantic import backend.data.graph from backend.data.api_key import APIKeyPermission, APIKeyWithoutHash class Methods(enum.Enum): SUBSCRIBE = "subscribe" UNSUBSCRIBE = "unsubscribe" EXECUTION_EVENT = "execution_event" ERROR = "error" ...
import enum import typing import pydantic import backend.data.graph from backend.data.api_key import APIKeyPermission, APIKeyWithoutHash class Methods(enum.Enum): SUBSCRIBE = "subscribe" UNSUBSCRIBE = "unsubscribe" EXECUTION_EVENT = "execution_event" ERROR = "error" class WsMessage(pydantic.BaseMo...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp import tempfile from unittest import TestCase from unittest.mock import Mock import torch import torch.nn as nn from torch.utils.data import Dataset from mmengine.hooks import EMAHook from mmengine.model import BaseModel, ExponentialMovingAverage f...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp import tempfile from unittest import TestCase from unittest.mock import Mock import torch import torch.nn as nn from torch.utils.data import Dataset from mmengine.hooks import EMAHook from mmengine.model import ExponentialMovingAverage from mmengin...
# Copyright (c) OpenMMLab. All rights reserved. import os from typing import Optional import torch try: import torch_npu # noqa: F401 import torch_npu.npu.utils as npu_utils # Enable operator support for dynamic shape and # binary operator support on the NPU. npu_jit_compile = bool(os.getenv('NP...
# Copyright (c) OpenMMLab. All rights reserved. import os from typing import Optional import torch try: import torch_npu # noqa: F401 import torch_npu.npu.utils as npu_utils # Enable operator support for dynamic shape and # binary operator support on the NPU. npu_jit_compile = bool(os.getenv('NP...
from argparse import Namespace from copy import deepcopy from typing import TYPE_CHECKING, Type from hubble.executor.helper import is_valid_huburi from hubble.executor.hubio import HubIO from jina.enums import PodRoleType from jina.orchestrate.pods import Pod from jina.orchestrate.pods.container import ContainerPod ...
from argparse import Namespace from copy import deepcopy from typing import TYPE_CHECKING, Type from hubble.executor.helper import is_valid_huburi from hubble.executor.hubio import HubIO from jina.enums import PodRoleType from jina.orchestrate.pods import Pod from jina.orchestrate.pods.container import ContainerPod ...
import json from typing import Any, Type, TypeVar, overload import jsonschema from fastapi.encoders import jsonable_encoder from .type import type_match def to_dict(data) -> dict: return jsonable_encoder(data) def dumps(data) -> str: return json.dumps(jsonable_encoder(data)) T = TypeVar("T") @overload...
import json from typing import Any, Type, TypeVar, overload from fastapi.encoders import jsonable_encoder from .type import type_match def to_dict(data) -> dict: return jsonable_encoder(data) def dumps(data) -> str: return json.dumps(jsonable_encoder(data)) T = TypeVar("T") @overload def loads(data: s...
import warnings from typing import Callable, Union from torch.ao.pruning.sparsifier.base_sparsifier import BaseSparsifier from .base_scheduler import BaseScheduler __all__ = ["LambdaSL"] class LambdaSL(BaseScheduler): """Sets the sparsity level of each parameter group to the final sl times a given functio...
# mypy: allow-untyped-defs import warnings from .base_scheduler import BaseScheduler __all__ = ["LambdaSL"] class LambdaSL(BaseScheduler): """Sets the sparsity level of each parameter group to the final sl times a given function. When last_epoch=-1, sets initial sl as zero. Args: sparsifier (Ba...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '2.15.1' short_version = __version__ def parse_version_info(version_str): version_info = [] for x in version_str.split('.'): if x.isdigit(): version_info.append(int(x)) elif x.find('rc') != -1: patch_version...
# Copyright (c) Open-MMLab. All rights reserved. __version__ = '2.15.1' short_version = __version__ def parse_version_info(version_str): version_info = [] for x in version_str.split('.'): if x.isdigit(): version_info.append(int(x)) elif x.find('rc') != -1: patch_versio...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from pathlib import Path import pytest import torch import numpy as np import torchvision.models.video as models from torchvision import transforms from jina import Document, DocumentArray, Executor from ...v...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import os import pytest import torch import numpy as np import torchvision.models.video as models from torchvision import transforms from jina import Document, DocumentArray try: from video_torch_encoder ...
import types from typing import TYPE_CHECKING from docarray.utils._internal.misc import ( _get_path_from_docarray_root_level, import_library, ) if TYPE_CHECKING: from docarray.index.backends.elastic import ElasticDocIndex # noqa: F401 from docarray.index.backends.elasticv7 import ElasticV7DocIndex #...
import types from typing import TYPE_CHECKING from docarray.utils._internal.misc import ( _get_path_from_docarray_root_level, import_library, ) if TYPE_CHECKING: from docarray.index.backends.elastic import ElasticDocIndex # noqa: F401 from docarray.index.backends.elasticv7 import ElasticV7DocIndex #...
_base_ = './decoupled-solo_r50_fpn_3x_coco.py' # model settings model = dict( mask_head=dict( type='DecoupledSOLOLightHead', num_classes=80, in_channels=256, stacked_convs=4, feat_channels=256, strides=[8, 8, 16, 32, 32], scale_ranges=((1, 64), (32, 128), (64...
_base_ = './decoupled-solo_r50_fpn_3x_coco.py' # model settings model = dict( mask_head=dict( type='DecoupledSOLOLightHead', num_classes=80, in_channels=256, stacked_convs=4, feat_channels=256, strides=[8, 8, 16, 32, 32], scale_ranges=((1, 64), (32, 128), (64...
import warnings from typing import Any, List import torch from torchvision.transforms import functional as _F @torch.jit.unused def to_tensor(inpt: Any) -> torch.Tensor: """[DEPREACTED] Use to_image() and to_dtype() instead.""" warnings.warn( "The function `to_tensor(...)` is deprecated and will be ...
import warnings from typing import Any, List import torch from torchvision.transforms import functional as _F @torch.jit.unused def to_tensor(inpt: Any) -> torch.Tensor: """[BETA] [DEPREACTED] Use to_image() and to_dtype() instead.""" warnings.warn( "The function `to_tensor(...)` is deprecated and w...
"""JSON Reader.""" import re import defusedxml.ElementTree as ET # safe XML parsing import xml.etree.ElementTree as _XmlET # for type annotations only from pathlib import Path from typing import Dict, List, Optional from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document ...
"""JSON Reader.""" import re import xml.etree.ElementTree as ET from pathlib import Path from typing import Dict, List, Optional from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document def _get_leaf_nodes_up_to_level(root: ET.Element, level: int) -> List[ET.Element]: ""...
from __future__ import annotations from copy import deepcopy import pytest from sentence_transformers import CrossEncoder @pytest.fixture() def distilroberta_base_ce_model() -> CrossEncoder: return CrossEncoder("distilroberta-base", num_labels=1) @pytest.fixture(scope="session") def _reranker_bert_tiny_model...
from __future__ import annotations import pytest from sentence_transformers import CrossEncoder @pytest.fixture() def distilroberta_base_ce_model() -> CrossEncoder: return CrossEncoder("distilroberta-base", num_labels=1) @pytest.fixture() def reranker_bert_tiny_model() -> CrossEncoder: return CrossEncoder...
from __future__ import annotations from typing import Any, List, Optional, Tuple, Union import PIL.Image import torch from torchvision.transforms import InterpolationMode from ._datapoint import _FillTypeJIT, Datapoint class Mask(Datapoint): @classmethod def _wrap(cls, tensor: torch.Tensor) -> Mask: ...
from __future__ import annotations from typing import Any, List, Optional, Tuple, Union import PIL.Image import torch from torchvision.transforms import InterpolationMode from ._datapoint import Datapoint, FillTypeJIT class Mask(Datapoint): @classmethod def _wrap(cls, tensor: torch.Tensor) -> Mask: ...
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
from typing import Optional import numpy as np import pytest from pydantic import BaseModel, ValidationError from typing_extensions import TypedDict from docarray import BaseDoc, DocArray from docarray.documents import AudioDoc, ImageDoc, TextDoc from docarray.documents.helper import ( create_doc, create_doc_...
from typing import Optional import numpy as np import pytest from pydantic import BaseModel, ValidationError from typing_extensions import TypedDict from docarray import BaseDocument, DocumentArray from docarray.documents import AudioDoc, ImageDoc, TextDoc from docarray.documents.helper import ( create_doc, c...
import os import numpy as np import pytest as pytest from jina import Document, DocumentArray cur_dir = os.path.dirname(os.path.abspath(__file__)) compose_yml = os.path.abspath(os.path.join(cur_dir, 'docker-compose.yml')) @pytest.mark.parametrize('docker_compose', [compose_yml], indirect=['docker_compose']) def tes...
import os import pytest as pytest from jina import Document, DocumentArray cur_dir = os.path.dirname(os.path.abspath(__file__)) compose_yml = os.path.abspath(os.path.join(cur_dir, 'docker-compose.yml')) @pytest.mark.parametrize('docker_compose', [compose_yml], indirect=['docker_compose']) def test_connection(indexe...
import asyncio from math import ceil import pytest from docarray import Document from jina.clients.request.asyncio import request_generator NUM_INPUT_DOCS = 30 REQUEST_SIZE = 10 @pytest.mark.asyncio async def test_asyncio_req_generator(): async def input_function(): data = [Document() for _ in range(NU...
import asyncio from math import ceil import pytest from docarray import Document from jina.clients.request.asyncio import request_generator NUM_INPUT_DOCS = 30 REQUEST_SIZE = 10 @pytest.mark.asyncio async def test_asyncio_req_generator(): async def input_function(): data = [Document() for _ in range(NU...
# flake8: noqa: F401 r""" This file is in the process of migration to `torch/ao/quantization`, and is kept here for compatibility while the migration process is ongoing. If you are adding a new entry/functionality, please, add it to the `torch/ao/quantization/fuse_modules.py`, while adding an import statement here. """...
# flake8: noqa: F401 r""" This file is in the process of migration to `torch/ao/quantization`, and is kept here for compatibility while the migration process is ongoing. If you are adding a new entry/functionality, please, add it to the `torch/ao/quantization/fuse_modules.py`, while adding an import statement here. """...
from typing import Dict, List, Optional, Set import pytest from docarray import BaseDoc, DocList from docarray.documents import ImageDoc from docarray.utils.reduce import reduce, reduce_all class InnerDoc(BaseDoc): integer: int inner_list: List class MMDoc(BaseDoc): text: str = '' price: int = 0 ...
from typing import Dict, List, Optional, Set import pytest from docarray import BaseDoc, DocList from docarray.documents import ImageDoc from docarray.utils.reduce import reduce, reduce_all class InnerDoc(BaseDoc): integer: int inner_list: List class MMDoc(BaseDoc): text: str = '' price: int = 0 ...
from typing import Any, Dict, Iterator import torch from ..utils import _log_api_usage_once try: from ._load_gpu_decoder import _HAS_GPU_VIDEO_DECODER except ModuleNotFoundError: _HAS_GPU_VIDEO_DECODER = False from ._video_opt import ( _HAS_VIDEO_OPT, _probe_video_from_file, _probe_video_from_mem...
from typing import Any, Dict, Iterator import torch from ..utils import _log_api_usage_once from ._video_opt import ( _HAS_VIDEO_OPT, _probe_video_from_file, _probe_video_from_memory, _read_video_from_file, _read_video_from_memory, _read_video_timestamps_from_file, _read_video_timestamps_...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools.dataforseo_api_search.tool import ( DataForSeoAPISearchResults, DataForSeoAPISearchRun, ) # Create a way to dynamically look up deprecated imports. # Used to conso...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools.dataforseo_api_search.tool import ( DataForSeoAPISearchResults, DataForSeoAPISearchRun, ) # Create a way to dynamically look up deprecated imports. # Used to conso...
import pytest from llama_index.postprocessor.nvidia_rerank import NVIDIARerank import respx @pytest.fixture(autouse=True) def mock_local_models(respx_mock: respx.MockRouter) -> None: respx_mock.get( "https://test_url/v1/models", json={ "data": [ {"id": "model1"}, ...
import pytest from llama_index.postprocessor.nvidia_rerank import NVIDIARerank from requests_mock import Mocker @pytest.fixture(autouse=True) def mock_local_models(requests_mock: Mocker) -> None: requests_mock.get( "https://test_url/v1/models", json={ "data": [ {"id": ...
import os from typing import Dict DEPLOYMENT_FILES = [ 'statefulset-executor', 'deployment-executor', 'deployment-gateway', 'deployment-uses-before', 'deployment-uses-after', 'deployment-uses-before-after', ] cur_dir = os.path.dirname(__file__) DEFAULT_RESOURCE_DIR = os.path.join( cur_dir,...
import os from typing import Dict DEPLOYMENT_FILES = [ 'statefulset-executor', 'deployment-executor', 'deployment-gateway', 'deployment-uses-before', 'deployment-uses-after', 'deployment-uses-before-after', ] cur_dir = os.path.dirname(__file__) DEFAULT_RESOURCE_DIR = os.path.join( cur_dir,...
# Copyright 2025 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# Copyright 2025 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
import os import urllib import pytest from pydantic import parse_obj_as, schema_json_of from docarray.base_document.io.json import orjson_dumps from docarray.typing import TextUrl REMOTE_TXT = 'https://de.wikipedia.org/wiki/Brixen' CUR_DIR = os.path.dirname(os.path.abspath(__file__)) LOCAL_TXT = os.path.join(CUR_DIR...
import os import urllib import pytest from pydantic import parse_obj_as, schema_json_of from docarray.base_document.io.json import orjson_dumps from docarray.typing import TextUrl REMOTE_TXT = 'https://de.wikipedia.org/wiki/Brixen' CUR_DIR = os.path.dirname(os.path.abspath(__file__)) LOCAL_TXT = os.path.join(CUR_DIR...
# Copyright 2025 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
# Copyright 2025 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
from torchvision.transforms import AutoAugmentPolicy, InterpolationMode # usort: skip from . import functional # usort: skip from ._transform import Transform # usort: skip from ._augment import CutMix, JPEG, MixUp, RandomErasing from ._auto_augment import AugMix, AutoAugment, RandAugment, TrivialAugmentWide from...
from torchvision.transforms import AutoAugmentPolicy, InterpolationMode # usort: skip from . import functional # usort: skip from ._transform import Transform # usort: skip from ._augment import CutMix, JPEG, MixUp, RandomErasing from ._auto_augment import AugMix, AutoAugment, RandAugment, TrivialAugmentWide from...
import pytest from backend.data import db from backend.executor import ExecutionScheduler from backend.server.model import CreateGraph from backend.usecases.sample import create_test_graph, create_test_user from backend.util.service import get_service_client from backend.util.test import SpinTestServer @pytest.mark....
import pytest from backend.data import db from backend.executor import ExecutionScheduler from backend.server.model import CreateGraph from backend.usecases.sample import create_test_graph, create_test_user from backend.util.service import get_service_client from backend.util.test import SpinTestServer @pytest.mark....