input
stringlengths
33
5k
output
stringlengths
32
5k
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.output_parsers.rail_parser import GuardrailsOutputParser # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handlin...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.output_parsers.rail_parser import GuardrailsOutputParser # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handlin...
# dataset settings dataset_type = 'VOCDataset' data_root = 'data/VOCdevkit/' # file_client_args = dict( # backend='petrel', # path_mapping=dict({ # './data/': 's3://openmmlab/datasets/detection/', # 'data/': 's3://openmmlab/datasets/detection/' # })) file_client_args = dict(backend='disk') ...
# dataset settings dataset_type = 'VOCDataset' data_root = 'data/VOCdevkit/' # file_client_args = dict( # backend='petrel', # path_mapping=dict({ # './data/': 's3://openmmlab/datasets/detection/', # 'data/': 's3://openmmlab/datasets/detection/' # })) file_client_args = dict(backend='disk') ...
# Copyright (c) OpenMMLab. All rights reserved. import bisect import os.path as osp import mmcv import torch.distributed as dist from mmcv.runner import DistEvalHook as BaseDistEvalHook from mmcv.runner import EvalHook as BaseEvalHook from torch.nn.modules.batchnorm import _BatchNorm def _calc_dynamic_intervals(star...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp import torch.distributed as dist from mmcv.runner import DistEvalHook as BaseDistEvalHook from mmcv.runner import EvalHook as BaseEvalHook from torch.nn.modules.batchnorm import _BatchNorm class EvalHook(BaseEvalHook): def _do_evaluate(self, ...
from docarray.typing.tensor.embedding import AnyEmbedding, NdArrayEmbedding from docarray.typing.tensor.image import ImageNdArray, ImageTensor from docarray.typing.tensor.ndarray import NdArray from docarray.typing.tensor.tensor import AnyTensor __all__ = [ 'NdArray', 'AnyTensor', 'AnyEmbedding', 'NdAr...
from docarray.typing.tensor.embedding import AnyEmbedding, NdArrayEmbedding from docarray.typing.tensor.ndarray import NdArray from docarray.typing.tensor.tensor import AnyTensor __all__ = [ 'NdArray', 'AnyTensor', 'AnyEmbedding', 'NdArrayEmbedding', ] try: import torch # noqa: F401 except Import...
"""Module for Jina Requests.""" from typing import ( TYPE_CHECKING, AsyncIterable, Dict, Iterable, Iterator, Optional, Tuple, Union, ) from jina.clients.request.helper import _new_data_request, _new_data_request_from_batch from jina.enums import DataInputType from jina.helper import ba...
"""Module for Jina Requests.""" from typing import ( TYPE_CHECKING, AsyncIterable, Dict, Iterable, Iterator, Optional, Tuple, Union, ) from jina.clients.request.helper import _new_data_request, _new_data_request_from_batch from jina.enums import DataInputType from jina.helper import ba...
# coding=utf-8 # Copyright 2025 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2024 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
_base_ = 'mask_rcnn_r50_fpn_syncbn-all_rpn-2conv_lsj_100e_coco.py' # Enable automatic-mixed-precision training with AmpOptimWrapper. optim_wrapper = dict(type='AmpOptimWrapper')
_base_ = 'mask_rcnn_r50_fpn_syncbn-all_rpn-2conv_lsj_100e_coco.py' # use FP16 fp16 = dict(loss_scale=512.)
from typing import Union import numpy as np import PIL.Image import torch from torchvision.prototype import datapoints from torchvision.transforms import functional as _F @torch.jit.unused def to_image_tensor(inpt: Union[torch.Tensor, PIL.Image.Image, np.ndarray]) -> datapoints.Image: if isinstance(inpt, np.ndar...
from typing import Union import numpy as np import PIL.Image import torch from torchvision.prototype import features from torchvision.transforms import functional as _F @torch.jit.unused def to_image_tensor(inpt: Union[torch.Tensor, PIL.Image.Image, np.ndarray]) -> features.Image: if isinstance(inpt, np.ndarray)...
_base_ = './mask-rcnn_r101_fpn_1x_coco.py' model = dict( # ResNeXt-101-32x8d model trained with Caffe2 at FB, # so the mean and std need to be changed. data_preprocessor=dict( mean=[103.530, 116.280, 123.675], std=[57.375, 57.120, 58.395], bgr_to_rgb=False), backbone=dict( ...
_base_ = './mask_rcnn_r101_fpn_1x_coco.py' model = dict( # ResNeXt-101-32x8d model trained with Caffe2 at FB, # so the mean and std need to be changed. data_preprocessor=dict( mean=[103.530, 116.280, 123.675], std=[57.375, 57.120, 58.395], bgr_to_rgb=False), backbone=dict( ...
from keras.src.utils.module_utils import tensorflow as tf def start_trace(logdir): tf.profiler.experimental.start(logdir=logdir) def stop_trace(save): tf.profiler.experimental.stop(save=save) def start_batch_trace(batch): batch_trace_context = tf.profiler.experimental.Trace( "Profiled batch", ...
import tensorflow as tf def start_trace(logdir): tf.profiler.experimental.start(logdir=logdir) def stop_trace(save): tf.profiler.experimental.stop(save=save)
"""Module to test base parser implementations.""" from langchain_core.exceptions import OutputParserException from langchain_core.language_models import GenericFakeChatModel from langchain_core.messages import AIMessage from langchain_core.output_parsers import ( BaseGenerationOutputParser, BaseTransformOutput...
"""Module to test base parser implementations.""" from typing import Optional as Optional from langchain_core.exceptions import OutputParserException from langchain_core.language_models import GenericFakeChatModel from langchain_core.messages import AIMessage from langchain_core.output_parsers import ( BaseGenera...
_base_ = '../cascade_rcnn/cascade-rcnn_r50_fpn_1x_coco.py' model = dict( backbone=dict( dcn=dict(type='DCN', deform_groups=1, fallback_on_stride=False), stage_with_dcn=(False, True, True, True)))
_base_ = '../cascade_rcnn/cascade_rcnn_r50_fpn_1x_coco.py' model = dict( backbone=dict( dcn=dict(type='DCN', deform_groups=1, fallback_on_stride=False), stage_with_dcn=(False, True, True, True)))
"""init.py.""" from llama_index.tools.code_interpreter.base import ( CodeInterpreterToolSpec, ) __all__ = ["CodeInterpreterToolSpec"]
"""init.py.""" from llama_index.tools.code_interpreter.base import ( CodeInterpreterToolSpec, ) __all__ = ["CodeInterpreterToolSpec"]
from .filtering import ( allpass_biquad, band_biquad, bandpass_biquad, bandreject_biquad, bass_biquad, biquad, contrast, dcshift, deemph_biquad, dither, equalizer_biquad, filtfilt, flanger, gain, highpass_biquad, lfilter, lowpass_biquad, overdrive,...
from .filtering import ( allpass_biquad, band_biquad, bandpass_biquad, bandreject_biquad, bass_biquad, biquad, contrast, dcshift, deemph_biquad, dither, equalizer_biquad, filtfilt, flanger, gain, highpass_biquad, lfilter, lowpass_biquad, overdrive,...
""" The pre-trained models produce embeddings of size 512 - 1024. However, when storing a large number of embeddings, this requires quite a lot of memory / storage. In this example, we reduce the dimensionality of the embeddings to e.g. 128 dimensions. This significantly reduces the required memory / storage while mai...
""" The pre-trained models produce embeddings of size 512 - 1024. However, when storing a large number of embeddings, this requires quite a lot of memory / storage. In this example, we reduce the dimensionality of the embeddings to e.g. 128 dimensions. This significantly reduces the required memory / storage while mai...
import dataclasses from collections import defaultdict from typing import TYPE_CHECKING, Type, List if TYPE_CHECKING: # pragma: no cover from docarray.typing import T from docarray.document.strawberry_type import StrawberryDocument class StrawberryMixin: """Provide helper functions to convert to/from a ...
import dataclasses from collections import defaultdict from typing import TYPE_CHECKING, Type, List if TYPE_CHECKING: from docarray.typing import T from docarray.document.strawberry_type import StrawberryDocument class StrawberryMixin: """Provide helper functions to convert to/from a Strawberry model""" ...
_base_ = './tood_r50_fpn_1x_coco.py' max_epochs = 24 # learning rate param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[16, 22], g...
_base_ = './tood_r50_fpn_1x_coco.py' max_epochs = 24 # learning rate param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[16, 22], g...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import io from typing import List import fitz import numpy as np import pdfplumber from jina import Document, DocumentArray, Executor, requests from jina.logging.logger import JinaLogger class PDFSegmenter(Exe...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import io from typing import List import fitz import numpy as np import pdfplumber from jina import Document, DocumentArray, Executor, requests from jina.logging.logger import JinaLogger class PDFSegmenter(Exe...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class FSAF(SingleStageDetector): """Implementation of `FSAF <https://arxiv.org/abs/1903.00621>`_""" def __init__(self, backbone, ...
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .single_stage import SingleStageDetector @DETECTORS.register_module() class FSAF(SingleStageDetector): """Implementation of `FSAF <https://arxiv.org/abs/1903.00621>`_""" def __init__(self, backbone, ...
import unittest import torch from mmengine.data import PixelData from mmengine.testing import assert_allclose from mmdet.models.seg_heads import PanopticFPNHead from mmdet.structures import DetDataSample class TestPanopticFPNHead(unittest.TestCase): def test_init_weights(self): head = PanopticFPNHead( ...
import unittest import torch from mmengine.data import PixelData from mmengine.testing import assert_allclose from mmdet.data_elements import DetDataSample from mmdet.models.seg_heads import PanopticFPNHead class TestPanopticFPNHead(unittest.TestCase): def test_init_weights(self): head = PanopticFPNHea...
import numpy as np from sklearn.datasets import make_classification import xgboost as xgb from xgboost.testing.updater import get_basescore def test_exp_family() -> None: X, y = make_classification(n_samples=128, n_classes=2, weights=[0.8, 0.2]) clf = xgb.train( {"objective": "binary:logistic"}, xgb....
import numpy as np from sklearn.datasets import make_classification import xgboost as xgb from xgboost.testing.updater import get_basescore def test_exp_family() -> None: X, y = make_classification(n_samples=128, n_classes=2, weights=[0.8, 0.2]) clf = xgb.train( {"objective": "binary:logistic"}, xgb....
import hashlib import logging from os import PathLike from pathlib import Path from typing import Union import torch from torchaudio._internal import download_url_to_file _LG = logging.getLogger(__name__) def _get_local_path(key): path = Path(torch.hub.get_dir()) / "torchaudio" / Path(key) path.parent.mkdir...
import hashlib import logging from os import PathLike from pathlib import Path from typing import Union import torch _LG = logging.getLogger(__name__) def _get_local_path(key): path = Path(torch.hub.get_dir()) / "torchaudio" / Path(key) path.parent.mkdir(parents=True, exist_ok=True) return path def _...
_base_ = '../rpn/rpn_r50_fpn_1x_coco.py' model = dict( rpn_head=dict( _delete_=True, type='GARPNHead', in_channels=256, feat_channels=256, approx_anchor_generator=dict( type='AnchorGenerator', octave_base_scale=8, scales_per_octave=3, ...
_base_ = '../rpn/rpn_r50_fpn_1x_coco.py' model = dict( rpn_head=dict( _delete_=True, type='GARPNHead', in_channels=256, feat_channels=256, approx_anchor_generator=dict( type='AnchorGenerator', octave_base_scale=8, scales_per_octave=3, ...
from __future__ import annotations import csv import logging import os from typing import TYPE_CHECKING import torch from torch.utils.data import DataLoader from sentence_transformers.evaluation.SentenceEvaluator import SentenceEvaluator from sentence_transformers.util import batch_to_device if TYPE_CHECKING: f...
import csv import logging import os from typing import TYPE_CHECKING, Dict import torch from torch.utils.data import DataLoader from sentence_transformers.evaluation.SentenceEvaluator import SentenceEvaluator from sentence_transformers.util import batch_to_device if TYPE_CHECKING: from sentence_transformers.Sent...
import torch import torch.nn.functional as F from ..utils import _log_api_usage_once def sigmoid_focal_loss( inputs: torch.Tensor, targets: torch.Tensor, alpha: float = 0.25, gamma: float = 2, reduction: str = "none", ) -> torch.Tensor: """ Loss used in RetinaNet for dense detection: http...
import torch import torch.nn.functional as F from ..utils import _log_api_usage_once def sigmoid_focal_loss( inputs: torch.Tensor, targets: torch.Tensor, alpha: float = 0.25, gamma: float = 2, reduction: str = "none", ) -> torch.Tensor: """ Loss used in RetinaNet for dense detection: http...
from datasets import Dataset from sentence_transformers.sparse_encoder import SparseEncoder, SparseEncoderTrainer, losses model = SparseEncoder("sparse-embedding/splade-distilbert-base-uncased-init") guide = SparseEncoder("naver/splade-cocondenser-ensembledistil") train_dataset = Dataset.from_dict( { "anc...
from datasets import Dataset from sentence_transformers.sparse_encoder import ( MLMTransformer, SparseCachedGISTEmbedLoss, SparseEncoder, SparseEncoderTrainer, SpladePooling, ) # Initialize the SPLADE model model_name = "naver/splade-cocondenser-ensembledistil" model = SparseEncoder( modules=[...
# Copyright (c) OpenMMLab. All rights reserved. from pathlib import Path import mmcv import torch from mmcv.runner import load_checkpoint from .. import build_detector from ..builder import DETECTORS from .single_stage import SingleStageDetector @DETECTORS.register_module() class KnowledgeDistillationSingleStageDet...
# Copyright (c) OpenMMLab. All rights reserved. import mmcv import torch from mmcv.runner import load_checkpoint from .. import build_detector from ..builder import DETECTORS from .single_stage import SingleStageDetector @DETECTORS.register_module() class KnowledgeDistillationSingleStageDetector(SingleStageDetector)...
import torch import torchaudio.prototype.transforms as T from torchaudio_unittest.common_utils import nested_params, TestBaseMixin, torch_script class Transforms(TestBaseMixin): @nested_params( ["Convolve", "FFTConvolve"], ["full", "valid", "same"], ) def test_Convolve(self, cls, mode): ...
import torch import torchaudio.prototype.transforms as T from torchaudio_unittest.common_utils import nested_params, TestBaseMixin, torch_script class Transforms(TestBaseMixin): @nested_params( ["Convolve", "FFTConvolve"], ["full", "valid", "same"], ) def test_Convolve(self, cls, mode): ...
_base_ = './tood_r50_fpn_1x_coco.py' max_epochs = 24 # learning rate param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[16, 22], g...
_base_ = './tood_r50_fpn_1x_coco.py' max_epochs = 24 # learning rate param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[16, 22], g...
"""Test Bedrock multi-modal LLM.""" import json from io import BytesIO import pytest from unittest.mock import patch, AsyncMock from llama_index.core.multi_modal_llms.base import MultiModalLLM from llama_index.multi_modal_llms.bedrock import BedrockMultiModal from llama_index.core.schema import ImageDocument def te...
"""Test Bedrock multi-modal LLM.""" import json from io import BytesIO import pytest from unittest.mock import patch, AsyncMock from llama_index.core.multi_modal_llms.base import MultiModalLLM from llama_index.multi_modal_llms.bedrock import BedrockMultiModal from llama_index.core.schema import ImageDocument def tes...
import os from typing import Optional, Type from langchain_core.callbacks import CallbackManagerForToolRun from langchain_core.tools import BaseTool from pydantic import BaseModel, Field from langchain_community.tools.file_management.utils import ( INVALID_PATH_TEMPLATE, BaseFileToolMixin, FileValidationE...
import os from typing import Optional, Type from langchain_core.callbacks import CallbackManagerForToolRun from langchain_core.tools import BaseTool from pydantic import BaseModel, Field from langchain_community.tools.file_management.utils import ( INVALID_PATH_TEMPLATE, BaseFileToolMixin, FileValidationE...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
import numpy as np from docarray.proto import DocumentProto, NdArrayProto, NodeProto from docarray.typing import Tensor def test_nested_item_proto(): NodeProto(text='hello') NodeProto(nested=DocumentProto()) def test_nested_optional_item_proto(): NodeProto() def test_ndarray(): nd_proto = NdArray...
import numpy as np from docarray.proto import DocumentProto, NdArrayProto, NodeProto from docarray.typing import Tensor def test_nested_item_proto(): NodeProto(text='hello') NodeProto(nested=DocumentProto()) def test_nested_optional_item_proto(): NodeProto() def test_ndarray(): nd_proto = NdArray...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Tuple, Union from torch import Tensor from mmdet.registry import MODELS from .convfc_bbox_head import ConvFCBBoxHead @MODELS.register_module() class SCNetBBoxHead(ConvFCBBoxHead): """BBox head for `SCNet <https://arxiv.org/abs/2012.10150>`_. ...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import MODELS from .convfc_bbox_head import ConvFCBBoxHead @MODELS.register_module() class SCNetBBoxHead(ConvFCBBoxHead): """BBox head for `SCNet <https://arxiv.org/abs/2012.10150>`_. This inherits ``ConvFCBBoxHead`` with modified forward() ...
# Copyright (c) OpenMMLab. All rights reserved. import logging from typing import List, Optional, Sequence import torch from torch.nn.parameter import Parameter from torch.nn.utils import clip_grad from mmengine.registry import HOOKS from .hook import Hook DATA_BATCH = Optional[Sequence[dict]] @HOOKS.register_modu...
# Copyright (c) OpenMMLab. All rights reserved. import logging from typing import List, Optional, Sequence import torch from torch.nn.parameter import Parameter from torch.nn.utils import clip_grad from mmengine.registry import HOOKS from .hook import Hook DATA_BATCH = Optional[Sequence[dict]] @HOOKS.register_modu...
# Copyright (c) OpenMMLab. All rights reserved. import mmcv import torch from mmdet.models.dense_heads.autoassign_head import AutoAssignHead from mmdet.models.dense_heads.paa_head import levels_to_images def test_autoassign_head_loss(): """Tests autoassign head loss when truth is empty and non-empty.""" s =...
import mmcv import torch from mmdet.models.dense_heads.autoassign_head import AutoAssignHead from mmdet.models.dense_heads.paa_head import levels_to_images def test_autoassign_head_loss(): """Tests autoassign head loss when truth is empty and non-empty.""" s = 256 img_metas = [{ 'img_shape': (s,...
from __future__ import annotations from dataclasses import dataclass from sentence_transformers.training_args import SentenceTransformerTrainingArguments @dataclass class SparseEncoderTrainingArguments(SentenceTransformerTrainingArguments): """ SparseEncoderTrainingArguments extends :class:`~transformers.Tr...
from __future__ import annotations from dataclasses import dataclass from sentence_transformers.training_args import SentenceTransformerTrainingArguments @dataclass class SparseEncoderTrainingArguments(SentenceTransformerTrainingArguments): """ SparseEncoderTrainingArguments extends :class:`~transformers.Tr...
#!/usr/bin/env python3 import numbers import random import warnings from torchvision.transforms import RandomCrop, RandomResizedCrop from . import _functional_video as F __all__ = [ "RandomCropVideo", "RandomResizedCropVideo", "CenterCropVideo", "NormalizeVideo", "ToTensorVideo", "RandomHor...
#!/usr/bin/env python3 import numbers import random import warnings from torchvision.transforms import RandomCrop, RandomResizedCrop from . import _functional_video as F __all__ = [ "RandomCropVideo", "RandomResizedCropVideo", "CenterCropVideo", "NormalizeVideo", "ToTensorVideo", "RandomHor...
from typing import Any, Dict, Sequence from llama_index.core.base.llms.types import ChatMessage LLAMA_MODELS = { "llama-2-13b-chat": 4096, "llama-2-70b-chat": 4096, } MISTRAL_MODELS = { "mistral-7b-instruct-v0-2": 32768, "mixtral-8x7b-instruct-v0-1": 32768, } GEMMA_MODELS = { "gemma-7b-it": 8192...
from typing import Any, Dict, Sequence from llama_index.core.base.llms.types import ChatMessage LLAMA_MODELS = { "llama-2-13b-chat": 4096, "llama-2-70b-chat": 4096, } MISTRAL_MODELS = { "mistral-7b-instruct-v0-2": 32768, "mixtral-8x7b-instruct-v0-1": 32768, } GEMMA_MODELS = { "gemma-7b-it": 8192...
from __future__ import annotations from collections.abc import Iterable from torch import Tensor from sentence_transformers import util from sentence_transformers.losses.CoSENTLoss import CoSENTLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseCoSENTLoss(CoSENTLoss): ...
from __future__ import annotations from sentence_transformers import util from sentence_transformers.losses.CoSENTLoss import CoSENTLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseCoSENTLoss(CoSENTLoss): def __init__(self, model: SparseEncoder, scale: float = 20.0, s...
from typing import Union import numpy as np import PIL.Image import torch from torchvision import tv_tensors from torchvision.transforms import functional as _F @torch.jit.unused def to_image(inpt: Union[torch.Tensor, PIL.Image.Image, np.ndarray]) -> tv_tensors.Image: """[BETA] See :class:`~torchvision.transform...
from typing import Union import numpy as np import PIL.Image import torch from torchvision import tv_tensors from torchvision.transforms import functional as _F @torch.jit.unused def to_image(inpt: Union[torch.Tensor, PIL.Image.Image, np.ndarray]) -> tv_tensors.Image: """[BETA] See :class:`~torchvision.transform...
from jina.serve.runtimes.gateway.grpc.gateway import GRPCGateway
import os from jina import __default_host__ from jina.excepts import PortAlreadyUsed from jina.helper import is_port_free from jina.serve.runtimes.gateway import GatewayRuntime from jina.serve.runtimes.gateway.grpc.gateway import GRPCGateway __all__ = ['GRPCGatewayRuntime'] class GRPCGatewayRuntime(GatewayRuntime):...
from __future__ import annotations import logging from typing import Any, Dict, Optional from langchain_core.callbacks import CallbackManagerForToolRun from langchain_core.tools import BaseTool from langchain_core.utils import get_from_dict_or_env from pydantic import model_validator logger = logging.getLogger(__nam...
from __future__ import annotations import logging from typing import Any, Dict, Optional from langchain_core.callbacks import CallbackManagerForToolRun from langchain_core.tools import BaseTool from langchain_core.utils import get_from_dict_or_env from pydantic import model_validator logger = logging.getLogger(__nam...
import pathlib from typing import Any, Dict, List, Union import torch from torchdata.datapipes.iter import CSVDictParser, IterDataPipe, Mapper from torchvision.datapoints import Image from torchvision.prototype.datapoints import Label from torchvision.prototype.datasets.utils import Dataset, KaggleDownloadResource, On...
import pathlib from typing import Any, Dict, List, Union import torch from torchdata.datapipes.iter import CSVDictParser, IterDataPipe, Mapper from torchvision.prototype.datapoints import Image, Label from torchvision.prototype.datasets.utils import Dataset, KaggleDownloadResource, OnlineResource from torchvision.prot...
# Copyright (c) OpenMMLab. All rights reserved. import torch import torch.nn as nn from mmcv.cnn import ConvModule, DepthwiseSeparableConvModule from mmengine.model import BaseModule from mmdet.registry import MODELS @MODELS.register_module() class SSDNeck(BaseModule): """Extra layers of SSD backbone to generate...
# Copyright (c) OpenMMLab. All rights reserved. import torch import torch.nn as nn from mmcv.cnn import ConvModule, DepthwiseSeparableConvModule from mmcv.runner import BaseModule from mmdet.registry import MODELS @MODELS.register_module() class SSDNeck(BaseModule): """Extra layers of SSD backbone to generate mu...
_base_ = [ '../_base_/models/mask-rcnn_r50_fpn.py', '../common/lsj-100e_coco-instance.py' ] image_size = (1024, 1024) batch_augments = [ dict(type='BatchFixedSizePad', size=image_size, pad_mask=True) ] model = dict(data_preprocessor=dict(batch_augments=batch_augments)) train_dataloader = dict(batch_size=8...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../common/lsj_100e_coco_instance.py' ] image_size = (1024, 1024) batch_augments = [ dict(type='BatchFixedSizePad', size=image_size, pad_mask=True) ] model = dict(data_preprocessor=dict(batch_augments=batch_augments)) train_dataloader = dict(batch_size=8...
import dataclasses from collections import defaultdict from typing import TYPE_CHECKING, Type, List if TYPE_CHECKING: from docarray.typing import T from docarray.document.strawberry_type import StrawberryDocument class StrawberryMixin: """Provide helper functions to convert to/from a Strawberry model""" ...
import dataclasses from collections import defaultdict from typing import TYPE_CHECKING, Type, List if TYPE_CHECKING: from docarray.typing import T from docarray.document.strawberry_type import StrawberryDocument class StrawberryMixin: """Provide helper functions to convert to/from a Strawberry model""" ...
# dataset settings dataset_type = 'CocoDataset' data_root = 'data/MOT17/' backend_args = None train_pipeline = [ dict(type='LoadImageFromFile', backend_args=backend_args, to_float32=True), dict(type='LoadAnnotations', with_bbox=True), dict( type='RandomResize', scale=(1088, 1088), r...
# dataset settings dataset_type = 'CocoDataset' data_root = 'data/MOT17/' train_pipeline = [ dict(type='LoadImageFromFile', to_float32=True), dict(type='LoadAnnotations', with_bbox=True), dict( type='RandomResize', scale=(1088, 1088), ratio_range=(0.8, 1.2), keep_ratio=True,...
import os import pathlib import sys def get_frontend_path() -> pathlib.Path: if getattr(sys, "frozen", False): # The application is frozen datadir = pathlib.Path(os.path.dirname(sys.executable)) / "example_files" else: # The application is not frozen # Change this bit to match ...
import os import pathlib import sys def get_secrets_path() -> pathlib.Path: return get_data_path() / "secrets" def get_config_path() -> pathlib.Path: return get_data_path() def get_frontend_path() -> pathlib.Path: if getattr(sys, "frozen", False): # The application is frozen datadir = ...
import os from pathlib import Path from typing import List, Optional, Tuple import requests def get_file_content(url: str, path: str) -> Tuple[str, int]: """Get the content of a file from the GitHub REST API.""" resp = requests.get(url + path) return resp.text, resp.status_code def get_file_content_byt...
import os from pathlib import Path from typing import List, Optional, Tuple import requests def get_file_content(url: str, path: str) -> Tuple[str, int]: """Get the content of a file from the GitHub REST API.""" resp = requests.get(url + path) return resp.text, resp.status_code def get_file_content_byt...
from typing import Any, Dict, Optional, Type from jina.jaml.parsers.base import BaseLegacyParser from jina.serve.gateway import BaseGateway class GatewayLegacyParser(BaseLegacyParser): """Legacy parser for gateway.""" def parse( self, cls: Type['BaseGateway'], data: Dict, run...
from typing import Any, Dict, Optional, Type from jina.jaml.parsers.base import BaseLegacyParser from jina.serve.gateway import BaseGateway class GatewayLegacyParser(BaseLegacyParser): """Legacy parser for gateway.""" def parse( self, cls: Type['BaseGateway'], data: Dict, run...
""" Feature engineering pipeline for categorical data ================================================= The script showcases how to keep the categorical data encoding consistent across training and inference. There are many ways to attain the same goal, this script can be used as a starting point. See Also -------- -...
""" Feature engineering pipeline for categorical data ================================================= The script showcases how to keep the categorical data encoding consistent across training and inference. There are many ways to attain the same goal, this script can be used as a starting point. See Also -------- -...
# -*- coding: utf-8 -*- # Generated by the protocol buffer compiler. DO NOT EDIT! # source: docarray.proto """Generated protocol buffer code.""" from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool from google.protobuf import symbol_database as _symbol_d...
# -*- coding: utf-8 -*- # Generated by the protocol buffer compiler. DO NOT EDIT! # source: docarray.proto """Generated protocol buffer code.""" from google.protobuf.internal import builder as _builder from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool...
# THIS FILE HAS BEEN AUTOGENERATED. To update: # 1. modify the `_deps` dict in setup.py # 2. run `make deps_table_update`` deps = { "Pillow": "Pillow>=10.0.1,<=15.0", "accelerate": "accelerate>=0.26.0", "av": "av", "beautifulsoup4": "beautifulsoup4", "blobfile": "blobfile", "codecarbon": "codeca...
# THIS FILE HAS BEEN AUTOGENERATED. To update: # 1. modify the `_deps` dict in setup.py # 2. run `make deps_table_update`` deps = { "Pillow": "Pillow>=10.0.1,<=15.0", "accelerate": "accelerate>=0.26.0", "av": "av", "beautifulsoup4": "beautifulsoup4", "blobfile": "blobfile", "codecarbon": "codeca...
_base_ = './cascade-mask-rcnn_r101_fpn_seesaw-loss_random-ms-2x_lvis-v1.py' # noqa: E501 model = dict( roi_head=dict( mask_head=dict( predictor_cfg=dict(type='NormedConv2d', tempearture=20))))
_base_ = './cascade_mask_rcnn_r101_fpn_random_seesaw_loss_mstrain_2x_lvis_v1.py' # noqa: E501 model = dict( roi_head=dict( mask_head=dict( predictor_cfg=dict(type='NormedConv2d', tempearture=20))))
from typing import List, Union class InputExample: """Structure for one input example with texts, the label and a unique id""" def __init__(self, guid: str = "", texts: List[str] = None, label: Union[int, float] = 0): """ Creates one InputExample with the given texts, guid and label ...
from typing import Union, List class InputExample: """ Structure for one input example with texts, the label and a unique id """ def __init__(self, guid: str = '', texts: List[str] = None, label: Union[int, float] = 0): """ Creates one InputExample with the given texts, guid and label...
from typing import Optional from docarray import Document, DocumentArray from pydantic import BaseModel from uvicorn import Config, Server from jina import Gateway, __default_host__ from jina.clients.request import request_generator class DummyResponseModel(BaseModel): arg1: Optional[str] arg2: Optional[str...
from typing import Optional from docarray import Document, DocumentArray from pydantic import BaseModel from uvicorn import Config, Server from jina import Gateway, __default_host__ from jina.clients.request import request_generator class DummyResponseModel(BaseModel): arg1: Optional[str] arg2: Optional[str...
import importlib.util from typing import List, Dict, Optional from llama_index.core.tools.tool_spec.base import BaseToolSpec class DuckDuckGoSearchToolSpec(BaseToolSpec): """DuckDuckGoSearch tool spec.""" spec_functions = ["duckduckgo_instant_search", "duckduckgo_full_search"] def __init__(self) -> None...
import importlib.util from typing import List, Dict, Optional from llama_index.core.tools.tool_spec.base import BaseToolSpec class DuckDuckGoSearchToolSpec(BaseToolSpec): """DuckDuckGoSearch tool spec.""" spec_functions = ["duckduckgo_instant_search", "duckduckgo_full_search"] def __init__(self) -> None...
from typing import Any, Union from torchvision import tv_tensors from torchvision.transforms.v2 import functional as F, Transform from torchvision.tv_tensors._bounding_boxes import CLAMPING_MODE_TYPE class ConvertBoundingBoxFormat(Transform): """Convert bounding box coordinates to the given ``format``, eg from "...
from typing import Any, Union from torchvision import tv_tensors from torchvision.transforms.v2 import functional as F, Transform from torchvision.tv_tensors._bounding_boxes import CLAMPING_MODE_TYPE class ConvertBoundingBoxFormat(Transform): """Convert bounding box coordinates to the given ``format``, eg from "...
# Copyright (c) OpenMMLab. All rights reserved. import torch.nn as nn from mmcv.cnn import (ConvModule, caffe2_xavier_init, constant_init, is_norm, normal_init) from torch.nn import BatchNorm2d from ..builder import NECKS class Bottleneck(nn.Module): """Bottleneck block for DilatedEncoder u...
# Copyright (c) OpenMMLab. All rights reserved. import torch.nn as nn from mmcv.cnn import (ConvModule, caffe2_xavier_init, constant_init, is_norm, normal_init) from torch.nn import BatchNorm2d from ..builder import NECKS class Bottleneck(nn.Module): """Bottleneck block for DilatedEncoder u...
import os from typing import Dict, List, Optional import torch from jina import DocumentArray, Executor, requests from jina.logging.logger import JinaLogger from jina_commons.batching import get_docs_batch_generator from transformers import CLIPTokenizer, CLIPModel class CLIPTextEncoder(Executor): """...
from jina import DocumentArray, Executor, requests import torch import clip from typing import Iterable, Optional, List from jina_commons.batching import get_docs_batch_generator class CLIPTextEncoder(Executor): """Encode text into embeddings using a CLIP model. :param model_name: The name of one of the pre-...
# Copyright 2020 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
# Copyright 2020 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
import os import shutil import subprocess import sys def _get_run_args(print_args: bool = True): from jina.helper import get_rich_console from jina.parsers import get_main_parser console = get_rich_console() silent_print = {'help', 'hub', 'export', 'auth', 'cloud', 'ping'} parser = get_main_par...
import os import shutil import subprocess import sys def _get_run_args(print_args: bool = True): from jina.helper import get_rich_console from jina.parsers import get_main_parser console = get_rich_console() silent_print = {'help', 'hub', 'export', 'auth', 'cloud', 'ping'} parser = get_main_par...
# Copyright (c) OpenMMLab. All rights reserved. import warnings import torch.nn as nn from mmcv.runner import BaseModule, auto_fp16 from mmdet.models.backbones import ResNet from mmdet.models.builder import SHARED_HEADS from mmdet.models.utils import ResLayer as _ResLayer @SHARED_HEADS.register_module() class ResLa...
# Copyright (c) OpenMMLab. All rights reserved. import warnings import torch.nn as nn from mmcv.runner import BaseModule, auto_fp16 from mmdet.models.backbones import ResNet from mmdet.models.builder import SHARED_HEADS from mmdet.models.utils import ResLayer as _ResLayer @SHARED_HEADS.register_module() class ResLa...
""" ========================= Multi-dimensional scaling ========================= An illustration of the metric and non-metric MDS on generated noisy data. """ # Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause # %% # Dataset preparation # ------------------- # # We start by uniformly ge...
""" ========================= Multi-dimensional scaling ========================= An illustration of the metric and non-metric MDS on generated noisy data. The reconstructed points using the metric MDS and non metric MDS are slightly shifted to avoid overlapping. """ # Authors: The scikit-learn developers # SPDX-Li...
from collections import defaultdict import torch import transforms as reference_transforms def get_modules(use_v2): # We need a protected import to avoid the V2 warning in case just V1 is used if use_v2: import torchvision.transforms.v2 import torchvision.tv_tensors return torchvisio...
from collections import defaultdict import torch import transforms as reference_transforms def get_modules(use_v2): # We need a protected import to avoid the V2 warning in case just V1 is used if use_v2: import torchvision.datapoints import torchvision.transforms.v2 return torchvisio...
"""Dict prompt template.""" import warnings from functools import cached_property from typing import Any, Literal, Optional from typing_extensions import override from langchain_core.load import dumpd from langchain_core.prompts.string import ( DEFAULT_FORMATTER_MAPPING, get_template_variables, ) from langch...
"""Dict prompt template.""" import warnings from functools import cached_property from typing import Any, Literal, Optional from typing_extensions import override from langchain_core.load import dumpd from langchain_core.prompts.string import ( DEFAULT_FORMATTER_MAPPING, get_template_variables, ) from langch...
from jina import Document, Flow from sentencizer import Sentencizer def test_exec(): f = Flow().add(uses=Sentencizer) with f: resp = f.post( on='/test', inputs=Document(text='Hello. World! Go? Back'), return_results=True, ) assert resp[0].docs[0].chu...
from jina import Document, Flow from ...sentencizer import Sentencizer def test_exec(): f = Flow().add(uses=Sentencizer) with f: resp = f.post( on='/test', inputs=Document(text='Hello. World! Go? Back'), return_results=True, ) assert resp[0].docs[0]...
from typing import Callable, TypeVar, ParamSpec import threading P = ParamSpec("P") R = TypeVar("R") def thread_cached(func: Callable[P, R]) -> Callable[P, R]: thread_local = threading.local() def wrapper(*args: P.args, **kwargs: P.kwargs) -> R: cache = getattr(thread_local, "cache", None) i...
import threading from functools import wraps from typing import Callable, ParamSpec, TypeVar T = TypeVar("T") P = ParamSpec("P") R = TypeVar("R") def thread_cached(func: Callable[P, R]) -> Callable[P, R]: thread_local = threading.local() @wraps(func) def wrapper(*args: P.args, **kwargs: P.kwargs) -> R: ...
"""Code Interpreter tool spec.""" import subprocess import sys from llama_index.core.tools.tool_spec.base import BaseToolSpec class CodeInterpreterToolSpec(BaseToolSpec): """ Code Interpreter tool spec. WARNING: This tool provides the Agent access to the `subprocess.run` command. Arbitrary code exe...
"""Code Interpreter tool spec.""" import subprocess import sys from llama_index.core.tools.tool_spec.base import BaseToolSpec class CodeInterpreterToolSpec(BaseToolSpec): """Code Interpreter tool spec. WARNING: This tool provides the Agent access to the `subprocess.run` command. Arbitrary code executio...
import warnings from typing import List, Optional, Type from jina.excepts import BadYAMLVersion from jina.jaml import JAMLCompatible from jina.jaml.parsers.base import VersionedYAMLParser from jina.serve.gateway import BaseGateway def _get_all_parser(cls: Type['JAMLCompatible']): """Get all parsers and legacy pa...
import warnings from typing import List, Optional, Type from jina.excepts import BadYAMLVersion from jina.jaml import JAMLCompatible from jina.jaml.parsers.base import VersionedYAMLParser def _get_all_parser(cls: Type['JAMLCompatible']): """Get all parsers and legacy parser of a class :param cls: target cla...
_base_ = './ga-retinanet_r101-caffe_fpn_1x_coco.py' train_pipeline = [ dict(type='LoadImageFromFile', backend_args={{_base_.backend_args}}), dict(type='LoadAnnotations', with_bbox=True), dict( type='RandomResize', scale=[(1333, 480), (1333, 960)], keep_ratio=True), dict(type='RandomFlip...
_base_ = './ga-retinanet_r101-caffe_fpn_1x_coco.py' train_pipeline = [ dict( type='LoadImageFromFile', file_client_args={{_base_.file_client_args}}), dict(type='LoadAnnotations', with_bbox=True), dict( type='RandomResize', scale=[(1333, 480), (1333, 960)], keep_ratio=True), ...
"""Standard LangChain interface tests.""" from langchain_core.language_models import BaseChatModel from langchain_tests.unit_tests.chat_models import ( ChatModelUnitTests, ) from langchain_groq import ChatGroq class TestGroqStandard(ChatModelUnitTests): @property def chat_model_class(self) -> type[BaseC...
"""Standard LangChain interface tests""" from langchain_core.language_models import BaseChatModel from langchain_tests.unit_tests.chat_models import ( ChatModelUnitTests, ) from langchain_groq import ChatGroq class TestGroqStandard(ChatModelUnitTests): @property def chat_model_class(self) -> type[BaseCh...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from parameterized import parameterized from mmdet.models.roi_heads import HybridTaskCascadeRoIHead # noqa from mmdet.registry import MODELS from mmdet.testing import demo_mm_inputs, demo_mm_proposals, get_roi_...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from parameterized import parameterized from mmdet.models.roi_heads import HybridTaskCascadeRoIHead # noqa from mmdet.registry import MODELS from mmdet.testing import demo_mm_inputs, demo_mm_proposals, get_roi_...
import json from jina.orchestrate.flow.base import Flow from jina.orchestrate.deployments import Deployment from jina.jaml import JAML from jina.logging.predefined import default_logger from jina.schemas import get_full_schema from jina_cli.export import api_to_dict def export_kubernetes(args): """Export to k8s ...
import json from jina.orchestrate.flow.base import Flow from jina.orchestrate.deployments import Deployment from jina.jaml import JAML from jina.logging.predefined import default_logger from jina.schemas import get_full_schema from jina_cli.export import api_to_dict def export_kubernetes(args): """Export to k8s ...
# Copyright (c) OpenMMLab. All rights reserved. from .bfp import BFP from .channel_mapper import ChannelMapper from .cspnext_pafpn import CSPNeXtPAFPN from .ct_resnet_neck import CTResNetNeck from .dilated_encoder import DilatedEncoder from .dyhead import DyHead from .fpg import FPG from .fpn import FPN from .fpn_caraf...
# Copyright (c) OpenMMLab. All rights reserved. from .bfp import BFP from .channel_mapper import ChannelMapper from .cspnext_pafpn import CSPNeXtPAFPN from .ct_resnet_neck import CTResNetNeck from .dilated_encoder import DilatedEncoder from .dyhead import DyHead from .fpg import FPG from .fpn import FPN from .fpn_caraf...
"""langchain-core version information and utilities.""" VERSION = "0.3.53"
"""langchain-core version information and utilities.""" VERSION = "0.3.52"
import os import pathlib from typing import Any, Callable, Optional, Union from .folder import default_loader from .utils import download_and_extract_archive, verify_str_arg from .vision import VisionDataset class DTD(VisionDataset): """`Describable Textures Dataset (DTD) <https://www.robots.ox.ac.uk/~vgg/data/...
import os import pathlib from typing import Any, Callable, Optional, Tuple, Union from .folder import default_loader from .utils import download_and_extract_archive, verify_str_arg from .vision import VisionDataset class DTD(VisionDataset): """`Describable Textures Dataset (DTD) <https://www.robots.ox.ac.uk/~vg...
from __future__ import annotations import math from pathlib import Path import numpy as np import pytest from tokenizers import Tokenizer from sentence_transformers import SentenceTransformer from sentence_transformers.models.StaticEmbedding import StaticEmbedding try: import model2vec except ImportError: m...
from __future__ import annotations from pathlib import Path import numpy as np import pytest from tokenizers import Tokenizer from sentence_transformers.models.StaticEmbedding import StaticEmbedding try: import model2vec except ImportError: model2vec = None skip_if_no_model2vec = pytest.mark.skipif(model2v...
# Copyright (c) OpenMMLab. All rights reserved. import torch from ..builder import BBOX_SAMPLERS from ..transforms import bbox2roi from .base_sampler import BaseSampler @BBOX_SAMPLERS.register_module() class OHEMSampler(BaseSampler): r"""Online Hard Example Mining Sampler described in `Training Region-based ...
# Copyright (c) OpenMMLab. All rights reserved. import torch from ..builder import BBOX_SAMPLERS from ..transforms import bbox2roi from .base_sampler import BaseSampler @BBOX_SAMPLERS.register_module() class OHEMSampler(BaseSampler): r"""Online Hard Example Mining Sampler described in `Training Region-based ...
# Copyright (c) OpenMMLab. All rights reserved. from .checkpoint_hook import CheckpointHook from .ema_hook import EMAHook from .empty_cache_hook import EmptyCacheHook from .hook import Hook from .iter_timer_hook import IterTimerHook from .logger_hook import LoggerHook from .naive_visualization_hook import NaiveVisualiz...
# Copyright (c) OpenMMLab. All rights reserved. from .checkpoint_hook import CheckpointHook from .ema_hook import EMAHook from .empty_cache_hook import EmptyCacheHook from .hook import Hook from .iter_timer_hook import IterTimerHook from .logger_hook import LoggerHook from .naive_visualization_hook import NaiveVisualiz...
import types from keras.src.activations.activations import celu from keras.src.activations.activations import elu from keras.src.activations.activations import exponential from keras.src.activations.activations import gelu from keras.src.activations.activations import hard_sigmoid from keras.src.activations.activation...
import types from keras.src.activations.activations import elu from keras.src.activations.activations import exponential from keras.src.activations.activations import gelu from keras.src.activations.activations import hard_sigmoid from keras.src.activations.activations import hard_silu from keras.src.activations.activ...
import os import urllib.parse from typing import Dict, Union, Optional from llama_index.core.base.llms.generic_utils import ( get_from_param_or_env, ) # Import SecretStr directly from pydantic # since there is not one in llama_index.core.bridge.pydantic from pydantic import SecretStr def resolve_watsonx_creden...
import os import urllib.parse from typing import Dict, Union, Optional from llama_index.core.base.llms.generic_utils import ( get_from_param_or_env, ) # Import SecretStr directly from pydantic # since there is not one in llama_index.core.bridge.pydantic from pydantic import SecretStr def resolve_watsonx_creden...
"""Test formatting functionality.""" from typing import Union import pytest from langchain_core.agents import AgentAction, AgentActionMessageLog, AgentFinish from langchain_core.documents import Document from langchain_core.messages import ( AIMessage, AIMessageChunk, ChatMessage, ChatMessageChunk, ...
"""Test formatting functionality.""" from typing import Union import pytest from langchain_core.agents import AgentAction, AgentActionMessageLog, AgentFinish from langchain_core.documents import Document from langchain_core.messages import ( AIMessage, AIMessageChunk, ChatMessage, ChatMessageChunk, ...
SYSTEM_MESSAGE_PREFIX = """Answer the following questions as best you can. You have access to the following tools:""" # noqa: E501 FORMAT_INSTRUCTIONS = """The way you use the tools is by specifying a json blob. Specifically, this json should have a `action` key (with the name of the tool to use) and a `action_input` ...
# flake8: noqa SYSTEM_MESSAGE_PREFIX = """Answer the following questions as best you can. You have access to the following tools:""" FORMAT_INSTRUCTIONS = """The way you use the tools is by specifying a json blob. Specifically, this json should have a `action` key (with the name of the tool to use) and a `action_input`...
# Copyright (c) OpenMMLab. All rights reserved. from .base_det_dataset import BaseDetDataset from .cityscapes import CityscapesDataset from .coco import CocoDataset from .coco_panoptic import CocoPanopticDataset from .crowdhuman import CrowdHumanDataset from .dataset_wrappers import MultiImageMixDataset from .deepfashi...
# Copyright (c) OpenMMLab. All rights reserved. from .base_det_dataset import BaseDetDataset from .cityscapes import CityscapesDataset from .coco import CocoDataset from .coco_panoptic import CocoPanopticDataset from .crowdhuman import CrowdHumanDataset from .dataset_wrappers import MultiImageMixDataset from .deepfashi...
""" Tests the correct computation of evaluation scores from BinaryClassificationEvaluator """ from __future__ import annotations import csv import gzip import os from pathlib import Path import pytest from torch.utils.data import DataLoader from sentence_transformers import ( InputExample, SentenceTransform...
""" Tests the correct computation of evaluation scores from BinaryClassificationEvaluator """ from __future__ import annotations import csv import gzip import os from pathlib import Path from torch.utils.data import DataLoader from sentence_transformers import ( InputExample, SentenceTransformer, evalua...
""" Internal helpers """ from collections.abc import Callable from functools import wraps from inspect import signature from types import ModuleType from typing import TypeVar _T = TypeVar("_T") def get_xp(xp: ModuleType) -> Callable[[Callable[..., _T]], Callable[..., _T]]: """ Decorator to automatically re...
""" Internal helpers """ from functools import wraps from inspect import signature def get_xp(xp): """ Decorator to automatically replace xp with the corresponding array module. Use like import numpy as np @get_xp(np) def func(x, /, xp, kwarg=None): return xp.func(x, kwarg=kwarg) ...
from __future__ import annotations from abc import ABC, abstractmethod from typing import TYPE_CHECKING, Optional from pydantic import BaseModel from langchain_core.runnables import run_in_executor if TYPE_CHECKING: from collections.abc import Sequence from langchain_core.callbacks import Callbacks fro...
from __future__ import annotations from abc import ABC, abstractmethod from collections.abc import Sequence from typing import Optional from pydantic import BaseModel from langchain_core.callbacks import Callbacks from langchain_core.documents import Document from langchain_core.runnables import run_in_executor cl...
""" This is a simple application for sparse encoder: Computing embeddings. we have multiple sentences and we want to compute their embeddings. The embeddings are sparse, meaning that most of the values are zero. The embeddings are stored in a sparse matrix format, which is more efficient for storage and computation. w...
""" This is a simple application for sparse encoder: Computing embeddings. we have multiple sentences and we want to compute their embeddings. The embeddings are sparse, meaning that most of the values are zero. The embeddings are stored in a sparse matrix format, which is more efficient for storage and computation. w...
import pytest from llama_index.core import MockEmbedding from llama_index.core.chat_engine.context import ( ContextChatEngine, ) from llama_index.core.indices import VectorStoreIndex from llama_index.core.llms.mock import MockLLM from llama_index.core.schema import Document SYSTEM_PROMPT = "Talk like a pirate." ...
import pytest from llama_index.core import MockEmbedding from llama_index.core.chat_engine.context import ( ContextChatEngine, ) from llama_index.core.indices import VectorStoreIndex from llama_index.core.llms.mock import MockLLM from llama_index.core.schema import Document SYSTEM_PROMPT = "Talk like a pirate." ...
import pytest from llama_index.llms.nvidia import NVIDIA as Interface from llama_index.llms.nvidia.base import BASE_URL from pytest_httpx import HTTPXMock UNKNOWN_URLS = [ "https://test_url/v1", "https://test_url/v1/", "http://test_url/v1", "http://test_url/v1/", ] @pytest.fixture() def mock_unknown...
import pytest from llama_index.llms.nvidia import NVIDIA as Interface from llama_index.llms.nvidia.base import BASE_URL, KNOWN_URLS from pytest_httpx import HTTPXMock UNKNOWN_URLS = [ "https://test_url/v1", "https://test_url/v1/", "http://test_url/v1", "http://test_url/v1/", ] @pytest.fixture() def ...
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .detr import DETR @DETECTORS.register_module() class DeformableDETR(DETR): def __init__(self, *args, **kwargs): super(DETR, self).__init__(*args, **kwargs)
from ..builder import DETECTORS from .detr import DETR @DETECTORS.register_module() class DeformableDETR(DETR): def __init__(self, *args, **kwargs): super(DETR, self).__init__(*args, **kwargs)
import os from unittest.mock import patch import pytest from langchain_openai import OpenAIEmbeddings os.environ["OPENAI_API_KEY"] = "foo" def test_openai_invalid_model_kwargs() -> None: with pytest.raises(ValueError): OpenAIEmbeddings(model_kwargs={"model": "foo"}) def test_openai_incorrect_field() ...
import os from unittest.mock import patch import pytest from langchain_openai import OpenAIEmbeddings os.environ["OPENAI_API_KEY"] = "foo" def test_openai_invalid_model_kwargs() -> None: with pytest.raises(ValueError): OpenAIEmbeddings(model_kwargs={"model": "foo"}) def test_openai_incorrect_field() ...
"""Simple Reader that reads transcript of youtube video.""" import re from typing import Any, List, Optional from youtube_transcript_api import YouTubeTranscriptApi from llama_index.core.readers.base import BasePydanticReader from llama_index.core.schema import Document from llama_index.readers.youtube_transcript.uti...
"""Simple Reader that reads transcript of youtube video.""" import re from typing import Any, List, Optional from youtube_transcript_api import YouTubeTranscriptApi from llama_index.core.readers.base import BasePydanticReader from llama_index.core.schema import Document from llama_index.readers.youtube_transcript.uti...
"""Test the criteria eval chain.""" import pytest from langchain.evaluation.criteria.eval_chain import ( _SUPPORTED_CRITERIA, Criteria, CriteriaEvalChain, CriteriaResultOutputParser, LabeledCriteriaEvalChain, ) from langchain.evaluation.schema import StringEvaluator from tests.unit_tests.llms.fake...
"""Test the criteria eval chain.""" import pytest from langchain.evaluation.criteria.eval_chain import ( _SUPPORTED_CRITERIA, Criteria, CriteriaEvalChain, CriteriaResultOutputParser, LabeledCriteriaEvalChain, ) from langchain.evaluation.schema import StringEvaluator from tests.unit_tests.llms.fake...
from typing import TYPE_CHECKING, NamedTuple, TypeVar import numpy as np from pydantic import parse_obj_as from docarray.typing import NdArray from docarray.typing.url.url_3d.url_3d import Url3D if TYPE_CHECKING: from docarray.proto import NodeProto T = TypeVar('T', bound='Mesh3DUrl') class Mesh3DLoadResult(N...
from typing import TYPE_CHECKING, Tuple, TypeVar import numpy as np from docarray.typing.url.url_3d.url_3d import Url3D if TYPE_CHECKING: from docarray.proto import NodeProto T = TypeVar('T', bound='Mesh3DUrl') class Mesh3DUrl(Url3D): """ URL to a .obj, .glb, or .ply file containing 3D mesh informatio...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '0.8.4' def parse_version_info(version_str): """Parse the version information. Args: version_str (str): version string like '0.1.0'. Returns: tuple: version information contains major, minor, micro version. """ versio...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '0.8.3' def parse_version_info(version_str): """Parse the version information. Args: version_str (str): version string like '0.1.0'. Returns: tuple: version information contains major, minor, micro version. """ versio...
"""DocumentFilter that uses an LLM chain to extract the relevant parts of documents.""" from __future__ import annotations from collections.abc import Sequence from typing import Any, Callable, Optional, cast from langchain_core.callbacks import Callbacks from langchain_core.documents import BaseDocumentCompressor, ...
"""DocumentFilter that uses an LLM chain to extract the relevant parts of documents.""" from __future__ import annotations from collections.abc import Sequence from typing import Any, Callable, Optional, cast from langchain_core.callbacks import Callbacks from langchain_core.documents import BaseDocumentCompressor, ...
import numpy as np import pytest import torch from pydantic.tools import parse_obj_as, schema_json_of from docarray import BaseDoc from docarray.base_doc.io.json import orjson_dumps from docarray.typing import AnyEmbedding, NdArrayEmbedding, TorchEmbedding from docarray.utils._internal.misc import is_tf_available tf_...
import numpy as np import pytest from pydantic.tools import parse_obj_as, schema_json_of from docarray.base_doc.io.json import orjson_dumps from docarray.typing import AnyEmbedding @pytest.mark.proto def test_proto_embedding(): embedding = parse_obj_as(AnyEmbedding, np.zeros((3, 224, 224))) embedding._to_no...