input
stringlengths
33
5k
output
stringlengths
32
5k
import functools import os import os.path import pathlib from typing import Any, BinaryIO, Collection, Dict, List, Optional, Tuple, Union from torchdata.datapipes.iter import FileLister, FileOpener, Filter, IterDataPipe, Mapper from torchvision.prototype.datasets.utils import EncodedData, EncodedImage from torchvision...
import functools import os import os.path import pathlib from typing import Any, BinaryIO, Collection, Dict, List, Optional, Tuple, Union from torchdata.datapipes.iter import FileLister, FileOpener, Filter, IterDataPipe, Mapper from torchvision.prototype.datasets.utils._internal import hint_sharding, hint_shuffling fr...
from hubble.executor.hubio import HubIO from jina.orchestrate.pods.factory import PodFactory from jina.parsers import set_pod_parser def test_container_pod(mocker, monkeypatch): mock = mocker.Mock() def _mock_pull(self): return 'docker://jinahub/dummy_executor' monkeypatch.setattr(HubIO, 'pull'...
from jina.orchestrate.pods.factory import PodFactory from jina.hubble.hubio import HubIO from jina.parsers import set_pod_parser def test_container_pod(mocker, monkeypatch): mock = mocker.Mock() def _mock_pull(self): return 'docker://jinahub/dummy_executor' monkeypatch.setattr(HubIO, 'pull', _m...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp import shutil import time from unittest import TestCase from unittest.mock import Mock import torch from mmengine.structures import InstanceData from mmdet.engine.hooks import DetVisualizationHook from mmdet.structures import DetDataSample from mmd...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp import shutil import time from unittest import TestCase from unittest.mock import Mock import torch from mmengine.data import InstanceData from mmdet.engine.hooks import DetVisualizationHook from mmdet.structures import DetDataSample from mmdet.vis...
# Copyright (c) OpenMMLab. All rights reserved. from .coco_api import COCO, COCOeval __all__ = ['COCO', 'COCOeval']
from .coco_api import COCO, COCOeval __all__ = ['COCO', 'COCOeval']
__version__ = '0.1.0' from docarray.array import DocumentArray from docarray.document.document import BaseDocument as Document from docarray.predefined_document import Image, Text __all__ = ['Document', 'DocumentArray', 'Image', 'Text']
__version__ = '0.18.2' import os from docarray.document import Document from docarray.array import DocumentArray from docarray.dataclasses import dataclass, field from docarray.helper import login, logout if 'DA_RICH_HANDLER' in os.environ: from rich.traceback import install install()
from keras.src import ops from keras.src.api_export import keras_export from keras.src.layers.attention.attention import Attention @keras_export("keras.layers.AdditiveAttention") class AdditiveAttention(Attention): """Additive attention layer, a.k.a. Bahdanau-style attention. Inputs are a list with 2 or 3 el...
from keras.src import ops from keras.src.api_export import keras_export from keras.src.layers.attention.attention import Attention @keras_export("keras.layers.AdditiveAttention") class AdditiveAttention(Attention): """Additive attention layer, a.k.a. Bahdanau-style attention. Inputs are a list with 2 or 3 el...
# training schedule for 2x train_cfg = dict(type='EpochBasedTrainLoop', max_epochs=24, val_interval=1) val_cfg = dict(type='ValLoop') test_cfg = dict(type='TestLoop') # learning rate param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='Mu...
# training schedule for 2x train_cfg = dict(by_epoch=True, max_epochs=24) val_cfg = dict(interval=1) test_cfg = dict() # learning rate param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=24, ...
import torch from torch import nn from typing import List import os import json class LSTM(nn.Module): """ Bidirectional LSTM running over word embeddings. """ def __init__( self, word_embedding_dimension: int, hidden_dim: int, num_layers: int = 1, dropout: flo...
import torch from torch import nn from typing import List import os import json class LSTM(nn.Module): """ Bidirectional LSTM running over word embeddings. """ def __init__( self, word_embedding_dimension: int, hidden_dim: int, num_layers: int = 1, dropout: flo...
from enum import Enum # --8<-- [start:ProviderName] class ProviderName(str, Enum): ANTHROPIC = "anthropic" APOLLO = "apollo" COMPASS = "compass" DISCORD = "discord" D_ID = "d_id" E2B = "e2b" EXA = "exa" FAL = "fal" GITHUB = "github" GOOGLE = "google" GOOGLE_MAPS = "google_m...
from enum import Enum # --8<-- [start:ProviderName] class ProviderName(str, Enum): ANTHROPIC = "anthropic" COMPASS = "compass" DISCORD = "discord" D_ID = "d_id" E2B = "e2b" EXA = "exa" FAL = "fal" GITHUB = "github" GOOGLE = "google" GOOGLE_MAPS = "google_maps" GROQ = "groq"...
# Copyright (c) OpenMMLab. All rights reserved. import torch from .base_data_element import BaseDataElement class LabelData(BaseDataElement): """Data structure for label-level annnotations or predictions.""" @staticmethod def onehot_to_label(onehot: torch.Tensor) -> torch.Tensor: """Convert the...
# Copyright (c) OpenMMLab. All rights reserved. import torch from .base_data_element import BaseDataElement class LabelData(BaseDataElement): """Data structure for label-level annnotations or predictions.""" @staticmethod def onehot_to_label(onehot: torch.Tensor) -> torch.Tensor: """Convert the...
"""Prompt display utils.""" from llama_index.core.prompts.mixin import PromptDictType # define prompt viewing function def display_prompt_dict(prompts_dict: PromptDictType) -> None: """ Display prompt dict. Args: prompts_dict: prompt dict """ from IPython.display import Markdown, displa...
"""Prompt display utils.""" from llama_index.core.prompts.mixin import PromptDictType # define prompt viewing function def display_prompt_dict(prompts_dict: PromptDictType) -> None: """ Display prompt dict. Args: prompts_dict: prompt dict """ from IPython.display import Markdown, displa...
"""Chain-of-Abstraction Output Parser.""" import asyncio import json import networkx as nx import re from collections import defaultdict from typing import Dict, Tuple from llama_index.core.tools import AsyncBaseTool, ToolOutput from llama_index.core.types import BaseOutputParser class ChainOfAbstractionParser(Base...
"""Chain-of-Abstraction Output Parser.""" import asyncio import json import networkx as nx import re from collections import defaultdict from typing import Dict, Tuple from llama_index.core.tools import AsyncBaseTool, ToolOutput from llama_index.core.types import BaseOutputParser class ChainOfAbstractionParser(Base...
import logging import os from typing import Optional from jina.importer import ImportExtensions from jina.serve.gateway import BaseGateway from jina.serve.runtimes.gateway.websocket.app import get_fastapi_app class WebSocketGateway(BaseGateway): """WebSocket Gateway implementation""" def __init__( s...
import logging import os from typing import Optional from jina import __default_host__ from jina.importer import ImportExtensions from jina.serve.gateway import BaseGateway from jina.serve.runtimes.gateway.websocket.app import get_fastapi_app class WebSocketGateway(BaseGateway): """WebSocket Gateway implementati...
import unittest import torch from mmengine.structures import PixelData from mmengine.testing import assert_allclose from mmdet.models.seg_heads import PanopticFPNHead from mmdet.structures import DetDataSample class TestPanopticFPNHead(unittest.TestCase): def test_init_weights(self): head = PanopticFPN...
import unittest import torch from mmengine.structures import PixelData from mmengine.testing import assert_allclose from mmdet.models.seg_heads import PanopticFPNHead from mmdet.structures import DetDataSample class TestPanopticFPNHead(unittest.TestCase): def test_init_weights(self): head = PanopticFPN...
"""Tests for the InMemoryStore class.""" import pytest from langchain_core.stores import InMemoryStore from langchain_tests.integration_tests.base_store import ( BaseStoreAsyncTests, BaseStoreSyncTests, ) class TestInMemoryStore(BaseStoreSyncTests[str]): @pytest.fixture def three_values(self) -> tup...
"""Tests for the InMemoryStore class.""" import pytest from langchain_core.stores import InMemoryStore from langchain_tests.integration_tests.base_store import ( BaseStoreAsyncTests, BaseStoreSyncTests, ) class TestInMemoryStore(BaseStoreSyncTests): @pytest.fixture def three_values(self) -> tuple[st...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.api.optimizers import legacy from keras.api.optimizers import schedules from keras.src.optimizers import deserialize from keras.src.optimizers import get from keras.src.optimizers import ...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.api.optimizers import legacy from keras.api.optimizers import schedules from keras.src.optimizers import deserialize from keras.src.optimizers import get from keras.src.optimizers import ...
""" This examples trains a CrossEncoder for the NLI task. A CrossEncoder takes a sentence pair as input and outputs a label. Here, it learns to predict the labels: "contradiction": 0, "entailment": 1, "neutral": 2. It does NOT produce a sentence embedding and does NOT work for individual sentences. Usage: python trai...
""" This examples trains a CrossEncoder for the NLI task. A CrossEncoder takes a sentence pair as input and outputs a label. Here, it learns to predict the labels: "contradiction": 0, "entailment": 1, "neutral": 2. It does NOT produce a sentence embedding and does NOT work for individual sentences. Usage: python trai...
_base_ = './mask-rcnn_r50_fpn_gn-ws-all_2x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://jhu/resnet101_gn_ws')))
_base_ = './mask_rcnn_r50_fpn_gn_ws-all_2x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://jhu/resnet101_gn_ws')))
"""Semi-supervised learning algorithms. These algorithms utilize small amounts of labeled data and large amounts of unlabeled data for classification tasks. """ # Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from ._label_propagation import LabelPropagation, LabelSpreading from ._self_...
"""Semi-supervised learning algorithms. These algorithms utilize small amounts of labeled data and large amounts of unlabeled data for classification tasks. """ # Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from ._label_propagation import LabelPropagation, LabelSpreading from ._self_...
"""Hatena Blog reader.""" from typing import Dict, List from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document ATOM_PUB_ENTRY_URL = "{root_endpoint}/entry" class Article: def __init__(self) -> None: self.title = "" self.content = "" self.publis...
"""Hatena Blog reader.""" from typing import Dict, List from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document ATOM_PUB_ENTRY_URL = "{root_endpoint}/entry" class Article: def __init__(self) -> None: self.title = "" self.content = "" self.publis...
from typing import Dict, List, Optional, Callable from jina.importer import ImportExtensions from jina.types.request.data import DataRequest from jina import DocumentArray from jina._docarray import docarray_v2 if docarray_v2: from docarray import DocList def get_fastapi_app( request_models_map: Dict, ...
from typing import Dict, List, Optional, Callable from jina.importer import ImportExtensions from jina.types.request.data import DataRequest from jina import DocumentArray def get_fastapi_app( request_models_map: Dict, caller: Callable, **kwargs ): """ Get the app from FastAPI as the ...
# Copyright (c) OpenMMLab. All rights reserved. from .approx_max_iou_assigner import ApproxMaxIoUAssigner from .assign_result import AssignResult from .atss_assigner import ATSSAssigner from .base_assigner import BaseAssigner from .center_region_assigner import CenterRegionAssigner from .grid_assigner import GridAssign...
# Copyright (c) OpenMMLab. All rights reserved. from .approx_max_iou_assigner import ApproxMaxIoUAssigner from .assign_result import AssignResult from .atss_assigner import ATSSAssigner from .base_assigner import BaseAssigner from .center_region_assigner import CenterRegionAssigner from .grid_assigner import GridAssign...
from abc import ABC, abstractmethod from typing import Optional, Union from .. import Dataset, DatasetDict, Features, IterableDataset, IterableDatasetDict, NamedSplit from ..utils.typing import NestedDataStructureLike, PathLike class AbstractDatasetReader(ABC): def __init__( self, path_or_paths: ...
from abc import ABC, abstractmethod from typing import Optional, Union from .. import Dataset, DatasetDict, Features, IterableDataset, IterableDatasetDict, NamedSplit from ..utils.typing import NestedDataStructureLike, PathLike class AbstractDatasetReader(ABC): def __init__( self, path_or_paths: ...
# coding=utf-8 # Copyright 2025 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2025 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
"""Test chat model integration using standard integration tests.""" from typing import Type from langchain_tests.integration_tests import ChatModelIntegrationTests from langchain_ollama.chat_models import ChatOllama class TestChatOllama(ChatModelIntegrationTests): @property def chat_model_class(self) -> Ty...
"""Test chat model integration using standard integration tests.""" from typing import Type import pytest from langchain_core.language_models import BaseChatModel from langchain_tests.integration_tests import ChatModelIntegrationTests from langchain_ollama.chat_models import ChatOllama class TestChatOllama(ChatMod...
import json import logging import os from typing import Dict, Optional import fsspec from llama_index.core.storage.kvstore.types import ( DEFAULT_COLLECTION, BaseInMemoryKVStore, ) logger = logging.getLogger(__name__) DATA_TYPE = Dict[str, Dict[str, dict]] class SimpleKVStore(BaseInMemoryKVStore): """ ...
import json import logging import os from typing import Dict, Optional import fsspec from llama_index.core.storage.kvstore.types import ( DEFAULT_COLLECTION, BaseInMemoryKVStore, ) logger = logging.getLogger(__name__) DATA_TYPE = Dict[str, Dict[str, dict]] class SimpleKVStore(BaseInMemoryKVStore): """S...
import re from io import BytesIO from pathlib import Path from typing import Any, Type import numpy as np import pytest from langchain_core.documents.base import Blob from langchain_core.language_models import FakeMessagesListChatModel from langchain_core.messages import ChatMessage from langchain_community.document_...
import re from pathlib import Path from typing import Any, Type import pytest from langchain_core.documents.base import Blob from langchain_core.language_models import FakeMessagesListChatModel from langchain_core.messages import ChatMessage from langchain_community.document_loaders.parsers.images import ( LLMIma...
import logging from datasets import load_dataset from sentence_transformers.sparse_encoder import ( MLMTransformer, SparseEncoder, SparseRerankingEvaluator, SpladePooling, ) logging.basicConfig(format="%(asctime)s - %(message)s", datefmt="%Y-%m-%d %H:%M:%S", level=logging.INFO) # Initialize the SPLA...
from datasets import load_dataset from sentence_transformers.sparse_encoder import ( MLMTransformer, SparseEncoder, SparseRerankingEvaluator, SpladePooling, ) # Initialize the SPLADE model model_name = "naver/splade-cocondenser-ensembledistil" model = SparseEncoder( modules=[ MLMTransforme...
import inspect import re from typing import Dict, List, Tuple from huggingface_hub.utils import insecure_hashlib from .arrow import arrow from .audiofolder import audiofolder from .cache import cache from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parq...
import inspect import re from typing import Dict, List, Tuple from huggingface_hub.utils import insecure_hashlib from .arrow import arrow from .audiofolder import audiofolder from .cache import cache from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parq...
from __future__ import annotations from collections.abc import Sequence from copy import deepcopy from typing import Any, Optional, Union from langchain_core._api.deprecation import deprecated from langchain_core.callbacks import Callbacks from langchain_core.documents import BaseDocumentCompressor, Document from lan...
from __future__ import annotations from collections.abc import Sequence from copy import deepcopy from typing import Any, Optional, Union from langchain_core._api.deprecation import deprecated from langchain_core.callbacks import Callbacks from langchain_core.documents import BaseDocumentCompressor, Document from lan...
from typing import TYPE_CHECKING, Any, List, Tuple, Type, TypeVar, Union import numpy as np from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.ndarray import NdArray from docarray.typing.tensor.video.video_tensor_mixin import VideoTensorMixin T = TypeVar('T', bound='VideoNdArray')...
from typing import TYPE_CHECKING, Any, List, Tuple, Type, TypeVar, Union import numpy as np from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.ndarray import NdArray from docarray.typing.tensor.video.video_tensor_mixin import VideoTensorMixin T = TypeVar('T', bound='VideoNdArray')...
import numpy as np import pytest from absl.testing import parameterized from keras.src import backend from keras.src import layers from keras.src import testing class UpSampling3dTest(testing.TestCase): @parameterized.product( data_format=["channels_first", "channels_last"], length_dim1=[2, 3], ...
import numpy as np import pytest from absl.testing import parameterized from keras.src import backend from keras.src import layers from keras.src import testing class UpSampling3dTest(testing.TestCase, parameterized.TestCase): @parameterized.product( data_format=["channels_first", "channels_last"], ...
""" =================================== Demo of DBSCAN clustering algorithm =================================== DBSCAN (Density-Based Spatial Clustering of Applications with Noise) finds core samples in regions of high density and expands clusters from them. This algorithm is good for data which contains clusters of s...
""" =================================== Demo of DBSCAN clustering algorithm =================================== DBSCAN (Density-Based Spatial Clustering of Applications with Noise) finds core samples in regions of high density and expands clusters from them. This algorithm is good for data which contains clusters of s...
_base_ = './faster-rcnn_s50_fpn_syncbn-backbone+head_ms-range-1x_coco.py' model = dict( backbone=dict( stem_channels=128, depth=101, init_cfg=dict(type='Pretrained', checkpoint='open-mmlab://resnest101')))
_base_ = './faster_rcnn_s50_fpn_syncbn-backbone+head_mstrain-range_1x_coco.py' model = dict( backbone=dict( stem_channels=128, depth=101, init_cfg=dict(type='Pretrained', checkpoint='open-mmlab://resnest101')))
"""General node utils.""" import logging import uuid from typing import List, Optional, Protocol, runtime_checkable from llama_index.core.schema import ( BaseNode, Document, ImageDocument, ImageNode, NodeRelationship, TextNode, ) from llama_index.core.utils import truncate_text logger = loggi...
"""General node utils.""" import logging import uuid from typing import List, Optional, Protocol, runtime_checkable from llama_index.core.schema import ( BaseNode, Document, ImageDocument, ImageNode, NodeRelationship, TextNode, ) from llama_index.core.utils import truncate_text logger = loggi...
from workflows.types import StopEventT, RunResultT # noqa
from typing import Any, TypeVar, Union from .events import StopEvent StopEventT = TypeVar("StopEventT", bound=StopEvent) # TODO: When releasing 1.0, remove support for Any # and enforce usage of StopEventT RunResultT = Union[StopEventT, Any]
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import os import librosa from jina import Flow, Document, DocumentArray from ... import AudioCLIPEncoder cur_dir = os.path.dirname(os.path.abspath(__file__)) def test_flow_from_yml(): doc = DocumentArray...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import os import librosa from jina import Flow, Document, DocumentArray cur_dir = os.path.dirname(os.path.abspath(__file__)) def test_flow_from_yml(): doc = DocumentArray([Document()]) with Flow.load...
import re from typing import TYPE_CHECKING, Any, Dict, Union if TYPE_CHECKING: from sentence_transformers.SentenceTransformer import SentenceTransformer class SentenceEvaluator: """ Base class for all evaluators Extend this class and implement __call__ for custom evaluators. """ def __init_...
import re from typing import TYPE_CHECKING, Any, Dict, Union if TYPE_CHECKING: from sentence_transformers.SentenceTransformer import SentenceTransformer class SentenceEvaluator: """ Base class for all evaluators Extend this class and implement __call__ for custom evaluators. """ def __init_...
_base_ = './cascade-mask-rcnn_convnext-t-p4-w7_fpn_4conv1fc-giou_amp-ms-crop-3x_coco.py' # noqa # please install mmpretrain # import mmpretrain.models to trigger register_module in mmpretrain custom_imports = dict( imports=['mmpretrain.models'], allow_failed_imports=False) checkpoint_file = 'https://download.open...
_base_ = './cascade-mask-rcnn_convnext-t-p4-w7_fpn_4conv1fc-giou_amp-ms-crop-3x_coco.py' # noqa # TODO: delete custom_imports after mmcls supports auto import # please install mmcls>=1.0 # import mmcls.models to trigger register_module in mmcls custom_imports = dict(imports=['mmcls.models'], allow_failed_imports=Fals...
from typing import TYPE_CHECKING, Dict, Iterable from sentence_transformers.evaluation.SentenceEvaluator import SentenceEvaluator if TYPE_CHECKING: from sentence_transformers.SentenceTransformer import SentenceTransformer class SequentialEvaluator(SentenceEvaluator): """ This evaluator allows that multi...
from typing import TYPE_CHECKING, Dict, Iterable from sentence_transformers.evaluation.SentenceEvaluator import SentenceEvaluator if TYPE_CHECKING: from sentence_transformers.SentenceTransformer import SentenceTransformer class SequentialEvaluator(SentenceEvaluator): """ This evaluator allows that multi...
from typing import Any from langchain_core.documents import Document from langchain_core.messages import AIMessage, AIMessageChunk, HumanMessage class AnyStr(str): __slots__ = () def __eq__(self, other: object) -> bool: return isinstance(other, str) # The code below creates version of pydantic mod...
from typing import Any from langchain_core.documents import Document from langchain_core.messages import AIMessage, AIMessageChunk, HumanMessage class AnyStr(str): __slots__ = () def __eq__(self, other: Any) -> bool: return isinstance(other, str) # The code below creates version of pydantic models...
from typing import TypeVar from docarray.document.base_node import BaseNode from .ndarray import Embedding, Tensor from .url import ImageUrl T = TypeVar('T') __all__ = ['Tensor', 'Embedding', 'BaseNode']
from typing import ( Union, TYPE_CHECKING, TypeVar, Sequence, Optional, List, Dict, Generator, Iterable, Tuple, ForwardRef, ) if TYPE_CHECKING: # pragma: no cover import scipy.sparse import tensorflow import torch import numpy as np from PIL.Image import...
# coding=utf-8 # Copyright 2025 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
# coding=utf-8 # Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
# Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from functools import update_wrapper, wraps from types import MethodType class _AvailableIfDescriptor: """Implements a conditional property using the descriptor protocol. Using this class to create a decorator will raise an ``Att...
# Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from functools import update_wrapper, wraps from types import MethodType class _AvailableIfDescriptor: """Implements a conditional property using the descriptor protocol. Using this class to create a decorator will raise an ``Att...
from sentence_transformers import losses, SentenceTransformer, util class AnglELoss(losses.CoSENTLoss): def __init__(self, model: SentenceTransformer, scale: float = 20.0): """ This class implements AnglE (Angle Optimized) loss. This is a modification of :class:`CoSENTLoss`, designed to ad...
from sentence_transformers import losses, SentenceTransformer, util class AnglELoss(losses.CoSENTLoss): def __init__(self, model: SentenceTransformer, scale: float = 20.0): """ This class implements AnglE (Angle Optimized) loss. This is a modification of :class:`CoSENTLoss`, designed to ad...
import gc import unittest import pytest import torch from diffusers import ( StableDiffusionUpscalePipeline, ) from diffusers.utils import load_image from diffusers.utils.testing_utils import ( backend_empty_cache, enable_full_determinism, numpy_cosine_similarity_distance, require_torch_accelerato...
import gc import unittest import pytest import torch from diffusers import ( StableDiffusionUpscalePipeline, ) from diffusers.utils import load_image from diffusers.utils.testing_utils import ( backend_empty_cache, enable_full_determinism, numpy_cosine_similarity_distance, require_torch_accelerato...
""" This is a simple application for sentence embeddings: clustering Sentences are mapped to sentence embeddings and then k-mean clustering is applied. """ from sklearn.cluster import KMeans from sentence_transformers import SentenceTransformer embedder = SentenceTransformer("all-MiniLM-L6-v2") # Corpus with examp...
""" This is a simple application for sentence embeddings: clustering Sentences are mapped to sentence embeddings and then k-mean clustering is applied. """ from sentence_transformers import SentenceTransformer from sklearn.cluster import KMeans embedder = SentenceTransformer("all-MiniLM-L6-v2") # Corpus with exampl...
"""**Callback handlers** allow listening to events in LangChain. **Class hierarchy:** .. code-block:: BaseCallbackHandler --> <name>CallbackHandler # Example: AimCallbackHandler """ from typing import TYPE_CHECKING from langchain_core._import_utils import import_attr if TYPE_CHECKING: from langchain_core...
"""**Callback handlers** allow listening to events in LangChain. **Class hierarchy:** .. code-block:: BaseCallbackHandler --> <name>CallbackHandler # Example: AimCallbackHandler """ from typing import TYPE_CHECKING from langchain_core._import_utils import import_attr if TYPE_CHECKING: from langchain_core...
from docarray.document.any_document import AnyDocument from docarray.document.base_node import BaseNode from docarray.document.document import BaseDocument __all__ = ['AnyDocument', 'BaseDocument', 'BaseNode']
from docarray.document.any_document import AnyDocument from docarray.document.document import BaseDocument __all__ = ['AnyDocument', 'BaseDocument']
import torchaudio _LAZILY_IMPORTED = [ "StreamReader", "StreamReaderSourceStream", "StreamReaderSourceAudioStream", "StreamReaderSourceVideoStream", "StreamReaderOutputStream", ] def __getattr__(name: str): if name in _LAZILY_IMPORTED: torchaudio._extension._init_ffmpeg() fr...
_INITIALIZED = False _LAZILY_IMPORTED = [ "StreamReader", "StreamReaderSourceStream", "StreamReaderSourceAudioStream", "StreamReaderSourceVideoStream", "StreamReaderOutputStream", ] def _init_extension(): import torch import torchaudio try: torchaudio._extension._load_lib("lib...
"""LlamaPack class.""" from typing import Any, Dict from llama_index.core.llama_pack.base import BaseLlamaPack # backwards compatibility try: from llama_index.agent.legacy.openai_agent import OpenAIAgent except ImportError: from llama_index.agent.openai import OpenAIAgent class GmailOpenAIAgentPack(BaseLla...
"""LlamaPack class.""" from typing import Any, Dict from llama_index.core.llama_pack.base import BaseLlamaPack # backwards compatibility try: from llama_index.agent.legacy.openai_agent import OpenAIAgent except ImportError: from llama_index.agent.openai import OpenAIAgent class GmailOpenAIAgentPack(BaseLl...
from __future__ import annotations import functools import operator from typing import Any, TYPE_CHECKING import torch # NOTE: other files rely on the imports below from torch._dynamo import callback as compilation_callback # noqa: F401 from torch._inductor.runtime.cache_dir_utils import ( # noqa: F401 cache_d...
from __future__ import annotations import functools import operator from typing import Any, TYPE_CHECKING import torch # NOTE: other files rely on the imports below from torch._dynamo import callback as compilation_callback # noqa: F401 from torch._inductor.runtime.cache_dir_utils import ( # noqa: F401 cache_d...
from typing import List, Optional from llama_index.core.data_structs.data_structs import IndexStruct from llama_index.core.storage.index_store.types import BaseIndexStore from llama_index.core.storage.index_store.utils import ( index_struct_to_json, json_to_index_struct, ) from llama_index.core.storage.kvstore...
from typing import List, Optional from llama_index.core.data_structs.data_structs import IndexStruct from llama_index.core.storage.index_store.types import BaseIndexStore from llama_index.core.storage.index_store.utils import ( index_struct_to_json, json_to_index_struct, ) from llama_index.core.storage.kvstore...
_base_ = './detr_r50_8xb2-500e_coco.py' model = dict( backbone=dict( depth=18, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet18')), neck=dict(in_channels=[512]))
_base_ = './detr_r50_8xb2-500e_coco.py' model = dict( backbone=dict( depth=18, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet18')), bbox_head=dict(in_channels=512))
_base_ = './detr_r50_8xb2-150e_coco.py' # learning policy max_epochs = 500 train_cfg = dict( type='EpochBasedTrainLoop', max_epochs=max_epochs, val_interval=10) param_scheduler = [ dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[334], ...
_base_ = './detr_r50_8x2_150e_coco.py' # learning policy max_epochs = 500 train_cfg = dict( type='EpochBasedTrainLoop', max_epochs=max_epochs, val_interval=10) param_scheduler = [ dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[334], ...
_base_ = [ '../_base_/models/mask-rcnn_r50_fpn.py', '../common/lsj-100e_coco-instance.py' ] image_size = (1024, 1024) batch_augments = [ dict(type='BatchFixedSizePad', size=image_size, pad_mask=True) ] norm_cfg = dict(type='SyncBN', requires_grad=True) # Use MMSyncBN that handles empty tensor in head. It ca...
_base_ = [ '../_base_/models/mask-rcnn_r50_fpn.py', '../common/lsj-100e_coco-instance.py' ] image_size = (1024, 1024) batch_augments = [ dict(type='BatchFixedSizePad', size=image_size, pad_mask=True) ] norm_cfg = dict(type='SyncBN', requires_grad=True) # Use MMSyncBN that handles empty tensor in head. It ca...
import os from pathlib import Path from torchaudio.datasets.libritts import LIBRITTS from torchaudio_unittest.common_utils import ( get_whitenoise, normalize_wav, save_wav, TempDirMixin, TorchaudioTestCase, ) _UTTERANCE_IDS = [ [19, 198, "000000", "000000"], [26, 495, "000004", "000000"], ...
import os from pathlib import Path from torchaudio.datasets.libritts import LIBRITTS from torchaudio_unittest.common_utils import ( TempDirMixin, TorchaudioTestCase, get_whitenoise, save_wav, normalize_wav, ) _UTTERANCE_IDS = [ [19, 198, "000000", "000000"], [26, 495, "000004", "000000"], ...
""" This examples trains BERT (or any other transformer model like RoBERTa, DistilBERT etc.) for the STSbenchmark from scratch. It generates sentence embeddings that can be compared using cosine-similarity to measure the similarity. Usage: python training_nli.py OR python training_nli.py pretrained_transformer_model_...
""" This examples trains BERT (or any other transformer model like RoBERTa, DistilBERT etc.) for the STSbenchmark from scratch. It generates sentence embeddings that can be compared using cosine-similarity to measure the similarity. Usage: python training_nli.py OR python training_nli.py pretrained_transformer_model_...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import os.path as osp import numpy as np from mmengine.config import Config, DictAction from mmengine.utils import ProgressBar from mmdet.models.utils import mask2ndarray from mmdet.registry import DATASETS, VISUALIZERS from mmdet.structures.bbox import ...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import os.path as osp import numpy as np from mmengine.config import Config, DictAction from mmengine.utils import ProgressBar from mmdet.models.utils import mask2ndarray from mmdet.registry import DATASETS, VISUALIZERS from mmdet.structures.bbox import ...
_base_ = [ '../_base_/models/retinanet_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] # please install mmpretrain # import mmpretrain.models to trigger register_module in mmpretrain custom_imports = dict( imports=['mmpretrain....
_base_ = [ '../_base_/models/retinanet_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] # TODO: delete custom_imports after mmcls supports auto import # please install mmcls>=1.0 # import mmcls.models to trigger register_module in m...
from sentence_transformers import SentenceTransformer, losses, util class AnglELoss(losses.CoSENTLoss): def __init__(self, model: SentenceTransformer, scale: float = 20.0) -> None: """ This class implements AnglE (Angle Optimized) loss. This is a modification of :class:`CoSENTLoss`, design...
from sentence_transformers import SentenceTransformer, losses, util class AnglELoss(losses.CoSENTLoss): def __init__(self, model: SentenceTransformer, scale: float = 20.0) -> None: """ This class implements AnglE (Angle Optimized) loss. This is a modification of :class:`CoSENTLoss`, design...
import os from pathlib import Path from typing import Any, Callable, Optional, Union import numpy as np from PIL import Image from .utils import download_url from .vision import VisionDataset class USPS(VisionDataset): """`USPS <https://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/multiclass.html#usps>`_ Dat...
import os from pathlib import Path from typing import Any, Callable, Optional, Tuple, Union import numpy as np from PIL import Image from .utils import download_url from .vision import VisionDataset class USPS(VisionDataset): """`USPS <https://www.csie.ntu.edu.tw/~cjlin/libsvmtools/datasets/multiclass.html#usps...
import numpy as np import pytest import torch from docarray import BaseDoc, DocList from docarray.array import DocVec from docarray.typing import NdArray, TorchTensor @pytest.fixture() def batch(): class Image(BaseDoc): tensor: TorchTensor[3, 224, 224] batch = DocList[Image]([Image(tensor=torch.zero...
import numpy as np import pytest import torch from docarray import BaseDoc, DocList from docarray.array import DocVec from docarray.typing import NdArray, TorchTensor @pytest.fixture() def batch(): class Image(BaseDoc): tensor: TorchTensor[3, 224, 224] batch = DocList[Image]([Image(tensor=torch.zero...
_base_ = './solov2_r50_fpn_ms-3x_coco.py' # model settings model = dict( backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, dcn=dict(type='DCNv2', deformable_groups=1, fallback_on_stride=False), stage_with_dcn=(False, True, True, True), init_c...
_base_ = 'solov2_r50_fpn_mstrain_3x_coco.py' # model settings model = dict( backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, dcn=dict(type='DCNv2', deformable_groups=1, fallback_on_stride=False), stage_with_dcn=(False, True, True, True), ini...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import tempfile from collections import OrderedDict import torch from mmcv import Config def parse_config(config_strings): temp_file = tempfile.NamedTemporaryFile() config_path = f'{temp_file.name}.py' with open(config_path, 'w') as f: ...
import argparse import tempfile from collections import OrderedDict import torch from mmcv import Config def parse_config(config_strings): temp_file = tempfile.NamedTemporaryFile() config_path = f'{temp_file.name}.py' with open(config_path, 'w') as f: f.write(config_strings) config = Config....
# Copyright (c) OpenMMLab. All rights reserved. """Collecting some commonly used type hint in mmdetection.""" from typing import Dict, List, Optional, Tuple, Union import torch from mmengine.config import ConfigDict from mmengine.data import InstanceData, PixelData from ..bbox.samplers import SamplingResult from ..da...
# Copyright (c) OpenMMLab. All rights reserved. """Collecting some commonly used type hint in mmdetection.""" from typing import Dict, List, Optional, Tuple, Union import torch from mmengine.config import ConfigDict from mmengine.data import InstanceData from ..bbox.samplers import SamplingResult from ..data_structur...
import os from typing import Literal, Optional, overload import nomic # type: ignore[import] from langchain_core.embeddings import Embeddings from nomic import embed class NomicEmbeddings(Embeddings): """NomicEmbeddings embedding model. Example: .. code-block:: python from langchain_n...
import os from typing import Literal, Optional, overload import nomic # type: ignore[import] from langchain_core.embeddings import Embeddings from nomic import embed class NomicEmbeddings(Embeddings): """NomicEmbeddings embedding model. Example: .. code-block:: python from langchain_no...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import subprocess import pytest from jina import Document, DocumentArray, Flow from laser_encoder import LaserEncoder _EMBEDDING_DIM = 1024 @pytest.mark.parametrize('request_size', [1, 10, 50, 100]) def test_...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import subprocess import pytest from jina import Document, DocumentArray, Flow from ...laser_encoder import LaserEncoder _EMBEDDING_DIM = 1024 @pytest.mark.parametrize('request_size', [1, 10, 50, 100]) def t...
import os import shutil from typing import Sequence import pytest from llama_index.core.schema import BaseNode, TextNode from llama_index.core.vector_stores import VectorStoreQuery from llama_index.vector_stores.objectbox import ObjectBoxVectorStore EMBEDDING_DIM = 3 @pytest.fixture() def vectorstore(): obx = O...
import os import shutil from typing import Sequence import pytest from llama_index.core.schema import TextNode, BaseNode from llama_index.core.vector_stores import VectorStoreQuery from llama_index.vector_stores.objectbox import ObjectBoxVectorStore EMBEDDING_DIM = 3 @pytest.fixture() def vectorstore(): obx =...
import logging from argparse import ArgumentParser import sentencepiece as spm import torch import torchaudio from transforms import get_data_module logger = logging.getLogger(__name__) def compute_word_level_distance(seq1, seq2): return torchaudio.functional.edit_distance(seq1.lower().split(), seq2.lower().sp...
import logging from argparse import ArgumentParser import sentencepiece as spm import torch import torchaudio from transforms import get_data_module logger = logging.getLogger(__name__) def compute_word_level_distance(seq1, seq2): return torchaudio.functional.edit_distance(seq1.lower().split(), seq2.lower().sp...
"""**Chat Models** are a variation on language models. While Chat Models use language models under the hood, the interface they expose is a bit different. Rather than expose a "text in, text out" API, they expose an interface where "chat messages" are the inputs and outputs. **Class hierarchy:** .. code-block:: ...
"""**Chat Models** are a variation on language models. While Chat Models use language models under the hood, the interface they expose is a bit different. Rather than expose a "text in, text out" API, they expose an interface where "chat messages" are the inputs and outputs. **Class hierarchy:** .. code-block:: ...
# Copyright (c) OpenMMLab. All rights reserved. from mmengine.dist import all_reduce_params, is_distributed from mmengine.registry import HOOKS from .hook import Hook @HOOKS.register_module() class SyncBuffersHook(Hook): """Synchronize model buffers such as running_mean and running_var in BN at the end of eac...
# Copyright (c) OpenMMLab. All rights reserved. from mmengine.dist import all_reduce_params, is_distributed from mmengine.registry import HOOKS from .hook import Hook @HOOKS.register_module() class SyncBuffersHook(Hook): """Synchronize model buffers such as running_mean and running_var in BN at the end of eac...
from torchvision.transforms import AutoAugmentPolicy, InterpolationMode # usort: skip from . import functional # usort: skip from ._transform import Transform # usort: skip from ._presets import StereoMatching # usort: skip from ._augment import RandomCutmix, RandomErasing, RandomMixup, SimpleCopyPaste from ._au...
from torchvision.transforms import AutoAugmentPolicy, InterpolationMode # usort: skip from . import functional # usort: skip from ._transform import Transform # usort: skip from ._presets import StereoMatching # usort: skip from ._augment import RandomCutmix, RandomErasing, RandomMixup, SimpleCopyPaste from ._au...
# Copyright (c) OpenMMLab. All rights reserved. import math import torch from torch.utils.data import DistributedSampler as _DistributedSampler from mmdet.core.utils import sync_random_seed from mmdet.utils import get_device class DistributedSampler(_DistributedSampler): def __init__(self, data...
# Copyright (c) OpenMMLab. All rights reserved. import math import torch from torch.utils.data import DistributedSampler as _DistributedSampler from mmdet.core.utils import sync_random_seed class DistributedSampler(_DistributedSampler): def __init__(self, dataset, num_replicas...
import unittest from transformers.testing_utils import Expectations class ExpectationsTest(unittest.TestCase): def test_expectations(self): # We use the expectations below to make sure the right expectations are found for the right devices. # Each value is just a unique ID. expectations =...
import unittest from transformers.testing_utils import Expectations class ExpectationsTest(unittest.TestCase): def test_expectations(self): expectations = Expectations( { (None, None): 1, ("cuda", 8): 2, ("cuda", 7): 3, ("rocm", ...
# model settings norm_cfg = dict(type='BN', requires_grad=False) model = dict( type='FasterRCNN', backbone=dict( type='ResNet', depth=50, num_stages=3, strides=(1, 2, 2), dilations=(1, 1, 1), out_indices=(2, ), frozen_stages=1, norm_cfg=norm_cfg, ...
# model settings norm_cfg = dict(type='BN', requires_grad=False) model = dict( type='FasterRCNN', backbone=dict( type='ResNet', depth=50, num_stages=3, strides=(1, 2, 2), dilations=(1, 1, 1), out_indices=(2, ), frozen_stages=1, norm_cfg=norm_cfg, ...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '3.0.0rc6' short_version = __version__ def parse_version_info(version_str): """Parse a version string into a tuple. Args: version_str (str): The version string. Returns: tuple[int | str]: The version info, e.g., "1.3.0" is par...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '3.0.0rc5' short_version = __version__ def parse_version_info(version_str): """Parse a version string into a tuple. Args: version_str (str): The version string. Returns: tuple[int | str]: The version info, e.g., "1.3.0" is par...
import os import warnings from modulefinder import Module import torch # Don't re-order these, we need to load the _C extension (done when importing # .extensions) before entering _meta_registrations. from .extension import _HAS_OPS # usort:skip from torchvision import _meta_registrations, datasets, io, models, ops,...
import os import warnings from modulefinder import Module import torch from torchvision import _meta_registrations, datasets, io, models, ops, transforms, utils from .extension import _HAS_OPS try: from .version import __version__ # noqa: F401 except ImportError: pass # Check if torchvision is being impor...
from typing import Any, Optional from unittest import mock import pytest from langchain_community.tools.databricks._execution import ( DEFAULT_EXECUTE_FUNCTION_ARGS, EXECUTE_FUNCTION_ARG_NAME, execute_function, ) @pytest.mark.requires("databricks.sdk") @pytest.mark.parametrize( ("parameters", "execu...
from unittest import mock import pytest from langchain_community.tools.databricks._execution import ( DEFAULT_EXECUTE_FUNCTION_ARGS, EXECUTE_FUNCTION_ARG_NAME, execute_function, ) @pytest.mark.requires("databricks.sdk") @pytest.mark.parametrize( ("parameters", "execute_params"), [ ({"a":...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase from unittest.mock import patch import numpy as np from mmengine.dataset import DefaultSampler from torch.utils.data import Dataset from mmdet.datasets.samplers import AspectRatioBatchSampler class DummyDataset(Dataset): def __init_...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase from unittest.mock import patch import numpy as np from mmengine.data import DefaultSampler from torch.utils.data import Dataset from mmdet.datasets.samplers import AspectRatioBatchSampler class DummyDataset(Dataset): def __init__(s...
def __getattr__(name: str = "") -> None: """Raise an error on import since is deprecated.""" msg = ( "This module has been moved to langchain-experimental. " "For more details: https://github.com/langchain-ai/langchain/discussions/11352." "To access this code, install it with `pip instal...
def __getattr__(name: str = "") -> None: """Raise an error on import since is deprecated.""" raise AttributeError( "This module has been moved to langchain-experimental. " "For more details: https://github.com/langchain-ai/langchain/discussions/11352." "To access this code, install it wi...
import os import pytest from google.ai.generativelanguage_v1beta.types import ( FunctionCallingConfig, ToolConfig, ) from llama_index.core.base.llms.base import BaseLLM from llama_index.core.base.llms.types import ChatMessage, ImageBlock, MessageRole from llama_index.core.prompts.base import ChatPromptTemplate...
import os from llama_index.core.tools.function_tool import FunctionTool import pytest from llama_index.core.base.llms.base import BaseLLM from llama_index.core.base.llms.types import ChatMessage, ImageBlock, MessageRole from llama_index.llms.gemini import Gemini from llama_index.llms.gemini.utils import chat_message_t...
# -*- coding: utf-8 -*- """ Audio Feature Augmentation ========================== """ # When running this tutorial in Google Colab, install the required packages # with the following. # !pip install torchaudio librosa import torch import torchaudio import torchaudio.transforms as T print(torch.__version__) print(tor...
# -*- coding: utf-8 -*- """ Audio Feature Augmentation ========================== """ # When running this tutorial in Google Colab, install the required packages # with the following. # !pip install torchaudio librosa import torch import torchaudio import torchaudio.transforms as T print(torch.__version__) print(tor...
from enum import Enum from typing import Callable, Union from numpy import ndarray from torch import Tensor from .util import ( cos_sim, manhattan_sim, euclidean_sim, dot_score, pairwise_cos_sim, pairwise_manhattan_sim, pairwise_euclidean_sim, pairwise_dot_score, ) class SimilarityFun...
from enum import Enum from typing import Callable, Union from numpy import ndarray from torch import Tensor from .util import ( cos_sim, manhattan_sim, euclidean_sim, dot_score, pairwise_cos_sim, pairwise_manhattan_sim, pairwise_euclidean_sim, pairwise_dot_score, ) class SimilarityFun...
from typing import Optional, Dict, List, Set, Tuple import numpy as np import pytest import torch from docarray import DocumentArray from docarray.base_document import BaseDocument from docarray.typing import NdArray, TorchTensor @pytest.mark.proto def test_proto_simple(): class CustomDoc(BaseDocument): ...
from typing import Optional, Dict, List, Set, Tuple import numpy as np import pytest import torch from docarray import DocumentArray from docarray.base_document import BaseDocument from docarray.typing import NdArray, TorchTensor @pytest.mark.proto def test_proto_simple(): class CustomDoc(BaseDocument): ...
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def mock_emitted_deprecation_warnings(monkeypatch): monkeypatch.setattr("datasets.utils.deprecation_utils._emitted_deprecation_warnings", set()) # Used by list_metrics @pytest.fixture def mock_hfh(monkeypatch): cla...
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def mock_emitted_deprecation_warnings(monkeypatch): monkeypatch.setattr("datasets.utils.deprecation_utils._emitted_deprecation_warnings", set()) # Used by list_metrics @pytest.fixture def mock_hfh(monkeypatch): cla...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp import tempfile from unittest import TestCase import torch from torch import Tensor from mmengine.evaluator import DumpResults from mmengine.fileio import load class TestDumpResults(TestCase): def test_init(self): with self.assertRai...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp import tempfile from unittest import TestCase import torch from torch import Tensor from mmengine.evaluator import DumpResults from mmengine.fileio import load class TestDumpResults(TestCase): def test_init(self): with self.assertRai...
from typing import Any, Optional from backend.util.request import requests class GetRequest: @classmethod def get_request( cls, url: str, headers: Optional[dict] = None, json: bool = False ) -> Any: if headers is None: headers = {} response = requests.get(url, headers=...
from typing import Any, Optional import requests class GetRequest: @classmethod def get_request( cls, url: str, headers: Optional[dict] = None, json: bool = False ) -> Any: if headers is None: headers = {} response = requests.get(url, headers=headers) response....
import socket from dataclasses import asdict import numpy as np import pytest from loky import get_reusable_executor import xgboost as xgb from xgboost import RabitTracker, build_info, federated from xgboost import testing as tm from xgboost.collective import Config def run_rabit_worker(rabit_env: dict, world_size:...
import socket import sys from threading import Thread import numpy as np import pytest from loky import get_reusable_executor import xgboost as xgb from xgboost import RabitTracker, build_info, federated from xgboost import testing as tm def run_rabit_worker(rabit_env: dict, world_size: int) -> int: with xgb.co...
from .conv_emformer import ConvEmformer from .rnnt import conformer_rnnt_base, conformer_rnnt_model __all__ = [ "conformer_rnnt_base", "conformer_rnnt_model", "ConvEmformer", ]
from .conv_emformer import ConvEmformer from .conv_tasnet import conv_tasnet_base from .rnnt import conformer_rnnt_base, conformer_rnnt_model __all__ = [ "conformer_rnnt_base", "conformer_rnnt_model", "conv_tasnet_base", "ConvEmformer", ]
"""Module containing the base parser for arguments of Jina.""" import argparse from jina.parsers.helper import _chf def set_base_parser(): """Set the base parser :return: the parser """ from jina import __version__ from jina.helper import colored, format_full_version_info, get_full_version ...
"""Module containing the base parser for arguments of Jina.""" import argparse from jina.parsers.helper import _chf def set_base_parser(): """Set the base parser :return: the parser """ from jina import __version__ from jina.helper import colored, format_full_version_info, get_full_version ...
"""Callback Handler that tracks AIMessage.usage_metadata.""" import threading from collections.abc import Generator from contextlib import contextmanager from contextvars import ContextVar from typing import Any, Optional from langchain_core._api import beta from langchain_core.callbacks import BaseCallbackHandler fr...
"""Callback Handler that tracks AIMessage.usage_metadata.""" import threading from collections.abc import Generator from contextlib import contextmanager from contextvars import ContextVar from typing import Any, Optional from langchain_core._api import beta from langchain_core.callbacks import BaseCallbackHandler fr...
import os import numpy as np import pytest from jina import Document, DocumentArray from .. import NumpySearcher TOP_K = 5 cur_dir = os.path.dirname(os.path.abspath(__file__)) @pytest.fixture() def query_docs(): chunks = DocumentArray([Document(embedding=np.random.random(7))]) root_doc = Document(embedding...
import os import numpy as np import pytest from jina import Document, DocumentArray from .. import NumpySearcher TOP_K = 5 cur_dir = os.path.dirname(os.path.abspath(__file__)) def test_query_vector(tmpdir): runtime = { 'workspace': str(tmpdir), 'name': 'searcher', 'pea_id': 0, '...
from backend.blocks.jina._auth import ( JinaCredentials, JinaCredentialsField, JinaCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField from backend.util.request import Requests class JinaChunkingBlock(Block): clas...
from backend.blocks.jina._auth import ( JinaCredentials, JinaCredentialsField, JinaCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField from backend.util.request import requests class JinaChunkingBlock(Block): clas...
# Copyright (c) OpenMMLab. All rights reserved. import torch from mmdet.registry import TASK_UTILS from mmdet.structures.bbox import HorizontalBoxes, get_box_tensor from .base_bbox_coder import BaseBBoxCoder @TASK_UTILS.register_module() class YOLOBBoxCoder(BaseBBoxCoder): """YOLO BBox coder. Following `YOL...
# Copyright (c) OpenMMLab. All rights reserved. import torch from mmdet.models.utils.misc import get_box_tensor from mmdet.registry import TASK_UTILS from mmdet.structures.bbox import HorizontalBoxes from .base_bbox_coder import BaseBBoxCoder @TASK_UTILS.register_module() class YOLOBBoxCoder(BaseBBoxCoder): """Y...
# Copyright (c) OpenMMLab. All rights reserved. import os import pytest import torch import torch.distributed as torch_dist import torch.multiprocessing as mp import mmengine.dist as dist def _test_get_backend_non_dist(): assert dist.get_backend() is None def _test_get_world_size_non_dist(): assert dist.g...
# Copyright (c) OpenMMLab. All rights reserved. import os import pytest import torch import torch.distributed as torch_dist import torch.multiprocessing as mp import mmengine.dist as dist def _test_get_backend_non_dist(): assert dist.get_backend() is None def _test_get_world_size_non_dist(): assert dist.g...
from typing import Any, Callable, Dict, Optional, Sequence from llama_index.core.base.llms.types import ChatMessage, LLMMetadata from llama_index.core.callbacks import CallbackManager from llama_index.core.constants import DEFAULT_NUM_OUTPUTS, DEFAULT_TEMPERATURE from llama_index.core.base.llms.generic_utils import ge...
from typing import Any, Callable, Dict, Optional, Sequence from llama_index.core.base.llms.types import ChatMessage, LLMMetadata from llama_index.core.callbacks import CallbackManager from llama_index.core.constants import DEFAULT_NUM_OUTPUTS, DEFAULT_TEMPERATURE from llama_index.core.base.llms.generic_utils import ge...
from typing import TYPE_CHECKING, Any, Dict, Type, TypeVar from docarray.document.abstract_document import AbstractDocument from docarray.document.base_node import BaseNode if TYPE_CHECKING: from docarray.proto import DocumentProto, NodeProto try: import torch # noqa: F401 except ImportError: torch_imp...
from typing import TYPE_CHECKING, Any, Dict, Type, TypeVar from docarray.document.abstract_document import AbstractDocument from docarray.document.base_node import BaseNode if TYPE_CHECKING: from docarray.proto import DocumentProto, NodeProto try: import torch # noqa: F401 except ImportError: torch_imp...
"""Pydantic v1 compatibility shim.""" from langchain_core._api import warn_deprecated try: from pydantic.v1.main import * # noqa: F403 except ImportError: from pydantic.main import * # type: ignore[assignment,no-redef] # noqa: F403 warn_deprecated( "0.3.0", removal="1.0.0", alternative="pydanti...
"""Pydantic v1 compatibility shim.""" from langchain_core._api import warn_deprecated try: from pydantic.v1.main import * # noqa: F403 except ImportError: from pydantic.main import * # noqa: F403 warn_deprecated( "0.3.0", removal="1.0.0", alternative="pydantic.v1 or pydantic", message=( ...
"""FastAPI framework, high performance, easy to learn, fast to code, ready for production""" __version__ = "0.115.11" from starlette import status as status from .applications import FastAPI as FastAPI from .background import BackgroundTasks as BackgroundTasks from .datastructures import UploadFile as UploadFile fro...
"""FastAPI framework, high performance, easy to learn, fast to code, ready for production""" __version__ = "0.115.10" from starlette import status as status from .applications import FastAPI as FastAPI from .background import BackgroundTasks as BackgroundTasks from .datastructures import UploadFile as UploadFile fro...