input
stringlengths
33
5k
output
stringlengths
32
5k
"""Evaluation metrics for cluster analysis results. - Supervised evaluation uses a ground truth class values for each sample. - Unsupervised evaluation does not use ground truths and measures the "quality" of the model itself. """ # Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from ...
"""Evaluation metrics for cluster analysis results. - Supervised evaluation uses a ground truth class values for each sample. - Unsupervised evaluation does not use ground truths and measures the "quality" of the model itself. """ # Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from ...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase import pytest import torch from mmengine.structures import LabelData class TestLabelData(TestCase): def test_label_to_onehot(self): item = torch.tensor([1], dtype=torch.int64) num_classes = 10 onehot = LabelDa...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase import pytest import torch from mmengine.structures import LabelData class TestLabelData(TestCase): def test_label_to_onehot(self): item = torch.tensor([1], dtype=torch.int64) num_classes = 10 onehot = LabelDa...
"""Chat Message.""" from typing import Any, Literal from typing_extensions import override from langchain_core.messages.base import ( BaseMessage, BaseMessageChunk, merge_content, ) from langchain_core.utils._merge import merge_dicts class ChatMessage(BaseMessage): """Message that can be assigned a...
"""Chat Message.""" from typing import Any, Literal from typing_extensions import override from langchain_core.messages.base import ( BaseMessage, BaseMessageChunk, merge_content, ) from langchain_core.utils._merge import merge_dicts class ChatMessage(BaseMessage): """Message that can be assigned a...
from typing import Optional import numpy as np import pytest import torch from pydantic.tools import parse_obj_as, schema_json_of from docarray import BaseDocument from docarray.base_document.io.json import orjson_dumps from docarray.typing import AudioTorchTensor, AudioUrl from docarray.utils.misc import is_tf_avail...
from typing import Optional import numpy as np import pytest import torch from pydantic.tools import parse_obj_as, schema_json_of from docarray import BaseDocument from docarray.base_document.io.json import orjson_dumps from docarray.typing import AudioTorchTensor, AudioUrl from docarray.utils.misc import is_tf_avail...
"""Configuration for unit tests.""" from collections.abc import Iterator, Sequence from importlib import util from uuid import UUID import pytest from blockbuster import BlockBuster, blockbuster_ctx from pytest_mock import MockerFixture @pytest.fixture(autouse=True) def blockbuster() -> Iterator[BlockBuster]: w...
"""Configuration for unit tests.""" from collections.abc import Iterator, Sequence from importlib import util from uuid import UUID import pytest from blockbuster import BlockBuster, blockbuster_ctx from pytest_mock import MockerFixture @pytest.fixture(autouse=True) def blockbuster() -> Iterator[BlockBuster]: w...
import numpy as np import pytest from keras.src import backend from keras.src import layers from keras.src import ops from keras.src import regularizers from keras.src import testing class LayerNormalizationTest(testing.TestCase): @pytest.mark.requires_trainable_backend def test_ln_basics(self): self...
import numpy as np import pytest from keras.src import backend from keras.src import layers from keras.src import ops from keras.src import regularizers from keras.src import testing class LayerNormalizationTest(testing.TestCase): @pytest.mark.requires_trainable_backend def test_ln_basics(self): self...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class TOOD(SingleStageDetector): r"""Implementation of `TOOD: Task-aligned One-stage Object Detection. <https://arxiv.org/abs/2108.07755>`_.""" def __i...
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .single_stage import SingleStageDetector @DETECTORS.register_module() class TOOD(SingleStageDetector): r"""Implementation of `TOOD: Task-aligned One-stage Object Detection. <https://arxiv.org/abs/2108.07755>`_.""" def __...
# Copyright 2018 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable ...
# Copyright 2018 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable ...
# TODO: Remove this config after benchmarking all related configs _base_ = 'fcos_r50_caffe_fpn_gn-head_1x_coco.py' # dataset settings train_dataloader = dict(batch_size=4, num_workers=4)
# TODO: Remove this config after benchmarking all related configs _base_ = 'fcos_r50_caffe_fpn_gn-head_1x_coco.py' data = dict(samples_per_gpu=4, workers_per_gpu=4)
_base_ = 'ssj_270k_coco_instance.py' # dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' image_size = (1024, 1024) file_client_args = dict(backend='disk') # comment out the code below to use different file client # file_client_args = dict( # backend='petrel', # path_mapping=dict({ # ...
_base_ = '../_base_/default_runtime.py' # dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' img_norm_cfg = dict( mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True) image_size = (1024, 1024) file_client_args = dict(backend='disk') # Standard Scale Jittering (SSJ) resizes...
_base_ = [ 'mmpretrain::_base_/datasets/imagenet_bs256_rsb_a12.py', 'mmpretrain::_base_/schedules/imagenet_bs2048_rsb.py', 'mmpretrain::_base_/default_runtime.py' ] model = dict( type='ImageClassifier', backbone=dict( type='mmdet.CSPNeXt', arch='P5', out_indices=(4, ), ...
_base_ = [ 'mmcls::_base_/datasets/imagenet_bs256_rsb_a12.py', 'mmcls::_base_/schedules/imagenet_bs2048_rsb.py', 'mmcls::_base_/default_runtime.py' ] model = dict( type='ImageClassifier', backbone=dict( type='mmdet.CSPNeXt', arch='P5', out_indices=(4, ), expand_ratio...
import asyncio from langchain_core.callbacks import ( AsyncCallbackManagerForRetrieverRun, CallbackManagerForRetrieverRun, ) from langchain_core.documents import Document from langchain_core.retrievers import BaseRetriever class MergerRetriever(BaseRetriever): """Retriever that merges the results of mult...
import asyncio from langchain_core.callbacks import ( AsyncCallbackManagerForRetrieverRun, CallbackManagerForRetrieverRun, ) from langchain_core.documents import Document from langchain_core.retrievers import BaseRetriever class MergerRetriever(BaseRetriever): """Retriever that merges the results of mult...
# Licensed to the LF AI & Data foundation under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the "License"); # you may not use this fil...
from typing import Optional from rich.progress import ( BarColumn, MofNCompleteColumn, Progress, SpinnerColumn, Text, TextColumn, TimeElapsedColumn, TimeRemainingColumn, ) class _QPSColumn(TextColumn): def render(self, task) -> Text: if task.speed: _text = f'{t...
import numpy as np import pytest from pydantic import parse_obj_as from docarray.base_doc.doc import BaseDoc from docarray.documents import Mesh3D from tests import TOYDATA_DIR LOCAL_OBJ_FILE = str(TOYDATA_DIR / 'tetrahedron.obj') REMOTE_OBJ_FILE = 'https://people.sc.fsu.edu/~jburkardt/data/obj/al.obj' @pytest.mark...
import numpy as np import pytest from pydantic import parse_obj_as from docarray.base_doc.doc import BaseDoc from docarray.documents import Mesh3D from tests import TOYDATA_DIR LOCAL_OBJ_FILE = str(TOYDATA_DIR / 'tetrahedron.obj') REMOTE_OBJ_FILE = 'https://people.sc.fsu.edu/~jburkardt/data/obj/al.obj' @pytest.mark...
_base_ = [ '../_base_/models/cascade-mask-rcnn_r50_fpn.py', '../_base_/datasets/lvis_v1_instance.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvi...
_base_ = [ '../_base_/models/cascade-mask-rcnn_r50_fpn.py', '../_base_/datasets/lvis_v1_instance.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvi...
import logging from typing import Any, Callable, List from llama_index.core.node_parser.interface import TextSplitter logger = logging.getLogger(__name__) def truncate_text(text: str, text_splitter: TextSplitter) -> str: """Truncate text to fit within the chunk size. Args: text (str): The text to t...
import logging from typing import Any, Callable, List from llama_index.core.node_parser.interface import TextSplitter logger = logging.getLogger(__name__) def truncate_text(text: str, text_splitter: TextSplitter) -> str: """Truncate text to fit within the chunk size.""" chunks = text_splitter.split_text(tex...
import pathlib from typing import Any, BinaryIO, Dict, List, Tuple, Union import numpy as np from torchdata.datapipes.iter import IterDataPipe, Mapper, UnBatcher from torchvision.prototype.datapoints import Image, Label from torchvision.prototype.datasets.utils import Dataset, HttpResource, OnlineResource from torchvi...
import pathlib from typing import Any, BinaryIO, Dict, List, Tuple, Union import numpy as np from torchdata.datapipes.iter import IterDataPipe, Mapper, UnBatcher from torchvision.prototype.datasets.utils import Dataset, HttpResource, OnlineResource from torchvision.prototype.datasets.utils._internal import hint_shardi...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/lvis_v1_instance.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( roi_head=dict( bbox_head=dict(num_classes=1203), mask_head=dict(num_classes=1203)), test_cfg=dict( rcnn=d...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/lvis_v1_instance.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( roi_head=dict( bbox_head=dict(num_classes=1203), mask_head=dict(num_classes=1203)), test_cfg=dict( rcnn=d...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.document_loaders.baiducloud_bos_file import ( BaiduBOSFileLoader, ) # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.document_loaders.baiducloud_bos_file import ( BaiduBOSFileLoader, ) # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation...
_base_ = [ '../_base_/models/mask-rcnn_r50_fpn.py', '../_base_/datasets/coco_instance.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] # please install mmpretrain # import mmpretrain.models to trigger register_module in mmpretrain custom_imports = dict( imports=['mmpretrain.m...
_base_ = [ '../_base_/models/mask-rcnn_r50_fpn.py', '../_base_/datasets/coco_instance.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] # TODO: delete custom_imports after mmcls supports auto import # please install mmcls>=1.0 # import mmcls.models to trigger register_module in mm...
# Copyright (c) OpenMMLab. All rights reserved. from abc import ABCMeta, abstractmethod from typing import Dict, List, Tuple, Union import torch.nn.functional as F from mmengine.model import BaseModule from torch import Tensor from mmdet.core.utils import ConfigType, OptMultiConfig, SampleList from mmdet.registry imp...
# Copyright (c) OpenMMLab. All rights reserved. from abc import ABCMeta, abstractmethod from typing import Dict, List, Tuple, Union import torch.nn.functional as F from mmengine.model import BaseModule from torch import Tensor from mmdet.core.utils import ConfigType, OptMultiConfig, SampleList from mmdet.registry imp...
""" A quantized model executes some or all of the operations with integers rather than floating point values. This allows for a more compact models and the use of high performance vectorized operations on many hardware platforms. As a result, you get about 40% smaller and faster models. The speed-up depends on your CP...
""" A quantized model executes some or all of the operations with integers rather than floating point values. This allows for a more compact models and the use of high performance vectorized operations on many hardware platforms. As a result, you get about 40% smaller and faster models. The speed-up depends on your CP...
import importlib import pytest from fastapi.testclient import TestClient from ...utils import needs_py39 @pytest.fixture( name="client", params=[ "tutorial005", pytest.param("tutorial005_py39", marks=needs_py39), ], ) def get_client(request: pytest.FixtureRequest): mod = importlib.im...
from fastapi.testclient import TestClient from docs_src.extra_models.tutorial005 import app client = TestClient(app) def test_get_items(): response = client.get("/keyword-weights/") assert response.status_code == 200, response.text assert response.json() == {"foo": 2.3, "bar": 3.4} def test_openapi_sc...
from typing import Any, Dict, Iterable import torch from torch import Tensor, nn from sentence_transformers.SentenceTransformer import SentenceTransformer from sentence_transformers.util import fullname class CosineSimilarityLoss(nn.Module): def __init__(self, model: SentenceTransformer, loss_fct=nn.MSELoss(), ...
import torch from torch import nn, Tensor from typing import Iterable, Dict from ..SentenceTransformer import SentenceTransformer class CosineSimilarityLoss(nn.Module): def __init__(self, model: SentenceTransformer, loss_fct=nn.MSELoss(), cos_score_transformation=nn.Identity()): """ CosineSimilari...
import numpy as np import pytest from docarray.utils.misc import is_tf_available tf_available = is_tf_available() if tf_available: import tensorflow as tf from docarray.computation.tensorflow_backend import TensorFlowCompBackend from docarray.typing import TensorFlowTensor @pytest.mark.tensorflow @pyte...
import numpy as np import pytest try: import tensorflow as tf from docarray.computation.tensorflow_backend import TensorFlowCompBackend from docarray.typing import TensorFlowTensor except (ImportError, TypeError): pass @pytest.mark.tensorflow @pytest.mark.parametrize( 'shape,result', [ ...
from typing import List, Optional, Type from langchain_core.callbacks import CallbackManagerForToolRun from pydantic import BaseModel, Field from langchain_community.tools.office365.base import O365BaseTool class CreateDraftMessageSchema(BaseModel): """Input for SendMessageTool.""" body: str = Field( ...
from typing import List, Optional, Type from langchain_core.callbacks import CallbackManagerForToolRun from pydantic import BaseModel, Field from langchain_community.tools.office365.base import O365BaseTool class CreateDraftMessageSchema(BaseModel): """Input for SendMessageTool.""" body: str = Field( ...
# mypy: allow-untyped-defs import torch.distributed as dist from torch._C._distributed_c10d import FakeProcessGroup class FakeStore(dist.Store): """ A fake store is a fake Key-Value store simply for initialization usage the of fake process group, one can either use FakeStore or HashStore. """ def _...
# mypy: allow-untyped-defs import torch.distributed as dist from torch._C._distributed_c10d import FakeProcessGroup class FakeStore(dist.Store): """ A fake store is a fake Key-Value store simply for initialization usage the of fake process group, one can either use FakeStore or HashStore. """ def _...
"""This is now a no-op and can be safely removed from your code. It used to enable the use of :class:`~sklearn.ensemble.HistGradientBoostingClassifier` and :class:`~sklearn.ensemble.HistGradientBoostingRegressor` when they were still :term:`experimental`, but these estimators are now stable and can be imported normall...
"""This is now a no-op and can be safely removed from your code. It used to enable the use of :class:`~sklearn.ensemble.HistGradientBoostingClassifier` and :class:`~sklearn.ensemble.HistGradientBoostingRegressor` when they were still :term:`experimental`, but these estimators are now stable and can be imported normall...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import RedditSearchRun, RedditSearchSchema # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handling option...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import RedditSearchRun, RedditSearchSchema # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handling option...
from __future__ import annotations from collections.abc import Iterable from typing import Any import torch from torch import Tensor, nn from sentence_transformers import util from sentence_transformers.SentenceTransformer import SentenceTransformer class MultipleNegativesSymmetricRankingLoss(nn.Module): def _...
from __future__ import annotations from typing import Any, Iterable import torch from torch import Tensor, nn from sentence_transformers import util from sentence_transformers.SentenceTransformer import SentenceTransformer class MultipleNegativesSymmetricRankingLoss(nn.Module): def __init__(self, model: Senten...
import enum from typing import Any, Optional import pydantic from backend.data.api_key import APIKeyPermission, APIKeyWithoutHash from backend.data.graph import Graph class WSMethod(enum.Enum): SUBSCRIBE_GRAPH_EXEC = "subscribe_graph_execution" SUBSCRIBE_GRAPH_EXECS = "subscribe_graph_executions" UNSUBS...
import enum from typing import Any, Optional import pydantic from backend.data.api_key import APIKeyPermission, APIKeyWithoutHash from backend.data.graph import Graph class WSMethod(enum.Enum): SUBSCRIBE_GRAPH_EXEC = "subscribe_graph_execution" UNSUBSCRIBE = "unsubscribe" GRAPH_EXECUTION_EVENT = "graph_...
import itertools from dataclasses import dataclass from typing import Optional import pyarrow as pa import datasets from datasets.table import table_cast logger = datasets.utils.logging.get_logger(__name__) @dataclass class ArrowConfig(datasets.BuilderConfig): """BuilderConfig for Arrow.""" features: Opt...
import itertools from dataclasses import dataclass from typing import Optional import pyarrow as pa import datasets from datasets.table import table_cast logger = datasets.utils.logging.get_logger(__name__) @dataclass class ArrowConfig(datasets.BuilderConfig): """BuilderConfig for Arrow.""" features: Opt...
from langchain_core.utils.strings import comma_list, stringify_dict, stringify_value __all__ = ["comma_list", "stringify_dict", "stringify_value"]
from langchain_core.utils.strings import comma_list, stringify_dict, stringify_value __all__ = ["stringify_value", "stringify_dict", "comma_list"]
# Copyright (c) OpenMMLab. All rights reserved. from mmengine.config import ConfigDict from mmdet.core.utils import OptConfigType, OptMultiConfig from mmdet.registry import MODELS from .two_stage import TwoStageDetector @MODELS.register_module() class PointRend(TwoStageDetector): """PointRend: Image Segmentation...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import MODELS from .two_stage import TwoStageDetector @MODELS.register_module() class PointRend(TwoStageDetector): """PointRend: Image Segmentation as Rendering This detector is the implementation of `PointRend <https://arxiv.org/abs/191...
from __future__ import annotations import csv import logging import os from scipy.stats import pearsonr, spearmanr from sentence_transformers import InputExample logger = logging.getLogger(__name__) class CECorrelationEvaluator: """ This evaluator can be used with the CrossEncoder class. Given sentence pa...
from __future__ import annotations import csv import logging import os from scipy.stats import pearsonr, spearmanr from sentence_transformers import InputExample logger = logging.getLogger(__name__) class CECorrelationEvaluator: """ This evaluator can be used with the CrossEncoder class. Given sentence pa...
import importlib import pytest from fastapi.testclient import TestClient from pytest import MonkeyPatch from ...utils import needs_pydanticv1, needs_pydanticv2 @pytest.fixture( name="app", params=[ pytest.param("tutorial001", marks=needs_pydanticv2), pytest.param("tutorial001_pv1", marks=nee...
from fastapi.testclient import TestClient from pytest import MonkeyPatch from ...utils import needs_pydanticv2 @needs_pydanticv2 def test_settings(monkeypatch: MonkeyPatch): monkeypatch.setenv("ADMIN_EMAIL", "admin@example.com") from docs_src.settings.tutorial001 import app client = TestClient(app) ...
import numpy as np import pytest from pydantic import parse_obj_as from docarray.computation.numpy_backend import NumpyCompBackend from docarray.typing import NdArray def test_to_device(): with pytest.raises(NotImplementedError): NumpyCompBackend.to_device(np.random.rand(10, 3), 'meta') @pytest.mark.pa...
import numpy as np import pytest from pydantic import parse_obj_as from docarray.computation.numpy_backend import NumpyCompBackend from docarray.typing import NdArray def test_to_device(): with pytest.raises(NotImplementedError): NumpyCompBackend.to_device(np.random.rand(10, 3), 'meta') @pytest.mark.pa...
import numpy as np import torch from docarray import BaseDocument from docarray.typing import AnyTensor, NdArray, TorchTensor def test_set_tensor(): class MyDocument(BaseDocument): tensor: AnyTensor d = MyDocument(tensor=np.zeros((3, 224, 224))) assert isinstance(d.tensor, NdArray) assert i...
import numpy as np import torch from docarray import Document from docarray.typing import AnyTensor, NdArray, TorchTensor def test_set_tensor(): class MyDocument(Document): tensor: AnyTensor d = MyDocument(tensor=np.zeros((3, 224, 224))) assert isinstance(d.tensor, NdArray) assert isinstanc...
import re from typing import Union from langchain_core.agents import AgentAction, AgentFinish from langchain_core.exceptions import OutputParserException from langchain.agents.agent import AgentOutputParser from langchain.agents.mrkl.prompt import FORMAT_INSTRUCTIONS FINAL_ANSWER_ACTION = "Final Answer:" MISSING_ACT...
import re from typing import Union from langchain_core.agents import AgentAction, AgentFinish from langchain_core.exceptions import OutputParserException from langchain.agents.agent import AgentOutputParser from langchain.agents.mrkl.prompt import FORMAT_INSTRUCTIONS FINAL_ANSWER_ACTION = "Final Answer:" MISSING_ACT...
import http.client import json from typing import Optional def list_packages(*, contains: Optional[str] = None) -> list[str]: conn = http.client.HTTPSConnection("api.github.com") try: headers = { "Accept": "application/vnd.github+json", "X-GitHub-Api-Version": "2022-11-28", ...
import http.client import json from typing import Optional def list_packages(*, contains: Optional[str] = None): conn = http.client.HTTPSConnection("api.github.com") headers = { "Accept": "application/vnd.github+json", "X-GitHub-Api-Version": "2022-11-28", "User-Agent": "langchain-cli...
from backend.blocks.hubspot._auth import ( HubSpotCredentials, HubSpotCredentialsField, HubSpotCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField from backend.util.request import requests class HubSpotContactBlock(Bl...
from backend.blocks.hubspot._auth import ( HubSpotCredentials, HubSpotCredentialsField, HubSpotCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField from backend.util.request import requests class HubSpotContactBlock(Bl...
from typing import Dict, Tuple, Optional, List import numpy as np from jina import Executor, DocumentArray, requests, Document from jina.types.arrays.memmap import DocumentArrayMemmap from jina_commons import get_logger class SimpleIndexer(Executor): """ A simple indexer that stores all the Document data tog...
from typing import Dict, Tuple, Optional, List import numpy as np from jina import Executor, DocumentArray, requests, Document from jina.types.arrays.memmap import DocumentArrayMemmap class SimpleIndexer(Executor): """ A simple indexer that stores all the Document data together, in a DocumentArrayMemmap ...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
from keras.src.api_export import keras_export from keras.src.layers.pooling.base_pooling import BasePooling @keras_export(["keras.layers.MaxPooling2D", "keras.layers.MaxPool2D"]) class MaxPooling2D(BasePooling): """Max pooling operation for 2D spatial data. Downsamples the input along its spatial dimensions ...
from keras.src.api_export import keras_export from keras.src.layers.pooling.base_pooling import BasePooling @keras_export(["keras.layers.MaxPooling2D", "keras.layers.MaxPool2D"]) class MaxPooling2D(BasePooling): """Max pooling operation for 2D spatial data. Downsamples the input along its spatial dimensions ...
# Copyright 2021 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
# Copyright 2021 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
from typing import Optional import pytest from docarray import BaseDoc, DocArray from docarray.documents import ImageDoc from docarray.helper import ( _access_path_dict_to_nested_dict, _access_path_to_dict, _dict_to_access_paths, _is_access_path_valid, _update_nested_dicts, get_paths, ) @pyt...
from typing import Optional import pytest from docarray import BaseDocument, DocumentArray from docarray.documents import ImageDoc from docarray.helper import ( _access_path_dict_to_nested_dict, _access_path_to_dict, _dict_to_access_paths, _is_access_path_valid, _update_nested_dicts, get_paths...
"""Generate SQL queries using LlamaIndex.""" import argparse import json import logging import os import re from typing import Any, cast from llama_index import LLMPredictor, SQLDatabase from llama_index.indices import SQLStructStoreIndex from llama_index.llms.openai import OpenAI from sqlalchemy import create_engine...
"""Generate SQL queries using LlamaIndex.""" import argparse import json import logging import os import re from typing import Any, cast from llama_index import LLMPredictor, SQLDatabase from llama_index.indices import SQLStructStoreIndex from llama_index.llms.openai import OpenAI from sqlalchemy import create_engine,...
"""**sys_info** prints information about the system and langchain packages for debugging purposes.""" # noqa: E501 from collections.abc import Sequence def _get_sub_deps(packages: Sequence[str]) -> list[str]: """Get any specified sub-dependencies.""" from importlib import metadata sub_deps = set() ...
"""**sys_info** prints information about the system and langchain packages for debugging purposes.""" # noqa: E501 from collections.abc import Sequence def _get_sub_deps(packages: Sequence[str]) -> list[str]: """Get any specified sub-dependencies.""" from importlib import metadata sub_deps = set() ...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase import torch from mmengine import Config from mmengine.data import InstanceData from mmdet import * # noqa from mmdet.models.dense_heads import DDODHead class TestDDODHead(TestCase): def test_ddod_head_loss(self): """Tests d...
# Copyright (c) OpenMMLab. All rights reserved. import mmcv import torch from mmdet.models.dense_heads import DDODHead def test_ddod_head_loss(): """Tests ddod head loss when truth is empty and non-empty.""" s = 256 img_metas = [{ 'img_shape': (s, s, 3), 'scale_factor': 1, 'pad_sh...
from keras.src import ops from keras.src.api_export import keras_export from keras.src.layers.layer import Layer @keras_export("keras.layers.UnitNormalization") class UnitNormalization(Layer): """Unit normalization layer. Normalize a batch of inputs so that each input in the batch has a L2 norm equal to ...
from keras.src import ops from keras.src.api_export import keras_export from keras.src.layers.layer import Layer @keras_export("keras.layers.UnitNormalization") class UnitNormalization(Layer): """Unit normalization layer. Normalize a batch of inputs so that each input in the batch has a L2 norm equal to ...
__version__ = '0.14.1' import os from docarray.document import Document from docarray.array import DocumentArray from docarray.dataclasses import dataclass, field if 'DA_RICH_HANDLER' in os.environ: from rich.traceback import install install()
__version__ = '0.14.0' import os from docarray.document import Document from docarray.array import DocumentArray from docarray.dataclasses import dataclass, field if 'DA_RICH_HANDLER' in os.environ: from rich.traceback import install install()
from typing import Optional import pytest from docarray import BaseDocument from docarray.documents import Image from docarray.helper import ( _access_path_dict_to_nested_dict, _access_path_to_dict, _dict_to_access_paths, _is_access_path_valid, _update_nested_dicts, ) @pytest.fixture() def neste...
from typing import Optional import pytest from docarray import BaseDocument from docarray.documents import Image from docarray.helper import ( _access_path_to_dict, _dict_to_access_paths, _update_nested_dicts, is_access_path_valid, ) @pytest.fixture() def nested_doc(): class Inner(BaseDocument):...
from .conformer import Conformer from .conv_tasnet import ConvTasNet from .deepspeech import DeepSpeech from .emformer import Emformer from .rnnt import emformer_rnnt_base, emformer_rnnt_model, RNNT from .rnnt_decoder import Hypothesis, RNNTBeamSearch from .tacotron2 import Tacotron2 from .wav2letter import Wav2Letter ...
from .conformer import Conformer from .conv_tasnet import ConvTasNet from .deepspeech import DeepSpeech from .emformer import Emformer from .rnnt import RNNT, emformer_rnnt_base, emformer_rnnt_model from .rnnt_decoder import Hypothesis, RNNTBeamSearch from .tacotron2 import Tacotron2 from .wav2letter import Wav2Letter ...
"""**Graphs** provide a natural language interface to graph databases.""" from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.graphs import ( ArangoGraph, FalkorDBGraph, HugeGraph, KuzuGraph, MemgraphG...
"""**Graphs** provide a natural language interface to graph databases.""" from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.graphs import ( ArangoGraph, FalkorDBGraph, HugeGraph, KuzuGraph, MemgraphG...
# coding=utf-8 # Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
# coding=utf-8 # Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
# Copyright (c) OpenMMLab. All rights reserved. from .conditional_detr_layers import (ConditionalDetrTransformerDecoder, ConditionalDetrTransformerDecoderLayer) from .dab_detr_layers import (DABDetrTransformerDecoder, DABDetrTransformerDecoderLayer, ...
# Copyright (c) OpenMMLab. All rights reserved. from .conditional_detr_transformer import ( ConditionalDetrTransformerDecoder, ConditionalDetrTransformerDecoderLayer) from .deformable_detr_transformer import ( DeformableDetrTransformerDecoder, DeformableDetrTransformerDecoderLayer, DeformableDetrTransformer...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
# Copyright (c) OpenMMLab. All rights reserved. from .anchor_free_head import AnchorFreeHead from .anchor_head import AnchorHead from .atss_head import ATSSHead from .autoassign_head import AutoAssignHead from .cascade_rpn_head import CascadeRPNHead, StageCascadeRPNHead from .centernet_head import CenterNetHead from .c...
# Copyright (c) OpenMMLab. All rights reserved. from .anchor_free_head import AnchorFreeHead from .anchor_head import AnchorHead from .atss_head import ATSSHead from .autoassign_head import AutoAssignHead from .cascade_rpn_head import CascadeRPNHead, StageCascadeRPNHead from .centernet_head import CenterNetHead from .c...
from typing import Dict import torch.nn.functional as F from torch import Tensor, nn class Normalize(nn.Module): """This layer normalizes embeddings to unit length""" def __init__(self): super(Normalize, self).__init__() def forward(self, features: Dict[str, Tensor]): features.update({"...
from torch import Tensor from torch import nn from typing import Dict import torch.nn.functional as F class Normalize(nn.Module): """ This layer normalizes embeddings to unit length """ def __init__(self): super(Normalize, self).__init__() def forward(self, features: Dict[str, Tensor]): ...
from operator import itemgetter from typing import Sequence, Iterable from docarray.array.storage.base.getsetdel import BaseGetSetDelMixin from docarray.array.storage.base.helper import Offset2ID from docarray import Document class GetSetDelMixin(BaseGetSetDelMixin): """Implement required and derived functions t...
from operator import itemgetter from typing import Sequence, Iterable from ..base.getsetdel import BaseGetSetDelMixin from ..base.helper import Offset2ID from .... import Document class GetSetDelMixin(BaseGetSetDelMixin): """Implement required and derived functions that power `getitem`, `setitem`, `delitem`""" ...
_base_ = './mask-rcnn_r50_fpn_2x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
_base_ = './mask_rcnn_r50_fpn_2x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
from contextlib import contextmanager from threading import Lock from typing import TYPE_CHECKING, Any from expiringdict import ExpiringDict if TYPE_CHECKING: from redis import Redis from redis.lock import Lock as RedisLock class RedisKeyedMutex: """ This class provides a mutex that can be locked an...
from contextlib import contextmanager from threading import Lock from typing import TYPE_CHECKING, Any from expiringdict import ExpiringDict if TYPE_CHECKING: from redis import Redis from redis.lock import Lock as RedisLock class RedisKeyedMutex: """ This class provides a mutex that can be locked an...
# coding=utf-8 # Copyright 2025 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
# coding=utf-8 # Copyright 2025 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
from enum import Enum from typing import Dict, Iterable import torch.nn.functional as F from torch import Tensor, nn from sentence_transformers.SentenceTransformer import SentenceTransformer class TripletDistanceMetric(Enum): """The metric for the triplet loss""" COSINE = lambda x, y: 1 - F.cosine_similari...
from torch import nn, Tensor from typing import Iterable, Dict import torch.nn.functional as F from enum import Enum from ..SentenceTransformer import SentenceTransformer class TripletDistanceMetric(Enum): """ The metric for the triplet loss """ COSINE = lambda x, y: 1 - F.cosine_similarity(x, y) ...
"""Experiment with different models.""" from __future__ import annotations from collections.abc import Sequence from typing import Optional from langchain_core.language_models.llms import BaseLLM from langchain_core.prompts.prompt import PromptTemplate from langchain_core.utils.input import get_color_mapping, print_...
"""Experiment with different models.""" from __future__ import annotations from collections.abc import Sequence from typing import Optional from langchain_core.language_models.llms import BaseLLM from langchain_core.prompts.prompt import PromptTemplate from langchain_core.utils.input import get_color_mapping, print_...
# Copyright (c) OpenMMLab. All rights reserved. from functools import partial import numpy as np import torch from six.moves import map, zip from ..mask.structures import BitmapMasks, PolygonMasks def multi_apply(func, *args, **kwargs): """Apply function to a list of arguments. Note: This function ...
# Copyright (c) OpenMMLab. All rights reserved. from functools import partial import numpy as np import torch from six.moves import map, zip from ..mask.structures import BitmapMasks, PolygonMasks def multi_apply(func, *args, **kwargs): """Apply function to a list of arguments. Note: This function ...
# Copyright 2022 The Music Spectrogram Diffusion Authors. # Copyright 2025 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache...
# Copyright 2022 The Music Spectrogram Diffusion Authors. # Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache...
import warnings from typing import Any from langchain_core.memory import BaseMemory from pydantic import field_validator from langchain.memory.chat_memory import BaseChatMemory class CombinedMemory(BaseMemory): """Combining multiple memories' data together.""" memories: list[BaseMemory] """For tracking...
import warnings from typing import Any from langchain_core.memory import BaseMemory from pydantic import field_validator from langchain.memory.chat_memory import BaseChatMemory class CombinedMemory(BaseMemory): """Combining multiple memories' data together.""" memories: list[BaseMemory] """For tracking...
import asyncio from contextlib import asynccontextmanager from typing import TYPE_CHECKING, Any from expiringdict import ExpiringDict if TYPE_CHECKING: from redis.asyncio import Redis as AsyncRedis from redis.asyncio.lock import Lock as AsyncRedisLock class AsyncRedisKeyedMutex: """ This class provi...
from contextlib import contextmanager from threading import Lock from typing import TYPE_CHECKING, Any from expiringdict import ExpiringDict if TYPE_CHECKING: from redis import Redis from redis.lock import Lock as RedisLock class RedisKeyedMutex: """ This class provides a mutex that can be locked an...
from torchvision.transforms import AutoAugmentPolicy, InterpolationMode # usort: skip from . import functional # usort: skip from ._transform import Transform # usort: skip from ._augment import CutMix, MixUp, RandomErasing from ._auto_augment import AugMix, AutoAugment, RandAugment, TrivialAugmentWide from ._col...
from torchvision.transforms import AutoAugmentPolicy, InterpolationMode # usort: skip from . import functional # usort: skip from ._transform import Transform # usort: skip from ._augment import CutMix, MixUp, RandomErasing from ._auto_augment import AugMix, AutoAugment, RandAugment, TrivialAugmentWide from ._col...
_base_ = '../cascade_rcnn/cascade-mask-rcnn_r50_fpn_1x_coco.py' train_pipeline = [ dict( type='LoadImageFromFile', file_client_args={{_base_.file_client_args}}), dict( type='InstaBoost', action_candidate=('normal', 'horizontal', 'skip'), action_prob=(1, 0, 0), sc...
_base_ = '../cascade_rcnn/cascade_mask_rcnn_r50_fpn_1x_coco.py' train_pipeline = [ dict( type='LoadImageFromFile', file_client_args={{_base_.file_client_args}}), dict( type='InstaBoost', action_candidate=('normal', 'horizontal', 'skip'), action_prob=(1, 0, 0), sc...
import pytest @pytest.mark.compile def test_placeholder() -> None: """Used for compiling integration tests without running any real tests."""
import pytest @pytest.mark.compile def test_placeholder() -> None: """Used for compiling integration tests without running any real tests.""" pass
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.callbacks.infino_callback import InfinoCallbackHandler # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handling ...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.callbacks.infino_callback import InfinoCallbackHandler # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handling ...
# dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' # Example to use different file client # Method 1: simply set the data root and let the file I/O module # automatically infer from prefix (not support LMDB and Memcache yet) # data_root = 's3://openmmlab/datasets/detection/coco/' # Method 2: Us...
# dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' # file_client_args = dict( # backend='petrel', # path_mapping=dict({ # './data/': 's3://openmmlab/datasets/detection/', # 'data/': 's3://openmmlab/datasets/detection/' # })) file_client_args = dict(backend='disk') tra...
from typing import Union, Iterable from docarray.array.storage.base.seqlike import BaseSequenceLikeMixin from docarray.array.memory import DocumentArrayInMemory from docarray import Document class SequenceLikeMixin(BaseSequenceLikeMixin): """Implement sequence-like methods""" def extend(self, values: Iterab...
from typing import Union, Iterable from ..base.seqlike import BaseSequenceLikeMixin from ...memory import DocumentArrayInMemory from .... import Document class SequenceLikeMixin(BaseSequenceLikeMixin): """Implement sequence-like methods""" def extend(self, values: Iterable['Document']) -> None: docs...
import numpy as np import pytest from tensorflow import data as tf_data from keras.src import backend from keras.src import layers from keras.src import testing from keras.src.ops import convert_to_tensor class StringLookupTest(testing.TestCase): # TODO: increase coverage. Most features aren't being tested. ...
import numpy as np import pytest from tensorflow import data as tf_data from keras.src import backend from keras.src import layers from keras.src import testing from keras.src.ops import convert_to_tensor class StringLookupTest(testing.TestCase): # TODO: increase coverage. Most features aren't being tested. ...
"""Prompt display utils.""" from llama_index.core.prompts.mixin import PromptDictType # define prompt viewing function def display_prompt_dict(prompts_dict: PromptDictType) -> None: """ Display prompt dict. Args: prompts_dict: prompt dict """ from IPython.display import Markdown, displa...
"""Prompt display utils.""" from llama_index.core.prompts.mixin import PromptDictType # define prompt viewing function def display_prompt_dict(prompts_dict: PromptDictType) -> None: """Display prompt dict. Args: prompts_dict: prompt dict """ from IPython.display import Markdown, display ...
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
# Copyright (c) OpenMMLab. All rights reserved. from .activations import SiLU from .bbox_nms import fast_nms, multiclass_nms from .brick_wrappers import AdaptiveAvgPool2d, adaptive_avg_pool2d from .conv_upsample import ConvUpsample from .csp_layer import CSPLayer from .dropblock import DropBlock from .ema import ExpMom...
# Copyright (c) OpenMMLab. All rights reserved. from .activations import SiLU from .bbox_nms import fast_nms, multiclass_nms from .brick_wrappers import AdaptiveAvgPool2d, adaptive_avg_pool2d from .conv_upsample import ConvUpsample from .csp_layer import CSPLayer from .dropblock import DropBlock from .ema import ExpMom...
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .single_stage import SingleStageDetector @DETECTORS.register_module() class ATSS(SingleStageDetector): """Implementation of `ATSS <https://arxiv.org/abs/1912.02424>`_.""" def __init__(self, backbone, ...
from ..builder import DETECTORS from .single_stage import SingleStageDetector @DETECTORS.register_module() class ATSS(SingleStageDetector): """Implementation of `ATSS <https://arxiv.org/abs/1912.02424>`_.""" def __init__(self, backbone, neck, bbox_head, ...
from llama_index.core.schema import NodeRelationship, RelatedNodeInfo, TextNode from llama_index.vector_stores.qdrant import QdrantVectorStore import qdrant_client import pytest_asyncio @pytest_asyncio.fixture async def vector_store() -> QdrantVectorStore: client = qdrant_client.QdrantClient(":memory:") aclie...
from llama_index.core.schema import NodeRelationship, RelatedNodeInfo, TextNode from llama_index.vector_stores.qdrant import QdrantVectorStore import qdrant_client import pytest_asyncio @pytest_asyncio.fixture async def vector_store() -> QdrantVectorStore: client = qdrant_client.QdrantClient(":memory:") aclie...
"""Copyright 2024, XGBoost contributors""" import dask import pytest from distributed import Client from xgboost import testing as tm from xgboost.testing import dask as dtm pytestmark = [ pytest.mark.skipif(**tm.no_dask()), pytest.mark.skipif(**tm.no_dask_cuda()), tm.timeout(120), ] @pytest.mark.filte...
"""Copyright 2024, XGBoost contributors""" import dask import pytest from distributed import Client from xgboost.testing import dask as dtm @pytest.mark.filterwarnings("error") def test_no_group_split(local_cuda_client: Client) -> None: with dask.config.set( { "array.backend": "cupy", ...
# Copyright (c) OpenMMLab. All rights reserved. from .checkloss_hook import CheckInvalidLossHook from .ema import ExpMomentumEMAHook, LinearMomentumEMAHook from .memory_profiler_hook import MemoryProfilerHook from .set_epoch_info_hook import SetEpochInfoHook from .sync_norm_hook import SyncNormHook from .sync_random_si...
# Copyright (c) OpenMMLab. All rights reserved. from .checkloss_hook import CheckInvalidLossHook from .ema import ExpMomentumEMAHook, LinearMomentumEMAHook from .set_epoch_info_hook import SetEpochInfoHook from .sync_norm_hook import SyncNormHook from .sync_random_size_hook import SyncRandomSizeHook from .yolox_lrupdat...
import numpy as np import pytest from pydantic.tools import parse_obj_as, schema_json_of from docarray.base_document.io.json import orjson_dumps from docarray.typing import Mesh3DUrl, NdArray from docarray.typing.url.url_3d.mesh_url import Mesh3DLoadResult from tests import TOYDATA_DIR MESH_FILES = { 'obj': str(T...
import numpy as np import pytest from pydantic.tools import parse_obj_as, schema_json_of from docarray.base_document.io.json import orjson_dumps from docarray.typing import Mesh3DUrl from tests import TOYDATA_DIR MESH_FILES = { 'obj': str(TOYDATA_DIR / 'tetrahedron.obj'), 'glb': str(TOYDATA_DIR / 'test.glb'),...
# coding: utf-8 """Find the path to xgboost dynamic library files.""" import os import platform import sys from typing import List class XGBoostLibraryNotFound(Exception): """Error thrown by when xgboost is not found""" def find_lib_path() -> List[str]: """Find the path to xgboost dynamic library files. ...
# coding: utf-8 """Find the path to xgboost dynamic library files.""" import os import platform import sys from typing import List class XGBoostLibraryNotFound(Exception): """Error thrown by when xgboost is not found""" def find_lib_path() -> List[str]: """Find the path to xgboost dynamic library files. ...
_base_ = './yolox_s_8x8_300e_coco.py' # model settings model = dict( random_size_range=(10, 20), backbone=dict(deepen_factor=0.33, widen_factor=0.375), neck=dict(in_channels=[96, 192, 384], out_channels=96), bbox_head=dict(in_channels=96, feat_channels=96)) img_scale = (640, 640) train_pipeline = [ ...
_base_ = './yolox_s_8x8_300e_coco.py' # model settings model = dict( backbone=dict(deepen_factor=0.33, widen_factor=0.375), neck=dict(in_channels=[96, 192, 384], out_channels=96), bbox_head=dict(in_channels=96, feat_channels=96)) # dataset settings img_norm_cfg = dict( mean=[123.675, 116.28, 103.53], ...
from langchain_core.agents import AgentAction from langchain.agents.conversational.output_parser import ConvoOutputParser def test_normal_output_parsing() -> None: _test_convo_output( """ Action: my_action Action Input: my action input """, "my_action", "my action input", ) def test...
from langchain_core.agents import AgentAction from langchain.agents.conversational.output_parser import ConvoOutputParser def test_normal_output_parsing() -> None: _test_convo_output( """ Action: my_action Action Input: my action input """, "my_action", "my action input", ) def test...
_base_ = [ '../common/ms-poly_3x_coco-instance.py', '../_base_/models/mask-rcnn_r50_fpn.py' ] model = dict( backbone=dict( type='ResNeXt', depth=101, groups=32, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=di...
_base_ = [ '../common/mstrain-poly_3x_coco_instance.py', '../_base_/models/mask_rcnn_r50_fpn.py' ] model = dict( backbone=dict( type='ResNeXt', depth=101, groups=32, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_c...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.dtype_policies.dtype_policy import DTypePolicy as DTypePolicy from keras.src.dtype_policies.dtype_policy import DTypePolicy as Policy from keras.src.dtype_policies.dtype_policy import...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.dtype_policies.dtype_policy import DTypePolicy from keras.src.dtype_policies.dtype_policy import DTypePolicy as Policy from keras.src.dtype_policies.dtype_policy import dtype_policy f...
_base_ = './vfnet_r50_fpn_ms-2x_coco.py' model = dict( backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), norm_eval=True, ...
_base_ = './vfnet_r50_fpn_mstrain_2x_coco.py' model = dict( backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), norm_eval=True...
from typing import TYPE_CHECKING, Dict, Iterable from sentence_transformers.evaluation.SentenceEvaluator import SentenceEvaluator if TYPE_CHECKING: from sentence_transformers.SentenceTransformer import SentenceTransformer class SequentialEvaluator(SentenceEvaluator): """ This evaluator allows that multi...
from sentence_transformers import SentenceTransformer from . import SentenceEvaluator from typing import Iterable class SequentialEvaluator(SentenceEvaluator): """ This evaluator allows that multiple sub-evaluators are passed. When the model is evaluated, the data is passed sequentially to all sub-evaluat...
# dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' # file_client_args = dict( # backend='petrel', # path_mapping=dict({ # './data/': 's3://openmmlab/datasets/detection/', # 'data/': 's3://openmmlab/datasets/detection/' # })) file_client_args = dict(backend='disk') tra...
# dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' # file_client_args = dict( # backend='petrel', # path_mapping=dict({ # './data/': 's3://openmmlab/datasets/detection/', # 'data/': 's3://openmmlab/datasets/detection/' # })) file_client_args = dict(backend='disk') tra...
"""**Messages** are objects used in prompts and chat conversations. **Class hierarchy:** .. code-block:: BaseMessage --> SystemMessage, AIMessage, HumanMessage, ChatMessage, FunctionMessage, ToolMessage --> BaseMessageChunk --> SystemMessageChunk, AIMessageChunk, HumanMessageChunk, ChatMessageChu...
"""**Messages** are objects used in prompts and chat conversations. **Class hierarchy:** .. code-block:: BaseMessage --> SystemMessage, AIMessage, HumanMessage, ChatMessage, FunctionMessage, ToolMessage --> BaseMessageChunk --> SystemMessageChunk, AIMessageChunk, HumanMessageChunk, ChatMessageChu...
""" This is a simple application for sparse encoder: Computing embeddings. we have multiple sentences and we want to compute their embeddings. The embeddings are sparse, meaning that most of the values are zero. The embeddings are stored in a sparse matrix format, which is more efficient for storage and computation. w...
""" This is a simple application for sparse encoder: Computing embeddings. we have multiple sentences and we want to compute their embeddings. The embeddings are sparse, meaning that most of the values are zero. The embeddings are stored in a sparse matrix format, which is more efficient for storage and computation. w...
# Copyright (c) OpenMMLab. All rights reserved. import copy import os.path as osp import unittest import numpy as np import torch from mmengine.structures import InstanceData, PixelData from mmdet.datasets.transforms import PackDetInputs from mmdet.structures import DetDataSample from mmdet.structures.mask import Bit...
# Copyright (c) OpenMMLab. All rights reserved. import copy import os.path as osp import unittest import numpy as np import torch from mmengine.structures import InstanceData, PixelData from mmdet.datasets.transforms import PackDetInputs from mmdet.structures import DetDataSample from mmdet.structures.mask import Bit...
from tqdm import tqdm from typing import Any, Sequence from llama_index.core.schema import TransformComponent, BaseNode, NodeRelationship from llama_index.core.graph_stores.types import Relation, KG_NODES_KEY, KG_RELATIONS_KEY def get_node_rel_string(relationship: NodeRelationship) -> str: return str(relationshi...
from tqdm import tqdm from typing import Any, Sequence from llama_index.core.schema import TransformComponent, BaseNode, NodeRelationship from llama_index.core.graph_stores.types import Relation, KG_NODES_KEY, KG_RELATIONS_KEY def get_node_rel_string(relationship: NodeRelationship) -> str: return str(relationshi...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.chat_message_histories import CassandraChatMessageHistory # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handli...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.chat_message_histories import CassandraChatMessageHistory # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handli...
import logging from typing import List, Optional from llama_index.core.schema import Document from llama_index.readers.box import BoxReaderBase from llama_index.readers.box.BoxAPI.box_api import ( get_box_files_details, get_box_folder_files_details, get_files_ai_extract_data, box_check_connection, ) f...
import logging from typing import List, Optional from llama_index.core.schema import Document from llama_index.readers.box import BoxReaderBase from llama_index.readers.box.BoxAPI.box_api import ( get_box_files_details, get_box_folder_files_details, get_files_ai_extract_data, box_check_connection, ) f...
# Copyright (c) OpenMMLab. All rights reserved. import torch from mmdet.registry import TASK_UTILS from .base_bbox_coder import BaseBBoxCoder @TASK_UTILS.register_module() class YOLOBBoxCoder(BaseBBoxCoder): """YOLO BBox coder. Following `YOLO <https://arxiv.org/abs/1506.02640>`_, this coder divide imag...
# Copyright (c) OpenMMLab. All rights reserved. import mmcv import torch from mmdet.registry import TASK_UTILS from .base_bbox_coder import BaseBBoxCoder @TASK_UTILS.register_module() class YOLOBBoxCoder(BaseBBoxCoder): """YOLO BBox coder. Following `YOLO <https://arxiv.org/abs/1506.02640>`_, this coder div...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/lvis_v1_instance.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] model = dict( roi_head=dict( bbox_head=dict( num_classes=1203, cls_predictor_cfg=dict(type='NormedLinear', ...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/lvis_v1_instance.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] model = dict( roi_head=dict( bbox_head=dict( num_classes=1203, cls_predictor_cfg=dict(type='NormedLinear', ...