input
stringlengths
33
5k
output
stringlengths
32
5k
"""Prompt class.""" from llama_index.core.base.llms.types import ChatMessage, MessageRole from llama_index.core.prompts.base import ( BasePromptTemplate, ChatPromptTemplate, LangchainPromptTemplate, Prompt, PromptTemplate, PromptType, SelectorPromptTemplate, ) from llama_index.core.prompts....
"""Prompt class.""" from llama_index.core.base.llms.types import ChatMessage, MessageRole from llama_index.core.prompts.base import ( BasePromptTemplate, ChatPromptTemplate, LangchainPromptTemplate, Prompt, PromptTemplate, PromptType, SelectorPromptTemplate, ) from llama_index.core.prompts....
# Copyright (c) OpenMMLab. All rights reserved. import copy import platform import random import numpy as np import torch from mmdet.registry import DATASETS, TRANSFORMS if platform.system() != 'Windows': # https://github.com/pytorch/pytorch/issues/973 import resource rlimit = resource.getrlimit(resource...
# Copyright (c) OpenMMLab. All rights reserved. import copy import platform import random import numpy as np import torch from mmdet.registry import DATASETS, TRANSFORMS if platform.system() != 'Windows': # https://github.com/pytorch/pytorch/issues/973 import resource rlimit = resource.getrlimit(resource...
__version__ = '0.13.4' import os from .document import Document from .array import DocumentArray from .dataclasses import dataclass, field if 'DA_NO_RICH_HANDLER' not in os.environ: from rich.traceback import install install()
__version__ = '0.13.3' import os from .document import Document from .array import DocumentArray from .dataclasses import dataclass, field if 'DA_NO_RICH_HANDLER' not in os.environ: from rich.traceback import install install()
from __future__ import annotations from .CSRLoss import CSRLoss, CSRReconstructionLoss from .RegularizerLoss import FlopsLoss, IDFFlopsLoss from .SparseAnglELoss import SparseAnglELoss from .SparseCachedGISTEmbedLoss import SparseCachedGISTEmbedLoss from .SparseCachedMultipleNegativesRankingLoss import SparseCachedMul...
from __future__ import annotations from .CSRLoss import CSRLoss, CSRReconstructionLoss from .FlopsLoss import FlopsLoss from .SparseAnglELoss import SparseAnglELoss from .SparseCachedGISTEmbedLoss import SparseCachedGISTEmbedLoss from .SparseCachedMultipleNegativesRankingLoss import SparseCachedMultipleNegativesRankin...
"""Tests related to the `DataIter` interface.""" import numpy as np import xgboost from xgboost import testing as tm def run_mixed_sparsity(device: str) -> None: """Check QDM with mixed batches.""" X_0, y_0, _ = tm.make_regression(128, 16, False) if device.startswith("cuda"): X_1, y_1 = tm.make_...
"""Tests related to the `DataIter` interface.""" import numpy as np import xgboost from xgboost import testing as tm def run_mixed_sparsity(device: str) -> None: """Check QDM with mixed batches.""" X_0, y_0, _ = tm.make_regression(128, 16, False) if device.startswith("cuda"): X_1, y_1 = tm.make_...
_base_ = './mask_rcnn_swin-t-p4-w7_fpn_ms-crop-3x_coco.py' # Enable automatic-mixed-precision training with AmpOptimWrapper. optim_wrapper = dict(type='AmpOptimWrapper')
_base_ = './mask_rcnn_swin-t-p4-w7_fpn_ms-crop-3x_coco.py' # you need to set mode='dynamic' if you are using pytorch<=1.5.0 fp16 = dict(loss_scale=dict(init_scale=512))
from typing import Any from collections import deque from llama_index.core.bridge.pydantic import BaseModel from llama_index.core.llms.custom import CustomLLM from llama_index.core.base.llms.types import ChatMessage, MessageRole from llama_index.core.llms import LLMMetadata, CompletionResponse, CompletionResponseGen f...
from typing import Any from collections import deque from llama_index.core.bridge.pydantic import BaseModel from llama_index.core.llms.custom import CustomLLM from llama_index.core.base.llms.types import ChatMessage, MessageRole from llama_index.core.llms import LLMMetadata, CompletionResponse, CompletionResponseGen f...
# Copyright (c) OpenMMLab. All rights reserved. from mmengine.utils.dl_utils import TORCH_VERSION from mmengine.utils.version_utils import digit_version from .averaged_model import (BaseAveragedModel, ExponentialMovingAverage, MomentumAnnealingEMA, StochasticWeightAverage) from .base_model ...
# Copyright (c) OpenMMLab. All rights reserved. from mmengine.utils.parrots_wrapper import TORCH_VERSION from mmengine.utils.version_utils import digit_version from .averaged_model import (BaseAveragedModel, ExponentialMovingAverage, MomentumAnnealingEMA, StochasticWeightAverage) from .base...
import types from typing_extensions import TYPE_CHECKING from docarray.typing.tensor.video.video_ndarray import VideoNdArray from docarray.typing.tensor.video.video_tensor import VideoTensor from docarray.utils._internal.misc import ( _get_path_from_docarray_root_level, import_library, ) if TYPE_CHECKING: ...
from docarray.typing.tensor.video.video_ndarray import VideoNdArray __all__ = ['VideoNdArray'] from docarray.utils._internal.misc import is_tf_available, is_torch_available torch_available = is_torch_available() if torch_available: from docarray.typing.tensor.video.video_torch_tensor import VideoTorchTensor # n...
from keras.src import backend from keras.src import ops from keras.src.api_export import keras_export from keras.src.layers.layer import Layer @keras_export("keras.layers.Masking") class Masking(Layer): """Masks a sequence by using a mask value to skip timesteps. For each timestep in the input tensor (dimens...
from keras.src import backend from keras.src import ops from keras.src.api_export import keras_export from keras.src.layers.layer import Layer @keras_export("keras.layers.Masking") class Masking(Layer): """Masks a sequence by using a mask value to skip timesteps. For each timestep in the input tensor (dimens...
from typing import Any, Literal from pydantic import SecretStr from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import ( APIKeyCredentials, CredentialsField, CredentialsMetaInput, SchemaField, ) from backend.integrations.providers import ProviderNam...
from typing import Any, Literal from pydantic import SecretStr from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import ( APIKeyCredentials, CredentialsField, CredentialsMetaInput, SchemaField, ) from backend.util.request import requests TEST_CREDEN...
import asyncio from typing import Any, AsyncGenerator, List, Optional from llama_index.core.workflow.context import Context from llama_index.core.workflow.errors import WorkflowDone from llama_index.core.workflow.events import Event, StopEvent from .types import RunResultT from .utils import BUSY_WAIT_DELAY class W...
import asyncio from typing import Any, AsyncGenerator, Optional from llama_index.core.workflow.context import Context from llama_index.core.workflow.errors import WorkflowDone from llama_index.core.workflow.events import Event, StopEvent from .types import RunResultT from .utils import BUSY_WAIT_DELAY class Workflo...
# Copyright (c) OpenMMLab. All rights reserved. from .build_functions import (build_from_cfg, build_model_from_cfg, build_runner_from_cfg, build_scheduler_from_cfg) from .default_scope import DefaultScope from .registry import Registry from .root import (DATA_SAMPLERS, DATASETS, EVALUATOR,...
# Copyright (c) OpenMMLab. All rights reserved. from .build_functions import (build_from_cfg, build_model_from_cfg, build_runner_from_cfg, build_scheduler_from_cfg) from .default_scope import DefaultScope from .registry import Registry from .root import (DATA_SAMPLERS, DATASETS, EVALUATOR,...
from __future__ import annotations from torch import Tensor, nn from sentence_transformers.cross_encoder.CrossEncoder import CrossEncoder from sentence_transformers.util import fullname class MSELoss(nn.Module): def __init__(self, model: CrossEncoder, activation_fct: nn.Module = nn.Identity(), **kwargs) -> None...
from __future__ import annotations from torch import Tensor, nn from sentence_transformers.cross_encoder.CrossEncoder import CrossEncoder from sentence_transformers.util import fullname class MSELoss(nn.Module): def __init__(self, model: CrossEncoder, activation_fct: nn.Module = nn.Identity(), **kwargs) -> None...
"""Standard LangChain interface tests""" import os from typing import Type from langchain_core.language_models import BaseChatModel from langchain_tests.integration_tests import ChatModelIntegrationTests from langchain_openai import AzureChatOpenAI OPENAI_API_VERSION = os.environ.get("AZURE_OPENAI_API_VERSION", "")...
"""Standard LangChain interface tests""" import os from typing import Type import pytest from langchain_core.language_models import BaseChatModel from langchain_tests.integration_tests import ChatModelIntegrationTests from langchain_openai import AzureChatOpenAI OPENAI_API_VERSION = os.environ.get("AZURE_OPENAI_API...
"""Base class for Amadeus tools.""" from __future__ import annotations from typing import TYPE_CHECKING from langchain_core.tools import BaseTool from pydantic import Field from langchain_community.tools.amadeus.utils import authenticate if TYPE_CHECKING: from amadeus import Client class AmadeusBaseTool(Base...
"""Base class for Amadeus tools.""" from __future__ import annotations from typing import TYPE_CHECKING from langchain_core.tools import BaseTool from pydantic import Field from langchain_community.tools.amadeus.utils import authenticate if TYPE_CHECKING: from amadeus import Client class AmadeusBaseTool(Base...
import re from setuptools import find_packages, setup # type: ignore from pkg_resources import DistributionNotFound, get_distribution def readme(): with open('README.md', encoding='utf-8') as f: content = f.read() return content version_file = 'mmengine/version.py' def choose_requirement(primary...
from setuptools import find_packages, setup # type: ignore def readme(): with open('README.md', encoding='utf-8') as f: content = f.read() return content version_file = 'mmengine/version.py' def get_version(): with open(version_file, 'r') as f: exec(compile(f.read(), version_file, 'ex...
import logging from typing import Any, Callable, List from llama_index.core.node_parser.interface import TextSplitter from llama_index.core.utils import globals_helper logger = logging.getLogger(__name__) def truncate_text(text: str, text_splitter: TextSplitter) -> str: """Truncate text to fit within the chunk ...
import logging from typing import Any, Callable, List from llama_index.core.node_parser.interface import TextSplitter logger = logging.getLogger(__name__) def truncate_text(text: str, text_splitter: TextSplitter) -> str: """Truncate text to fit within the chunk size. Args: text (str): The text to t...
""" This script contains an example how to perform semantic search with Seismic. For more information, please refer to the documentation: https://github.com/TusKANNy/seismic/blob/main/docs/Guidelines.md All you need is installing the `pyseismic-lsr` package: ``` pip install pyseismic-lsr ``` """ import time from dat...
""" This script contains an example how to perform semantic search with Seismic. For more information, please refer to the documentation: https://github.com/TusKANNy/seismic/blob/main/docs/Guidelines.md All you need is installing the `pyseismic-lsr` package: ``` pip install pyseismic-lsr ``` """ import time from dat...
__version__ = "3.1.0.dev0" __MODEL_HUB_ORGANIZATION__ = "sentence-transformers" import importlib import os from sentence_transformers.cross_encoder.CrossEncoder import CrossEncoder from sentence_transformers.datasets import ParallelSentencesDataset, SentencesDataset from sentence_transformers.LoggingHandler import Lo...
__version__ = "3.0.0.dev0" __MODEL_HUB_ORGANIZATION__ = "sentence-transformers" import importlib import os from sentence_transformers.cross_encoder.CrossEncoder import CrossEncoder from sentence_transformers.datasets import ParallelSentencesDataset, SentencesDataset from sentence_transformers.LoggingHandler import Lo...
_base_ = [ '../_base_/models/retinanet_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] norm_cfg = dict(type='BN', requires_grad=True) model = dict( data_preprocessor=dict( type='DetDataPreprocessor', mean=[123.6...
_base_ = [ '../_base_/models/retinanet_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] norm_cfg = dict(type='BN', requires_grad=True) model = dict( data_preprocessor=dict( type='DetDataPreprocessor', mean=[123.6...
import time from functools import partial from huggingface_hub import HfApi, hf_hub_url from huggingface_hub.hf_api import RepoFile from packaging import version from requests import ConnectionError, HTTPError from .. import config from . import logging logger = logging.get_logger(__name__) # Retry `preupload_lfs_...
import time from functools import partial from huggingface_hub import HfApi, hf_hub_url from huggingface_hub.hf_api import RepoFile from packaging import version from requests import ConnectionError, HTTPError from .. import config from . import logging logger = logging.get_logger(__name__) # Retry `preupload_lfs_...
import json import logging from abc import ABC, abstractmethod from datetime import datetime from typing import Any, AsyncGenerator, Generator, Generic, TypeVar from pydantic import BaseModel from redis.asyncio.client import PubSub as AsyncPubSub from redis.client import PubSub from backend.data import redis logger ...
import json import logging from abc import ABC, abstractmethod from datetime import datetime from typing import Any, AsyncGenerator, Generator, Generic, TypeVar from pydantic import BaseModel from redis.asyncio.client import PubSub as AsyncPubSub from redis.client import PubSub from backend.data import redis from bac...
_base_ = '../mask_rcnn/mask-rcnn_r50_fpn_1x_coco.py' model = dict( type='MaskScoringRCNN', roi_head=dict( type='MaskScoringRoIHead', mask_iou_head=dict( type='MaskIoUHead', num_convs=4, num_fcs=2, roi_feat_size=14, in_channels=256, ...
_base_ = '../mask_rcnn/mask_rcnn_r50_fpn_1x_coco.py' model = dict( type='MaskScoringRCNN', roi_head=dict( type='MaskScoringRoIHead', mask_iou_head=dict( type='MaskIoUHead', num_convs=4, num_fcs=2, roi_feat_size=14, in_channels=256, ...
_base_ = [ '../_base_/models/faster-rcnn_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] checkpoint = 'https://download.pytorch.org/models/resnet50-11ad3fa6.pth' model = dict( backbone=dict(init_cfg=dict(type='Pretrained', chec...
_base_ = [ '../_base_/models/faster-rcnn_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] checkpoint = 'https://download.pytorch.org/models/resnet50-11ad3fa6.pth' model = dict( backbone=dict(init_cfg=dict(type='Pretrained', chec...
_base_ = './gfl_r50_fpn_ms-2x_coco.py' model = dict( backbone=dict( type='ResNet', depth=101, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), dcn=dict(type='DCN', deform_groups=1, fallback_on_stride=False)...
_base_ = './gfl_r50_fpn_mstrain_2x_coco.py' model = dict( backbone=dict( type='ResNet', depth=101, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), dcn=dict(type='DCN', deform_groups=1, fallback_on_stride=F...
from docarray import BaseDoc from docarray.typing import ImageUrl def test_set_image_url(): class MyDocument(BaseDoc): image_url: ImageUrl d = MyDocument(image_url="https://jina.ai/img.png") assert isinstance(d.image_url, ImageUrl) assert d.image_url == "https://jina.ai/img.png"
from docarray import BaseDocument from docarray.typing import ImageUrl def test_set_image_url(): class MyDocument(BaseDocument): image_url: ImageUrl d = MyDocument(image_url="https://jina.ai/img.png") assert isinstance(d.image_url, ImageUrl) assert d.image_url == "https://jina.ai/img.png"
# coding: utf-8 import logging import numpy as np import lightgbm as lgb def test_register_logger(tmp_path): logger = logging.getLogger("LightGBM") logger.setLevel(logging.DEBUG) formatter = logging.Formatter('%(levelname)s | %(message)s') log_filename = tmp_path / "LightGBM_test_logger.log" fil...
# coding: utf-8 import logging import numpy as np import lightgbm as lgb def test_register_logger(tmp_path): logger = logging.getLogger("LightGBM") logger.setLevel(logging.DEBUG) formatter = logging.Formatter('%(levelname)s | %(message)s') log_filename = tmp_path / "LightGBM_test_logger.log" fil...
# coding=utf-8 # Copyright 2025 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
# coding=utf-8 # Copyright 2024 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
"""Fake Chat Model wrapper for testing purposes.""" import json from typing import Any, Optional from langchain_core.callbacks import ( AsyncCallbackManagerForLLMRun, CallbackManagerForLLMRun, ) from langchain_core.language_models.chat_models import SimpleChatModel from langchain_core.messages import AIMessag...
"""Fake Chat Model wrapper for testing purposes.""" import json from typing import Any, Dict, List, Optional from langchain_core.callbacks import ( AsyncCallbackManagerForLLMRun, CallbackManagerForLLMRun, ) from langchain_core.language_models.chat_models import SimpleChatModel from langchain_core.messages imp...
""" This file evaluates CrossEncoder on the TREC 2019 Deep Learning (DL) Track: https://arxiv.org/abs/2003.07820 TREC 2019 DL is based on the corpus of MS Marco. MS Marco provides a sparse annotation, i.e., usually only a single passage is marked as relevant for a given query. Many other highly relevant passages are n...
""" This file evaluates CrossEncoder on the TREC 2019 Deep Learning (DL) Track: https://arxiv.org/abs/2003.07820 TREC 2019 DL is based on the corpus of MS Marco. MS Marco provides a sparse annotation, i.e., usually only a single passage is marked as relevant for a given query. Many other highly relevant passages are n...
"""Torch backend APIs. # Note on device placement Torch has a different device placement style compared to TF and JAX. In short, variables/tensors are not created on GPU by default, and the GPU cannot directly communicate with the CPU. To bring Torch behavior in line with TF and JAX automated device placement, we are...
"""Torch backend APIs. # Note on device placement Torch has a different device placement style compared to TF and JAX. In short, variables/tensors are not created on GPU by default, and the GPU cannot directly communicate with the CPU. To bring Torch behavior in line with TF and JAX automated device placement, we are...
# Copyright (c) OpenMMLab. All rights reserved. import mmcv import mmengine from mmengine.utils import digit_version from .version import __version__, version_info mmcv_minimum_version = '2.0.0rc4' mmcv_maximum_version = '3.0.0' mmcv_version = digit_version(mmcv.__version__) mmengine_minimum_version = '0.7.1' mmengi...
# Copyright (c) OpenMMLab. All rights reserved. import mmcv import mmengine from mmengine.utils import digit_version from .version import __version__, version_info mmcv_minimum_version = '2.0.0rc4' mmcv_maximum_version = '2.1.0' mmcv_version = digit_version(mmcv.__version__) mmengine_minimum_version = '0.7.1' mmengi...
"""Test for Serializable base class""" import json import os from typing import Any from unittest.mock import patch import pytest from langchain_core.load.dump import dumps from langchain_core.load.serializable import Serializable from pydantic import ConfigDict, Field, model_validator class Person(Serializable): ...
"""Test for Serializable base class""" import json import os from typing import Any, Dict, List from unittest.mock import patch import pytest from langchain_core.load.dump import dumps from langchain_core.load.serializable import Serializable from pydantic import ConfigDict, Field, model_validator class Person(Seri...
"""Score functions, performance metrics, pairwise metrics and distance computations.""" # Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from . import cluster from ._classification import ( accuracy_score, balanced_accuracy_score, brier_score_loss, class_likelihood_ratios...
"""Score functions, performance metrics, pairwise metrics and distance computations.""" # Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from . import cluster from ._classification import ( accuracy_score, balanced_accuracy_score, brier_score_loss, class_likelihood_ratios...
from enum import Enum from typing import Any, Optional from langchain_core.callbacks import ( AsyncCallbackManagerForRetrieverRun, CallbackManagerForRetrieverRun, ) from langchain_core.documents import Document from langchain_core.retrievers import BaseRetriever from langchain_core.stores import BaseStore, Byt...
from enum import Enum from typing import Any, Optional from langchain_core.callbacks import ( AsyncCallbackManagerForRetrieverRun, CallbackManagerForRetrieverRun, ) from langchain_core.documents import Document from langchain_core.retrievers import BaseRetriever from langchain_core.stores import BaseStore, Byt...
"""Ollama specific chat model integration tests""" from typing import Annotated, Optional import pytest from pydantic import BaseModel, Field from typing_extensions import TypedDict from langchain_ollama import ChatOllama @pytest.mark.parametrize(("method"), [("function_calling"), ("json_schema")]) def test_struct...
"""Ollama specific chat model integration tests""" from typing import List, Optional import pytest from pydantic import BaseModel, Field from typing_extensions import Annotated, TypedDict from langchain_ollama import ChatOllama @pytest.mark.parametrize(("method"), [("function_calling"), ("json_schema")]) def test_...
_base_ = ['./mask2former_r50_lsj_8x2_50e_coco.py'] model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
_base_ = './mask2former_r50_lsj_8x2_50e_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
import pytest from xgboost import testing as tm from xgboost.testing.ordinal import ( run_cat_container, run_cat_container_iter, run_cat_container_mixed, run_cat_invalid, run_cat_leaf, run_cat_predict, run_cat_shap, run_cat_thread_safety, run_specified_cat, ) pytestmark = pytest.ma...
import pytest from xgboost import testing as tm from xgboost.testing.ordinal import ( run_cat_container, run_cat_container_iter, run_cat_container_mixed, run_cat_invalid, run_cat_leaf, run_cat_predict, run_cat_shap, run_cat_thread_safety, ) pytestmark = pytest.mark.skipif(**tm.no_multi...
# coding=utf-8 # Copyright 2025 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
# coding=utf-8 # Copyright 2024 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
import re from typing import TYPE_CHECKING, Dict, Iterable, Optional, Tuple from langchain_core._api import beta from langchain_core.documents import Document from langchain_community.graph_vectorstores.links import get_links if TYPE_CHECKING: import graphviz def _escape_id(id: str) -> str: return id.repla...
import re from typing import TYPE_CHECKING, Dict, Iterable, Optional, Tuple from langchain_core._api import beta from langchain_core.documents import Document from langchain_community.graph_vectorstores.links import get_links if TYPE_CHECKING: import graphviz def _escape_id(id: str) -> str: return id.repla...
import os from typing import Union from uuid import UUID from pydantic import BaseModel, Field from docarray.document.abstract_document import AbstractDocument from docarray.document.base_node import BaseNode from .mixins import ProtoMixin class BaseDocument(BaseModel, ProtoMixin, AbstractDocument, BaseNode): ...
import os from typing import Union from uuid import UUID from pydantic import BaseModel, Field from docarray.document.abstract_document import AbstractDocument from docarray.document.base_node import BaseNode from .mixins import ProtoMixin class BaseDocument(BaseModel, ProtoMixin, AbstractDocument, BaseNode): ...
from enum import Enum from typing import Any, Dict, Iterable import torch.nn.functional as F from torch import Tensor, nn from sentence_transformers.SentenceTransformer import SentenceTransformer class TripletDistanceMetric(Enum): """The metric for the triplet loss""" COSINE = lambda x, y: 1 - F.cosine_sim...
from enum import Enum from typing import Dict, Iterable import torch.nn.functional as F from torch import Tensor, nn from sentence_transformers.SentenceTransformer import SentenceTransformer class TripletDistanceMetric(Enum): """The metric for the triplet loss""" COSINE = lambda x, y: 1 - F.cosine_similari...
"""Load Documents from a set of persistent Steamship Files.""" from typing import List, Optional from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class SteamshipFileReader(BaseReader): """ Reads persistent Steamship Files and converts them to Documents. ...
"""Load Documents from a set of persistent Steamship Files.""" from typing import List, Optional from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class SteamshipFileReader(BaseReader): """ Reads persistent Steamship Files and converts them to Documents. A...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.regularizers import deserialize as deserialize from keras.src.regularizers import get as get from keras.src.regularizers import serialize as serialize from keras.src.regularizers.regu...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.regularizers import deserialize from keras.src.regularizers import get from keras.src.regularizers import serialize from keras.src.regularizers.regularizers import L1 from keras.src.r...
# Copyright (c) OpenMMLab. All rights reserved. from .collect_env import collect_env from .compat_config import compat_cfg from .dist_utils import (DistOptimizerHook, all_reduce_dict, allreduce_grads, reduce_mean, sync_random_seed) from .logger import get_caller_name, get_root_logger, log_img_s...
# Copyright (c) OpenMMLab. All rights reserved. from .collect_env import collect_env from .compat_config import compat_cfg from .dist_utils import (DistOptimizerHook, all_reduce_dict, allreduce_grads, reduce_mean, sync_random_seed) from .logger import get_caller_name, get_root_logger, log_img_s...
from __future__ import annotations try: from typing import Self except ImportError: from typing_extensions import Self import torch import transformers from PIL import Image from sentence_transformers.models.Asym import InputModule class CLIPModel(InputModule): save_in_root: bool = True def __init...
from __future__ import annotations import torch import transformers from PIL import Image from torch import nn class CLIPModel(nn.Module): save_in_root: bool = True def __init__(self, model_name: str = "openai/clip-vit-base-patch32", processor_name=None) -> None: super().__init__() if proce...
from typing import Any, Optional from langchain_core.language_models import BaseLanguageModel from langchain_core.memory import BaseMemory from langchain_core.messages import SystemMessage from langchain_core.prompts.chat import MessagesPlaceholder from langchain_core.tools import BaseTool from langchain.agents.agent...
from typing import Any, Optional from langchain_core.language_models import BaseLanguageModel from langchain_core.memory import BaseMemory from langchain_core.messages import SystemMessage from langchain_core.prompts.chat import MessagesPlaceholder from langchain_core.tools import BaseTool from langchain.agents.agent...
from typing import Annotated, Optional import typer from langchain_cli._version import __version__ from langchain_cli.namespaces import app as app_namespace from langchain_cli.namespaces import integration as integration_namespace from langchain_cli.namespaces import template as template_namespace from langchain_cli....
from typing import Optional import typer from typing_extensions import Annotated from langchain_cli._version import __version__ from langchain_cli.namespaces import app as app_namespace from langchain_cli.namespaces import integration as integration_namespace from langchain_cli.namespaces import template as template_...
""" ========================= Tensor transforms and JIT ========================= .. note:: Try on `collab <https://colab.research.google.com/github/pytorch/vision/blob/gh-pages/main/_generated_ipynb_notebooks/plot_scripted_tensor_transforms.ipynb>`_ or :ref:`go to the end <sphx_glr_download_auto_examples_othe...
""" ========================= Tensor transforms and JIT ========================= .. note:: Try on `collab <https://colab.research.google.com/github/pytorch/vision/blob/gh-pages/main/_generated_ipynb_notebooks/plot_scripted_tensor_transforms.ipynb>`_ or :ref:`go to the end <sphx_glr_download_auto_examples_othe...
# Copyright (c) OpenMMLab. All rights reserved. from .vis_backend import (BaseVisBackend, LocalVisBackend, TensorboardVisBackend, WandbVisBackend) from .visualizer import Visualizer __all__ = [ 'Visualizer', 'BaseVisBackend', 'LocalVisBackend', 'WandbVisBackend', 'TensorboardVisBacken...
# Copyright (c) OpenMMLab. All rights reserved. from .visualizer import Visualizer from .writer import (BaseWriter, ComposedWriter, LocalWriter, TensorboardWriter, WandbWriter) __all__ = [ 'Visualizer', 'BaseWriter', 'LocalWriter', 'WandbWriter', 'TensorboardWriter', 'ComposedWriter' ]
import torch import torchaudio.prototype.functional as F from parameterized import parameterized from torch.autograd import gradcheck, gradgradcheck from torchaudio_unittest.common_utils import nested_params, TestBaseMixin class AutogradTestImpl(TestBaseMixin): @nested_params( [F.convolve, F.fftconvolve],...
import torch import torchaudio.prototype.functional as F from parameterized import parameterized from torch.autograd import gradcheck, gradgradcheck from torchaudio_unittest.common_utils import nested_params, TestBaseMixin class AutogradTestImpl(TestBaseMixin): @nested_params( [F.convolve, F.fftconvolve],...
# Copyright (c) OpenMMLab. All rights reserved. from mmcv.parallel import is_module_wrapper from mmcv.runner.hooks import Hook from mmdet.registry import HOOKS @HOOKS.register_module() class YOLOXModeSwitchHook(Hook): """Switch the mode of YOLOX during training. This hook turns off the mosaic and mixup data...
# Copyright (c) OpenMMLab. All rights reserved. from mmcv.parallel import is_module_wrapper from mmcv.runner.hooks import HOOKS, Hook @HOOKS.register_module() class YOLOXModeSwitchHook(Hook): """Switch the mode of YOLOX during training. This hook turns off the mosaic and mixup data augmentation and switches ...
from keras.src import tree from keras.src.api_export import keras_export from keras.src.backend import KerasTensor from keras.src.layers.layer import Layer @keras_export("keras.layers.Identity") class Identity(Layer): """Identity layer. This layer should be used as a placeholder when no operation is to be ...
from keras.src import tree from keras.src.api_export import keras_export from keras.src.backend import KerasTensor from keras.src.layers.layer import Layer @keras_export("keras.layers.Identity") class Identity(Layer): """Identity layer. This layer should be used as a placeholder when no operation is to be ...
# flake8: noqa import torchaudio from . import utils from .utils import _is_backend_dispatcher_enabled, get_audio_backend, list_audio_backends, set_audio_backend if _is_backend_dispatcher_enabled(): from torchaudio._backend.utils import get_info_func, get_load_func, get_save_func torchaudio.info = get_info_f...
# flake8: noqa import torchaudio from torchaudio._backend.utils import get_info_func, get_load_func, get_save_func from . import utils from .utils import _is_backend_dispatcher_enabled, get_audio_backend, list_audio_backends, set_audio_backend if _is_backend_dispatcher_enabled(): torchaudio.info = get_info_func(...
"""Util that Searches calendar events in Office 365. Free, but setup is required. See link below. https://learn.microsoft.com/en-us/graph/auth/ """ from datetime import datetime as dt from typing import Any, Dict, List, Optional, Type from langchain_core.callbacks import CallbackManagerForToolRun from pydantic impor...
"""Util that Searches calendar events in Office 365. Free, but setup is required. See link below. https://learn.microsoft.com/en-us/graph/auth/ """ from datetime import datetime as dt from typing import Any, Dict, List, Optional, Type from langchain_core.callbacks import CallbackManagerForToolRun from pydantic impor...
""" ============================================== Plot randomly generated multilabel dataset ============================================== This illustrates the :func:`~sklearn.datasets.make_multilabel_classification` dataset generator. Each sample consists of counts of two features (up to 50 in total), which are dif...
""" ============================================== Plot randomly generated multilabel dataset ============================================== This illustrates the :func:`~sklearn.datasets.make_multilabel_classification` dataset generator. Each sample consists of counts of two features (up to 50 in total), which are dif...
#!/usr/bin/env python # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless...
#!/usr/bin/env python # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless...
_base_ = '../_base_/default_runtime.py' # dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' # Example to use different file client # Method 1: simply set the data root and let the file I/O module # automatically infer from prefix (not support LMDB and Memcache yet) # data_root = 's3://openmmlab/d...
_base_ = '../_base_/default_runtime.py' # dataset settings dataset_type = 'CocoDataset' data_root = 'data/coco/' # file_client_args = dict( # backend='petrel', # path_mapping=dict({ # './data/': 's3://openmmlab/datasets/detection/', # 'data/': 's3://openmmlab/datasets/detection/' # })) file...
import os # DO NOT EDIT. Generated by api_gen.sh from keras.api import DTypePolicy from keras.api import FloatDTypePolicy from keras.api import Function from keras.api import Initializer from keras.api import Input from keras.api import InputSpec from keras.api import KerasTensor from keras.api import Layer from keras...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ import os # Import everything from /api/ into keras. from keras.api import * # noqa: F403 from keras.api import __version__ # Import * ignores names start with "_". # Add everything in /api/ to ...
import os.path from pathlib import Path from typing import Callable, Optional, Union import numpy as np import torch from torchvision.datasets.utils import download_url, verify_str_arg from torchvision.datasets.vision import VisionDataset class MovingMNIST(VisionDataset): """`MovingMNIST <http://www.cs.toronto.e...
import os.path from typing import Callable, Optional import numpy as np import torch from torchvision.datasets.utils import download_url, verify_str_arg from torchvision.datasets.vision import VisionDataset class MovingMNIST(VisionDataset): """`MovingMNIST <http://www.cs.toronto.edu/~nitish/unsupervised_video/>`...
_base_ = './vfnet_r50-mdconv-c3-c5_fpn_ms-2x_coco.py' model = dict( backbone=dict( type='ResNet', depth=101, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), norm_eval=True, style='pytorch', ...
_base_ = './vfnet_r50_fpn_mdconv_c3-c5_mstrain_2x_coco.py' model = dict( backbone=dict( type='ResNet', depth=101, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), norm_eval=True, style='pytorch', ...
from __future__ import annotations from typing import Any from langchain_text_splitters.base import TextSplitter class NLTKTextSplitter(TextSplitter): """Splitting text using NLTK package.""" def __init__( self, separator: str = "\n\n", language: str = "english", *, ...
from __future__ import annotations from typing import Any from langchain_text_splitters.base import TextSplitter class NLTKTextSplitter(TextSplitter): """Splitting text using NLTK package.""" def __init__( self, separator: str = "\n\n", language: str = "english", *, ...
from langchain_anthropic.chat_models import ( ChatAnthropic, ChatAnthropicMessages, convert_to_anthropic_tool, ) from langchain_anthropic.llms import Anthropic, AnthropicLLM __all__ = [ "ChatAnthropicMessages", "ChatAnthropic", "convert_to_anthropic_tool", "Anthropic", "AnthropicLLM", ]...
from langchain_anthropic.chat_models import ChatAnthropic, ChatAnthropicMessages from langchain_anthropic.llms import Anthropic, AnthropicLLM __all__ = ["ChatAnthropicMessages", "ChatAnthropic", "Anthropic", "AnthropicLLM"]
_base_ = './gfl_r50_fpn_1x_coco.py' max_epochs = 24 # learning policy param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[16, 22], ...
_base_ = './gfl_r50_fpn_1x_coco.py' max_epochs = 24 # learning policy param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_epochs, by_epoch=True, milestones=[16, 22], ...
import os import shutil import subprocess import sys def _get_run_args(print_args: bool = True): from jina.helper import get_rich_console from jina.parsers import get_main_parser console = get_rich_console() silent_print = {'help', 'hub', 'export', 'auth', 'cloud', 'ping'} parser = get_main_par...
import os import shutil import subprocess import sys def _get_run_args(print_args: bool = True): from jina.helper import get_rich_console from jina.parsers import get_main_parser console = get_rich_console() silent_print = {'help', 'hub', 'export', 'auth', 'cloud'} parser = get_main_parser() ...
import pathlib from typing import Any, Dict, List, Tuple, Union import torch from torchdata.datapipes.iter import CSVParser, IterDataPipe, Mapper from torchvision.prototype.datasets.utils import Dataset, HttpResource, OnlineResource from torchvision.prototype.datasets.utils._internal import hint_sharding, hint_shuffli...
import pathlib from typing import Any, Dict, List, Tuple, Union import torch from torchdata.datapipes.iter import CSVParser, IterDataPipe, Mapper from torchvision.datapoints import Image from torchvision.prototype.datapoints import OneHotLabel from torchvision.prototype.datasets.utils import Dataset, HttpResource, Onl...
from ._multi_channel import MVDR, PSD, RTFMVDR, SoudenMVDR from ._transforms import ( AddNoise, AmplitudeToDB, ComputeDeltas, Convolve, Deemphasis, Fade, FFTConvolve, FrequencyMasking, GriffinLim, InverseMelScale, InverseSpectrogram, LFCC, Loudness, MelScale, ...
from ._multi_channel import MVDR, PSD, RTFMVDR, SoudenMVDR from ._transforms import ( AmplitudeToDB, ComputeDeltas, Fade, FrequencyMasking, GriffinLim, InverseMelScale, InverseSpectrogram, LFCC, Loudness, MelScale, MelSpectrogram, MFCC, MuLawDecoding, MuLawEncodin...
""" Custom hook to customize the behavior of Hatchling. Here, we customize the tag of the generated wheels. """ from typing import Any, Dict from hatchling.builders.hooks.plugin.interface import BuildHookInterface from packaging.tags import platform_tags def get_tag() -> str: """Get appropriate wheel tag accord...
""" Custom hook to customize the behavior of Hatchling. Here, we customize the tag of the generated wheels. """ import sysconfig from typing import Any, Dict from hatchling.builders.hooks.plugin.interface import BuildHookInterface def get_tag() -> str: """Get appropriate wheel tag according to system""" tag...
# Copyright (c) OpenMMLab. All rights reserved. import pytest import torch from mmcv import ConfigDict from mmcv.cnn import build_plugin_layer from mmdet.models.plugins import DropBlock def test_dropblock(): feat = torch.rand(1, 1, 11, 11) drop_prob = 1.0 dropblock = DropBlock(drop_prob, block_size=11, w...
# Copyright (c) OpenMMLab. All rights reserved. import pytest import torch from mmdet.models.plugins import DropBlock def test_dropblock(): feat = torch.rand(1, 1, 11, 11) drop_prob = 1.0 dropblock = DropBlock(drop_prob, block_size=11, warmup_iters=0) out_feat = dropblock(feat) assert (out_feat =...
_base_ = './libra-faster-rcnn_r50_fpn_1x_coco.py' model = dict( backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), style='pyt...
_base_ = './libra_faster_rcnn_r50_fpn_1x_coco.py' model = dict( backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), style='pyt...
from __future__ import annotations import os from copy import deepcopy import numpy as np import pytest from tokenizers import Tokenizer from sentence_transformers import SentenceTransformer from sentence_transformers.models import Pooling, StaticEmbedding, Transformer from sentence_transformers.util import is_datas...
from __future__ import annotations import os from copy import deepcopy import numpy as np import pytest from tokenizers import Tokenizer from sentence_transformers import SentenceTransformer from sentence_transformers.models import Pooling, StaticEmbedding, Transformer from sentence_transformers.util import is_datas...
"""Filter that uses an LLM to drop documents that aren't relevant to the query.""" from collections.abc import Sequence from typing import Any, Callable, Optional from langchain_core.callbacks.manager import Callbacks from langchain_core.documents import Document from langchain_core.language_models import BaseLanguag...
"""Filter that uses an LLM to drop documents that aren't relevant to the query.""" from typing import Any, Callable, Dict, Optional, Sequence from langchain_core.callbacks.manager import Callbacks from langchain_core.documents import Document from langchain_core.language_models import BaseLanguageModel from langchain...
from __future__ import annotations import json import os import torch from safetensors.torch import load_model as load_safetensors_model from safetensors.torch import save_model as save_safetensors_model from torch import Tensor, nn from sentence_transformers.util import fullname, import_from_string class Dense(nn...
from __future__ import annotations import json import os import torch from safetensors.torch import load_model as load_safetensors_model from safetensors.torch import save_model as save_safetensors_model from torch import Tensor, nn from sentence_transformers.util import fullname, import_from_string class Dense(nn...
from abc import ABC from typing import TYPE_CHECKING, Any, Dict, Optional, Type, TypeVar, Union from docarray.typing.proto_register import _register_proto from docarray.typing.url.any_url import AnyUrl if TYPE_CHECKING: import trimesh from pydantic import BaseConfig from pydantic.fields import ModelField ...
from abc import ABC from typing import TYPE_CHECKING, Any, Dict, Optional, Type, TypeVar, Union import numpy as np from docarray.typing.proto_register import _register_proto from docarray.typing.url.any_url import AnyUrl if TYPE_CHECKING: import trimesh from pydantic import BaseConfig from pydantic.field...
import logging import os import sys from torchaudio._internal.module_utils import eval_env, fail_with_message, is_module_available, no_op try: from .fb import _init_ffmpeg except ImportError: from .utils import _init_ffmpeg from .utils import _check_cuda_version, _fail_since_no_ffmpeg, _fail_since_no_sox, _in...
import logging import os import sys from torchaudio._internal.module_utils import eval_env, fail_with_message, is_module_available, no_op try: from .fb import _init_ffmpeg except ImportError: from .utils import _init_ffmpeg from .utils import _check_cuda_version, _fail_since_no_ffmpeg, _init_dll_path, _init_s...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import AINValueOps from langchain_community.tools.ainetwork.value import ValueSchema # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for rais...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import AINValueOps from langchain_community.tools.ainetwork.value import ValueSchema # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for rais...
# Copyright (c) OpenMMLab. All rights reserved. from .activations import SiLU from .bbox_nms import fast_nms, multiclass_nms from .brick_wrappers import AdaptiveAvgPool2d, adaptive_avg_pool2d from .conv_upsample import ConvUpsample from .csp_layer import CSPLayer from .dropblock import DropBlock from .ema import ExpMom...
# Copyright (c) OpenMMLab. All rights reserved. from .activations import SiLU from .bbox_nms import fast_nms, multiclass_nms from .brick_wrappers import AdaptiveAvgPool2d, adaptive_avg_pool2d from .conv_upsample import ConvUpsample from .csp_layer import CSPLayer from .dropblock import DropBlock from .ema import ExpMom...
from dataclasses import dataclass, field from typing import Any, Callable, Dict, List import torch @dataclass class SentenceTransformerDataCollator: """Collator for a SentenceTransformers model. This encodes the text columns to {column}_input_ids and {column}_attention_mask columns. This works with the t...
from dataclasses import dataclass, field from typing import Any, Callable, Dict, List import torch @dataclass class SentenceTransformerDataCollator: """Collator for a SentenceTransformers model. This encodes the text columns to {column}_input_ids and {column}_attention_mask columns. This works with the t...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from pathlib import Path from typing import Dict import numpy as np import pytest from jina import Document, DocumentArray, Executor from ...image_tf_encoder import ImageTFEncoder input_dim = 336 target_output...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from pathlib import Path from typing import Dict import numpy as np from jina import DocumentArray, Document, Executor from ...image_tf_encoder import ImageTFEncoder input_dim = 336 target_output_dim = 1280 ...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.core.utils import ConfigType, OptConfigType, OptMultiConfig from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class RetinaNet(SingleStageDetector): """Implementation of `RetinaNet <https://arxiv.org/...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.core.utils import ConfigType, OptConfigType, OptMultiConfig from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class RetinaNet(SingleStageDetector): """Implementation of `RetinaNet <https://arxiv.org/...
# dataset settings dataset_type = 'MOTChallengeDataset' data_root = 'data/MOT17/' resized_shape = (1088, 1088) # data pipeline train_pipeline = [ dict( type='UniformRefFrameSample', num_ref_imgs=1, frame_range=10, filter_key_img=True), dict( type='TransformBroadcaster', ...
# dataset settings dataset_type = 'MOTChallengeDataset' data_root = 'data/MOT17/' resized_shape = (1088, 1088) # data pipeline train_pipeline = [ dict( type='UniformSample', num_ref_imgs=1, frame_range=10, filter_key_img=True), dict( type='TransformBroadcaster', ...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
_base_ = [ '../_base_/models/faster_rcnn_r50_fpn.py', '../_base_/datasets/cityscapes_detection.py', '../_base_/default_runtime.py' ] model = dict( backbone=dict(init_cfg=None), roi_head=dict( bbox_head=dict( type='Shared2FCBBoxHead', in_channels=256, fc_ou...
_base_ = [ '../_base_/models/faster_rcnn_r50_fpn.py', '../_base_/datasets/cityscapes_detection.py', '../_base_/default_runtime.py' ] model = dict( backbone=dict(init_cfg=None), roi_head=dict( bbox_head=dict( type='Shared2FCBBoxHead', in_channels=256, fc_ou...
"""Test embeddings base module.""" import pytest from langchain.embeddings.base import ( _SUPPORTED_PROVIDERS, _infer_model_and_provider, _parse_model_string, ) def test_parse_model_string() -> None: """Test parsing model strings into provider and model components.""" assert _parse_model_string(...
"""Test embeddings base module.""" import pytest from langchain.embeddings.base import ( _SUPPORTED_PROVIDERS, _infer_model_and_provider, _parse_model_string, ) def test_parse_model_string() -> None: """Test parsing model strings into provider and model components.""" assert _parse_model_string(...
import os from . import InputExample class LabelSentenceReader: """Reads in a file that has at least two columns: a label and a sentence. This reader can for example be used with the BatchHardTripletLoss. Maps labels automatically to integers """ def __init__(self, folder, label_col_idx=0, sente...
from . import InputExample import csv import gzip import os class LabelSentenceReader: """Reads in a file that has at least two columns: a label and a sentence. This reader can for example be used with the BatchHardTripletLoss. Maps labels automatically to integers""" def __init__(self, folder, label_c...
# Licensed to the LF AI & Data foundation under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the "License"); # you may not use this fil...
import numpy as np from docarray import BaseDoc from docarray.array.doc_vec.doc_vec import DocVec from docarray.typing import AnyTensor, NdArray def test_da_init(): class MyDoc(BaseDoc): tensor: AnyTensor name: str docs = [MyDoc(tensor=np.zeros(10), name='hello') for _ in range(4)] da =...
# Copyright (c) OpenMMLab. All rights reserved. import warnings import torch.nn as nn from mmcv.cnn import VGG from mmengine.model import BaseModule from mmdet.registry import MODELS from ..necks import ssd_neck @MODELS.register_module() class SSDVGG(VGG, BaseModule): """VGG Backbone network for single-shot-det...
# Copyright (c) OpenMMLab. All rights reserved. import warnings import torch.nn as nn from mmcv.cnn import VGG from mmcv.runner import BaseModule from mmdet.registry import MODELS from ..necks import ssd_neck @MODELS.register_module() class SSDVGG(VGG, BaseModule): """VGG Backbone network for single-shot-detect...
"""Test EdenAi's object detection Tool . In order to run this test, you need to have an EdenAI api key. You can get it by registering for free at https://app.edenai.run/user/register. A test key can be found at https://app.edenai.run/admin/account/settings by clicking on the 'sandbox' toggle. (calls will be free, and ...
"""Test EdenAi's object detection Tool . In order to run this test, you need to have an EdenAI api key. You can get it by registering for free at https://app.edenai.run/user/register. A test key can be found at https://app.edenai.run/admin/account/settings by clicking on the 'sandbox' toggle. (calls will be free, and ...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
# Copyright (c) OpenMMLab. All rights reserved. from .auto_augment import (AutoAugment, BrightnessTransform, ColorTransform, ContrastTransform, EqualizeTransform, Rotate, Shear, Translate) from .compose import Compose from .formatting import (Collect, DefaultFormatB...
# Copyright (c) OpenMMLab. All rights reserved. from .auto_augment import (AutoAugment, BrightnessTransform, ColorTransform, ContrastTransform, EqualizeTransform, Rotate, Shear, Translate) from .compose import Compose from .formatting import (Collect, DefaultFormatB...
from llama_index.core.instrumentation.events.base import BaseEvent class StreamChatStartEvent(BaseEvent): """ StreamChatStartEvent. Fired at the start of writing to the stream chat-engine queue. """ @classmethod def class_name(cls) -> str: """Class name.""" return "StreamChat...
from llama_index.core.instrumentation.events.base import BaseEvent class StreamChatStartEvent(BaseEvent): """StreamChatStartEvent. Fired at the start of writing to the stream chat-engine queue. """ @classmethod def class_name(cls) -> str: """Class name.""" return "StreamChatStart...
_INITIALIZED = False _LAZILY_IMPORTED = [ "Hypothesis", "CTCDecoder", "ctc_decoder", "lexicon_decoder", "download_pretrained_files", ] def _init_extension(): import torchaudio torchaudio._extension._load_lib("libtorchaudio_decoder") global _INITIALIZED _INITIALIZED = True def _...
import torchaudio try: torchaudio._extension._load_lib("libtorchaudio_decoder") from .ctc_decoder import Hypothesis, CTCDecoder, ctc_decoder, lexicon_decoder, download_pretrained_files except ImportError as err: raise ImportError( "flashlight decoder bindings are required to use this functionality....
"""Init file.""" from llama_index.readers.papers.arxiv.base import ArxivReader from llama_index.readers.papers.pubmed.base import PubmedReader __all__ = ["ArxivReader", "PubmedReader"]
"""Init file.""" from llama_index.readers.papers.arxiv.base import ArxivReader from llama_index.readers.papers.pubmed.base import PubmedReader __all__ = ["ArxivReader", "PubmedReader"]
# Copyright (c) OpenMMLab. All rights reserved. from mmcv.cnn import ConvModule, Linear from mmengine.model import ModuleList from torch import Tensor from mmdet.core.utils import MultiConfig from mmdet.registry import MODELS from .fcn_mask_head import FCNMaskHead @MODELS.register_module() class CoarseMaskHead(FCNMa...
# Copyright (c) OpenMMLab. All rights reserved. from mmcv.cnn import ConvModule, Linear from mmcv.runner import ModuleList, auto_fp16 from mmdet.registry import MODELS from .fcn_mask_head import FCNMaskHead @MODELS.register_module() class CoarseMaskHead(FCNMaskHead): """Coarse mask head used in PointRend. C...
from dataclasses import dataclass, field from typing import Any, Dict, Type import pytest from pydantic import Field from docarray import BaseDoc from docarray.index.abstract import BaseDocIndex from docarray.typing import NdArray pytestmark = pytest.mark.index class SimpleDoc(BaseDoc): tens: NdArray[10] = Fie...
from dataclasses import dataclass, field from typing import Any, Dict, Type import pytest from pydantic import Field from docarray import BaseDoc from docarray.index.abstract import BaseDocIndex from docarray.typing import NdArray pytestmark = pytest.mark.index class SimpleDoc(BaseDoc): tens: NdArray[10] = Fie...
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
""" Example of using callbacks with Dask ==================================== """ from typing import Any import numpy as np from dask.distributed import Client, LocalCluster from dask_ml.datasets import make_regression from dask_ml.model_selection import train_test_split import xgboost as xgb import xgboost.dask as ...
""" Example of using callbacks with Dask ==================================== """ import numpy as np from dask.distributed import Client, LocalCluster from dask_ml.datasets import make_regression from dask_ml.model_selection import train_test_split import xgboost as xgb import xgboost.dask as dxgb from xgboost.dask i...
import base64 from os.path import exists from typing import Any, Dict, List, Optional from urllib.parse import urlparse import requests from langchain_core.embeddings import Embeddings from langchain_core.utils import convert_to_secret_str, get_from_dict_or_env from pydantic import BaseModel, ConfigDict, SecretStr, mo...
import base64 from os.path import exists from typing import Any, Dict, List, Optional from urllib.parse import urlparse import requests from langchain_core.embeddings import Embeddings from langchain_core.utils import convert_to_secret_str, get_from_dict_or_env from pydantic import BaseModel, ConfigDict, SecretStr, mo...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase import torch from mmengine.structures import InstanceData from mmdet.models.task_modules.assigners import TaskAlignedAssigner class TestTaskAlignedAssigner(TestCase): def test_task_aligned_assigner(self): with self.assertRai...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase import torch from mmengine.data import InstanceData from mmdet.models.task_modules.assigners import TaskAlignedAssigner class TestTaskAlignedAssigner(TestCase): def test_task_aligned_assigner(self): with self.assertRaises(As...