input
stringlengths
33
5k
output
stringlengths
32
5k
from pydantic import parse_obj_as from docarray import BaseDocument from docarray.documents import TextDoc def test_simple_init(): t = TextDoc(text='hello') assert t.text == 'hello' def test_str_init(): t = parse_obj_as(TextDoc, 'hello') assert t.text == 'hello' def test_doc(): class MyDoc(Ba...
from pydantic import parse_obj_as from docarray import BaseDocument from docarray.documents import Text def test_simple_init(): t = Text(text='hello') assert t.text == 'hello' def test_str_init(): t = parse_obj_as(Text, 'hello') assert t.text == 'hello' def test_doc(): class MyDoc(BaseDocumen...
import torch import torchaudio.prototype.functional as F from torchaudio_unittest.common_utils import nested_params, TorchaudioTestCase class BatchConsistencyTest(TorchaudioTestCase): @nested_params( [F.convolve, F.fftconvolve], ["full", "valid", "same"], ) def test_convolve(self, fn, mode...
import torch import torchaudio.prototype.functional as F from torchaudio_unittest.common_utils import nested_params, TorchaudioTestCase class BatchConsistencyTest(TorchaudioTestCase): @nested_params( [F.convolve, F.fftconvolve], ) def test_convolve(self, fn): leading_dims = (2, 3) ...
from __future__ import annotations import os import pytest from sentence_transformers import SentenceTransformer from sentence_transformers.models import Pooling, Transformer from sentence_transformers.util import is_datasets_available from tests.utils import SafeTemporaryDirectory if is_datasets_available(): f...
from __future__ import annotations import os import pytest from sentence_transformers import CrossEncoder, SentenceTransformer from sentence_transformers.models import Pooling, Transformer from sentence_transformers.util import is_datasets_available from tests.utils import SafeTemporaryDirectory if is_datasets_avai...
_base_ = './retinanet_r50_caffe_fpn_mstrain_1x_coco.py' # training schedule for 2x train_cfg = dict(max_epochs=24) # learning rate policy param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=24, ...
_base_ = './retinanet_r50_caffe_fpn_mstrain_1x_coco.py' # learning policy lr_config = dict(step=[16, 23]) runner = dict(type='EpochBasedRunner', max_epochs=24)
import torch __all__ = ["DeepSpeech"] class FullyConnected(torch.nn.Module): """ Args: n_feature: Number of input features n_hidden: Internal hidden unit size. """ def __init__(self, n_feature: int, n_hidden: int, dropout: float, relu_max_clip: int = 20) -> None: super(FullyC...
import torch __all__ = ["DeepSpeech"] class FullyConnected(torch.nn.Module): """ Args: n_feature: Number of input features n_hidden: Internal hidden unit size. """ def __init__(self, n_feature: int, n_hidden: int, dropout: float, relu_max_clip: int = 20) -> None: super(FullyC...
""" This example starts multiple processes (1 per GPU), which encode sentences in parallel. This gives a near linear speed-up when encoding large text collections. It also demonstrates how to stream data which is helpful in case you don't want to wait for an extremely large dataset to download, or if you want to limit ...
""" This example starts multiple processes (1 per GPU), which encode sentences in parallel. This gives a near linear speed-up when encoding large text collections. It also demonstrates how to stream data which is helpful in case you don't want to wait for an extremely large dataset to download, or if you want to limit ...
"""Utils for jupyter notebook.""" import os from io import BytesIO from typing import Any, Dict, List, Tuple import matplotlib.pyplot as plt import requests from IPython.display import Markdown, display from llama_index.core.base.response.schema import Response from llama_index.core.img_utils import b64_2_img from ll...
"""Utils for jupyter notebook.""" import os from io import BytesIO from typing import Any, Dict, List, Tuple import matplotlib.pyplot as plt import requests from IPython.display import Markdown, display from llama_index.core.base.response.schema import Response from llama_index.core.img_utils import b64_2_img from lla...
# Copyright (c) OpenMMLab. All rights reserved. from .local_visualizer import DetLocalVisualizer from .palette import get_palette, jitter_color, palette_val __all__ = ['palette_val', 'get_palette', 'DetLocalVisualizer', 'jitter_color']
# Copyright (c) OpenMMLab. All rights reserved. from .local_visualizer import DetLocalVisualizer from .palette import get_palette, palette_val __all__ = ['palette_val', 'get_palette', 'DetLocalVisualizer']
import numpy as np import pytest import xgboost as xgb from xgboost import testing as tm from xgboost.testing.updater import get_basescore rng = np.random.RandomState(1994) class TestEarlyStopping: @pytest.mark.skipif(**tm.no_sklearn()) def test_early_stopping_nonparallel(self): from sklearn.dataset...
import numpy as np import pytest import xgboost as xgb from xgboost import testing as tm from xgboost.testing.updater import get_basescore rng = np.random.RandomState(1994) class TestEarlyStopping: @pytest.mark.skipif(**tm.no_sklearn()) def test_early_stopping_nonparallel(self): from sklearn.dataset...
# Copyright (c) OpenMMLab. All rights reserved. from mmengine.utils.parrots_wrapper import TORCH_VERSION from mmengine.utils.version_utils import digit_version from .averaged_model import (BaseAveragedModel, ExponentialMovingAverage, MomentumAnnealingEMA, StochasticWeightAverage) from .base...
# Copyright (c) OpenMMLab. All rights reserved. from mmengine.utils.parrots_wrapper import TORCH_VERSION from mmengine.utils.version_utils import digit_version from .averaged_model import (ExponentialMovingAverage, MomentumAnnealingEMA, StochasticWeightAverage) from .base_model import BaseD...
_base_ = './fcos_hrnetv2p_w18_gn-head_4x4_1x_coco.py' # learning policy max_epochs = 24 train_cfg = dict(max_epochs=max_epochs) param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_epochs, ...
_base_ = './fcos_hrnetv2p_w18_gn-head_4x4_1x_coco.py' # learning policy lr_config = dict(step=[16, 22]) runner = dict(type='EpochBasedRunner', max_epochs=24)
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import JsonGetValueTool, JsonListKeysTool from langchain_community.tools.json.tool import JsonSpec # Create a way to dynamically look up deprecated imports. # Used to consolidate ...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import JsonGetValueTool, JsonListKeysTool from langchain_community.tools.json.tool import JsonSpec # Create a way to dynamically look up deprecated imports. # Used to consolidate ...
from __future__ import annotations try: from typing import Self except ImportError: from typing_extensions import Self from torch import Tensor, nn from sentence_transformers.models.Module import Module class LayerNorm(Module): config_keys: list[str] = ["dimension"] def __init__(self, dimension: i...
from __future__ import annotations import json import os import torch from safetensors.torch import load_model as load_safetensors_model from safetensors.torch import save_model as save_safetensors_model from torch import Tensor, nn class LayerNorm(nn.Module): def __init__(self, dimension: int): super()...
__version__ = '0.12.9' import os from .document import Document from .array import DocumentArray from .dataclasses import dataclass, field if 'DA_NO_RICH_HANDLER' not in os.environ: from rich.traceback import install install()
__version__ = '0.12.8' import os from .document import Document from .array import DocumentArray from .dataclasses import dataclass, field if 'DA_NO_RICH_HANDLER' not in os.environ: from rich.traceback import install install()
from typing import Optional from docarray import Document, DocumentArray from pydantic import BaseModel from uvicorn import Config, Server from jina import Gateway class DummyResponseModel(BaseModel): arg1: Optional[str] arg2: Optional[str] arg3: Optional[str] class ProcessedResponseModel(BaseModel): ...
from typing import Optional from docarray import Document, DocumentArray from pydantic import BaseModel from uvicorn import Config, Server from jina import Gateway, __default_host__ from jina.clients.request import request_generator class DummyResponseModel(BaseModel): arg1: Optional[str] arg2: Optional[str...
from abc import ABC from typing import TYPE_CHECKING, Any, Dict, Optional, Type, TypeVar, Union from docarray.typing.proto_register import _register_proto from docarray.typing.url.any_url import AnyUrl from docarray.utils._internal.misc import import_library if TYPE_CHECKING: import trimesh from pydantic impo...
from abc import ABC from typing import TYPE_CHECKING, Any, Dict, Optional, Type, TypeVar, Union from docarray.typing.proto_register import _register_proto from docarray.typing.url.any_url import AnyUrl if TYPE_CHECKING: import trimesh from pydantic import BaseConfig from pydantic.fields import ModelField ...
# Copyright (c) OpenMMLab. All rights reserved. from .batch_sampler import (AspectRatioBatchSampler, MultiDataAspectRatioBatchSampler, TrackAspectRatioBatchSampler) from .class_aware_sampler import ClassAwareSampler from .multi_data_sampler import MultiDataSampler...
# Copyright (c) OpenMMLab. All rights reserved. from .batch_sampler import (AspectRatioBatchSampler, TrackAspectRatioBatchSampler) from .class_aware_sampler import ClassAwareSampler from .multi_source_sampler import GroupMultiSourceSampler, MultiSourceSampler from .track_img_sampler import T...
import asyncio from datetime import datetime, timedelta, timezone from typing import Any import feedparser import pydantic from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField class RSSEntry(pydantic.BaseModel): title: str link: str des...
import time from datetime import datetime, timedelta, timezone from typing import Any import feedparser import pydantic from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField class RSSEntry(pydantic.BaseModel): title: str link: str descri...
"""Utilities for image processing.""" from typing import Any def __getattr__(name: str) -> Any: if name in {"encode_image", "image_to_data_url"}: msg = ( f"'{name}' has been removed for security reasons.\n\n" f"Usage of this utility in environments with user-input paths is a " ...
"""Utilities for image processing.""" from typing import Any def __getattr__(name: str) -> Any: if name in ("encode_image", "image_to_data_url"): msg = ( f"'{name}' has been removed for security reasons.\n\n" f"Usage of this utility in environments with user-input paths is a " ...
"""Run smoke tests""" import os from pathlib import Path import torch import torchvision from torchvision.io import read_image from torchvision.models import resnet50, ResNet50_Weights SCRIPT_DIR = Path(__file__).parent def smoke_test_torchvision() -> None: print( "Is torchvision useable?", all...
"""Run smoke tests""" import os import torchvision from torchvision.io import read_image image_path = os.path.join( os.path.dirname(os.path.abspath(__file__)), "assets", "encode_jpeg", "grace_hopper_517x606.jpg" ) print("torchvision version is ", torchvision.__version__) img = read_image(image_path)
""" This examples trains a CrossEncoder for the Quora Duplicate Questions Detection task. A CrossEncoder takes a sentence pair as input and outputs a label. Here, it output a continuous labels 0...1 to indicate the similarity between the input pair. It does NOT produce a sentence embedding and does NOT work for indivi...
""" This examples trains a CrossEncoder for the Quora Duplicate Questions Detection task. A CrossEncoder takes a sentence pair as input and outputs a label. Here, it output a continuous labels 0...1 to indicate the similarity between the input pair. It does NOT produce a sentence embedding and does NOT work for indivi...
""" A demo for multi-output regression ================================== The demo is adopted from scikit-learn: https://scikit-learn.org/stable/auto_examples/ensemble/plot_random_forest_regression_multioutput.html#sphx-glr-auto-examples-ensemble-plot-random-forest-regression-multioutput-py See :doc:`/tutorials/mult...
""" A demo for multi-output regression ================================== The demo is adopted from scikit-learn: https://scikit-learn.org/stable/auto_examples/ensemble/plot_random_forest_regression_multioutput.html#sphx-glr-auto-examples-ensemble-plot-random-forest-regression-multioutput-py See :doc:`/tutorials/mult...
"""Shopify tool spec.""" from llama_index.core.tools.tool_spec.base import BaseToolSpec class ShopifyToolSpec(BaseToolSpec): """Shopify tool spec.""" spec_functions = ["run_graphql_query"] def __init__(self, shop_url: str, api_version: str, admin_api_key: str): # Currently only supports Admin A...
"""Shopify tool spec.""" from llama_index.core.tools.tool_spec.base import BaseToolSpec class ShopifyToolSpec(BaseToolSpec): """Shopify tool spec.""" spec_functions = ["run_graphql_query"] def __init__(self, shop_url: str, api_version: str, admin_api_key: str): # Currently only supports Admin ...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.ops.image import affine_transform from keras.src.ops.image import crop_images from keras.src.ops.image import extract_patches from keras.src.ops.image import gaussian_blur from keras....
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.ops.image import affine_transform from keras.src.ops.image import crop_images from keras.src.ops.image import extract_patches from keras.src.ops.image import hsv_to_rgb from keras.src...
from datetime import datetime import pytest from jina import Document, DocumentArray, Flow @pytest.mark.parametrize('protocol', ['grpc', 'http', 'websocket']) def test_invalid_input_raise(protocol): f = Flow(protocol=protocol).add() with pytest.raises(BaseException): with f: da = Documen...
from datetime import datetime import pytest from jina import Document, DocumentArray, Flow class MyOwnException(Exception): pass @pytest.mark.parametrize('protocol', ['grpc', 'http', 'websocket']) def test_invalid_input_raise(protocol): f = Flow(protocol=protocol).add() try: with f: ...
import logging from datasets import load_dataset from sentence_transformers import SparseEncoder from sentence_transformers.sparse_encoder.evaluation import SparseTripletEvaluator logging.basicConfig(format="%(message)s", level=logging.INFO) # Load a model model = SparseEncoder("naver/splade-cocondenser-ensembledis...
import logging from datasets import load_dataset from sentence_transformers import SparseEncoder from sentence_transformers.sparse_encoder.evaluation import SparseTripletEvaluator logging.basicConfig(format="%(message)s", level=logging.INFO) # Load a model model = SparseEncoder("naver/splade-cocondenser-ensembledis...
"""Tests for the minimum dependencies in README.rst and pyproject.toml""" import os import re from collections import defaultdict from pathlib import Path import pytest import sklearn from sklearn._min_dependencies import dependent_packages from sklearn.utils.fixes import parse_version min_depencies_tag_to_packages...
"""Tests for the minimum dependencies in README.rst and pyproject.toml""" import os import re from collections import defaultdict from pathlib import Path import pytest import sklearn from sklearn._min_dependencies import dependent_packages from sklearn.utils.fixes import parse_version min_depencies_tag_to_packages...
"""Prompts for scoring the outputs of a models for a given question. This prompt is used to score the responses and evaluate how it follows the instructions and answers the question. The prompt is based on the paper from Zheng, et. al. https://arxiv.org/abs/2306.05685 """ from langchain_core.prompts.chat import ChatP...
"""Prompts for scoring the outputs of a models for a given question. This prompt is used to score the responses and evaluate how it follows the instructions and answers the question. The prompt is based on the paper from Zheng, et. al. https://arxiv.org/abs/2306.05685 """ # flake8: noqa from langchain_core.prompts.ch...
""" Audio Datasets ============== **Author**: `Moto Hira <moto@meta.com>`__ ``torchaudio`` provides easy access to common, publicly accessible datasets. Please refer to the official documentation for the list of available datasets. """ import torch import torchaudio print(torch.__version__) print(torchaudio.__versi...
# -*- coding: utf-8 -*- """ Audio Datasets ============== **Author**: `Moto Hira <moto@meta.com>`__ ``torchaudio`` provides easy access to common, publicly accessible datasets. Please refer to the official documentation for the list of available datasets. """ # When running this tutorial in Google Colab, install the...
"""Test PandasDataframeParser""" from typing import Any import pandas as pd from langchain_core.exceptions import OutputParserException from langchain.output_parsers.pandas_dataframe import PandasDataFrameOutputParser df = pd.DataFrame( { "chicken": [1, 2, 3, 4], "veggies": [5, 4, 3, 2], ...
"""Test PandasDataframeParser""" from typing import Any import pandas as pd from langchain_core.exceptions import OutputParserException from langchain.output_parsers.pandas_dataframe import PandasDataFrameOutputParser df = pd.DataFrame( { "chicken": [1, 2, 3, 4], "veggies": [5, 4, 3, 2], ...
"""Test ChatDeepSeek chat model.""" from typing import Optional import pytest from langchain_core.language_models import BaseChatModel from langchain_core.messages import AIMessageChunk, BaseMessageChunk from langchain_core.tools import BaseTool from langchain_tests.integration_tests import ChatModelIntegrationTests ...
"""Test ChatDeepSeek chat model.""" from typing import Optional, Type import pytest from langchain_core.language_models import BaseChatModel from langchain_core.messages import AIMessageChunk, BaseMessageChunk from langchain_core.tools import BaseTool from langchain_tests.integration_tests import ChatModelIntegration...
import numpy as np import pytest from keras.src import testing from keras.src.layers.activations import leaky_relu class LeakyReLUTest(testing.TestCase): @pytest.mark.requires_trainable_backend def test_leaky_relu(self): self.run_layer_test( leaky_relu.LeakyReLU, init_kwargs={...
import numpy as np import pytest from keras.src import testing from keras.src.layers.activations import leaky_relu class LeakyReLUTest(testing.TestCase): @pytest.mark.requires_trainable_backend def test_leaky_relu(self): self.run_layer_test( leaky_relu.LeakyReLU, init_kwargs={...
import numpy as np from mmdet.core.evaluation.mean_ap import (eval_map, tpfp_default, tpfp_imagenet, tpfp_openimages) det_bboxes = np.array([ [0, 0, 10, 10], [10, 10, 20, 20], [32, 32, 38, 42], ]) gt_bboxes = np.array([[0, 0, 10, 20], [0, 10, 10, 19], [10, 10, 20...
import numpy as np from mmdet.core.evaluation.mean_ap import eval_map, tpfp_default, tpfp_imagenet det_bboxes = np.array([ [0, 0, 10, 10], [10, 10, 20, 20], [32, 32, 38, 42], ]) gt_bboxes = np.array([[0, 0, 10, 20], [0, 10, 10, 19], [10, 10, 20, 20]]) gt_ignore = np.array([[5, 5, 10, 20], [6, 10, 10, 19]]...
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from typing import Dict, List, Optional import numpy as np from annoy import AnnoyIndex from jina import Document, DocumentArray, Executor, requests from jina_commons import get_logger from jina_commons.indexers.dump...
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from typing import Optional, List, Union, Dict import numpy as np from annoy import AnnoyIndex from jina import Executor, requests, DocumentArray, Document from jina_commons import get_logger from jina_commons.index...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import logging import os import os.path as osp from mmengine.config import Config, DictAction from mmengine.logging import print_log from mmengine.registry import RUNNERS from mmengine.runner import Runner from mmdet.utils import setup_cache_size_limit_o...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import logging import os import os.path as osp from mmengine.config import Config, DictAction from mmengine.logging import print_log from mmengine.registry import RUNNERS from mmengine.runner import Runner def parse_args(): parser = argparse.Argumen...
from langchain_core.callbacks.base import ( AsyncCallbackHandler, BaseCallbackHandler, BaseCallbackManager, CallbackManagerMixin, ChainManagerMixin, LLMManagerMixin, RetrieverManagerMixin, RunManagerMixin, ToolManagerMixin, ) __all__ = [ "AsyncCallbackHandler", "BaseCallback...
from langchain_core.callbacks.base import ( AsyncCallbackHandler, BaseCallbackHandler, BaseCallbackManager, CallbackManagerMixin, ChainManagerMixin, LLMManagerMixin, RetrieverManagerMixin, RunManagerMixin, ToolManagerMixin, ) __all__ = [ "RetrieverManagerMixin", "LLMManagerM...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from pathlib import Path from typing import List import pytest from jina import Document, DocumentArray, Executor from spacy_text_encoder import SpacyTextEncoder _EMBEDDING_DIM = 96 @pytest.fixture(scope='ses...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from pathlib import Path from typing import List import pytest from jina import Document, DocumentArray, Executor from ...spacy_text_encoder import SpacyTextEncoder _EMBEDDING_DIM = 96 @pytest.fixture(scope=...
from keras.src.api_export import keras_export # Unique source of truth for the version number. __version__ = "3.3.2" @keras_export("keras.version") def version(): return __version__
from keras.src.api_export import keras_export # Unique source of truth for the version number. __version__ = "3.3.1" @keras_export("keras.version") def version(): return __version__
""" This scripts runs the evaluation (dev & test) for the AskUbuntu dataset Usage: python eval_askubuntu.py [sbert_model_name_or_path] """ import gzip import logging import os import sys from sentence_transformers import LoggingHandler, SentenceTransformer, evaluation, util #### Just some code to print debug inform...
""" This scripts runs the evaluation (dev & test) for the AskUbuntu dataset Usage: python eval_askubuntu.py [sbert_model_name_or_path] """ from sentence_transformers import SentenceTransformer, LoggingHandler from sentence_transformers import util, evaluation import logging import os import gzip import sys #### Just...
import gc import unittest from transformers import AutoModelForCausalLM, AutoTokenizer, CompressedTensorsConfig from transformers.testing_utils import backend_empty_cache, require_compressed_tensors, require_torch, torch_device from transformers.utils import is_torch_available if is_torch_available(): import tor...
import gc import unittest from transformers import AutoModelForCausalLM, AutoTokenizer, CompressedTensorsConfig from transformers.testing_utils import require_compressed_tensors, require_torch from transformers.utils import is_torch_available if is_torch_available(): import torch @require_compressed_tensors @r...
""" Copyright (c) 2013, Triad National Security, LLC All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and ...
""" Copyright (c) 2013, Triad National Security, LLC All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and ...
from __future__ import annotations from typing import List from langchain_core.tools import BaseTool from langchain_core.tools.base import BaseToolkit from langchain_community.tools.azure_ai_services import ( AzureAiServicesDocumentIntelligenceTool, AzureAiServicesImageAnalysisTool, AzureAiServicesSpeech...
from __future__ import annotations from typing import List from langchain_core.tools import BaseTool from langchain_core.tools.base import BaseToolkit from langchain_community.tools.azure_ai_services import ( AzureAiServicesDocumentIntelligenceTool, AzureAiServicesImageAnalysisTool, AzureAiServicesSpeech...
# coding: utf-8 """Find the path to xgboost dynamic library files.""" import os import platform import sys from typing import List class XGBoostLibraryNotFound(Exception): """Error thrown by when xgboost is not found""" def is_sphinx_build() -> bool: """`XGBOOST_BUILD_DOC` is used by the sphinx conf.py to ...
# coding: utf-8 """Find the path to xgboost dynamic library files.""" import os import platform import sys from typing import List class XGBoostLibraryNotFound(Exception): """Error thrown by when xgboost is not found""" def find_lib_path() -> List[str]: """Find the path to xgboost dynamic library files. ...
import os import numpy as np import pytest import torch from pydantic import parse_obj_as from docarray import BaseDoc from docarray.typing.tensor.audio.audio_ndarray import AudioNdArray from docarray.typing.tensor.audio.audio_torch_tensor import AudioTorchTensor from docarray.utils._internal.misc import is_tf_availa...
import os import numpy as np import pytest import torch from pydantic import parse_obj_as from docarray import BaseDoc from docarray.typing.tensor.audio.audio_ndarray import AudioNdArray from docarray.typing.tensor.audio.audio_torch_tensor import AudioTorchTensor from docarray.utils.misc import is_tf_available tf_av...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Optional import torch.nn as nn from mmengine.hooks import Hook from mmengine.model import is_model_wrapper from mmengine.runner import Runner from mmdet.registry import HOOKS @HOOKS.register_module() class MeanTeacherHook(Hook): """Mean Teacher ...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Optional, Sequence import torch.nn as nn from mmengine.hooks import Hook from mmengine.model import is_model_wrapper from mmengine.runner import Runner from mmdet.registry import HOOKS @HOOKS.register_module() class MeanTeacherHook(Hook): """Mea...
# Copyright (c) OpenMMLab. All rights reserved. from math import ceil from unittest import TestCase import torch from mmengine import Config from mmengine.structures import InstanceData from mmdet import * # noqa from mmdet.models.dense_heads import PISARetinaHead class TestPISARetinaHead(TestCase): def test_...
# Copyright (c) OpenMMLab. All rights reserved. from math import ceil from unittest import TestCase import torch from mmengine import Config from mmengine.data import InstanceData from mmdet import * # noqa from mmdet.models.dense_heads import PISARetinaHead class TestPISARetinaHead(TestCase): def test_pisa_r...
import torch import torchaudio.prototype.functional as F from parameterized import parameterized from torchaudio_unittest.common_utils import TestBaseMixin, torch_script class TorchScriptConsistencyTestImpl(TestBaseMixin): def _assert_consistency(self, func, inputs, shape_only=False): inputs_ = [] ...
import torch import torchaudio.prototype.functional as F from parameterized import parameterized from torchaudio_unittest.common_utils import TestBaseMixin, torch_script class TorchScriptConsistencyTestImpl(TestBaseMixin): def _assert_consistency(self, func, inputs, shape_only=False): inputs_ = [] ...
_base_ = './cascade_mask_rcnn_convnext-t_p4_w7_fpn_giou_4conv1f_fp16_ms-crop_3x_coco.py' # noqa # please install mmcls>=0.22.0 # import mmcls.models to trigger register_module in mmcls custom_imports = dict(imports=['mmcls.models'], allow_failed_imports=False) checkpoint_file = 'https://download.openmmlab.com/mmclass...
_base_ = './cascade_mask_rcnn_convnext-t_p4_w7_fpn_giou_4conv1f_fp16_ms-crop_3x_coco.py' # noqa # please install mmcls>=0.22.0 # import mmcls.models to trigger register_module in mmcls custom_imports = dict(imports=['mmcls.models'], allow_failed_imports=False) checkpoint_file = 'https://download.openmmlab.com/mmclass...
from __future__ import annotations import logging from typing import TYPE_CHECKING, Any, Callable from sentence_transformers.evaluation import InformationRetrievalEvaluator if TYPE_CHECKING: import numpy as np from torch import Tensor from sentence_transformers.similarity_functions import SimilarityFunc...
from __future__ import annotations import logging from typing import TYPE_CHECKING, Any, Callable from sentence_transformers.evaluation import InformationRetrievalEvaluator if TYPE_CHECKING: import numpy as np from torch import Tensor from sentence_transformers.similarity_functions import SimilarityFunc...
import json import os import subprocess import pytest from jina.checker import NetworkChecker from jina.jaml import JAML from jina.orchestrate.pods.factory import PodFactory from jina.parsers import set_deployment_parser from jina.parsers.ping import set_ping_parser from jina_cli.autocomplete import ac_table from jin...
import json import os import subprocess import pytest from jina.checker import NetworkChecker from jina.jaml import JAML from jina.orchestrate.pods.factory import PodFactory from jina.parsers import set_deployment_parser from jina.parsers.ping import set_ping_parser from jina_cli.autocomplete import ac_table from jin...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] # model settings model = dict( type='NASFCOS', data_preprocessor=dict( type='DetDataPreprocessor', mean=[103.530, 116.280, 123.675], std=[1.0, 1.0, 1.0], ...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] # model settings preprocess_cfg = dict( mean=[103.530, 116.280, 123.675], std=[1.0, 1.0, 1.0], to_rgb=False, pad_size_divisor=32) model = dict( type='NASFCOS', prepr...
_base_ = './mask-rcnn_r50_fpn_instaboost-4x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
_base_ = './mask_rcnn_r50_fpn_instaboost_4x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
# Copyright (c) OpenMMLab. All rights reserved. from .atss import ATSS from .autoassign import AutoAssign from .base import BaseDetector from .base_detr import DetectionTransformer from .boxinst import BoxInst from .cascade_rcnn import CascadeRCNN from .centernet import CenterNet from .condinst import CondInst from .co...
# Copyright (c) OpenMMLab. All rights reserved. from .atss import ATSS from .autoassign import AutoAssign from .base import BaseDetector from .base_detr import DetectionTransformer from .boxinst import BoxInst from .cascade_rcnn import CascadeRCNN from .centernet import CenterNet from .condinst import CondInst from .co...
_base_ = '../mask_rcnn/mask-rcnn_r50_fpn_1x_coco.py' norm_cfg = dict(type='GN', num_groups=32, requires_grad=True) model = dict( data_preprocessor=dict( mean=[103.530, 116.280, 123.675], std=[1.0, 1.0, 1.0], bgr_to_rgb=False), backbone=dict( norm_cfg=norm_cfg, init_cfg=di...
_base_ = '../mask_rcnn/mask_rcnn_r50_fpn_1x_coco.py' norm_cfg = dict(type='GN', num_groups=32, requires_grad=True) model = dict( data_preprocessor=dict( mean=[103.530, 116.280, 123.675], std=[1.0, 1.0, 1.0], bgr_to_rgb=False), backbone=dict( norm_cfg=norm_cfg, init_cfg=di...
"""Default query for EmptyIndex.""" from typing import Any, List, Optional from llama_index.core.base.base_retriever import BaseRetriever from llama_index.core.callbacks.base import CallbackManager from llama_index.core.indices.empty.base import EmptyIndex from llama_index.core.prompts import BasePromptTemplate from ...
"""Default query for EmptyIndex.""" from typing import Any, List, Optional from llama_index.core.base.base_retriever import BaseRetriever from llama_index.core.callbacks.base import CallbackManager from llama_index.core.indices.empty.base import EmptyIndex from llama_index.core.prompts import BasePromptTemplate from l...
from __future__ import annotations import csv import logging import os import numpy as np from sklearn.metrics import average_precision_score from sentence_transformers import InputExample from sentence_transformers.evaluation import BinaryClassificationEvaluator logger = logging.getLogger(__name__) class CEBinar...
from __future__ import annotations import csv import logging import os import numpy as np from sklearn.metrics import average_precision_score from sentence_transformers import InputExample from sentence_transformers.evaluation import BinaryClassificationEvaluator logger = logging.getLogger(__name__) class CEBinar...
import copy from dataclasses import dataclass from pathlib import Path from typing import Dict, Optional, Union @dataclass class DownloadConfig: """Configuration for our cached path manager. Attributes: cache_dir (`str` or `Path`, *optional*): Specify a cache directory to save the file to...
import copy from dataclasses import dataclass from pathlib import Path from typing import Dict, Optional, Union @dataclass class DownloadConfig: """Configuration for our cached path manager. Attributes: cache_dir (`str` or `Path`, *optional*): Specify a cache directory to save the file to...
from collections import defaultdict import torch import transforms as reference_transforms def get_modules(use_v2): # We need a protected import to avoid the V2 warning in case just V1 is used if use_v2: import torchvision.datapoints import torchvision.transforms.v2 return torchvisio...
from collections import defaultdict import torch import transforms as reference_transforms def get_modules(use_v2): # We need a protected import to avoid the V2 warning in case just V1 is used if use_v2: import torchvision.datapoints import torchvision.transforms.v2 return torchvisio...
from enum import Enum # --8<-- [start:ProviderName] class ProviderName(str, Enum): ANTHROPIC = "anthropic" COMPASS = "compass" DISCORD = "discord" D_ID = "d_id" E2B = "e2b" EXA = "exa" FAL = "fal" GITHUB = "github" GOOGLE = "google" GOOGLE_MAPS = "google_maps" GROQ = "groq"...
from enum import Enum # --8<-- [start:ProviderName] class ProviderName(str, Enum): ANTHROPIC = "anthropic" DISCORD = "discord" D_ID = "d_id" E2B = "e2b" EXA = "exa" FAL = "fal" GITHUB = "github" GOOGLE = "google" GOOGLE_MAPS = "google_maps" GROQ = "groq" HUBSPOT = "hubspot"...
def __getattr__(name: str = "") -> None: """Raise an error on import since is deprecated.""" msg = ( "This module has been moved to langchain-experimental. " "For more details: https://github.com/langchain-ai/langchain/discussions/11352." "To access this code, install it with `pip instal...
def __getattr__(name: str = "") -> None: """Raise an error on import since is deprecated.""" raise AttributeError( "This module has been moved to langchain-experimental. " "For more details: https://github.com/langchain-ai/langchain/discussions/11352." "To access this code, install it wi...
# Copyright (c) OpenMMLab. All rights reserved. from .builder import OPTIMIZER_CONSTRUCTORS, OPTIMIZERS, build_optimizer from .default_constructor import DefaultOptimizerConstructor __all__ = [ 'OPTIMIZER_CONSTRUCTORS', 'OPTIMIZERS', 'DefaultOptimizerConstructor', 'build_optimizer' ]
# Copyright (c) OpenMMLab. All rights reserved. from .builder import (OPTIMIZER_CONSTRUCTORS, OPTIMIZERS, build_optimizer, build_optimizer_constructor) from .default_constructor import DefaultOptimizerConstructor __all__ = [ 'OPTIMIZER_CONSTRUCTORS', 'OPTIMIZERS', 'DefaultOptimizerConstructor...
# Copyright (c) OpenMMLab. All rights reserved. import unittest import torch import torch.nn as nn from mmengine.runner import autocast from mmengine.utils import TORCH_VERSION, digit_version class TestAmp(unittest.TestCase): def test_autocast(self): if not torch.cuda.is_available(): if dig...
# Copyright (c) OpenMMLab. All rights reserved. import unittest import torch import torch.nn as nn from mmengine.runner import autocast from mmengine.utils import TORCH_VERSION, digit_version class TestAmp(unittest.TestCase): def test_autocast(self): if not torch.cuda.is_available(): if dig...
"""CouchDB client.""" import json import logging from typing import Dict, List, Optional import couchdb3 from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class SimpleCouchDBReader(BaseReader): """ Simple CouchDB reader. Concatenates each CouchDB doc into...
"""CouchDB client.""" import json import logging from typing import Dict, List, Optional import couchdb3 from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class SimpleCouchDBReader(BaseReader): """Simple CouchDB reader. Concatenates each CouchDB doc into Docu...
from keras.src import ops from keras.src.api_export import keras_export from keras.src.optimizers import optimizer @keras_export(["keras.optimizers.Adadelta"]) class Adadelta(optimizer.Optimizer): """Optimizer that implements the Adadelta algorithm. Adadelta optimization is a stochastic gradient descent meth...
from keras.src import ops from keras.src.api_export import keras_export from keras.src.optimizers import optimizer @keras_export(["keras.optimizers.Adadelta"]) class Adadelta(optimizer.Optimizer): """Optimizer that implements the Adadelta algorithm. Adadelta optimization is a stochastic gradient descent meth...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
# Copyright (c) OpenMMLab. All rights reserved. from .ade20k import (ADE20KInstanceDataset, ADE20KPanopticDataset, ADE20KSegDataset) from .base_det_dataset import BaseDetDataset from .base_semseg_dataset import BaseSegDataset from .base_video_dataset import BaseVideoDataset from .cityscapes import ...
# Copyright (c) OpenMMLab. All rights reserved. from .ade20k import (ADE20KInstanceDataset, ADE20KPanopticDataset, ADE20KSegDataset) from .base_det_dataset import BaseDetDataset from .base_semseg_dataset import BaseSegDataset from .base_video_dataset import BaseVideoDataset from .cityscapes import ...
from typing import TYPE_CHECKING, Tuple, TypeVar import numpy as np from docarray.typing.url.url_3d.url_3d import Url3D if TYPE_CHECKING: from docarray.proto import NodeProto T = TypeVar('T', bound='Mesh3DUrl') class Mesh3DUrl(Url3D): """ URL to a .obj, .glb, or .ply file containing 3D mesh informatio...
from typing import TYPE_CHECKING, Tuple, TypeVar import numpy as np from docarray.typing.url.url_3d.url_3d import Url3D if TYPE_CHECKING: from docarray.proto import NodeProto T = TypeVar('T', bound='Mesh3DUrl') class Mesh3DUrl(Url3D): """ URL to a .obj, .glb, or .ply file containing 3D mesh informatio...
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
_base_ = 'faster-rcnn_r50-caffe_fpn_1x_coco.py' max_iter = 90000 param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=max_iter, by_epoch=False, milestones=[60000, 80000], ...
_base_ = 'faster-rcnn_r50-caffe_fpn_1x_coco.py' # learning policy lr_config = dict( policy='step', warmup='linear', warmup_iters=500, warmup_ratio=0.001, step=[60000, 80000]) # Runner type runner = dict(_delete_=True, type='IterBasedRunner', max_iters=90000) checkpoint_config = dict(interval=1000...
from pathlib import Path import pytest import numpy as np import paddlehub as hub from jina import Document, DocumentArray, Executor from ...text_paddle import TextPaddleEncoder @pytest.fixture(scope='function') def model(): return hub.Module(name='ernie_tiny') @pytest.fixture(scope='function') def content():...
import pytest import numpy as np import paddlehub as hub from jina.executors import BaseExecutor from jina import Document, DocumentArray @pytest.fixture(scope='function') def model(): return hub.Module(name='ernie_tiny') @pytest.fixture(scope='function') def content(): return 'hello world' @pytest.fixtur...
import numpy as np from absl.testing import parameterized from keras.src import backend from keras.src import testing from keras.src.utils import numerical_utils NUM_CLASSES = 5 class TestNumericalUtils(testing.TestCase): @parameterized.parameters( [ ((1,), (1, NUM_CLASSES)), ((3...
import numpy as np from absl.testing import parameterized from keras.src import backend from keras.src import testing from keras.src.utils import numerical_utils NUM_CLASSES = 5 class TestNumericalUtils(testing.TestCase, parameterized.TestCase): @parameterized.parameters( [ ((1,), (1, NUM_CL...
from functools import partial from inspect import isclass from typing import Any, Union, cast from pydantic import BaseModel from langchain_core.language_models import FakeListChatModel from langchain_core.load.dump import dumps from langchain_core.load.load import loads from langchain_core.messages import HumanMessa...
from functools import partial from inspect import isclass from typing import Any, Union, cast from pydantic import BaseModel from langchain_core.language_models import FakeListChatModel from langchain_core.load.dump import dumps from langchain_core.load.load import loads from langchain_core.messages import HumanMessa...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
# Copyright (c) OpenMMLab. All rights reserved. import copy import inspect from typing import List import torch import torch.nn as nn from mmengine.registry import OPTIMIZER_CONSTRUCTORS, OPTIMIZERS def register_torch_optimizers() -> List[str]: """Register optimizers in ``torch.optim`` to the ``OPTIMIZERS`` reg...
# Copyright (c) OpenMMLab. All rights reserved. import copy import inspect from typing import List, Optional import torch import torch.nn as nn from mmengine.registry import OPTIMIZER_CONSTRUCTORS, OPTIMIZERS def register_torch_optimizers() -> List[str]: """Register optimizers in ``torch.optim`` to the ``OPTIMI...
from __future__ import annotations import collections import json import os import string from collections.abc import Iterable from .WordTokenizer import ENGLISH_STOP_WORDS, WordTokenizer class WhitespaceTokenizer(WordTokenizer): """ Simple and fast white-space tokenizer. Splits sentence based on white spac...
from __future__ import annotations import collections import json import os import string from typing import Iterable from .WordTokenizer import ENGLISH_STOP_WORDS, WordTokenizer class WhitespaceTokenizer(WordTokenizer): """ Simple and fast white-space tokenizer. Splits sentence based on white spaces. P...
import pytest import os from typing import Generator # this fixture is used to mask the NVIDIA_API_KEY environment variable and restore it # after the test. it also returns the value of the NVIDIA_API_KEY environment variable # before it was masked so that it can be used in the test. @pytest.fixture() def masked_en...
import pytest import os from typing import Generator # this fixture is used to mask the NVIDIA_API_KEY environment variable and restore it # after the test. it also returns the value of the NVIDIA_API_KEY environment variable # before it was masked so that it can be used in the test. @pytest.fixture() def masked_en...
from backend.blocks.hubspot._auth import ( HubSpotCredentials, HubSpotCredentialsField, HubSpotCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField from backend.util.request import Requests class HubSpotCompanyBlock(Bl...
from backend.blocks.hubspot._auth import ( HubSpotCredentials, HubSpotCredentialsField, HubSpotCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField from backend.util.request import requests class HubSpotCompanyBlock(Bl...
"""Generation output schema.""" from __future__ import annotations from typing import Any, Literal, Optional from langchain_core.load import Serializable from langchain_core.utils._merge import merge_dicts class Generation(Serializable): """A single text generation output. Generation represents the respon...
from __future__ import annotations from typing import Any, Literal, Optional from langchain_core.load import Serializable from langchain_core.utils._merge import merge_dicts class Generation(Serializable): """A single text generation output. Generation represents the response from an "old-fashioned" LLM th...
"""Chat Message.""" from typing import Any, Literal from typing_extensions import override from langchain_core.messages.base import ( BaseMessage, BaseMessageChunk, merge_content, ) from langchain_core.utils._merge import merge_dicts class ChatMessage(BaseMessage): """Message that can be assigned a...
from typing import Any, Literal from langchain_core.messages.base import ( BaseMessage, BaseMessageChunk, merge_content, ) from langchain_core.utils._merge import merge_dicts class ChatMessage(BaseMessage): """Message that can be assigned an arbitrary speaker (i.e. role).""" role: str """The...
import asyncio import logging import os from jina import __default_host__ from jina.importer import ImportExtensions from jina.serve.runtimes.gateway import GatewayRuntime from jina.serve.runtimes.gateway.http.app import get_fastapi_app __all__ = ['HTTPGatewayRuntime'] class HTTPGatewayRuntime(GatewayRuntime): ...
import asyncio import logging import os from jina import __default_host__ from jina.importer import ImportExtensions from jina.serve.runtimes.gateway import GatewayRuntime from jina.serve.runtimes.gateway.http.app import get_fastapi_app __all__ = ['HTTPGatewayRuntime'] class HTTPGatewayRuntime(GatewayRuntime): ...
import gc import asyncio from llama_index.core.memory import ChatMemoryBuffer from llama_index.core.base.llms.types import ( ChatMessage, CompletionResponse, CompletionResponseGen, ) from typing import Any from llama_index.core.llms.callbacks import llm_completion_callback from llama_index.core.llms.mock im...
from llama_index.core.base.llms.types import ChatMessage, MessageRole from llama_index.core.chat_engine.simple import SimpleChatEngine def test_simple_chat_engine() -> None: engine = SimpleChatEngine.from_defaults() engine.reset() response = engine.chat("Test message 1") assert str(response) == "user...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/coco_instance.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( data_preprocessor=dict( # The mean and std are used in PyCls when training RegNets mean=[103.53, 116.28, 123.675...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/coco_instance.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( backbone=dict( _delete_=True, type='RegNet', arch='regnetx_3.2gf', out_indices=(0, 1, 2, 3), ...
import os import sysconfig from typing import Optional from torch.utils._triton import has_triton def enable_triton(lib_dir: Optional[str] = None) -> dict[str, str]: """ Enable NVSHMEM device functions for Triton. It performs a NVSHMEM device-side initialization on the kernel module created by Triton. ...
import os import sysconfig from typing import Optional from torch.utils._triton import has_triton def enable_triton(lib_dir: Optional[str] = None) -> dict[str, str]: """ Enable NVSHMEM device functions for Triton. It performs a NVSHMEM device-side initialization on the kernel module created by Triton. ...
_base_ = './cascade-rcnn_r50_fpn_20e_coco.py' model = dict( type='CascadeRCNN', backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True)...
_base_ = './cascade_rcnn_r50_fpn_20e_coco.py' model = dict( type='CascadeRCNN', backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True)...
# Copyright (c) OpenMMLab. All rights reserved. from mmengine.registry import HOOKS from .hook import Hook @HOOKS.register_module() class DistSamplerSeedHook(Hook): """Data-loading sampler for distributed training. When distributed training, it is only useful in conjunction with :obj:`EpochBasedRunner`, ...
# Copyright (c) OpenMMLab. All rights reserved. from mmengine.registry import HOOKS from .hook import Hook @HOOKS.register_module() class DistSamplerSeedHook(Hook): """Data-loading sampler for distributed training. When distributed training, it is only useful in conjunction with :obj:`EpochBasedRunner`, ...
from typing import Dict, Set from fastapi import WebSocket from backend.data.execution import ( ExecutionEventType, GraphExecutionEvent, NodeExecutionEvent, ) from backend.server.model import WSMessage, WSMethod _EVENT_TYPE_TO_METHOD_MAP: dict[ExecutionEventType, WSMethod] = { ExecutionEventType.GRAP...
from typing import Dict, Set from fastapi import WebSocket from backend.data.execution import ( ExecutionEventType, GraphExecutionEvent, NodeExecutionEvent, ) from backend.server.model import WSMessage, WSMethod _EVENT_TYPE_TO_METHOD_MAP: dict[ExecutionEventType, WSMethod] = { ExecutionEventType.GRAP...
import os import urllib import urllib.parse import urllib.request from typing import TYPE_CHECKING, Optional, Type, TypeVar from pydantic import AnyUrl as BaseAnyUrl from pydantic import errors, parse_obj_as from docarray.typing.abstract_type import AbstractType from docarray.typing.proto_register import _register_pr...
from typing import TYPE_CHECKING, Optional, Type, TypeVar from pydantic import AnyUrl as BaseAnyUrl from pydantic import errors, parse_obj_as from docarray.typing.abstract_type import AbstractType from docarray.typing.proto_register import _register_proto if TYPE_CHECKING: from pydantic.networks import Parts ...
""" ============================================= A demo of the Spectral Biclustering algorithm ============================================= This example demonstrates how to generate a checkerboard dataset and bicluster it using the :class:`~sklearn.cluster.SpectralBiclustering` algorithm. The spectral biclustering a...
""" ============================================= A demo of the Spectral Biclustering algorithm ============================================= This example demonstrates how to generate a checkerboard dataset and bicluster it using the :class:`~sklearn.cluster.SpectralBiclustering` algorithm. The spectral biclustering a...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] # model settings preprocess_cfg = dict( mean=[102.9801, 115.9465, 122.7717], std=[1.0, 1.0, 1.0], to_rgb=False, pad_size_divisor=32) model = dict( type='FCOS', prepr...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] # model settings preprocess_cfg = dict( mean=[102.9801, 115.9465, 122.7717], std=[1.0, 1.0, 1.0], to_rgb=False, pad_size_divisor=32) model = dict( type='FCOS', prepr...
import defusedxml.ElementTree as ET import pytest from llama_index.readers.file.xml import XMLReader # Sample XML data for testing SAMPLE_XML = """<?xml version="1.0" encoding="UTF-8"?> <data> <item type="fruit"> <name>Apple</name> <color>Red</color> <price>1.20</price> </item> <it...
import xml.etree.ElementTree as ET import pytest from llama_index.readers.file.xml import XMLReader # Sample XML data for testing SAMPLE_XML = """<?xml version="1.0" encoding="UTF-8"?> <data> <item type="fruit"> <name>Apple</name> <color>Red</color> <price>1.20</price> </item> <ite...
"""Callback Handler that tracks AIMessage.usage_metadata.""" import threading from collections.abc import Generator from contextlib import contextmanager from contextvars import ContextVar from typing import Any, Optional from langchain_core._api import beta from langchain_core.callbacks import BaseCallbackHandler fr...
"""Callback Handler that tracks AIMessage.usage_metadata.""" import threading from collections.abc import Generator from contextlib import contextmanager from contextvars import ContextVar from typing import Any, Optional from langchain_core._api import beta from langchain_core.callbacks import BaseCallbackHandler fr...
"""This modules defines all kinds of exceptions raised in Jina.""" from typing import Set, Union import grpc.aio class BaseJinaException(BaseException): """A base class for all exceptions raised by Jina""" class RuntimeFailToStart(SystemError, BaseJinaException): """When pod/deployment is failed to started...
"""This modules defines all kinds of exceptions raised in Jina.""" from typing import Set, Union import grpc.aio class BaseJinaException(BaseException): """A base class for all exceptions raised by Jina""" class RuntimeFailToStart(SystemError, BaseJinaException): """When pod/deployment is failed to started...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Optional, Tuple import torch from torch import Tensor from mmdet.structures.bbox import BaseBoxes def anchor_inside_flags(flat_anchors: Tensor, valid_flags: Tensor, img_shape: Tuple[int], ...
# Copyright (c) OpenMMLab. All rights reserved. import torch from mmdet.structures.bbox import BaseBoxes def anchor_inside_flags(flat_anchors, valid_flags, img_shape, allowed_border=0): """Check whether the anchors are inside the border. ...
from __future__ import annotations import collections import json import logging import os import string from typing import Iterable from transformers.utils.import_utils import NLTK_IMPORT_ERROR, is_nltk_available from .WordTokenizer import ENGLISH_STOP_WORDS, WordTokenizer logger = logging.getLogger(__name__) cl...
from __future__ import annotations import collections import json import logging import os import string from typing import Iterable from transformers.utils.import_utils import NLTK_IMPORT_ERROR, is_nltk_available from .WordTokenizer import ENGLISH_STOP_WORDS, WordTokenizer logger = logging.getLogger(__name__) cl...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( type='CenterNet', # use caffe img_norm data_preprocessor=dict( type='DetDataPreprocessor', mean=[103.530, 116.280, 123.675], std=[1.0, 1.0,...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( type='CenterNet', # use caffe img_norm data_preprocessor=dict( type='DetDataPreprocessor', mean=[103.530, 116.280, 123.675], std=[1.0, 1.0,...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from mmdet.datasets import CocoDataset class TestCocoDataset(unittest.TestCase): def test_coco_dataset(self): # test CocoDataset metainfo = dict(CLASSES=('bus', 'car'), task_name='new_task') dataset = CocoDataset( ...
# Copyright (c) OpenMMLab. All rights reserved. import pytest from mmdet.datasets import CocoDataset class TestCocoDataset: def test_coco_dataset(self): # test CocoDataset metainfo = dict(CLASSES=('bus', 'car'), task_name='new_task') dataset = CocoDataset( data_prefix=dict(im...
# Copyright 2025 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
_base_ = [ './faster-rcnn_r50_fpn.py', './mot_challenge.py', '../../../configs/_base_/default_runtime.py' ] model = dict( type='Tracktor', pretrains=dict( detector= # noqa: E251 'https://download.openmmlab.com/mmtracking/mot/faster_rcnn/faster-rcnn_r50_fpn_4e_mot17-half-64ee2ed4.pth', ...
_base_ = [ './faster_rcnn_r50_fpn.py', './mot_challenge.py', '../../../configs/_base_/default_runtime.py' ] model = dict( type='Tracktor', pretrains=dict( detector= # noqa: E251 'https://download.openmmlab.com/mmtracking/mot/faster_rcnn/faster-rcnn_r50_fpn_4e_mot17-half-64ee2ed4.pth', ...
import keras.src.backend from keras.src import tree from keras.src.layers.layer import Layer from keras.src.random.seed_generator import SeedGenerator from keras.src.utils import backend_utils from keras.src.utils import tracking class TFDataLayer(Layer): """Layer that can safely used in a tf.data pipeline. ...
import keras.src.backend from keras.src import tree from keras.src.layers.layer import Layer from keras.src.random.seed_generator import SeedGenerator from keras.src.utils import backend_utils from keras.src.utils import tracking class TFDataLayer(Layer): """Layer that can safely used in a tf.data pipeline. ...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase from unittest.mock import Mock import torch.nn as nn from torch.optim import SGD from mmengine.hooks import RuntimeInfoHook from mmengine.logging import MessageHub from mmengine.optim import OptimWrapper, OptimWrapperDict class TestRuntim...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase from unittest.mock import Mock import torch.nn as nn from torch.optim import SGD from mmengine.hooks import RuntimeInfoHook from mmengine.logging import MessageHub from mmengine.optim import OptimWrapper, OptimWrapperDict class TestRuntim...