input
stringlengths
33
5k
output
stringlengths
32
5k
# Copyright (c) OpenMMLab. All rights reserved. import os import subprocess import warnings from packaging.version import parse def digit_version(version_str: str, length: int = 4): """Convert a version string into a tuple of integers. This method is usually used for comparing two versions. For pre-release ...
# Copyright (c) OpenMMLab. All rights reserved. import os import subprocess import warnings from packaging.version import parse def digit_version(version_str: str, length: int = 4): """Convert a version string into a tuple of integers. This method is usually used for comparing two versions. For pre-release ...
_base_ = [ '../_base_/models/cascade-rcnn_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_20e.py', '../_base_/default_runtime.py' ]
_base_ = [ '../_base_/models/cascade_rcnn_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_20e.py', '../_base_/default_runtime.py' ]
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import numpy as np import pytest import torch from jina import DocumentArray, Document from ...sentence_encoder import TransformerSentenceEncoder def test_encoding_cpu(): enc = TransformerSentenceEncoder(d...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import numpy as np import pytest import torch from jina import DocumentArray, Document from jinahub.text.encoders.sentence_encoder import TransformerSentenceEncoder def test_encoding_cpu(): enc = Transform...
"""Callback Handler that writes to a file.""" from __future__ import annotations from pathlib import Path from typing import TYPE_CHECKING, Any, Optional, TextIO, cast from langchain_core.callbacks import BaseCallbackHandler from langchain_core.utils.input import print_text if TYPE_CHECKING: from langchain_core...
"""Callback Handler that writes to a file.""" from __future__ import annotations from pathlib import Path from typing import TYPE_CHECKING, Any, Optional, TextIO, cast from langchain_core.callbacks import BaseCallbackHandler from langchain_core.utils.input import print_text if TYPE_CHECKING: from langchain_core...
""" This example starts multiple processes (1 per GPU), which encode sentences in parallel. This gives a near linear speed-up when encoding large text collections. It also demonstrates how to stream data which is helpful in case you don't want to wait for an extremely large dataset to download, or if you want to limit ...
""" This example starts multiple processes (1 per GPU), which encode sentences in parallel. This gives a near linear speed-up when encoding large text collections. It also demonstrates how to stream data which is helpful in case you don't want to wait for an extremely large dataset to download, or if you want to limit ...
from __future__ import annotations from collections.abc import Iterable import torch from torch import Tensor, nn from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class FlopsLoss(nn.Module): def __init__(self, model: SparseEncoder, threshold: float = None) -> None: """ ...
from __future__ import annotations from collections.abc import Iterable import torch from torch import Tensor, nn from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class FlopsLoss(nn.Module): def __init__(self, model: SparseEncoder, threshold: float = None) -> None: """ ...
# Copyright (c) OpenMMLab. All rights reserved. import math from mmcv.cnn import build_conv_layer, build_norm_layer from ..builder import BACKBONES from .detectors_resnet import Bottleneck as _Bottleneck from .detectors_resnet import DetectoRS_ResNet class Bottleneck(_Bottleneck): expansion = 4 def __init_...
import math from mmcv.cnn import build_conv_layer, build_norm_layer from ..builder import BACKBONES from .detectors_resnet import Bottleneck as _Bottleneck from .detectors_resnet import DetectoRS_ResNet class Bottleneck(_Bottleneck): expansion = 4 def __init__(self, inplanes, ...
"""From https://github.com/SidU/teams-langchain-js/wiki/Connecting-IFTTT-Services. # Creating a webhook - Go to https://ifttt.com/create # Configuring the "If This" - Click on the "If This" button in the IFTTT interface. - Search for "Webhooks" in the search bar. - Choose the first option for "Receive a web request w...
"""From https://github.com/SidU/teams-langchain-js/wiki/Connecting-IFTTT-Services. # Creating a webhook - Go to https://ifttt.com/create # Configuring the "If This" - Click on the "If This" button in the IFTTT interface. - Search for "Webhooks" in the search bar. - Choose the first option for "Receive a web request w...
from pathlib import Path from typing import Any, List, Union from langchain_community.document_loaders.unstructured import UnstructuredFileLoader class UnstructuredHTMLLoader(UnstructuredFileLoader): """Load `HTML` files using `Unstructured`. You can run the loader in one of two modes: "single" and "element...
from pathlib import Path from typing import Any, List, Union from langchain_community.document_loaders.unstructured import UnstructuredFileLoader class UnstructuredHTMLLoader(UnstructuredFileLoader): """Load `HTML` files using `Unstructured`. You can run the loader in one of two modes: "single" and "element...
__version__ = '0.19.1' import os from docarray.document import Document from docarray.array import DocumentArray from docarray.dataclasses import dataclass, field from docarray.helper import login, logout if 'DA_RICH_HANDLER' in os.environ: from rich.traceback import install install()
__version__ = '0.19.0' import os from docarray.document import Document from docarray.array import DocumentArray from docarray.dataclasses import dataclass, field from docarray.helper import login, logout if 'DA_RICH_HANDLER' in os.environ: from rich.traceback import install install()
import requests from packaging import version from typing import Union, List, Optional from llama_index.core.base.llms.types import ( ChatResponse, ) def get_max_input_tokens(url: str) -> Union[int, None]: url = f"{url}/info" model_info = dict(requests.get(url).json()) tgi_version = model_info.get("ve...
import requests from packaging import version from typing import Union, List, Optional from llama_index.core.base.llms.types import ( ChatResponse, ) def get_max_input_tokens(url: str) -> Union[int, None]: url = f"{url}/info" model_info = dict(requests.get(url).json()) tgi_version = model_info.get("ve...
# Copyright 2024 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
# Copyright 2024 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
_base_ = './mask-rcnn_r101_fpn_1x_coco.py' model = dict( # ResNeXt-101-32x8d model trained with Caffe2 at FB, # so the mean and std need to be changed. data_preprocessor=dict( mean=[103.530, 116.280, 123.675], std=[57.375, 57.120, 58.395], bgr_to_rgb=False), backbone=dict( ...
_base_ = './mask-rcnn_r101_fpn_1x_coco.py' model = dict( # ResNeXt-101-32x8d model trained with Caffe2 at FB, # so the mean and std need to be changed. data_preprocessor=dict( mean=[103.530, 116.280, 123.675], std=[57.375, 57.120, 58.395], bgr_to_rgb=False), backbone=dict( ...
import logging from datasets import load_dataset from sentence_transformers import SparseEncoder from sentence_transformers.sparse_encoder.evaluation import SparseTranslationEvaluator logging.basicConfig(format="%(message)s", level=logging.INFO) # Load a model, not mutilingual but hope to see some on the hub soon m...
import logging from datasets import load_dataset from sentence_transformers import SparseEncoder from sentence_transformers.sparse_encoder.evaluation import SparseTranslationEvaluator logging.basicConfig(format="%(message)s", level=logging.INFO) # Load a model, not mutilingual but hope to see some on the hub soon m...
import dataclasses from collections import defaultdict from typing import TYPE_CHECKING, Type, List if TYPE_CHECKING: from docarray.typing import T from docarray.document.strawberry_type import StrawberryDocument class StrawberryMixin: """Provide helper functions to convert to/from a Strawberry model""" ...
import dataclasses from collections import defaultdict from typing import TYPE_CHECKING, Type, List if TYPE_CHECKING: from ...typing import T from ..strawberry_type import StrawberryDocument class StrawberryMixin: """Provide helper functions to convert to/from a Strawberry model""" def to_strawberry...
# Copyright (c) OpenMMLab. All rights reserved. from .checkloss_hook import CheckInvalidLossHook from .mean_teacher_hook import MeanTeacherHook from .memory_profiler_hook import MemoryProfilerHook from .num_class_check_hook import NumClassCheckHook from .set_epoch_info_hook import SetEpochInfoHook from .sync_norm_hook ...
# Copyright (c) OpenMMLab. All rights reserved. from .checkloss_hook import CheckInvalidLossHook from .memory_profiler_hook import MemoryProfilerHook from .num_class_check_hook import NumClassCheckHook from .set_epoch_info_hook import SetEpochInfoHook from .sync_norm_hook import SyncNormHook from .visualization_hook im...
from enum import Enum from typing import Literal from pydantic import BaseModel, SecretStr from backend.data.model import APIKeyCredentials, CredentialsField, CredentialsMetaInput from backend.integrations.providers import ProviderName Slant3DCredentialsInput = CredentialsMetaInput[ Literal[ProviderName.SLANT3D]...
from enum import Enum from typing import Literal from pydantic import BaseModel, SecretStr from backend.data.model import APIKeyCredentials, CredentialsField, CredentialsMetaInput Slant3DCredentialsInput = CredentialsMetaInput[Literal["slant3d"], Literal["api_key"]] def Slant3DCredentialsField() -> Slant3DCredenti...
import sys from collections.abc import Mapping import numpy as np import pyarrow as pa from .. import config from ..utils.py_utils import map_nested from .formatting import Formatter class NumpyFormatter(Formatter[Mapping, np.ndarray, Mapping]): def __init__(self, features=None, **np_array_kwargs): supe...
import sys import numpy as np import pyarrow as pa from .. import config from ..utils.py_utils import map_nested from .formatting import Formatter class NumpyFormatter(Formatter[dict, np.ndarray, dict]): def __init__(self, features=None, decoded=True, **np_array_kwargs): super().__init__(features=featur...
import csv import gzip import logging import math import os from datetime import datetime from torch.utils.data import DataLoader from sentence_transformers import InputExample, LoggingHandler, SentenceTransformer, losses, models, util from sentence_transformers.evaluation import EmbeddingSimilarityEvaluator #### Ju...
from torch.utils.data import DataLoader import math from sentence_transformers import models, losses from sentence_transformers import LoggingHandler, SentenceTransformer, util, InputExample from sentence_transformers.evaluation import EmbeddingSimilarityEvaluator import logging from datetime import datetime import os ...
import datetime import prisma.fields import prisma.models import pytest import backend.server.v2.library.model as library_model from backend.util import json @pytest.mark.asyncio async def test_agent_preset_from_db(): # Create mock DB agent db_agent = prisma.models.AgentPreset( id="test-agent-123", ...
import datetime import prisma.fields import prisma.models import backend.server.v2.library.model as library_model def test_agent_preset_from_db(): # Create mock DB agent db_agent = prisma.models.AgentPreset( id="test-agent-123", createdAt=datetime.datetime.now(), updatedAt=datetime.d...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import MODELS from .two_stage import TwoStageDetector @MODELS.register_module() class MaskScoringRCNN(TwoStageDetector): """Mask Scoring RCNN. https://arxiv.org/abs/1903.00241 """ def __init__(self, backbone, ...
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .two_stage import TwoStageDetector @DETECTORS.register_module() class MaskScoringRCNN(TwoStageDetector): """Mask Scoring RCNN. https://arxiv.org/abs/1903.00241 """ def __init__(self, backbone, ...
# Copyright (c) OpenMMLab. All rights reserved. from .coco_api import COCO, COCOeval, COCOPanoptic from .panoptic_evaluation import pq_compute_multi_core, pq_compute_single_core __all__ = [ 'COCO', 'COCOeval', 'pq_compute_multi_core', 'pq_compute_single_core', 'COCOPanoptic' ]
# Copyright (c) OpenMMLab. All rights reserved. from .coco_api import COCO, COCOeval from .panoptic_evaluation import pq_compute_multi_core, pq_compute_single_core __all__ = [ 'COCO', 'COCOeval', 'pq_compute_multi_core', 'pq_compute_single_core' ]
""" This script contains an example how to perform semantic search with Elasticsearch. You need Elasticsearch up and running locally: https://www.elastic.co/guide/en/elasticsearch/reference/current/run-elasticsearch-locally.html Further, you need the Python Elasticsearch Client installed: https://elasticsearch-py.rea...
""" This script contains an example how to perform semantic search with Elasticsearch. You need Elasticsearch up and running locally: https://www.elastic.co/guide/en/elasticsearch/reference/current/run-elasticsearch-locally.html Further, you need the Python Elasticsearch Client installed: https://elasticsearch-py.rea...
_base_ = './fcos_r50_caffe_fpn_gn-head_1x_coco.py' # model settings preprocess_cfg = dict( mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True, pad_size_divisor=32) model = dict( preprocess_cfg=preprocess_cfg, backbone=dict( type='ResNeXt', depth=101, ...
_base_ = './fcos_r50_caffe_fpn_gn-head_1x_coco.py' # model settings preprocess_cfg = dict( mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rgb=True, pad_size_divisor=32) model = dict( preprocess_cfg=preprocess_cfg, backbone=dict( type='ResNeXt', depth=101, ...
from __future__ import annotations from typing import Any, List, Literal, Optional from langchain_core.embeddings import Embeddings from langchain_community.vectorstores.docarray.base import ( DocArrayIndex, _check_docarray_import, ) class DocArrayHnswSearch(DocArrayIndex): """`HnswLib` storage using `...
from __future__ import annotations from typing import Any, List, Literal, Optional from langchain_core.embeddings import Embeddings from langchain_community.vectorstores.docarray.base import ( DocArrayIndex, _check_docarray_import, ) class DocArrayHnswSearch(DocArrayIndex): """`HnswLib` storage using `...
"""langchain-core version information and utilities.""" VERSION = "0.3.54"
"""langchain-core version information and utilities.""" VERSION = "0.3.53"
import torch from torchaudio.models import Conformer from torchaudio_unittest.common_utils import TestBaseMixin, torch_script class ConformerTestImpl(TestBaseMixin): def _gen_model(self): conformer = ( Conformer( input_dim=80, num_heads=4, ffn_di...
import torch from torchaudio.models import Conformer from torchaudio_unittest.common_utils import TestBaseMixin, torch_script class ConformerTestImpl(TestBaseMixin): def _gen_model(self): conformer = ( Conformer( input_dim=80, num_heads=4, ffn_di...
"""Firebase Realtime Database Loader.""" from typing import Dict, List, Optional from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class FirebaseRealtimeDatabaseReader(BaseReader): """ Firebase Realtime Database reader. Retrieves data from Firebase Realti...
"""Firebase Realtime Database Loader.""" from typing import Dict, List, Optional from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class FirebaseRealtimeDatabaseReader(BaseReader): """Firebase Realtime Database reader. Retrieves data from Firebase Realtime Da...
from functools import lru_cache as _lru_cache from typing import Optional, TYPE_CHECKING import torch from torch.library import Library as _Library __all__ = ["is_built", "is_available", "is_macos13_or_newer", "is_macos_or_newer"] def is_built() -> bool: r"""Return whether PyTorch is built with MPS support. ...
# mypy: allow-untyped-defs from functools import lru_cache as _lru_cache from typing import Optional, TYPE_CHECKING import torch from torch.library import Library as _Library __all__ = ["is_built", "is_available", "is_macos13_or_newer", "is_macos_or_newer"] def is_built() -> bool: r"""Return whether PyTorch is...
__version__ = '0.39.2' import logging from docarray.array import DocList, DocVec from docarray.base_doc.doc import BaseDoc from docarray.utils._internal.misc import _get_path_from_docarray_root_level __all__ = ['BaseDoc', 'DocList', 'DocVec'] logger = logging.getLogger('docarray') handler = logging.StreamHandler()...
__version__ = '0.39.1' import logging from docarray.array import DocList, DocVec from docarray.base_doc.doc import BaseDoc from docarray.utils._internal.misc import _get_path_from_docarray_root_level __all__ = ['BaseDoc', 'DocList', 'DocVec'] logger = logging.getLogger('docarray') handler = logging.StreamHandler()...
""" Script to generate meta.json to store metadata for a nightly build of XGBoost Python package. """ import json import pathlib from argparse import ArgumentParser def main(args): wheel_path = pathlib.Path(args.wheel_path).expanduser().resolve() if not wheel_path.exists(): raise ValueError(f"Wheel ca...
""" Script to generate meta.json to store metadata for a nightly build of XGBoost Python package. """ import json import pathlib from argparse import ArgumentParser def main(args): wheel_path = pathlib.Path(args.wheel_path).expanduser().resolve() if not wheel_path.exists(): raise ValueError(f"Wheel ca...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import MODELS from mmdet.utils import ConfigType, OptConfigType, OptMultiConfig from .two_stage import TwoStageDetector @MODELS.register_module() class GridRCNN(TwoStageDetector): """Grid R-CNN. This detector is the implementation of: - ...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.core.utils import ConfigType, OptConfigType, OptMultiConfig from mmdet.registry import MODELS from .two_stage import TwoStageDetector @MODELS.register_module() class GridRCNN(TwoStageDetector): """Grid R-CNN. This detector is the implementation of: ...
_base_ = './rtmdet_s_8xb32-300e_coco.py' checkpoint = 'https://download.openmmlab.com/mmdetection/v3.0/rtmdet/cspnext_rsb_pretrain/cspnext-tiny_imagenet_600e.pth' # noqa model = dict( backbone=dict( deepen_factor=0.167, widen_factor=0.375, init_cfg=dict( type='Pretrained', pre...
_base_ = './rtmdet_s_8xb32-300e_coco.py' checkpoint = 'https://download.openmmlab.com/mmdetection/v3.0/rtmdet/cspnext_rsb_pretrain/cspnext-tiny_imagenet_600e.pth' # noqa model = dict( backbone=dict( deepen_factor=0.167, widen_factor=0.375, init_cfg=dict( type='Pretrained', pre...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.api.optimizers import legacy from keras.api.optimizers import schedules from keras.src.optimizers import deserialize from keras.src.optimizers import get from keras.src.optimizers import ...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.api.optimizers import legacy from keras.api.optimizers import schedules from keras.src.optimizers import deserialize from keras.src.optimizers import get from keras.src.optimizers import ...
import numpy as np import pytest from keras.src import backend from keras.src import layers from keras.src import testing class GaussianNoiseTest(testing.TestCase): @pytest.mark.requires_trainable_backend def test_gaussian_noise_basics(self): self.run_layer_test( layers.GaussianNoise, ...
import numpy as np import pytest from keras.src import backend from keras.src import layers from keras.src import testing class GaussianNoiseTest(testing.TestCase): @pytest.mark.requires_trainable_backend def test_gaussian_noise_basics(self): self.run_layer_test( layers.GaussianNoise, ...
import pytest from xgboost import testing as tm from xgboost.testing.ordinal import ( run_cat_container, run_cat_container_iter, run_cat_container_mixed, ) pytestmark = pytest.mark.skipif(**tm.no_multiple(tm.no_arrow(), tm.no_pandas())) def test_cat_container() -> None: run_cat_container("cpu") de...
import pytest from xgboost import testing as tm from xgboost.testing.ordinal import run_cat_container, run_cat_container_mixed pytestmark = pytest.mark.skipif(**tm.no_multiple(tm.no_arrow(), tm.no_pandas())) def test_cat_container() -> None: run_cat_container("cpu") def test_cat_container_mixed() -> None: ...
# Copyright 2020 The HuggingFace Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to...
# Copyright 2020 The HuggingFace Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from mmengine.config import ConfigDict from mmengine.data import InstanceData from parameterized import parameterized from mmdet.models.roi_heads.mask_heads import MaskIoUHead from mmdet.models.utils import unpa...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from mmengine.config import ConfigDict from mmengine.data import InstanceData from parameterized import parameterized from mmdet.data_elements.mask import mask_target from mmdet.models.roi_heads.mask_heads impor...
import json import re from typing import TypeVar import yaml from langchain_core.exceptions import OutputParserException from langchain_core.output_parsers import BaseOutputParser from pydantic import BaseModel, ValidationError from langchain.output_parsers.format_instructions import YAML_FORMAT_INSTRUCTIONS T = Typ...
import json import re from typing import TypeVar import yaml from langchain_core.exceptions import OutputParserException from langchain_core.output_parsers import BaseOutputParser from pydantic import BaseModel, ValidationError from langchain.output_parsers.format_instructions import YAML_FORMAT_INSTRUCTIONS T = Typ...
import time import pytest from jina import Document, DocumentArray, Executor, Flow, requests @pytest.mark.parametrize( 'shards, expected_response', [(1, ['slow', 'fast']), (2, ['fast', 'slow'])] ) def test_non_blocking_gateway(shards, expected_response): class FastSlowExecutor(Executor): def __init_...
import time import pytest from jina import Document, DocumentArray, Executor, Flow, requests @pytest.mark.parametrize( 'shards, expected_response', [(1, ['slow', 'fast']), (2, ['fast', 'slow'])] ) def test_non_blocking_gateway(shards, expected_response): class FastSlowExecutor(Executor): def __init_...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.llms import AI21 from langchain_community.llms.ai21 import AI21PenaltyData # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation ...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.llms import AI21 from langchain_community.llms.ai21 import AI21PenaltyData # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation ...
import pathlib from argparse import ArgumentParser import sentencepiece as spm from lightning import ConformerRNNTModule from pytorch_lightning import seed_everything, Trainer from pytorch_lightning.callbacks import LearningRateMonitor, ModelCheckpoint from pytorch_lightning.plugins import DDPPlugin from transforms i...
import pathlib from argparse import ArgumentParser from lightning import ConformerRNNTModule from pytorch_lightning import seed_everything, Trainer from pytorch_lightning.callbacks import LearningRateMonitor, ModelCheckpoint from pytorch_lightning.plugins import DDPPlugin from transforms import get_data_module def r...
"""Optimization related classes and functions.""" import logging from typing import Any, Dict, List, Optional, Literal from llama_index.core.bridge.pydantic import Field, PrivateAttr from llama_index.core.postprocessor.types import BaseNodePostprocessor from llama_index.core.schema import NodeWithScore, QueryBundle, ...
"""Optimization related classes and functions.""" import logging from typing import Any, Dict, List, Optional from llama_index.core.bridge.pydantic import Field, PrivateAttr from llama_index.core.postprocessor.types import BaseNodePostprocessor from llama_index.core.schema import MetadataMode, NodeWithScore, QueryBun...
""" Checkpoint functionality for machine learning models. This module provides classes for saving and loading model checkpoints in a distributed training environment. It includes functionality for coordinating checkpoint operations across multiple processes and customizing the checkpoint process through hooks. Key co...
""" Checkpoint functionality for machine learning models. This module provides classes for saving and loading model checkpoints in a distributed training environment. It includes functionality for coordinating checkpoint operations across multiple processes and customizing the checkpoint process through hooks. Key co...
# flake8: noqa: F401 r""" This file is in the process of migration to `torch/ao/quantization`, and is kept here for compatibility while the migration process is ongoing. If you are adding a new entry/functionality, please, add it to the `torch/ao/quantization/fuse_modules.py`, while adding an import statement here. """...
# flake8: noqa: F401 r""" This file is in the process of migration to `torch/ao/quantization`, and is kept here for compatibility while the migration process is ongoing. If you are adding a new entry/functionality, please, add it to the `torch/ao/quantization/fuse_modules.py`, while adding an import statement here. """...
import warnings from typing import Any, Union import numpy as np import PIL.Image import torch from torchvision.transforms import functional as _F from torchvision.transforms.v2 import Transform class ToTensor(Transform): """[DEPRECATED] Use ``v2.Compose([v2.ToImage(), v2.ToDtype(torch.float32, scale=True)])`` ...
import warnings from typing import Any, Dict, Union import numpy as np import PIL.Image import torch from torchvision.transforms import functional as _F from torchvision.transforms.v2 import Transform class ToTensor(Transform): """[DEPRECATED] Use ``v2.Compose([v2.ToImage(), v2.ToDtype(torch.float32, scale=True...
from functools import wraps from typing import Any, Callable, Concatenate, Coroutine, ParamSpec, TypeVar, cast from backend.data.credit import get_user_credit_model from backend.data.execution import ( ExecutionResult, NodeExecutionEntry, RedisExecutionEventBus, create_graph_execution, get_executio...
from functools import wraps from typing import Any, Callable, Concatenate, Coroutine, ParamSpec, TypeVar, cast from backend.data.credit import get_user_credit_model from backend.data.execution import ( ExecutionResult, NodeExecutionEntry, RedisExecutionEventBus, create_graph_execution, get_executio...
from importlib import import_module from .logging import get_logger logger = get_logger(__name__) class _PatchedModuleObj: """Set all the modules components as attributes of the _PatchedModuleObj object.""" def __init__(self, module, attrs=None): attrs = attrs or [] if module is not None: ...
from importlib import import_module from .logging import get_logger logger = get_logger(__name__) class _PatchedModuleObj: """Set all the modules components as attributes of the _PatchedModuleObj object.""" def __init__(self, module, attrs=None): attrs = attrs or [] if module is not None: ...
# coding: utf-8 """Get the most recent status of workflow for the current PR. [usage] python get_workflow_status.py TRIGGER_PHRASE TRIGGER_PHRASE: Code phrase that triggers workflow. """ import json from os import environ from sys import argv, exit from time import sleep try: from urllib import request excep...
# coding: utf-8 """Get the most recent status of workflow for the current PR. [usage] python get_workflow_status.py TRIGGER_PHRASE TRIGGER_PHRASE: Code phrase that triggers workflow. """ import json from os import environ from sys import argv, exit from time import sleep try: from urllib import request excep...
# Copyright (c) OpenMMLab. All rights reserved. from .backbones import * # noqa: F401,F403 from .builder import (BACKBONES, DETECTORS, HEADS, LOSSES, NECKS, ROI_EXTRACTORS, SHARED_HEADS, build_backbone, build_detector, build_head, build_loss, build_neck, ...
from .backbones import * # noqa: F401,F403 from .builder import (BACKBONES, DETECTORS, HEADS, LOSSES, NECKS, ROI_EXTRACTORS, SHARED_HEADS, build_backbone, build_detector, build_head, build_loss, build_neck, build_roi_extractor, build_shared_head) from ....
# ruff: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICE...
# ruff: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICE...
import logging from datasets import load_dataset from sentence_transformers import SparseEncoder from sentence_transformers.sparse_encoder.evaluation import SparseMSEEvaluator logging.basicConfig(format="%(message)s", level=logging.INFO) # Load a model student_model = SparseEncoder("prithivida/Splade_PP_en_v1") tea...
import logging from datasets import load_dataset from sentence_transformers import SparseEncoder from sentence_transformers.sparse_encoder.evaluation import SparseMSEEvaluator logging.basicConfig(format="%(message)s", level=logging.INFO) # Load a model student_model = SparseEncoder("prithivida/Splade_PP_en_v1") tea...
import grpc from grpc_health.v1 import health, health_pb2, health_pb2_grpc from grpc_reflection.v1alpha import reflection from pydantic import BaseModel from uvicorn import Config, Server from jina import Gateway from jina.constants import __default_host__ from jina.proto import jina_pb2, jina_pb2_grpc class DummyRe...
import grpc from grpc_health.v1 import health, health_pb2, health_pb2_grpc from grpc_reflection.v1alpha import reflection from pydantic import BaseModel from uvicorn import Config, Server from jina import Gateway, __default_host__ from jina.proto import jina_pb2, jina_pb2_grpc class DummyResponseModel(BaseModel): ...
import click from .cmd_exec import cmd_exec from .info import info @click.group(short_help="Manage packages in the monorepo") def pkg(): pass # pragma: no cover pkg.add_command(info) pkg.add_command(cmd_exec, name="exec")
import click from .cmd_exec import cmd_exec from .info import info @click.group(short_help="Manage packages in the monorepo") def pkg(): pass pkg.add_command(info) pkg.add_command(cmd_exec, name="exec")
from typing import Any, Mapping, Optional from llama_index.readers.airbyte_cdk.base import AirbyteCDKReader, RecordHandler class AirbyteSalesforceReader(AirbyteCDKReader): """ AirbyteSalesforceReader reader. Retrieve documents from Salesforce Args: config: The config object for the salesfor...
from typing import Any, Mapping, Optional from llama_index.readers.airbyte_cdk.base import AirbyteCDKReader, RecordHandler class AirbyteSalesforceReader(AirbyteCDKReader): """AirbyteSalesforceReader reader. Retrieve documents from Salesforce Args: config: The config object for the salesforce so...
# Licensed to the LF AI & Data foundation under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the "License"); # you may not use this fil...
# Licensed to the LF AI & Data foundation under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the "License"); # you may not use this fil...
from base64 import b64encode from urllib.parse import urlencode from backend.data.model import OAuth2Credentials from backend.integrations.providers import ProviderName from backend.util.request import requests from .base import BaseOAuthHandler class NotionOAuthHandler(BaseOAuthHandler): """ Based on the d...
from base64 import b64encode from urllib.parse import urlencode from backend.data.model import OAuth2Credentials from backend.util.request import requests from .base import BaseOAuthHandler class NotionOAuthHandler(BaseOAuthHandler): """ Based on the documentation at https://developers.notion.com/docs/autho...
_base_ = './mask_rcnn_r50_fpn_1x_coco.py' preprocess_cfg = dict( mean=[103.530, 116.280, 123.675], std=[1.0, 1.0, 1.0], to_rgb=False, pad_size_divisor=32) model = dict( # use caffe img_norm preprocess_cfg=preprocess_cfg, backbone=dict( norm_cfg=dict(requires_grad=False), styl...
_base_ = './mask_rcnn_r50_fpn_1x_coco.py' model = dict( backbone=dict( norm_cfg=dict(requires_grad=False), style='caffe', init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://detectron2/resnet50_caffe')), rpn_head=dict( loss_bbox=dict(type='SmoothL1L...
from __future__ import annotations __version__ = "3.5.0.dev0" __MODEL_HUB_ORGANIZATION__ = "sentence-transformers" import importlib import os from sentence_transformers.backend import ( export_dynamic_quantized_onnx_model, export_optimized_onnx_model, export_static_quantized_openvino_model, ) from senten...
from __future__ import annotations __version__ = "3.5.0.dev0" __MODEL_HUB_ORGANIZATION__ = "sentence-transformers" import importlib import os from sentence_transformers.backend import ( export_dynamic_quantized_onnx_model, export_optimized_onnx_model, export_static_quantized_openvino_model, ) from senten...
import sys from os import path from setuptools import find_packages from setuptools import setup if sys.version_info < (3, 7, 0): raise OSError(f'DocArray requires Python >=3.7, but yours is {sys.version}') try: pkg_name = 'docarray' libinfo_py = path.join(pkg_name, '__init__.py') libinfo_content = o...
import sys from os import path from setuptools import find_packages from setuptools import setup if sys.version_info < (3, 7, 0): raise OSError(f'DocArray requires Python >=3.7, but yours is {sys.version}') try: pkg_name = 'docarray' libinfo_py = path.join(pkg_name, '__init__.py') libinfo_content = o...
# Copyright (c) OpenMMLab. All rights reserved. import os from typing import Callable, Optional import torch.nn as nn from torch.nn.parallel import DistributedDataParallel from mmengine.device import get_device from mmengine.dist import init_dist, is_distributed, master_only from mmengine.model import convert_sync_ba...
# Copyright (c) OpenMMLab. All rights reserved. import os from typing import Callable, Optional import torch.nn as nn from torch.nn.parallel import DistributedDataParallel from mmengine.device import get_device from mmengine.dist import init_dist, is_distributed, master_only from mmengine.model import convert_sync_ba...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Optional, Sequence, Union import torch from mmengine.registry import HOOKS from ..device import is_cuda_available, is_musa_available from .hook import Hook DATA_BATCH = Optional[Union[dict, tuple, list]] @HOOKS.register_module() class EmptyCacheHoo...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Optional, Sequence, Union import torch from mmengine.registry import HOOKS from .hook import Hook DATA_BATCH = Optional[Union[dict, tuple, list]] @HOOKS.register_module() class EmptyCacheHook(Hook): """Releases all unoccupied cached GPU memory ...
from typing import TYPE_CHECKING, Any, Dict, Optional, TypeVar import numpy as np from pydantic import parse_obj_as from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.ndarray import NdArray from docarray.typing.url.url_3d.url_3d import Url3D if TYPE_CHECKING: from docarray.doc...
from typing import TYPE_CHECKING, Any, Dict, Optional, TypeVar import numpy as np from pydantic import parse_obj_as from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.ndarray import NdArray from docarray.typing.url.url_3d.url_3d import Url3D if TYPE_CHECKING: from docarray.doc...
""" ================================ ROC Curve with Visualization API ================================ Scikit-learn defines a simple API for creating visualizations for machine learning. The key features of this API is to allow for quick plotting and visual adjustments without recalculation. In this example, we will de...
""" ================================ ROC Curve with Visualization API ================================ Scikit-learn defines a simple API for creating visualizations for machine learning. The key features of this API is to allow for quick plotting and visual adjustments without recalculation. In this example, we will de...
"""Argparser module for WorkerRuntime""" from jina.parsers.helper import KVAppendAction, add_arg_group from jina.parsers.orchestrate.runtimes.grpc_channel import ( mixin_grpc_channel_options_parser, ) from jina.parsers.orchestrate.runtimes.runtime import mixin_base_runtime_parser def mixin_worker_runtime_parser(...
"""Argparser module for WorkerRuntime""" from jina.parsers.helper import KVAppendAction, add_arg_group from jina.parsers.orchestrate.runtimes.runtime import mixin_base_runtime_parser def mixin_worker_runtime_parser(parser): """Mixing in arguments required by :class:`WorkerRuntime` into the given parser. :par...
import pathlib from typing import Any, Union from torchdata.datapipes.iter import IterDataPipe, Mapper from torchvision.prototype.datasets.utils import Dataset, EncodedImage, HttpResource, OnlineResource from torchvision.prototype.datasets.utils._internal import hint_sharding, hint_shuffling from torchvision.prototype...
import pathlib from typing import Any, Dict, List, Tuple, Union from torchdata.datapipes.iter import IterDataPipe, Mapper from torchvision.prototype.datasets.utils import Dataset, EncodedImage, HttpResource, OnlineResource from torchvision.prototype.datasets.utils._internal import hint_sharding, hint_shuffling from to...
import json import os from typing import Dict import torch from torch import Tensor, nn class LayerNorm(nn.Module): def __init__(self, dimension: int): super(LayerNorm, self).__init__() self.dimension = dimension self.norm = nn.LayerNorm(dimension) def forward(self, features: Dict[st...
import torch from torch import Tensor from torch import nn from typing import Dict import os import json class LayerNorm(nn.Module): def __init__(self, dimension: int): super(LayerNorm, self).__init__() self.dimension = dimension self.norm = nn.LayerNorm(dimension) def forward(self, f...
import os # When using jax.experimental.enable_x64 in unit test, we want to keep the # default dtype with 32 bits, aligning it with Keras's default. os.environ["JAX_DEFAULT_DTYPE_BITS"] = "32" try: # When using torch and tensorflow, torch needs to be imported first, # otherwise it will segfault upon import. T...
import os # When using jax.experimental.enable_x64 in unit test, we want to keep the # default dtype with 32 bits, aligning it with Keras's default. os.environ["JAX_DEFAULT_DTYPE_BITS"] = "32" try: # When using torch and tensorflow, torch needs to be imported first, # otherwise it will segfault upon import. T...
# Copyright (c) OpenMMLab. All rights reserved. from .det_data_sample import DetDataSample, OptSampleList, SampleList from .track_data_sample import (OptTrackSampleList, TrackDataSample, TrackSampleList) __all__ = [ 'DetDataSample', 'SampleList', 'OptSampleList', 'TrackDataSample', ...
# Copyright (c) OpenMMLab. All rights reserved. from .det_data_sample import DetDataSample, OptSampleList, SampleList __all__ = ['DetDataSample', 'SampleList', 'OptSampleList']
import PIL.Image import torch from torchvision import tv_tensors from torchvision.transforms.functional import pil_to_tensor, to_pil_image from torchvision.utils import _log_api_usage_once from ._utils import _get_kernel, _register_kernel_internal def erase( inpt: torch.Tensor, i: int, j: int, h: in...
import PIL.Image import torch from torchvision import tv_tensors from torchvision.transforms.functional import pil_to_tensor, to_pil_image from torchvision.utils import _log_api_usage_once from ._utils import _get_kernel, _register_kernel_internal def erase( inpt: torch.Tensor, i: int, j: int, h: in...
from enum import Enum from fsspec import AbstractFileSystem from pathlib import Path from typing import Any, Dict, Iterable, Optional, Protocol, runtime_checkable import json import uuid from docling.document_converter import DocumentConverter from docling_core.types import DoclingDocument as DLDocument from llama_ind...
from enum import Enum import json from pathlib import Path from typing import Any, Dict, Iterable, Protocol, runtime_checkable import uuid from docling.document_converter import DocumentConverter from docling_core.types import DoclingDocument as DLDocument from llama_index.core.readers.base import BasePydanticReader f...
import os import pytest from catboost_ranker import CatboostRanker from jina import Flow @pytest.fixture def flow(): return Flow().add( uses=CatboostRanker, uses_with={ 'query_features': ['brand', 'price'], 'match_features': ['brand', 'price'], 'relevance_label...
import os import pytest from jina import Flow from ...catboost_ranker import CatboostRanker @pytest.fixture def flow(): return Flow().add( uses=CatboostRanker, uses_with={ 'query_features': ['brand', 'price'], 'match_features': ['brand', 'price'], 'relevance_l...
import numpy as np from .tensor import Tensor Embedding = Tensor
import numpy as np Tensor = np.ndarray Embedding = Tensor
"""Utilities for working with pydantic models. :private: """ def get_pydantic_major_version() -> int: """Get the major version of Pydantic.""" try: import pydantic return int(pydantic.__version__.split(".")[0]) except ImportError: return 0 PYDANTIC_MAJOR_VERSION = get_pydantic_...
""" Utilities for working with pydantic models. :private: """ def get_pydantic_major_version() -> int: """Get the major version of Pydantic.""" try: import pydantic return int(pydantic.__version__.split(".")[0]) except ImportError: return 0 PYDANTIC_MAJOR_VERSION = get_pydantic...
from collections.abc import Iterator from typing import Iterable class tracked_str(str): origins = {} def set_origin(self, origin: str): if super().__repr__() not in self.origins: self.origins[super().__repr__()] = origin def get_origin(self): return self.origins.get(super()....
from collections.abc import Iterator from typing import Iterable class tracked_str(str): origins = {} def set_origin(self, origin: str): if super().__repr__() not in self.origins: self.origins[super().__repr__()] = origin def get_origin(self): return self.origins.get(super()....
from typing import TYPE_CHECKING, Optional, Dict if TYPE_CHECKING: from ... import DocumentArray class PostMixin: """Helper functions for posting DocumentArray to Jina Flow.""" def post( self, host: str, show_progress: bool = False, batch_size: Optional[int] = None, ...
from typing import TYPE_CHECKING, Optional, Dict if TYPE_CHECKING: from ... import DocumentArray class PostMixin: """Helper functions for posting DocumentArray to Jina Flow.""" def post( self, host: str, show_progress: bool = False, batch_size: Optional[int] = None, ...
import os import warnings from modulefinder import Module import torch # Don't re-order these, we need to load the _C extension (done when importing # .extensions) before entering _meta_registrations. from .extension import _HAS_OPS # usort:skip from torchvision import _meta_registrations, datasets, io, models, ops,...
import os import warnings from modulefinder import Module import torch # Don't re-order these, we need to load the _C extension (done when importing # .extensions) before entering _meta_registrations. from .extension import _HAS_OPS # usort:skip from torchvision import _meta_registrations, datasets, io, models, ops,...
from jina.serve.runtimes.gateway.gateway import BaseGateway from jina.serve.runtimes.servers.websocket import WebSocketServer __all__ = ['WebSocketGateway'] class WebSocketGateway(WebSocketServer, BaseGateway): """ :class:`WebSocketGateway` is a WebSocketServer that can be loaded from YAML as any other Gatew...
from jina.serve.runtimes.gateway.gateway import BaseGateway from jina.serve.runtimes.servers.websocket import WebSocketServer __all__ = ['WebSocketGateway'] class WebSocketGateway(WebSocketServer, BaseGateway): """ :class:`WebSocketGateway` is a WebSocketServer that can be loaded from YAML as any other Gatew...
from typing import List import datasets from datasets.tasks import ImageClassification from ..folder_based_builder import folder_based_builder logger = datasets.utils.logging.get_logger(__name__) class ImageFolderConfig(folder_based_builder.FolderBasedBuilderConfig): """BuilderConfig for ImageFolder.""" ...
from typing import List import datasets from datasets.tasks import ImageClassification from ..folder_based_builder import folder_based_builder logger = datasets.utils.logging.get_logger(__name__) class ImageFolderConfig(folder_based_builder.FolderBasedBuilderConfig): """BuilderConfig for ImageFolder.""" ...
# Copyright (c) OpenMMLab. All rights reserved. from .base_boxes import BaseBoxes from .bbox_overlaps import bbox_overlaps from .box_type import (autocast_box_type, convert_box_type, get_box_type, register_box, register_box_converter) from .horizontal_boxes import HorizontalBoxes from .transforms...
# Copyright (c) OpenMMLab. All rights reserved. from .base_boxes import BaseBoxes from .bbox_overlaps import bbox_overlaps from .box_type import (autocast_box_type, convert_box_type, get_box_type, register_box, register_box_converter) from .horizontal_boxes import HorizontalBoxes from .transforms...
from pydantic import BaseModel from backend.data.block import ( Block, BlockCategory, BlockOutput, BlockSchema, BlockWebhookConfig, ) from backend.data.model import SchemaField from backend.integrations.providers import ProviderName from backend.util import settings from backend.util.settings impor...
from pydantic import BaseModel from backend.data.block import ( Block, BlockCategory, BlockOutput, BlockSchema, BlockWebhookConfig, ) from backend.data.model import SchemaField from backend.integrations.providers import ProviderName from backend.util import settings from backend.util.settings impor...
from typing import ( TYPE_CHECKING, Sequence, ) import numpy as np from docarray.helper import typename if TYPE_CHECKING: # pragma: no cover from docarray.typing import ( DocumentArrayIndexType, ) class DelItemMixin: """Provide help function to enable advanced indexing in `__delitem__`...
from typing import ( TYPE_CHECKING, Sequence, ) import numpy as np from docarray.helper import typename if TYPE_CHECKING: from docarray.typing import ( DocumentArrayIndexType, ) class DelItemMixin: """Provide help function to enable advanced indexing in `__delitem__`""" def __delit...
# Copyright 2025 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
# Copyright 2024 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
import types from typing import TYPE_CHECKING from docarray.index.backends.in_memory import InMemoryDocIndex from docarray.utils._internal.misc import ( _get_path_from_docarray_root_level, import_library, ) if TYPE_CHECKING: from docarray.index.backends.elastic import ElasticDocIndex # noqa: F401 fro...
import types from typing import TYPE_CHECKING from docarray.utils._internal.misc import ( _get_path_from_docarray_root_level, import_library, ) if TYPE_CHECKING: from docarray.index.backends.elastic import ElasticDocIndex # noqa: F401 from docarray.index.backends.elasticv7 import ElasticV7DocIndex #...
# model settings preprocess_cfg = dict( mean=[123.675, 116.28, 103.53], std=[1, 1, 1], to_rgb=True) input_size = 300 model = dict( type='SingleStageDetector', preprocess_cfg=preprocess_cfg, backbone=dict( type='SSDVGG', depth=16, with_last_pool=False, ceil_mode=True, ...
# model settings input_size = 300 model = dict( type='SingleStageDetector', backbone=dict( type='SSDVGG', depth=16, with_last_pool=False, ceil_mode=True, out_indices=(3, 4), out_feature_indices=(22, 34), init_cfg=dict( type='Pretrained', checkp...
# pyright: reportAttributeAccessIssue=false # pyright: reportUnknownArgumentType=false # pyright: reportUnknownMemberType=false # pyright: reportUnknownVariableType=false from __future__ import annotations import numpy as np # intersection of `np.linalg.__all__` on numpy 1.22 and 2.2, minus `_linalg.__all__` from nu...
from numpy.linalg import * # noqa: F403 from numpy.linalg import __all__ as linalg_all import numpy as _np from ..common import _linalg from .._internal import get_xp # These functions are in both the main and linalg namespaces from ._aliases import matmul, matrix_transpose, tensordot, vecdot # noqa: F401 import num...
import base64 import hashlib from datetime import datetime, timedelta, timezone import os import jwt from cryptography.hazmat.backends import default_backend from cryptography.hazmat.primitives.serialization import ( Encoding, PublicFormat, load_pem_private_key, ) SPCS_TOKEN_PATH = "/snowflake/session/toke...
import base64 import hashlib from datetime import datetime, timedelta, timezone import jwt from cryptography.hazmat.backends import default_backend from cryptography.hazmat.primitives.serialization import ( Encoding, PublicFormat, load_pem_private_key, ) def generate_sf_jwt(sf_account: str, sf_user: str,...
import tempfile import os import time import pytest cur_dir = os.path.dirname(os.path.abspath(__file__)) compose_yml = os.path.abspath( os.path.join(cur_dir, 'unit', 'array', 'docker-compose.yml') ) @pytest.fixture(autouse=True) def tmpfile(tmpdir): tmpfile = f'docarray_test_{next(tempfile._get_candidate_na...
import tempfile import os import time import pytest from elasticsearch import Elasticsearch cur_dir = os.path.dirname(os.path.abspath(__file__)) compose_yml = os.path.abspath( os.path.join(cur_dir, 'unit', 'array', 'docker-compose.yml') ) @pytest.fixture(autouse=True) def tmpfile(tmpdir): tmpfile = f'docarr...
from types import SimpleNamespace from jina.serve.executors import BaseExecutor def test_exec_from_python(): be = BaseExecutor(metas={'name': 'hello', 'random_name': 'random_value'}) assert be.metas.name == 'hello' assert be.metas.random_name == 'random_value' def test_runtime_args(): b = BaseExecu...
from types import SimpleNamespace from jina.serve.executors import BaseExecutor def test_exec_from_python(): be = BaseExecutor(metas={'name': 'hello', 'random_name': 'random_value'}) assert be.metas.name == 'hello' assert be.metas.random_name == 'random_value' def test_runtime_args(): b = BaseExecu...
# Copyright (c) OpenMMLab. All rights reserved. import numpy as np from ..builder import PIPELINES @PIPELINES.register_module() class InstaBoost: r"""Data augmentation method in `InstaBoost: Boosting Instance Segmentation Via Probability Map Guided Copy-Pasting <https://arxiv.org/abs/1908.07801>`_. ...
import numpy as np from ..builder import PIPELINES @PIPELINES.register_module() class InstaBoost: r"""Data augmentation method in `InstaBoost: Boosting Instance Segmentation Via Probability Map Guided Copy-Pasting <https://arxiv.org/abs/1908.07801>`_. Refer to https://github.com/GothicAi/Instaboost ...
""" this test check the docstring of all of our public API. It does it by checking the `__all__` of each of our namespace. to add a new namespace you need to * import it * add it to the `SUB_MODULE_TO_CHECK` list """ import pytest from mktestdocs import check_docstring, get_codeblock_members import docarray.data imp...
""" this test check the docstring of all of our public API. It does it by checking the `__all__` of each of our namespace. to add a new namespace you need to * import it * add it to the `SUB_MODULE_TO_CHECK` list """ import pytest from mktestdocs import check_docstring, get_codeblock_members import docarray.data imp...
from typing import Any, Dict, Union import torch from torchvision import transforms as _transforms from torchvision.prototype import datapoints from torchvision.prototype.transforms import functional as F, Transform from .utils import is_simple_tensor class ConvertBoundingBoxFormat(Transform): _transformed_typ...
from typing import Any, Dict, Union import torch from torchvision import transforms as _transforms from torchvision.prototype import datapoints from torchvision.prototype.transforms import functional as F, Transform from .utils import is_simple_tensor class ConvertBoundingBoxFormat(Transform): _transformed_typ...
from .clip_text import CLIPTextEncoder
from .clip_text import CLIPTextEncoder
import sqlite3 import warnings from dataclasses import dataclass, field, asdict from tempfile import NamedTemporaryFile from typing import ( Iterable, Dict, Optional, TYPE_CHECKING, Union, List, Tuple, ) from docarray.array.storage.sqlite.helper import initialize_table from docarray.array.s...
import sqlite3 import warnings from dataclasses import dataclass, field from tempfile import NamedTemporaryFile from typing import ( Iterable, Dict, Optional, TYPE_CHECKING, Union, List, Tuple, ) from docarray.array.storage.sqlite.helper import initialize_table from docarray.array.storage.b...
import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from data...
import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.load import dataset_module_factory, import_main_c...
from ._dsp import adsr_envelope, extend_pitch, oscillator_bank, sinc_impulse_response from .functional import add_noise, barkscale_fbanks, convolve, fftconvolve __all__ = [ "add_noise", "adsr_envelope", "barkscale_fbanks", "convolve", "extend_pitch", "fftconvolve", "oscillator_bank", "s...
from ._dsp import adsr_envelope, extend_pitch, oscillator_bank from .functional import add_noise, barkscale_fbanks, convolve, fftconvolve __all__ = [ "add_noise", "adsr_envelope", "barkscale_fbanks", "convolve", "extend_pitch", "fftconvolve", "oscillator_bank", ]
import multiprocessing from typing import TYPE_CHECKING, Optional, Union from .. import Dataset, Features, config from ..formatting import query_table from ..packaged_modules.sql.sql import Sql from ..utils import logging from .abc import AbstractDatasetInputStream if TYPE_CHECKING: import sqlite3 import sq...
import multiprocessing from typing import TYPE_CHECKING, Optional, Union from .. import Dataset, Features, config from ..formatting import query_table from ..packaged_modules.sql.sql import Sql from ..utils import logging from .abc import AbstractDatasetInputStream if TYPE_CHECKING: import sqlite3 import sq...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from parameterized import parameterized from mmdet.models.roi_heads import StandardRoIHead # noqa from mmdet.registry import MODELS from mmdet.testing import demo_mm_inputs, demo_mm_proposals, get_roi_head_cfg ...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from parameterized import parameterized from mmdet.models.roi_heads import StandardRoIHead # noqa from mmdet.registry import MODELS from mmdet.testing import demo_mm_inputs, demo_mm_proposals, get_roi_head_cfg ...
import csv import gzip import logging import os from datetime import datetime from torch.utils.data import DataLoader from sentence_transformers import InputExample, LoggingHandler, SentenceTransformer, datasets, losses, models, util from sentence_transformers.evaluation import EmbeddingSimilarityEvaluator #### Just...
from torch.utils.data import DataLoader import math from sentence_transformers import models, losses, datasets from sentence_transformers import LoggingHandler, SentenceTransformer, util, InputExample from sentence_transformers.evaluation import EmbeddingSimilarityEvaluator import logging from datetime import datetime ...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp import shutil import time from unittest import TestCase from unittest.mock import Mock import torch from mmengine.structures import InstanceData from mmdet.engine.hooks import DetVisualizationHook from mmdet.structures import DetDataSample from mmd...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp import shutil import time from unittest import TestCase from unittest.mock import Mock import torch from mmengine.structures import InstanceData from mmdet.engine.hooks import DetVisualizationHook from mmdet.structures import DetDataSample from mmd...