input
stringlengths
33
5k
output
stringlengths
32
5k
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.core.utils import ConfigType, OptConfigType, OptMultiConfig from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class CornerNet(SingleStageDetector): """CornerNet. This detector is the implementat...
# Copyright (c) OpenMMLab. All rights reserved. import torch from mmdet.core import bbox2result, bbox_mapping_back from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class CornerNet(SingleStageDetector): """CornerNet. This detector is the implementation...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.api import activations from keras.api import applications from keras.api import callbacks from keras.api import config from keras.api import constraints from keras.api import datasets fro...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.api import activations from keras.api import applications from keras.api import callbacks from keras.api import config from keras.api import constraints from keras.api import datasets fro...
import multiprocessing import time import pytest from docarray import Document, DocumentArray from docarray.array.mixins.post import _parse_host from docarray.helper import random_port @pytest.mark.parametrize( 'host, expected_on, expected_host, expected_port, expected_version, expected_scheme', [ (...
import multiprocessing import time import pytest from docarray import DocumentArray, Document from docarray.helper import random_port @pytest.mark.parametrize( 'conn_config', [ (dict(protocol='grpc'), 'grpc://127.0.0.1:$port/'), (dict(protocol='grpc'), 'grpc://127.0.0.1:$port'), (dic...
import inspect import re from typing import Dict, List from huggingface_hub.utils import insecure_hashlib from .arrow import arrow from .audiofolder import audiofolder from .cache import cache # noqa F401 from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from...
import inspect import re from typing import Dict, List from huggingface_hub.utils import insecure_hashlib from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql imp...
# Copyright (c) OpenMMLab. All rights reserved. from abc import ABCMeta, abstractmethod import torch.nn.functional as F from mmcv.runner import BaseModule, force_fp32 from ...core.utils import stack_batch from ..builder import build_loss from ..utils import interpolate_as class BaseSemanticHead(BaseModule, metaclas...
# Copyright (c) OpenMMLab. All rights reserved. from abc import ABCMeta, abstractmethod import torch.nn.functional as F from mmcv.runner import BaseModule, force_fp32 from ..builder import build_loss from ..utils import interpolate_as class BaseSemanticHead(BaseModule, metaclass=ABCMeta): """Base module of Sema...
import urllib.parse from typing import ClassVar, Optional from backend.data.model import OAuth2Credentials, ProviderName from backend.integrations.oauth.base import BaseOAuthHandler from backend.util.request import Requests class TodoistOAuthHandler(BaseOAuthHandler): PROVIDER_NAME = ProviderName.TODOIST DEF...
import urllib.parse from typing import ClassVar, Optional from backend.data.model import OAuth2Credentials, ProviderName from backend.integrations.oauth.base import BaseOAuthHandler from backend.util.request import Requests class TodoistOAuthHandler(BaseOAuthHandler): PROVIDER_NAME = ProviderName.TODOIST DEF...
class MonitoringMixin: """The Monitoring Mixin for pods""" def _setup_monitoring(self): """ Wait for the monitoring server to start """ if self.args.monitoring: from prometheus_client import CollectorRegistry self.metrics_registry = CollectorRegistry() ...
class MonitoringMixin: """The Monitoring Mixin for pods""" def _setup_monitoring(self): """ Wait for the monitoring server to start """ if self.args.monitoring: from prometheus_client import CollectorRegistry self.metrics_registry = CollectorRegistry() ...
# coding: utf-8 """Find the path to LightGBM dynamic library files.""" from pathlib import Path from platform import system from typing import List __all__: List[str] = [] def find_lib_path() -> List[str]: """Find the path to LightGBM library files. Returns ------- lib_path: list of str List ...
# coding: utf-8 """Find the path to LightGBM dynamic library files.""" from pathlib import Path from platform import system from typing import List __all__: List[str] = [] def find_lib_path() -> List[str]: """Find the path to LightGBM library files. Returns ------- lib_path: list of str List ...
"""Test functionality related to length based selector.""" import pytest from langchain_core.example_selectors import ( LengthBasedExampleSelector, ) from langchain_core.prompts import PromptTemplate EXAMPLES = [ {"question": "Question: who are you?\nAnswer: foo"}, {"question": "Question: who are you?\nA...
"""Test functionality related to length based selector.""" import pytest from langchain_core.example_selectors import ( LengthBasedExampleSelector, ) from langchain_core.prompts import PromptTemplate EXAMPLES = [ {"question": "Question: who are you?\nAnswer: foo"}, {"question": "Question: who are you?\nA...
""" This is a more complex example on performing clustering on large scale dataset. This examples find in a large set of sentences local communities, i.e., groups of sentences that are highly similar. You can freely configure the threshold what is considered as similar. A high threshold will only find extremely simila...
""" This is a more complex example on performing clustering on large scale dataset. This examples find in a large set of sentences local communities, i.e., groups of sentences that are highly similar. You can freely configure the threshold what is considered as similar. A high threshold will only find extremely simila...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class RetinaNet(SingleStageDetector): """Implementation of `RetinaNet <https://arxiv.org/abs/1708.02002>`_""" def __init__(self, backbone,...
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .single_stage import SingleStageDetector @DETECTORS.register_module() class RetinaNet(SingleStageDetector): """Implementation of `RetinaNet <https://arxiv.org/abs/1708.02002>`_""" def __init__(self, backbone...
"""Feishu docs reader.""" import json import os import time from typing import List import requests from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document # Copyright (2023) Bytedance Ltd. and/or its affiliates # # Licensed under the Apache License, Version 2.0 (the "Licens...
"""Feishu docs reader.""" import json import os import time from typing import List import requests from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document # Copyright (2023) Bytedance Ltd. and/or its affiliates # # Licensed under the Apache License, Version 2.0 (the "License...
from __future__ import annotations from collections.abc import Iterable from torch import Tensor from sentence_transformers.losses.TripletLoss import TripletDistanceMetric, TripletLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseTripletLoss(TripletLoss): def __init_...
from __future__ import annotations from sentence_transformers.losses.TripletLoss import TripletDistanceMetric, TripletLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseTripletLoss(TripletLoss): def __init__( self, model: SparseEncoder, distance_metric=TripletDi...
#!/usr/bin/env python3 """Generate feature statistics for training set. Example: python global_stats.py --model-type librispeech --dataset-path /home/librispeech """ import json import logging import pathlib from argparse import ArgumentParser, RawTextHelpFormatter import torch import torchaudio from common import (...
#!/usr/bin/env python3 """Generate feature statistics for training set. Example: python global_stats.py --model-type librispeech --dataset-path /home/librispeech """ import json import logging import pathlib from argparse import ArgumentParser, RawTextHelpFormatter import torch import torchaudio from common import (...
from __future__ import annotations from collections.abc import Iterable import torch from torch import Tensor, nn from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class FlopsLoss(nn.Module): def __init__(self, model: SparseEncoder, threshold: float = None) -> None: """ ...
from __future__ import annotations from collections.abc import Iterable import torch from torch import Tensor, nn from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class FlopsLoss(nn.Module): def __init__(self, model: SparseEncoder, threshold: float = None) -> None: """ ...
from __future__ import annotations import json import logging import os import torch from torch import Tensor, nn logger = logging.getLogger(__name__) class WordWeights(nn.Module): """This model can weight word embeddings, for example, with idf-values.""" def __init__(self, vocab: list[str], word_weights:...
import json import logging import os from typing import Dict, List import torch from torch import Tensor, nn logger = logging.getLogger(__name__) class WordWeights(nn.Module): """This model can weight word embeddings, for example, with idf-values.""" def __init__(self, vocab: List[str], word_weights: Dict[...
from __future__ import annotations from collections.abc import Iterable import torch import torch.nn as nn import torch.nn.functional as F from sentence_transformers.sparse_encoder import SparseEncoder def normalized_mean_squared_error(reconstruction: torch.Tensor, original_input: torch.Tensor) -> torch.Tensor: ...
from __future__ import annotations from collections.abc import Iterable import torch import torch.nn as nn import torch.nn.functional as F from sentence_transformers.sparse_encoder import SparseEncoder def normalized_mean_squared_error(reconstruction: torch.Tensor, original_input: torch.Tensor) -> torch.Tensor: ...
""" This file contains deprecated code that can only be used with the old `model.fit`-style Sentence Transformers v2.X training. It exists for backwards compatibility with the `model.old_fit` method, but will be removed in a future version. Nowadays, with Sentence Transformers v3+, it is recommended to use the `Senten...
""" This file contains deprecated code that can only be used with the old `model.fit`-style Sentence Transformers v2.X training. It exists for backwards compatibility with the `model.old_fit` method, but will be removed in a future version. Nowadays, with Sentence Transformers v3+, it is recommended to use the `Senten...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp from typing import Optional, Sequence, Tuple import cv2 import numpy as np from mmengine.data import BaseDataElement from mmengine.hooks import Hook from mmengine.registry import HOOKS from mmengine.utils.misc import tensor2imgs # TODO: Due to in...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp from typing import Optional, Sequence, Tuple import cv2 import numpy as np from mmengine.data import BaseDataElement from mmengine.hooks import Hook from mmengine.registry import HOOKS from mmengine.utils.misc import tensor2imgs # TODO: Due to in...
from collections import ChainMap from typing import ( TYPE_CHECKING, Any, Dict, Iterable, MutableMapping, Optional, Type, TypeVar, Union, ) from docarray.array.list_advance_indexing import ListAdvancedIndexing from docarray.typing import NdArray from docarray.typing.tensor.abstract_...
from collections import ChainMap from typing import ( TYPE_CHECKING, Any, Dict, Iterable, MutableMapping, Optional, Type, TypeVar, Union, ) from docarray.array.list_advance_indexing import ListAdvancedIndexing from docarray.typing import NdArray from docarray.typing.tensor.abstract_...
_base_ = 'mask-rcnn_r50_fpn_rpn-2conv_4conv1fc_syncbn-all_lsj-100e_coco.py' # Enable automatic-mixed-precision training with AmpOptimWrapper. optim_wrapper = dict(type='AmpOptimWrapper')
_base_ = 'mask_rcnn_r50_fpn_syncbn-all_rpn-2conv_lsj_100e_coco.py' # Enable automatic-mixed-precision training with AmpOptimWrapper. optim_wrapper = dict(type='AmpOptimWrapper')
import pathlib from typing import Any, Optional, Union from torchdata.datapipes.iter import CSVDictParser, Demultiplexer, Filter, IterDataPipe, Mapper, Zipper from torchvision.prototype.datasets.utils import Dataset, EncodedImage, HttpResource, OnlineResource from torchvision.prototype.datasets.utils._internal import ...
import pathlib from typing import Any, Dict, List, Optional, Tuple, Union from torchdata.datapipes.iter import CSVDictParser, Demultiplexer, Filter, IterDataPipe, Mapper, Zipper from torchvision.prototype.datasets.utils import Dataset, EncodedImage, HttpResource, OnlineResource from torchvision.prototype.datasets.util...
from dataclasses import dataclass, field from typing import Any, Dict, Type import pytest from pydantic import Field from docarray import BaseDoc from docarray.index.abstract import BaseDocIndex from docarray.typing import NdArray pytestmark = pytest.mark.index class SimpleDoc(BaseDoc): tens: NdArray[10] = Fie...
from dataclasses import dataclass, field from typing import Any, Dict, Type import pytest from pydantic import Field from docarray import BaseDoc from docarray.index.abstract import BaseDocIndex from docarray.typing import NdArray pytestmark = pytest.mark.index class SimpleDoc(BaseDoc): tens: NdArray[10] = Fie...
import numpy as np import pytest from absl.testing import parameterized from tensorflow import data as tf_data from keras.src import layers from keras.src import testing class MelSpectrogramTest(testing.TestCase): @pytest.mark.requires_trainable_backend def test_mel_spectrogram_basics(self): self.run...
import numpy as np import pytest from absl.testing import parameterized from tensorflow import data as tf_data from keras.src import layers from keras.src import testing class MelSpectrogramTest(testing.TestCase, parameterized.TestCase): @pytest.mark.requires_trainable_backend def test_mel_spectrogram_basics...
from llama_index.llms.bedrock.base import ( Bedrock, completion_response_to_chat_response, completion_with_retry, ) from llama_index.llms.bedrock.utils import ProviderType __all__ = [ "Bedrock", "completion_with_retry", "completion_response_to_chat_response", "ProviderType", ]
from llama_index.llms.bedrock.base import ( Bedrock, completion_response_to_chat_response, completion_with_retry, ) __all__ = ["Bedrock", "completion_with_retry", "completion_response_to_chat_response"]
"""Base argparser module for Pod and Deployment runtime""" import argparse import os from jina.enums import PollingType from jina.helper import random_identity from jina.parsers.helper import _SHOW_ALL_ARGS, add_arg_group def mixin_essential_parser(parser): """Mixing in arguments required by every module into th...
"""Base argparser module for Pod and Deployment runtime""" import argparse import os from jina.enums import PollingType from jina.helper import random_identity from jina.parsers.helper import _SHOW_ALL_ARGS, add_arg_group def mixin_essential_parser(parser): """Mixing in arguments required by every module into th...
from __future__ import annotations import logging from typing import Optional, Type from langchain_core.callbacks import CallbackManagerForToolRun from pydantic import BaseModel, Field, HttpUrl from langchain_community.tools.edenai.edenai_base_tool import EdenaiTool logger = logging.getLogger(__name__) class Invo...
from __future__ import annotations import logging from typing import Optional, Type from langchain_core.callbacks import CallbackManagerForToolRun from pydantic import BaseModel, Field, HttpUrl from langchain_community.tools.edenai.edenai_base_tool import EdenaiTool logger = logging.getLogger(__name__) class Invo...
import pathlib from typing import Any, Callable, Optional, TypeVar, Union from torchvision.prototype.datasets import home from torchvision.prototype.datasets.utils import Dataset from torchvision.prototype.utils._internal import add_suggestion T = TypeVar("T") D = TypeVar("D", bound=type[Dataset]) BUILTIN_INFOS: di...
import pathlib from typing import Any, Callable, Dict, List, Optional, Type, TypeVar, Union from torchvision.prototype.datasets import home from torchvision.prototype.datasets.utils import Dataset from torchvision.prototype.utils._internal import add_suggestion T = TypeVar("T") D = TypeVar("D", bound=Type[Dataset]) ...
import importlib import os from pathlib import Path import pytest from fastapi.testclient import TestClient from ...utils import needs_py39, needs_py310 @pytest.fixture( name="client", params=[ "tutorial002", pytest.param("tutorial002_py310", marks=needs_py310), "tutorial002_an", ...
import os from pathlib import Path from fastapi.testclient import TestClient from docs_src.background_tasks.tutorial002 import app client = TestClient(app) def test(): log = Path("log.txt") if log.is_file(): os.remove(log) # pragma: no cover response = client.post("/send-notification/foo@examp...
""" Tests the correct computation of evaluation scores from BinaryClassificationEvaluator """ from __future__ import annotations import csv import gzip import os from pathlib import Path from torch.utils.data import DataLoader from sentence_transformers import ( InputExample, SentenceTransformer, evalua...
""" Tests the correct computation of evaluation scores from BinaryClassificationEvaluator """ from __future__ import annotations import csv import gzip import os from torch.utils.data import DataLoader from sentence_transformers import ( InputExample, SentenceTransformer, evaluation, losses, uti...
from abc import abstractmethod import pytest from langchain_core.documents import Document from langchain_core.retrievers import BaseRetriever from langchain_tests.base import BaseStandardTests class RetrieversIntegrationTests(BaseStandardTests): """ Base class for retrievers integration tests. """ ...
from abc import abstractmethod from typing import Type import pytest from langchain_core.documents import Document from langchain_core.retrievers import BaseRetriever from langchain_tests.base import BaseStandardTests class RetrieversIntegrationTests(BaseStandardTests): """ Base class for retrievers integra...
from typing import TYPE_CHECKING, Dict, Iterable from sentence_transformers.evaluation.SentenceEvaluator import SentenceEvaluator if TYPE_CHECKING: from sentence_transformers.SentenceTransformer import SentenceTransformer class SequentialEvaluator(SentenceEvaluator): """ This evaluator allows that multi...
from typing import TYPE_CHECKING, Dict, Iterable from sentence_transformers.evaluation.SentenceEvaluator import SentenceEvaluator if TYPE_CHECKING: from sentence_transformers.SentenceTransformer import SentenceTransformer class SequentialEvaluator(SentenceEvaluator): """ This evaluator allows that multi...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Dict, Optional from torch import Tensor from mmdet.registry import MODELS from mmdet.structures import SampleList, TrackSampleList from mmdet.utils import OptConfigType, OptMultiConfig from .base import BaseMOTModel @MODELS.register_module() class B...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Dict, Optional from torch import Tensor from mmdet.registry import MODELS from mmdet.structures import SampleList, TrackSampleList from mmdet.utils import OptConfigType, OptMultiConfig from .base import BaseMOTModel @MODELS.register_module() class B...
from typing import Any, Dict, Union import torch from torchvision import transforms as _transforms from torchvision.prototype import datapoints from torchvision.prototype.transforms import functional as F, Transform from .utils import is_simple_tensor class ConvertBoundingBoxFormat(Transform): _transformed_typ...
from typing import Any, Dict, Union import torch from torchvision.prototype import datapoints from torchvision.prototype.transforms import functional as F, Transform from .utils import is_simple_tensor class ConvertBoundingBoxFormat(Transform): _transformed_types = (datapoints.BoundingBox,) def __init__(s...
from backend.data.credit import get_user_credit_model from backend.data.execution import ( ExecutionResult, NodeExecutionEntry, RedisExecutionEventBus, create_graph_execution, get_execution_results, get_incomplete_executions, get_latest_execution, update_execution_status, update_grap...
from backend.data.credit import get_user_credit_model from backend.data.execution import ( ExecutionResult, NodeExecutionEntry, RedisExecutionEventBus, create_graph_execution, get_execution_results, get_incomplete_executions, get_latest_execution, update_execution_status, update_grap...
"""Tests for tf.distribute related functionality under tf implementation.""" import numpy as np import pytest import tensorflow as tf from tensorflow.python.eager import context from keras.src import backend from keras.src import layers from keras.src import models from keras.src import testing from keras.src.backend...
"""Tests for tf.distribute related functionality under tf implementation.""" import numpy as np import pytest import tensorflow as tf from tensorflow.python.eager import context from keras.src import backend from keras.src import layers from keras.src import models from keras.src import testing from keras.src.backend...
from pathlib import Path def find_and_replace(source: str, replacements: dict[str, str]) -> str: rtn = source # replace keys in deterministic alphabetical order finds = sorted(replacements.keys()) for find in finds: replace = replacements[find] rtn = rtn.replace(find, replace) ret...
from pathlib import Path from typing import Dict def find_and_replace(source: str, replacements: Dict[str, str]) -> str: rtn = source # replace keys in deterministic alphabetical order finds = sorted(replacements.keys()) for find in finds: replace = replacements[find] rtn = rtn.replac...
import pytest from keras.src import backend from keras.src import testing class DeviceTest(testing.TestCase): @pytest.mark.skipif(backend.backend() != "tensorflow", reason="tf only") def test_tf_device_scope(self): import tensorflow as tf if not tf.config.list_physical_devices("GPU"): ...
import pytest from keras.src import backend from keras.src import testing class DeviceTest(testing.TestCase): @pytest.mark.skipif(backend.backend() != "tensorflow", reason="tf only") def test_tf_device_scope(self): import tensorflow as tf if not tf.config.list_physical_devices("GPU"): ...
# Copyright (c) OpenMMLab. All rights reserved. import logging import random from typing import List, Optional, Tuple import numpy as np import torch from mmengine.dist import get_rank, sync_random_seed from mmengine.logging import print_log from mmengine.utils import digit_version, is_list_of from mmengine.utils.dl_...
# Copyright (c) OpenMMLab. All rights reserved. from typing import List, Optional, Tuple from mmengine.utils import is_list_of def calc_dynamic_intervals( start_interval: int, dynamic_interval_list: Optional[List[Tuple[int, int]]] = None ) -> Tuple[List[int], List[int]]: """Calculate dynamic intervals. ...
# Copyright (c) OpenMMLab. All rights reserved. from .base_video_metric import BaseVideoMetric from .cityscapes_metric import CityScapesMetric from .coco_caption_metric import COCOCaptionMetric from .coco_metric import CocoMetric from .coco_occluded_metric import CocoOccludedSeparatedMetric from .coco_panoptic_metric i...
# Copyright (c) OpenMMLab. All rights reserved. from .base_video_metric import BaseVideoMetric from .cityscapes_metric import CityScapesMetric from .coco_caption_metric import COCOCaptionMetric from .coco_metric import CocoMetric from .coco_occluded_metric import CocoOccludedSeparatedMetric from .coco_panoptic_metric i...
__version__ = '0.13.16' import os from .document import Document from .array import DocumentArray from .dataclasses import dataclass, field if 'DA_NO_RICH_HANDLER' not in os.environ: from rich.traceback import install install()
__version__ = '0.13.15' import os from .document import Document from .array import DocumentArray from .dataclasses import dataclass, field if 'DA_NO_RICH_HANDLER' not in os.environ: from rich.traceback import install install()
""" NOTE: This file must be imported like ``import torch.distributed.fsdp._traversal_utils`` and not like ``from torch.distributed.fsdp._traversal_utils import ...`` to avoid circular imports. For brevity, we may import the file as ``traversal_utils``. """ import collections import torch.nn as nn from torch.distribut...
""" NOTE: This file must be imported like ``import torch.distributed.fsdp._traversal_utils`` and not like ``from torch.distirbuted.fsdp._traversal_utils import ...`` to avoid circular imports. For brevity, we may import the file as ``traversal_utils``. """ import collections import torch.nn as nn from torch.distribut...
from typing import Dict, Iterable import torch.nn.functional as F from torch import Tensor, nn from sentence_transformers.SentenceTransformer import SentenceTransformer from .ContrastiveLoss import SiameseDistanceMetric class OnlineContrastiveLoss(nn.Module): def __init__( self, model: SentenceTransfor...
from typing import Iterable, Dict import torch.nn.functional as F from torch import nn, Tensor from .ContrastiveLoss import SiameseDistanceMetric from sentence_transformers.SentenceTransformer import SentenceTransformer class OnlineContrastiveLoss(nn.Module): def __init__( self, model: SentenceTransformer...
# Copyright (c) OpenMMLab. All rights reserved. import warnings from abc import ABCMeta, abstractmethod from typing import Any, List, Optional, Sequence, Union from mmengine.dist import (broadcast_object_list, collect_results, is_main_process) class BaseMetric(metaclass=ABCMeta): """Ba...
# Copyright (c) OpenMMLab. All rights reserved. import warnings from abc import ABCMeta, abstractmethod from typing import Any, List, Optional, Sequence, Tuple, Union from mmengine.dist import (broadcast_object_list, collect_results, is_main_process) class BaseMetric(metaclass=ABCMeta): ...
import sys from dataclasses import dataclass from typing import TYPE_CHECKING, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast if TYPE_CHECKING: import sqlite3 i...
import sys from dataclasses import dataclass from typing import TYPE_CHECKING, Dict, List, Optional, Tuple, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast if TYPE_CHECKING: im...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import ( BaseSQLDatabaseTool, InfoSQLDatabaseTool, ListSQLDatabaseTool, QuerySQLCheckerTool, QuerySQLDataBaseTool, ) # Create a way to dyna...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.tools import ( BaseSQLDatabaseTool, InfoSQLDatabaseTool, ListSQLDatabaseTool, QuerySQLCheckerTool, QuerySQLDataBaseTool, ) # Create a way to dyna...
import os from typing import Dict DEPLOYMENT_FILES = [ 'deployment-executor', 'deployment-gateway', 'deployment-uses-before', 'deployment-uses-after', 'deployment-uses-before-after', ] cur_dir = os.path.dirname(__file__) DEFAULT_RESOURCE_DIR = os.path.join( cur_dir, '..', '..', '..', '..', 're...
import os from typing import Dict DEPLOYMENT_FILES = [ 'deployment', 'deployment-uses-before', 'deployment-uses-after', 'deployment-uses-before-after', ] cur_dir = os.path.dirname(__file__) DEFAULT_RESOURCE_DIR = os.path.join( cur_dir, '..', '..', '..', '..', 'resources', 'k8s', 'template' ) def...
from __future__ import annotations from pathlib import Path from unittest.mock import Mock, PropertyMock import pytest import torch from sentence_transformers import SentenceTransformer from sentence_transformers.evaluation import InformationRetrievalEvaluator from sentence_transformers.util import cos_sim @pytest...
from __future__ import annotations from pathlib import Path from unittest.mock import Mock, PropertyMock import pytest import torch from sentence_transformers import SentenceTransformer from sentence_transformers.evaluation import InformationRetrievalEvaluator from sentence_transformers.util import cos_sim @pytest...
from functools import partial from huggingface_hub import hf_hub_url hf_dataset_url = partial(hf_hub_url, repo_type="dataset")
import time from functools import partial from huggingface_hub import HfApi, hf_hub_url from huggingface_hub.hf_api import RepoFile from packaging import version from requests import ConnectionError, HTTPError from .. import config from . import logging logger = logging.get_logger(__name__) # Retry `preupload_lfs_...
import warnings from typing import Any, Dict, Union import numpy as np import PIL.Image import torch from torchvision.transforms import functional as _F from torchvision.transforms.v2 import Transform class ToTensor(Transform): """[BETA] Convert a PIL Image or ndarray to tensor and scale the values accordingly....
import warnings from typing import Any, Dict, Union import numpy as np import PIL.Image import torch from torchvision.transforms import functional as _F from torchvision.transforms.v2 import Transform class ToTensor(Transform): """[BETA] Convert a PIL Image or ndarray to tensor and scale the values accordingly....
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import os import subprocess from pathlib import Path import pytest TEST_DIR = os.path.dirname(os.path.abspath(__file__)) @pytest.fixture(scope='session') def docker_image_name() -> str: return Path(__file__).p...
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import subprocess from pathlib import Path import pytest @pytest.fixture(scope='session') def docker_image_name() -> str: return Path(__file__).parents[1].stem.lower() @pytest.fixture(scope='session') def bui...
import torch from torchvision import tv_tensors from torchvision.utils import _log_api_usage_once from ._utils import _get_kernel, _register_kernel_internal def uniform_temporal_subsample(inpt: torch.Tensor, num_samples: int) -> torch.Tensor: """See :class:`~torchvision.transforms.v2.UniformTemporalSubsample` ...
import torch from torchvision import tv_tensors from torchvision.utils import _log_api_usage_once from ._utils import _get_kernel, _register_kernel_internal def uniform_temporal_subsample(inpt: torch.Tensor, num_samples: int) -> torch.Tensor: """[BETA] See :class:`~torchvision.transforms.v2.UniformTemporalSubs...
# Copyright (c) OpenMMLab. All rights reserved. import copy import time from contextlib import contextmanager from typing import Generator, Optional from mmengine.utils.manager import ManagerMixin, _accquire_lock, _release_lock class DefaultScope(ManagerMixin): """Scope of current task used to reset the current ...
# Copyright (c) OpenMMLab. All rights reserved. import copy import time from contextlib import contextmanager from typing import Generator, Optional from mmengine.utils.manager import ManagerMixin, _accquire_lock, _release_lock class DefaultScope(ManagerMixin): """Scope of current task used to reset the current ...
_base_ = '../mask_rcnn/mask-rcnn_r101_fpn_1x_coco.py' model = dict( backbone=dict( dcn=dict(type='DCN', deform_groups=1, fallback_on_stride=False), stage_with_dcn=(False, True, True, True)))
_base_ = '../mask_rcnn/mask_rcnn_r101_fpn_1x_coco.py' model = dict( backbone=dict( dcn=dict(type='DCN', deform_groups=1, fallback_on_stride=False), stage_with_dcn=(False, True, True, True)))
"""PDF Table reader.""" from pathlib import Path from typing import Any, Dict, List, Optional import pandas as pd from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class PDFTableReader(BaseReader): """ PDF Table Reader. Reads table from PDF. Args: ...
"""PDF Table reader.""" from pathlib import Path from typing import Any, Dict, List, Optional import pandas as pd from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class PDFTableReader(BaseReader): """ PDF Table Reader. Reads table from PDF. Args: ...
from __future__ import annotations from .CSRLoss import CSRLoss, CSRReconstructionLoss from .RegularizerLoss import FlopsLoss, L0FlopsLoss from .SparseAnglELoss import SparseAnglELoss from .SparseCachedGISTEmbedLoss import SparseCachedGISTEmbedLoss from .SparseCachedMultipleNegativesRankingLoss import SparseCachedMult...
from __future__ import annotations from .CSRLoss import CSRLoss, CSRReconstructionLoss from .RegularizerLoss import FlopsLoss, IDFFlopsLoss from .SparseAnglELoss import SparseAnglELoss from .SparseCachedGISTEmbedLoss import SparseCachedGISTEmbedLoss from .SparseCachedMultipleNegativesRankingLoss import SparseCachedMul...
""" This script translates the queries in the MS MARCO dataset to the defined target languages. For machine translation, we use EasyNMT: https://github.com/UKPLab/EasyNMT You can install it via: pip install easynmt Usage: python translate_queries [target_language] """ import os from sentence_transformers import Logg...
""" This script translates the queries in the MS MARCO dataset to the defined target languages. For machine translation, we use EasyNMT: https://github.com/UKPLab/EasyNMT You can install it via: pip install easynmt Usage: python translate_queries [target_language] """ import os from sentence_transformers import Loggi...
from langchain_core.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForChainGroup, AsyncCallbackManagerForChainRun, AsyncCallbackManagerForLLMRun, AsyncCallbackManagerForRetrieverRun, AsyncCallbackManagerForToolRun, AsyncParentRunManager, AsyncRunManager, BaseRun...
from langchain_core.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForChainGroup, AsyncCallbackManagerForChainRun, AsyncCallbackManagerForLLMRun, AsyncCallbackManagerForRetrieverRun, AsyncCallbackManagerForToolRun, AsyncParentRunManager, AsyncRunManager, BaseRun...
import os from jina import Flow, Document, DocumentArray from ...tfidf_text_executor import TFIDFTextEncoder # is implicitly required cur_dir = os.path.dirname(os.path.abspath(__file__)) def test_flow_generates_embedding(): doc = DocumentArray([Document(text='Han likes eating pizza')]) with Flow.load_conf...
import os from jina import Flow, Document, DocumentArray from jinahub.encoder.tfidf_text_executor import TFIDFTextEncoder # is implicitly required cur_dir = os.path.dirname(os.path.abspath(__file__)) def test_flow_generates_embedding(): doc = DocumentArray([Document(text='Han likes eating pizza')]) with F...
from inspect import signature from typing import ( Any, Awaitable, Callable, List, Optional, Tuple, Type, Union, cast, get_origin, get_args, ) import typing from llama_index.core.bridge.pydantic import BaseModel, FieldInfo, create_model def create_schema_from_function( ...
from inspect import signature from typing import ( Any, Awaitable, Callable, List, Optional, Tuple, Type, Union, cast, get_origin, get_args, ) import typing from llama_index.core.bridge.pydantic import BaseModel, FieldInfo, create_model def create_schema_from_function( ...
# Copyright 2021 The HuggingFace Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to...
# Copyright 2021 The HuggingFace Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to...
from .torch_encoder import ImageTorchEncoder
from .torch_encoder import ImageTorchEncoder
import json import os from typing import Dict from torch import Tensor, nn class Dropout(nn.Module): """Dropout layer. Args: dropout: Sets a dropout value for dense layer. """ def __init__(self, dropout: float = 0.2): super(Dropout, self).__init__() self.dropout = dropout ...
from torch import Tensor from torch import nn from typing import Dict import os import json class Dropout(nn.Module): """Dropout layer. Args: dropout: Sets a dropout value for dense layer. """ def __init__(self, dropout: float = 0.2): super(Dropout, self).__init__() self.drop...
import json import pytest from langchain_core.agents import AgentFinish from langchain_core.exceptions import OutputParserException from langchain_core.messages import AIMessage, SystemMessage from langchain.agents.openai_functions_multi_agent.base import ( _FunctionsAgentAction, _parse_ai_message, ) # Test...
import json import pytest from langchain_core.agents import AgentFinish from langchain_core.exceptions import OutputParserException from langchain_core.messages import AIMessage, SystemMessage from langchain.agents.openai_functions_multi_agent.base import ( _FunctionsAgentAction, _parse_ai_message, ) # Test...
import datetime import prisma.fields import prisma.models import pytest import backend.server.v2.library.model as library_model from backend.util import json @pytest.mark.asyncio async def test_agent_preset_from_db(): # Create mock DB agent db_agent = prisma.models.AgentPreset( id="test-agent-123", ...
import datetime import prisma.fields import prisma.models import backend.server.v2.library.model as library_model def test_agent_preset_from_db(): # Create mock DB agent db_agent = prisma.models.AgentPreset( id="test-agent-123", createdAt=datetime.datetime.now(), updatedAt=datetime.d...
"""Standard LangChain interface tests""" from pathlib import Path from typing import Literal, cast from langchain_core.language_models import BaseChatModel from langchain_core.messages import AIMessage from langchain_tests.integration_tests import ChatModelIntegrationTests from langchain_anthropic import ChatAnthrop...
"""Standard LangChain interface tests""" from pathlib import Path from typing import Literal, cast from langchain_core.language_models import BaseChatModel from langchain_core.messages import AIMessage from langchain_tests.integration_tests import ChatModelIntegrationTests from langchain_anthropic import ChatAnthrop...
from abc import abstractmethod from typing import ( TYPE_CHECKING, TypeVar, Sequence, List, Union, Optional, Dict, ) from qdrant_client.http.models.models import Distance from docarray import Document, DocumentArray from docarray.math import ndarray from docarray.score import NamedScore if...
from abc import abstractmethod from typing import ( TYPE_CHECKING, TypeVar, Sequence, List, Union, Optional, Dict, ) from qdrant_client.http.models.models import Distance from docarray import Document, DocumentArray from docarray.math import ndarray from docarray.score import NamedScore if...
import csv import gzip import os from . import InputExample class STSDataReader: """Reads in the STS dataset. Each line contains two sentences (s1_col_idx, s2_col_idx) and one label (score_col_idx) Default values expects a tab separated file with the first & second column the sentence pair and third column ...
from . import InputExample import csv import gzip import os class STSDataReader: """ Reads in the STS dataset. Each line contains two sentences (s1_col_idx, s2_col_idx) and one label (score_col_idx) Default values expects a tab separated file with the first & second column the sentence pair and third col...
import os.path from pathlib import Path from typing import Any, Callable, Optional, Union import numpy as np from PIL import Image from .utils import check_integrity, download_url from .vision import VisionDataset class SEMEION(VisionDataset): r"""`SEMEION <http://archive.ics.uci.edu/ml/datasets/semeion+handwri...
import os.path from pathlib import Path from typing import Any, Callable, Optional, Tuple, Union import numpy as np from PIL import Image from .utils import check_integrity, download_url from .vision import VisionDataset class SEMEION(VisionDataset): r"""`SEMEION <http://archive.ics.uci.edu/ml/datasets/semeion+...
import os import pytest from langchain_openai import OpenAI os.environ["OPENAI_API_KEY"] = "foo" def test_openai_model_param() -> None: llm = OpenAI(model="foo") assert llm.model_name == "foo" llm = OpenAI(model_name="foo") # type: ignore[call-arg] assert llm.model_name == "foo" # Test standa...
import os from typing import List import pytest from langchain_openai import OpenAI os.environ["OPENAI_API_KEY"] = "foo" def test_openai_model_param() -> None: llm = OpenAI(model="foo") assert llm.model_name == "foo" llm = OpenAI(model_name="foo") # type: ignore[call-arg] assert llm.model_name == ...
# Copyright (c) OpenMMLab. All rights reserved. from .bbox_overlaps import bbox_overlaps from .cityscapes_utils import evaluateImgLists from .class_names import (cityscapes_classes, coco_classes, coco_panoptic_classes, dataset_aliases, get_classes, imagenet_det_classe...
# Copyright (c) OpenMMLab. All rights reserved. from .bbox_overlaps import bbox_overlaps from .class_names import (cityscapes_classes, coco_classes, coco_panoptic_classes, dataset_aliases, get_classes, imagenet_det_classes, imagenet_vid_classes, ...
import os import yaml from jina.serve.runtimes.gateway.gateway import BaseGateway, Gateway from jina.jaml import JAML class MyDummyGateway(Gateway): async def setup_server(self): self.server = 'dummy server' async def run_server(self): self.logger.info(self.server) async def shutdown(s...
import os import yaml from jina.serve.runtimes.gateway.gateway import BaseGateway, Gateway from jina.jaml import JAML class MyDummyGateway(Gateway): async def setup_server(self): self.server = 'dummy server' async def run_server(self): self.logger.info(self.server) async def shutdown(s...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Optional import torch.nn as nn import torch.nn.functional as F from torch import Tensor from mmdet.registry import MODELS from .utils import weighted_loss @weighted_loss def mse_loss(pred: Tensor, target: Tensor) -> Tensor: """A Wrapper of MSE l...
# Copyright (c) OpenMMLab. All rights reserved. import torch.nn as nn import torch.nn.functional as F from mmdet.registry import MODELS from .utils import weighted_loss @weighted_loss def mse_loss(pred, target): """Wrapper of mse loss.""" return F.mse_loss(pred, target, reduction='none') @MODELS.register_m...
_base_ = './retinanet_r50-caffe_fpn_ms-3x_coco.py' # learning policy model = dict( pretrained='open-mmlab://detectron2/resnet101_caffe', backbone=dict(depth=101))
_base_ = './retinanet_r50_caffe_fpn_mstrain_3x_coco.py' # learning policy model = dict( pretrained='open-mmlab://detectron2/resnet101_caffe', backbone=dict(depth=101))
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from pathlib import Path import pytest from jina import Executor from ...minranker import MinRanker def test_config(): ex = Executor.load_config(str(Path(__file__).parents[2] / 'config.yml')) assert ex.met...
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import pytest from ...minranker import MinRanker @pytest.mark.parametrize('default_traversal_paths', [['r'], ['c']]) def test_ranker(documents_chunk, documents_chunk_chunk, default_traversal_paths): ranker = Min...
import os import urllib import pytest from pydantic import parse_obj_as, schema_json_of from docarray.base_doc.io.json import orjson_dumps from docarray.typing import TextUrl from docarray.typing.url.mimetypes import ( OBJ_MIMETYPE, AUDIO_MIMETYPE, VIDEO_MIMETYPE, IMAGE_MIMETYPE, TEXT_MIMETYPE, ) ...
import os import urllib import pytest from pydantic import parse_obj_as, schema_json_of from docarray.base_doc.io.json import orjson_dumps from docarray.typing import TextUrl from tests import TOYDATA_DIR REMOTE_TEXT_FILE = 'https://de.wikipedia.org/wiki/Brixen' CUR_DIR = os.path.dirname(os.path.abspath(__file__)) L...
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import os import subprocess import numpy as np import pytest from jina import Document, DocumentArray, Flow from jina.executors.metas import get_default_metas from ...faiss_searcher import FaissSearcher def _get_d...
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import os import subprocess import numpy as np import pytest from jina import Document, DocumentArray, Flow from jina.executors.metas import get_default_metas from ...faiss_searcher import FaissSearcher def _get_d...
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .single_stage import SingleStageDetector @DETECTORS.register_module() class YOLOX(SingleStageDetector): r"""Implementation of `YOLOX: Exceeding YOLO Series in 2021 <https://arxiv.org/abs/2107.08430>`_""" def __init__(sel...
from ..builder import DETECTORS from .single_stage import SingleStageDetector @DETECTORS.register_module() class YOLOX(SingleStageDetector): r"""Implementation of `YOLOX: Exceeding YOLO Series in 2021 <https://arxiv.org/abs/2107.08430>`_""" def __init__(self, backbone, n...
""" This tool allows agents to interact with the pygithub library and operate on a GitHub repository. To use this tool, you must first set as environment variables: GITHUB_API_TOKEN GITHUB_REPOSITORY -> format: {owner}/{repo} """ from typing import Any, Optional, Type from langchain_core.callbacks import Ca...
""" This tool allows agents to interact with the pygithub library and operate on a GitHub repository. To use this tool, you must first set as environment variables: GITHUB_API_TOKEN GITHUB_REPOSITORY -> format: {owner}/{repo} """ from typing import Any, Optional, Type from langchain_core.callbacks import Ca...
import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def deprecate(*args, take_from: Optional[Union[Dict, Any]] = None, standard_warn=True, stacklevel=2): from .. import __version__ deprecated_kwargs = take_from values = () if not isinstance(args...
import inspect import warnings from typing import Any, Dict, Optional, Union from packaging import version def deprecate(*args, take_from: Optional[Union[Dict, Any]] = None, standard_warn=True, stacklevel=2): from .. import __version__ deprecated_kwargs = take_from values = () if not isinstance(args...
_base_ = [ '../_base_/models/fast-rcnn_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] dataset_type = 'CocoDataset' data_root = 'data/coco/' img_norm_cfg = dict( mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rg...
_base_ = [ '../_base_/models/fast_rcnn_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] dataset_type = 'CocoDataset' data_root = 'data/coco/' img_norm_cfg = dict( mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], to_rg...
import numpy as np from docarray.document import AnyDocument, BaseDocument from docarray.typing import NdArray def test_any_doc(): class InnerDocument(BaseDocument): text: str tensor: NdArray class CustomDoc(BaseDocument): inner: InnerDocument text: str doc = CustomDoc( ...
import numpy as np from docarray.document import AnyDocument, BaseDocument from docarray.typing import Tensor def test_any_doc(): class InnerDocument(BaseDocument): text: str tensor: Tensor class CustomDoc(BaseDocument): inner: InnerDocument text: str doc = CustomDoc( ...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.callbacks.mlflow_callback import ( MlflowCallbackHandler, MlflowLogger, analyze_text, construct_html_from_prompt_and_generation, ) # Create a way to dyna...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.callbacks.mlflow_callback import ( MlflowCallbackHandler, MlflowLogger, analyze_text, construct_html_from_prompt_and_generation, ) # Create a way to dyna...
# coding=utf-8 # Copyright 2025 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
# coding=utf-8 # Copyright 2024 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
""" This file runs Masked Language Model. You provide a training file. Each line is interpreted as a sentence / paragraph. Optionally, you can also provide a dev file. The fine-tuned model is stored in the output/model_name folder. Usage: python train_mlm.py model_name data/train_sentences.txt [data/dev_sentences.txt...
""" This file runs Masked Language Model. You provide a training file. Each line is interpreted as a sentence / paragraph. Optionally, you can also provide a dev file. The fine-tuned model is stored in the output/model_name folder. Usage: python train_mlm.py model_name data/train_sentences.txt [data/dev_sentences.txt...
# Copyright (c) OpenMMLab. All rights reserved. from .checkpoint_hook import CheckpointHook from .ema_hook import EMAHook from .empty_cache_hook import EmptyCacheHook from .hook import Hook from .iter_timer_hook import IterTimerHook from .logger_hook import LoggerHook from .naive_visualization_hook import NaiveVisualiz...
# Copyright (c) OpenMMLab. All rights reserved. from .checkpoint_hook import CheckpointHook from .ema_hook import EMAHook from .empty_cache_hook import EmptyCacheHook from .hook import Hook from .iter_timer_hook import IterTimerHook from .logger_hook import LoggerHook from .naive_visualization_hook import NaiveVisualiz...
""" This script contains an example how to perform semantic search with Elasticsearch. You need Elasticsearch up and running locally: https://www.elastic.co/guide/en/elasticsearch/reference/current/run-elasticsearch-locally.html Further, you need the Python Elasticsearch Client installed: https://elasticsearch-py.rea...
""" This script contains an example how to perform semantic search with Elasticsearch. You need Elasticsearch up and running locally: https://www.elastic.co/guide/en/elasticsearch/reference/current/run-elasticsearch-locally.html Further, you need the Python Elasticsearch Client installed: https://elasticsearch-py.rea...
import csv import os from pathlib import Path from typing import Union import torchaudio from torch.utils.data import Dataset class FluentSpeechCommands(Dataset): """Create *Fluent Speech Commands* :cite:`fluent` Dataset Args: root (str of Path): Path to the directory where the dataset is found. ...
import csv import os from pathlib import Path from typing import Union import torchaudio from torch.utils.data import Dataset class FluentSpeechCommands(Dataset): """Create *Fluent Speech Commands* [:footcite:`fluent`] Dataset Args: root (str of Path): Path to the directory where the dataset is foun...
# Copyright 2021 The HuggingFace Datasets Authors and the current dataset script contributor. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2....
# Copyright 2021 The HuggingFace Datasets Authors and the current dataset script contributor. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2....
# NOTE: # The entire `torchaudio.backend` module is deprecated. # New things should be added to `torchaudio._backend`. # Only things related to backward compatibility should be placed here. from . import common, no_backend, soundfile_backend, sox_io_backend # noqa __all__ = []
# NOTE: # The entire `torchaudio.backend` module is deprecated. # New things should be added to `torchaudio._backend`. # Only things related to backward compatibility should be placed here. def __getattr__(name: str): if name == "common": from . import _common return _common if name in ["no_...
""" ================================== Getting started with transforms v2 ================================== .. note:: Try on `collab <https://colab.research.google.com/github/pytorch/vision/blob/gh-pages/main/_generated_ipynb_notebooks/plot_transforms_v2.ipynb>`_ or :ref:`go to the end <sphx_glr_download_auto...
""" ================================== Getting started with transforms v2 ================================== Most computer vision tasks are not supported out of the box by ``torchvision.transforms`` v1, since it only supports images. ``torchvision.transforms.v2`` enables jointly transforming images, videos, bounding b...
# Licensed to the LF AI & Data foundation under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the "License"); # you may not use this fil...
import os import time import pytest import requests import weaviate HOST = "http://localhost:8080" cur_dir = os.path.dirname(os.path.abspath(__file__)) weaviate_yml = os.path.abspath(os.path.join(cur_dir, 'docker-compose.yml')) @pytest.fixture(scope='session', autouse=True) def start_storage(): os.system(f"do...
import os # DO NOT EDIT. Generated by api_gen.sh from keras.api import DTypePolicy from keras.api import FloatDTypePolicy from keras.api import Function from keras.api import Initializer from keras.api import Input from keras.api import InputSpec from keras.api import KerasTensor from keras.api import Layer from keras...
import os # DO NOT EDIT. Generated by api_gen.sh from keras.api import DTypePolicy from keras.api import FloatDTypePolicy from keras.api import Function from keras.api import Initializer from keras.api import Input from keras.api import InputSpec from keras.api import KerasTensor from keras.api import Layer from keras...
AUDIO_FILE_FORMATS = [ '3g2', '3ga', '3gp', 'aac', 'ac3', 'act', 'aiff', 'amr', 'ape', 'au', 'awb', 'dct', 'dsf', 'dvf', 'flac', 'gsm', 'iklax', 'ivs', 'm4a', 'm4b', 'm4p', 'mmf', 'mp2', 'mp3', 'mpc', 'msv', 'nsf...
AUDIO_FILE_FORMATS = [ '3g2', '3ga', '3gp', 'aac', 'ac3', 'act', 'aiff', 'amr', 'ape', 'au', 'awb', 'dct', 'dsf', 'dvf', 'flac', 'gsm', 'iklax', 'ivs', 'm4a', 'm4b', 'm4p', 'mmf', 'mp2', 'mp3', 'mpc', 'msv', 'nsf...
# Licensed to the LF AI & Data foundation under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the "License"); # you may not use this fil...
import os from functools import lru_cache from pathlib import Path @lru_cache(maxsize=None) def _get_cache_path() -> Path: """ Get the path to the cache directory. :return: The path to the cache directory. """ cache_path = Path.home() / '.cache' / 'docarray' if "DOCARRAY_CACHE" in os.environ:...
from typing import Union from torch import nn import transformers import torch from PIL import Image class CLIPModel(nn.Module): def __init__(self, model_name: str = "openai/clip-vit-base-patch32", processor_name=None): super(CLIPModel, self).__init__() if processor_name is None: proc...
from typing import Union from torch import nn import transformers import torch from PIL import Image class CLIPModel(nn.Module): def __init__(self, model_name: str = "openai/clip-vit-base-patch32", processor_name=None): super(CLIPModel, self).__init__() if processor_name is None: proc...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '2.17.0' short_version = __version__ def parse_version_info(version_str): version_info = [] for x in version_str.split('.'): if x.isdigit(): version_info.append(int(x)) elif x.find('rc') != -1: patch_version...
# Copyright (c) OpenMMLab. All rights reserved. __version__ = '2.16.0' short_version = __version__ def parse_version_info(version_str): version_info = [] for x in version_str.split('.'): if x.isdigit(): version_info.append(int(x)) elif x.find('rc') != -1: patch_version...
_base_ = '../dcn/cascade-mask-rcnn_x101-32x4d-dconv-c3-c5_fpn_1x_coco.py' model = dict( backbone=dict( norm_cfg=dict(type='SyncBN', requires_grad=True), norm_eval=False))
_base_ = '../dcn/cascade_mask_rcnn_x101_32x4d_fpn_dconv_c3-c5_1x_coco.py' model = dict( backbone=dict( norm_cfg=dict(type='SyncBN', requires_grad=True), norm_eval=False))
import numpy as np import pytest import torch from pydantic import parse_obj_as from docarray import BaseDocument from docarray.documents import ImageDoc from docarray.utils.misc import is_tf_available tf_available = is_tf_available() if tf_available: import tensorflow as tf import tensorflow._api.v2.experime...
import numpy as np import pytest import torch from pydantic import parse_obj_as from docarray import BaseDocument from docarray.documents import ImageDoc from docarray.utils.misc import is_tf_available tf_available = is_tf_available() if tf_available: import tensorflow as tf import tensorflow._api.v2.experime...
from typing import List import numpy as np import pytest from fastapi import FastAPI from httpx import AsyncClient from docarray import BaseDoc, DocArray from docarray.base_doc import DocArrayResponse from docarray.documents import ImageDoc, TextDoc from docarray.typing import NdArray @pytest.mark.asyncio async def...
import numpy as np import pytest from fastapi import FastAPI from httpx import AsyncClient from docarray import BaseDoc from docarray.base_doc import DocResponse from docarray.documents import ImageDoc, TextDoc from docarray.typing import NdArray @pytest.mark.asyncio async def test_fast_api(): class Mmdoc(BaseDo...