input
stringlengths
33
5k
output
stringlengths
32
5k
from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.audio.abstract_audio_tensor import AbstractAudioTensor from docarray.typing.tensor.ndarray import NdArray @_register_proto(proto_type_name='audio_ndarray') class AudioNdArray(AbstractAudioTensor, NdArray): """ Subclass of N...
from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.audio.abstract_audio_tensor import AbstractAudioTensor from docarray.typing.tensor.ndarray import NdArray @_register_proto(proto_type_name='audio_ndarray') class AudioNdArray(AbstractAudioTensor, NdArray): """ Subclass of N...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.callbacks.tracers.comet import ( CometTracer, import_comet_llm_api, ) # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising d...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.callbacks.tracers.comet import ( CometTracer, import_comet_llm_api, ) # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising d...
from typing import Optional, List, Dict, Any, TYPE_CHECKING, Union from pydantic import BaseModel, validator from ..math.ndarray import to_list if TYPE_CHECKING: from ..typing import ArrayType # this order must be preserved: https://pydantic-docs.helpmanual.io/usage/types/#unions _ProtoValueType = Optional[Unio...
from typing import Optional, List, Dict, Any, TYPE_CHECKING, Union from pydantic import BaseModel, validator from ..math.ndarray import to_list if TYPE_CHECKING: from ..typing import ArrayType # this order must be preserved: https://pydantic-docs.helpmanual.io/usage/types/#unions _ProtoValueType = Optional[Unio...
""" ========================= Tensor transforms and JIT ========================= This example illustrates various features that are now supported by the :ref:`image transformations <transforms>` on Tensor images. In particular, we show how image transforms can be performed on GPU, and how one can also script them usi...
""" ========================= Tensor transforms and JIT ========================= This example illustrates various features that are now supported by the :ref:`image transformations <transforms>` on Tensor images. In particular, we show how image transforms can be performed on GPU, and how one can also script them usi...
from typing import Annotated, Optional import typer from langchain_cli._version import __version__ from langchain_cli.namespaces import app as app_namespace from langchain_cli.namespaces import integration as integration_namespace from langchain_cli.namespaces import template as template_namespace from langchain_cli....
from typing import Annotated, Optional import typer from langchain_cli._version import __version__ from langchain_cli.namespaces import app as app_namespace from langchain_cli.namespaces import integration as integration_namespace from langchain_cli.namespaces import template as template_namespace from langchain_cli....
# Copyright (c) OpenMMLab. All rights reserved. import torch from mmdet.structures.bbox import BaseBoxes def anchor_inside_flags(flat_anchors, valid_flags, img_shape, allowed_border=0): """Check whether the anchors are inside the border. ...
# Copyright (c) OpenMMLab. All rights reserved. import torch def anchor_inside_flags(flat_anchors, valid_flags, img_shape, allowed_border=0): """Check whether the anchors are inside the border. Args: flat_anchors (torch.Tensor): ...
# Owner(s): ["oncall: distributed"] from unittest import mock import torch.distributed as c10d from torch.distributed.collective_utils import all_gather, broadcast from torch.testing._internal.common_distributed import MultiProcessTestCase from torch.testing._internal.common_utils import run_tests class TestCollect...
# Owner(s): ["oncall: distributed"] from unittest import mock import torch.distributed as c10d from torch.distributed.collective_utils import all_gather, broadcast from torch.testing._internal.common_distributed import MultiProcessTestCase from torch.testing._internal.common_utils import run_tests class TestCollect...
from typing import Any, Optional, Type, TypeVar, Union from docarray.base_doc import BaseDoc from docarray.typing import TextUrl from docarray.typing.tensor.embedding import AnyEmbedding T = TypeVar('T', bound='TextDoc') class TextDoc(BaseDoc): """ Document for handling text. It can contain: - a [...
from typing import Any, Optional, Type, TypeVar, Union from docarray.base_doc import BaseDoc from docarray.typing import TextUrl from docarray.typing.tensor.embedding import AnyEmbedding T = TypeVar('T', bound='TextDoc') class TextDoc(BaseDoc): """ Document for handling text. It can contain: - a [...
_base_ = [ '../_base_/models/mask-rcnn_r50_fpn.py', '../_base_/datasets/deepfashion.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( roi_head=dict( bbox_head=dict(num_classes=15), mask_head=dict(num_classes=15))) # runtime settings max_epochs = 15 train_c...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/deepfashion.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( roi_head=dict( bbox_head=dict(num_classes=15), mask_head=dict(num_classes=15))) # runtime settings max_epochs = 15 train_c...
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .single_stage import SingleStageDetector @DETECTORS.register_module() class RepPointsDetector(SingleStageDetector): """RepPoints: Point Set Representation for Object Detection. This detector is the implementation of: ...
from ..builder import DETECTORS from .single_stage import SingleStageDetector @DETECTORS.register_module() class RepPointsDetector(SingleStageDetector): """RepPoints: Point Set Representation for Object Detection. This detector is the implementation of: - RepPoints detector (https://arxiv.org/pdf...
import math import os import pytest import torch import torchvision from torchvision import _HAS_GPU_VIDEO_DECODER from torchvision.io import VideoReader try: import av except ImportError: av = None VIDEO_DIR = os.path.join(os.path.dirname(os.path.abspath(__file__)), "assets", "videos") @pytest.mark.skipif...
import math import os import pytest import torch from torchvision.io import _HAS_GPU_VIDEO_DECODER, VideoReader try: import av except ImportError: av = None VIDEO_DIR = os.path.join(os.path.dirname(os.path.abspath(__file__)), "assets", "videos") @pytest.mark.skipif(_HAS_GPU_VIDEO_DECODER is False, reason="...
# Copyright (c) OpenMMLab. All rights reserved. from .base_det_dataset import BaseDetDataset from .base_video_dataset import BaseVideoDataset from .cityscapes import CityscapesDataset from .coco import CocoDataset from .coco_panoptic import CocoPanopticDataset from .crowdhuman import CrowdHumanDataset from .dataset_wra...
# Copyright (c) OpenMMLab. All rights reserved. from .base_det_dataset import BaseDetDataset from .base_video_dataset import BaseVideoDataset from .cityscapes import CityscapesDataset from .coco import CocoDataset from .coco_panoptic import CocoPanopticDataset from .crowdhuman import CrowdHumanDataset from .dataset_wra...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from mmengine.config import ConfigDict from mmdet.structures import DetDataSample from mmdet.testing import demo_mm_inputs, get_detector_cfg from mmdet.utils import register_all_modules class TestCornerNet(Tes...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from mmengine.config import ConfigDict from mmdet.structures import DetDataSample from mmdet.testing import demo_mm_inputs, get_detector_cfg from mmdet.utils import register_all_modules class TestCornerNet(Tes...
import json import pytest from langchain_core.agents import AgentFinish from langchain_core.exceptions import OutputParserException from langchain_core.messages import AIMessage, SystemMessage from langchain.agents.openai_functions_multi_agent.base import ( _FunctionsAgentAction, _parse_ai_message, ) # Test...
import json import pytest from langchain_core.agents import AgentFinish from langchain_core.exceptions import OutputParserException from langchain_core.messages import AIMessage, SystemMessage from langchain.agents.openai_functions_multi_agent.base import ( _FunctionsAgentAction, _parse_ai_message, ) # Test...
"""The k-nearest neighbors algorithms.""" # Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from ._ball_tree import BallTree from ._base import VALID_METRICS, VALID_METRICS_SPARSE, sort_graph_by_row_values from ._classification import KNeighborsClassifier, RadiusNeighborsClassifier from ....
"""The k-nearest neighbors algorithms.""" # Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from ._ball_tree import BallTree from ._base import VALID_METRICS, VALID_METRICS_SPARSE, sort_graph_by_row_values from ._classification import KNeighborsClassifier, RadiusNeighborsClassifier from ....
import jax import jax.numpy as jnp import jax.scipy as jsp from keras.src.backend import config from keras.src.backend import standardize_dtype from keras.src.backend.common import dtypes from keras.src.backend.jax.core import cast from keras.src.backend.jax.core import convert_to_tensor def cholesky(a): out = j...
import jax import jax.numpy as jnp import jax.scipy as jsp from keras.src.backend import config from keras.src.backend import standardize_dtype from keras.src.backend.common import dtypes from keras.src.backend.jax.core import cast from keras.src.backend.jax.core import convert_to_tensor def cholesky(a): out = j...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
# Copyright (c) OpenMMLab. All rights reserved. from abc import ABCMeta, abstractmethod from mmcv.runner import BaseModule from ..builder import build_shared_head class BaseRoIHead(BaseModule, metaclass=ABCMeta): """Base class for RoIHeads.""" def __init__(self, bbox_roi_extractor=None, ...
from abc import ABCMeta, abstractmethod from mmcv.runner import BaseModule from ..builder import build_shared_head class BaseRoIHead(BaseModule, metaclass=ABCMeta): """Base class for RoIHeads.""" def __init__(self, bbox_roi_extractor=None, bbox_head=None, ...
from __future__ import annotations from collections.abc import Iterable import torch import torch.nn as nn from sentence_transformers.sparse_encoder.losses.CSRReconstructionLoss import CSRReconstructionLoss from sentence_transformers.sparse_encoder.losses.SparseMultipleNegativesRankingLoss import ( SparseMultipl...
from __future__ import annotations from collections.abc import Iterable import torch import torch.nn as nn from sentence_transformers.sparse_encoder.losses.CSRReconstructionLoss import CSRReconstructionLoss from sentence_transformers.sparse_encoder.losses.SparseMultipleNegativesRankingLoss import ( SparseMultipl...
import os import httpx import pytest from llama_index.core.base.embeddings.base import BaseEmbedding from llama_index.embeddings.cohere import CohereEmbedding def test_embedding_class(): emb = CohereEmbedding(api_key="token") assert isinstance(emb, BaseEmbedding) @pytest.mark.skipif( os.environ.get("C...
import os import httpx import pytest from llama_index.core.base.embeddings.base import BaseEmbedding from llama_index.embeddings.cohere import CohereEmbedding def test_embedding_class(): emb = CohereEmbedding(api_key="token") assert isinstance(emb, BaseEmbedding) @pytest.mark.skipif( os.environ.get("C...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from mmengine.config import ConfigDict from mmdet.core import DetDataSample from mmdet.testing import demo_mm_inputs, get_detector_cfg from mmdet.utils import register_all_modules class TestCornerNet(TestCase)...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from mmengine.config import ConfigDict from mmdet.core import DetDataSample from mmdet.testing import demo_mm_inputs, get_detector_cfg from mmdet.utils import register_all_modules register_all_modules() class...
""" This file loads sentences from a provided text file. It is expected, that the there is one sentence per line in that text file. CT will be training using these sentences. Checkpoints are stored every 500 steps to the output folder. Usage: python train_ct_from_file.py path/to/sentences.txt """ import gzip import...
""" This file loads sentences from a provided text file. It is expected, that the there is one sentence per line in that text file. CT will be training using these sentences. Checkpoints are stored every 500 steps to the output folder. Usage: python train_ct_from_file.py path/to/sentences.txt """ import math from s...
"""Standard LangChain interface tests""" from typing import Optional, Type import pytest from langchain_core.language_models import BaseChatModel from langchain_core.rate_limiters import InMemoryRateLimiter from langchain_core.tools import BaseTool from langchain_tests.integration_tests import ( ChatModelIntegrat...
"""Standard LangChain interface tests""" from typing import Optional, Type import pytest from langchain_core.language_models import BaseChatModel from langchain_core.rate_limiters import InMemoryRateLimiter from langchain_core.tools import BaseTool from langchain_tests.integration_tests import ( ChatModelIntegrat...
import logging import os from langchain_qdrant.qdrant import RetrievalMode from tests.integration_tests.common import qdrant_running_locally logger = logging.getLogger(__name__) def qdrant_locations(use_in_memory: bool = True) -> list[str]: locations = [] if use_in_memory: logger.info("Running Qdra...
import logging import os from typing import List from langchain_qdrant.qdrant import RetrievalMode from tests.integration_tests.common import qdrant_running_locally logger = logging.getLogger(__name__) def qdrant_locations(use_in_memory: bool = True) -> List[str]: locations = [] if use_in_memory: l...
from typing import Optional import numpy as np import pytest import torch from docarray import BaseDoc, DocList, DocVec from docarray.documents import ImageDoc from docarray.typing import NdArray, TorchTensor class MyDoc(BaseDoc): embedding: NdArray text: str image: ImageDoc def test_from_to_json_docl...
from docarray import BaseDoc, DocList from docarray.documents import ImageDoc from docarray.typing import NdArray class MyDoc(BaseDoc): embedding: NdArray text: str image: ImageDoc def test_from_to_json(): da = DocList[MyDoc]( [ MyDoc( embedding=[1, 2, 3, 4, 5], t...
from typing import Union import numpy as np import PIL.Image import torch from torchvision import tv_tensors from torchvision.transforms import functional as _F @torch.jit.unused def to_image(inpt: Union[torch.Tensor, PIL.Image.Image, np.ndarray]) -> tv_tensors.Image: """See :class:`~torchvision.transforms.v2.To...
from typing import Union import numpy as np import PIL.Image import torch from torchvision import tv_tensors from torchvision.transforms import functional as _F @torch.jit.unused def to_image(inpt: Union[torch.Tensor, PIL.Image.Image, np.ndarray]) -> tv_tensors.Image: """[BETA] See :class:`~torchvision.transform...
# Copyright (c) OpenMMLab. All rights reserved. from .anchor_free_head import AnchorFreeHead from .anchor_head import AnchorHead from .atss_head import ATSSHead from .autoassign_head import AutoAssignHead from .cascade_rpn_head import CascadeRPNHead, StageCascadeRPNHead from .centernet_head import CenterNetHead from .c...
# Copyright (c) OpenMMLab. All rights reserved. from .anchor_free_head import AnchorFreeHead from .anchor_head import AnchorHead from .atss_head import ATSSHead from .autoassign_head import AutoAssignHead from .cascade_rpn_head import CascadeRPNHead, StageCascadeRPNHead from .centernet_head import CenterNetHead from .c...
"""Base interface class for storing chat history per user.""" import asyncio from abc import abstractmethod from typing import List, Optional from llama_index.core.llms import ChatMessage from llama_index.core.schema import BaseComponent class BaseChatStore(BaseComponent): @classmethod def class_name(cls) ->...
"""Base interface class for storing chat history per user.""" from abc import abstractmethod from typing import List, Optional from llama_index.core.llms import ChatMessage from llama_index.core.schema import BaseComponent class BaseChatStore(BaseComponent): @classmethod def class_name(cls) -> str: "...
import os import warnings from modulefinder import Module import torch from torchvision import datasets, io, models, ops, transforms, utils from .extension import _HAS_OPS try: from .version import __version__ # noqa: F401 except ImportError: pass # Check if torchvision is being imported within the root f...
import os import warnings from modulefinder import Module import torch from torchvision import datasets, io, models, ops, transforms, utils from .extension import _HAS_OPS try: from .version import __version__ # noqa: F401 except ImportError: pass # Check if torchvision is being imported within the root f...
import numpy as np from tensorflow import data as tf_data from keras.src import backend from keras.src import layers from keras.src import testing class IntegerLookupTest(testing.TestCase): # TODO: increase coverage. Most features aren't being tested. def test_config(self): layer = layers.IntegerLoo...
import numpy as np from tensorflow import data as tf_data from keras.src import backend from keras.src import layers from keras.src import testing class IntegerLookupTest(testing.TestCase): # TODO: increase coverage. Most features aren't being tested. def test_config(self): layer = layers.IntegerLoo...
"""Module containing the base parser for arguments of Jina.""" import argparse from jina.parsers.helper import _chf def set_base_parser(): """Set the base parser :return: the parser """ from jina import __version__ from jina.helper import colored, format_full_version_info, get_full_version ...
"""Module containing the base parser for arguments of Jina.""" import argparse from jina.parsers.helper import _chf def set_base_parser(): """Set the base parser :return: the parser """ from jina import __version__ from jina.helper import colored, format_full_version_info, get_full_version ...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from typing import Iterable, Optional import torch from jina import DocumentArray, Executor, requests from .audio_clip.model import AudioCLIP class AudioCLIPTextEncoder(Executor): """ Encode text data...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from typing import Iterable, Optional import torch from jina import DocumentArray, Executor, requests from jina_commons.batching import get_docs_batch_generator from .audio_clip.model import AudioCLIP class A...
import os from pathlib import Path from typing import List, Tuple, Union from torch import Tensor from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.librispeech import load_librispeech_item from torchaudio.datasets.utils import extract_archive _ARCHIVE_NAME = "li...
import os from pathlib import Path from typing import List, Tuple, Union from torch import Tensor from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.librispeech import load_librispeech_item from torchaudio.datasets.utils import extract_archive _ARCHIVE_NAME = "li...
from __future__ import annotations import torch.nn as nn from sentence_transformers.losses.CosineSimilarityLoss import CosineSimilarityLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseCosineSimilarityLoss(CosineSimilarityLoss): def __init__( self, mod...
from __future__ import annotations import torch.nn as nn from sentence_transformers.losses.CosineSimilarityLoss import CosineSimilarityLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseCosineSimilarityLoss(CosineSimilarityLoss): def __init__( self, mod...
"""Toolkit for interacting with a vector store.""" from langchain_core.language_models import BaseLanguageModel from langchain_core.tools import BaseTool from langchain_core.tools.base import BaseToolkit from langchain_core.vectorstores import VectorStore from pydantic import BaseModel, ConfigDict, Field class Vecto...
"""Toolkit for interacting with a vector store.""" from langchain_core.language_models import BaseLanguageModel from langchain_core.tools import BaseTool from langchain_core.tools.base import BaseToolkit from langchain_core.vectorstores import VectorStore from pydantic import BaseModel, ConfigDict, Field class Vecto...
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from typing import Dict, List, Optional import numpy as np from annoy import AnnoyIndex from jina import Document, DocumentArray, Executor, requests from jina_commons import get_logger from jina_commons.indexers.dump...
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from typing import Dict, List, Optional import numpy as np from annoy import AnnoyIndex from jina import Document, DocumentArray, Executor, requests from jina_commons import get_logger from jina_commons.indexers.dump...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase import torch from mmengine.structures import InstanceData from mmdet.registry import MODELS from mmdet.structures import DetDataSample from mmdet.testing import get_detector_cfg from mmdet.utils import register_all_modules class TestDABDE...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase import torch from mmengine.structures import InstanceData from mmdet.models import build_detector from mmdet.structures import DetDataSample from mmdet.testing import get_detector_cfg from mmdet.utils import register_all_modules class Tes...
""" The system trains BERT (or any other transformer model like RoBERTa, DistilBERT etc.) on the SNLI + MultiNLI (AllNLI) dataset with softmax loss function. At every 1000 training steps, the model is evaluated on the STS benchmark dataset Usage: python training_nli.py OR python training_nli.py pretrained_transformer...
""" The system trains BERT (or any other transformer model like RoBERTa, DistilBERT etc.) on the SNLI + MultiNLI (AllNLI) dataset with softmax loss function. At every 1000 training steps, the model is evaluated on the STS benchmark dataset Usage: python training_nli.py OR python training_nli.py pretrained_transformer...
_base_ = ['../common/ms_3x_coco.py', '../_base_/models/faster-rcnn_r50_fpn.py'] model = dict( backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_...
_base_ = [ '../common/mstrain_3x_coco.py', '../_base_/models/faster_rcnn_r50_fpn.py' ] model = dict( backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN'...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from parameterized import parameterized from mmdet.registry import MODELS from mmdet.testing import demo_mm_inputs, demo_mm_proposals, get_roi_head_cfg from mmdet.utils import register_all_modules class TestPI...
# Copyright (c) OpenMMLab. All rights reserved. import unittest from unittest import TestCase import torch from parameterized import parameterized from mmdet.registry import MODELS from mmdet.testing import demo_mm_inputs, demo_mm_proposals, get_roi_head_cfg from mmdet.utils import register_all_modules class TestPI...
from datetime import timedelta from typing import Optional from torch._C._distributed_c10d import _DEFAULT_PG_TIMEOUT __all__ = ["default_pg_timeout", "default_pg_nccl_timeout"] # Default process group wide timeout, if applicable. # This only applies to the non-nccl backends # To make an attempt at backwards compat...
from datetime import timedelta from typing import Optional from torch._C._distributed_c10d import _DEFAULT_PG_TIMEOUT __all__ = ["default_pg_timeout", "default_pg_nccl_timeout"] # Default process group wide timeout, if applicable. # This only applies to the non-nccl backends # To make an attempt at backwards compat...
from typing import Any, ForwardRef, Optional from typing_extensions import get_origin from typing_inspect import get_args, is_typevar, is_union_type from docarray.typing.tensor.abstract_tensor import AbstractTensor def is_type_tensor(type_: Any) -> bool: """Return True if type is a type Tensor or an Optional Te...
from typing import Any, Optional from typing_extensions import get_origin from typing_inspect import get_args, is_typevar, is_union_type from docarray.typing.tensor.abstract_tensor import AbstractTensor from typing import ForwardRef def is_type_tensor(type_: Any) -> bool: """Return True if type is a type Tensor...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp import xml.etree.ElementTree as ET from mmengine.fileio import list_from_file from mmdet.registry import DATASETS from .xml_style import XMLDataset @DATASETS.register_module() class WIDERFaceDataset(XMLDataset): """Reader for the WIDER Face d...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp import xml.etree.ElementTree as ET from mmengine.fileio import list_from_file from mmdet.registry import DATASETS from .xml_style import XMLDataset @DATASETS.register_module() class WIDERFaceDataset(XMLDataset): """Reader for the WIDER Face d...
from typing import TYPE_CHECKING, List, Union from docarray.array.abstract_array import AbstractDocumentArray from docarray.document import BaseDocument if TYPE_CHECKING: from docarray.typing import NdArray, TorchTensor class GetAttributeArrayMixin(AbstractDocumentArray): """Helpers that provide attributes ...
from typing import List, Union from docarray.array.abstract_array import AbstractDocumentArray from docarray.document import BaseDocument class GetAttributeArrayMixin(AbstractDocumentArray): """Helpers that provide attributes getter in bulk""" def _get_documents_attribute( self, field: str ) -> ...
import os.path from pathlib import Path from typing import Callable, Optional, Union import numpy as np import torch from torchvision.datasets.utils import download_url, verify_str_arg from torchvision.datasets.vision import VisionDataset class MovingMNIST(VisionDataset): """`MovingMNIST <http://www.cs.toronto.e...
import os.path from pathlib import Path from typing import Callable, Optional, Union import numpy as np import torch from torchvision.datasets.utils import download_url, verify_str_arg from torchvision.datasets.vision import VisionDataset class MovingMNIST(VisionDataset): """`MovingMNIST <http://www.cs.toronto.e...
from typing import List, Sequence from llama_index.core.agent.workflow.base_agent import BaseWorkflowAgent from llama_index.core.agent.workflow.single_agent_workflow import SingleAgentRunnerMixin from llama_index.core.agent.workflow.workflow_events import ( AgentInput, AgentOutput, AgentStream, ToolCal...
from typing import List, Sequence from llama_index.core.agent.workflow.base_agent import BaseWorkflowAgent from llama_index.core.agent.workflow.workflow_events import ( AgentInput, AgentOutput, AgentStream, ToolCallResult, ) from llama_index.core.base.llms.types import ChatResponse from llama_index.cor...
from __future__ import annotations from collections.abc import Iterable from typing import Any import torch from torch import Tensor, nn from sentence_transformers.SentenceTransformer import SentenceTransformer from sentence_transformers.util import fullname class CosineSimilarityLoss(nn.Module): def __init__(...
from __future__ import annotations from collections.abc import Iterable from typing import Any import torch from torch import Tensor, nn from sentence_transformers.SentenceTransformer import SentenceTransformer from sentence_transformers.util import fullname class CosineSimilarityLoss(nn.Module): def __init__(...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.core import ConfigType, OptConfigType, OptMultiConfig from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class VFNet(SingleStageDetector): """Implementation of `VarifocalNet (VFNet).<https://arxiv...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class VFNet(SingleStageDetector): """Implementation of `VarifocalNet (VFNet).<https://arxiv.org/abs/2008.13367>`_""" def __init__(self, ...
_base_ = [ '../_base_/models/retinanet_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] # model settings norm_cfg = dict(type='GN', num_groups=32, requires_grad=True) model = dict( backbone=dict( depth=101, init_c...
_base_ = [ '../_base_/models/retinanet_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] # model settings norm_cfg = dict(type='GN', num_groups=32, requires_grad=True) model = dict( backbone=dict( depth=101, init_c...
# Copyright (c) Meta Platforms, Inc. and affiliates. # This software may be used and distributed according to the terms of the Llama 2 Community License Agreement. import fire from llama import Llama def main( ckpt_dir: str, tokenizer_path: str, temperature: float = 0.6, top_p: float = 0.9, max_...
# Copyright (c) Meta Platforms, Inc. and affiliates. # This software may be used and distributed according to the terms of the Llama 2 Community License Agreement. import fire from llama import Llama def main( ckpt_dir: str, tokenizer_path: str, temperature: float = 0.6, top_p: float = 0.9, max_...
import datetime from typing import List import prisma.enums import pydantic class Pagination(pydantic.BaseModel): total_items: int = pydantic.Field( description="Total number of items.", examples=[42] ) total_pages: int = pydantic.Field( description="Total number of pages.", examples=[97]...
import datetime from typing import List import prisma.enums import pydantic class Pagination(pydantic.BaseModel): total_items: int = pydantic.Field( description="Total number of items.", examples=[42] ) total_pages: int = pydantic.Field( description="Total number of pages.", examples=[97]...
# Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from .glm import ( GammaRegressor, PoissonRegressor, TweedieRegressor, _GeneralizedLinearRegressor, ) __all__ = [ "GammaRegressor", "PoissonRegressor", "TweedieRegressor", "_GeneralizedLinearRegressor", ]
# Authors: The scikit-learn developers # SPDX-License-Identifier: BSD-3-Clause from .glm import ( GammaRegressor, PoissonRegressor, TweedieRegressor, _GeneralizedLinearRegressor, ) __all__ = [ "_GeneralizedLinearRegressor", "PoissonRegressor", "GammaRegressor", "TweedieRegressor", ]
import os import re from typing import Type from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField class BlockInstallationBlock(Block): """ This block allows the verification and installation of other blocks in the system. NOTE: T...
import os import re from typing import Type from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField class BlockInstallationBlock(Block): """ This block allows the verification and installation of other blocks in the system. NOTE: T...
import importlib from base64 import b64encode import pytest from fastapi.testclient import TestClient from ...utils import needs_py39 @pytest.fixture( name="client", params=[ "tutorial006", "tutorial006_an", pytest.param("tutorial006_an_py39", marks=needs_py39), ], ) def get_clie...
from base64 import b64encode from fastapi.testclient import TestClient from docs_src.security.tutorial006 import app client = TestClient(app) def test_security_http_basic(): response = client.get("/users/me", auth=("john", "secret")) assert response.status_code == 200, response.text assert response.jso...
from __future__ import annotations from typing import Optional from ..common import ( DeviceOpOverrides, register_device_op_overrides, TritonScratchWorkspace, ) class XPUDeviceOpOverrides(DeviceOpOverrides): def import_get_raw_stream_as(self, name: str) -> str: return f"from torch._C import ...
from __future__ import annotations from typing import Optional from ..common import DeviceOpOverrides, register_device_op_overrides class XPUDeviceOpOverrides(DeviceOpOverrides): def import_get_raw_stream_as(self, name: str) -> str: return f"from torch._C import _xpu_getCurrentRawStream as {name}" ...
import functools import hashlib from typing import Any @functools.cache def has_triton_package() -> bool: try: import triton # noqa: F401 return True except ImportError: return False @functools.cache def _device_supports_tma() -> bool: import torch return ( torch.c...
import functools import hashlib from typing import Any @functools.cache def has_triton_package() -> bool: try: from triton.compiler.compiler import triton_key return triton_key is not None except ImportError: return False except RuntimeError: return False @functools.cach...
import json import os import requests import sys import time from typing import Dict, List, Tuple CHECK_INTERVAL = 30 def get_environment_variables() -> Tuple[str, str, str, str, str]: """Retrieve and return necessary environment variables.""" try: with open(os.environ["GITHUB_EVENT_PATH"]) as f: ...
import json import os import requests import sys import time from typing import Dict, List, Tuple CHECK_INTERVAL = 30 def get_environment_variables() -> Tuple[str, str, str, str, str]: """Retrieve and return necessary environment variables.""" try: with open(os.environ["GITHUB_EVENT_PATH"]) as f: ...
# Copyright (c) OpenMMLab. All rights reserved. from .atss import ATSS from .autoassign import AutoAssign from .base import BaseDetector from .base_detr import DetectionTransformer from .boxinst import BoxInst from .cascade_rcnn import CascadeRCNN from .centernet import CenterNet from .condinst import CondInst from .co...
# Copyright (c) OpenMMLab. All rights reserved. from .atss import ATSS from .autoassign import AutoAssign from .base import BaseDetector from .base_detr import DetectionTransformer from .boxinst import BoxInst from .cascade_rcnn import CascadeRCNN from .centernet import CenterNet from .condinst import CondInst from .co...
from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.audio.abstract_audio_tensor import AbstractAudioTensor from docarray.typing.tensor.ndarray import NdArray @_register_proto(proto_type_name='audio_ndarray') class AudioNdArray(AbstractAudioTensor, NdArray): """ Subclass of [...
from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.audio.abstract_audio_tensor import AbstractAudioTensor from docarray.typing.tensor.ndarray import NdArray @_register_proto(proto_type_name='audio_ndarray') class AudioNdArray(AbstractAudioTensor, NdArray): """ Subclass of N...
import os from pathlib import Path from typing import Any, Callable, Optional, Union from .folder import default_loader, ImageFolder from .utils import download_and_extract_archive class EuroSAT(ImageFolder): """RGB version of the `EuroSAT <https://github.com/phelber/eurosat>`_ Dataset. For the MS version o...
import os from pathlib import Path from typing import Callable, Optional, Union from .folder import ImageFolder from .utils import download_and_extract_archive class EuroSAT(ImageFolder): """RGB version of the `EuroSAT <https://github.com/phelber/eurosat>`_ Dataset. For the MS version of the dataset, see ...
from langchain_core.embeddings import Embeddings from langchain_tests.unit_tests.embeddings import EmbeddingsTests class EmbeddingsIntegrationTests(EmbeddingsTests): """Base class for embeddings integration tests. Test subclasses must implement the ``embeddings_class`` property to specify the embeddings...
from typing import List from langchain_core.embeddings import Embeddings from langchain_tests.unit_tests.embeddings import EmbeddingsTests class EmbeddingsIntegrationTests(EmbeddingsTests): """Base class for embeddings integration tests. Test subclasses must implement the ``embeddings_class`` property to s...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import mmcv from mmcv import Config, DictAction from mmdet.datasets import build_dataset from mmdet.utils import update_data_root def parse_args(): parser = argparse.ArgumentParser(description='Evaluate metric of the ' ...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import mmcv from mmcv import Config, DictAction from mmdet.datasets import build_dataset def parse_args(): parser = argparse.ArgumentParser(description='Evaluate metric of the ' 'results saved in pkl format') ...
_base_ = [ '../_base_/models/ssd300.py', '../_base_/datasets/voc0712.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] model = dict( bbox_head=dict( num_classes=20, anchor_generator=dict(basesize_ratio_range=(0.2, ...
_base_ = [ '../_base_/models/ssd300.py', '../_base_/datasets/voc0712.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] model = dict( bbox_head=dict( num_classes=20, anchor_generator=dict(basesize_ratio_range=(0.2, ...
""" This example uses average word embeddings (for example from GloVe). It adds two fully-connected feed-forward layers (dense layers) to create a Deep Averaging Network (DAN). If 'glove.6B.300d.txt.gz' does not exist, it tries to download it from our server. See https://public.ukp.informatik.tu-darmstadt.de/reimers/...
""" This example uses average word embeddings (for example from GloVe). It adds two fully-connected feed-forward layers (dense layers) to create a Deep Averaging Network (DAN). If 'glove.6B.300d.txt.gz' does not exist, it tries to download it from our server. See https://public.ukp.informatik.tu-darmstadt.de/reimers/...
from torch.hub import download_url_to_file, load_state_dict_from_url __all__ = [ "load_state_dict_from_url", "download_url_to_file", ]
from torch.hub import load_state_dict_from_url, download_url_to_file __all__ = [ "load_state_dict_from_url", "download_url_to_file", ]
import os import pytest from jina import Flow cur_dir = os.path.dirname(os.path.abspath(__file__)) def _validate_flow(f): graph_dict = f._get_graph_representation() addresses = f._get_deployments_addresses() for name, pod in f: if name != 'gateway': for n in pod.needs: ...
import os import pytest from jina import Flow cur_dir = os.path.dirname(os.path.abspath(__file__)) def _validate_flow(f): graph_dict = f._get_graph_representation() addresses = f._get_deployments_addresses() for name, pod in f: if name != 'gateway': for n in pod.needs: ...
import argparse from jina.enums import GatewayProtocolType from jina.helper import parse_host_scheme from jina.logging.predefined import default_logger class NetworkChecker: """Check if a Deployment is running or not.""" def __init__(self, args: 'argparse.Namespace'): """ Create a new :class...
import argparse from jina.enums import GatewayProtocolType from jina.helper import parse_host_scheme from jina.logging.predefined import default_logger class NetworkChecker: """Check if a Deployment is running or not.""" def __init__(self, args: 'argparse.Namespace'): """ Create a new :class...
import datetime import autogpt_libs.auth.depends import autogpt_libs.auth.middleware import fastapi import fastapi.testclient import pytest import pytest_mock import backend.server.v2.library.db import backend.server.v2.library.model import backend.server.v2.library.routes app = fastapi.FastAPI() app.include_router(...
import autogpt_libs.auth.depends import autogpt_libs.auth.middleware import fastapi import fastapi.testclient import pytest import pytest_mock import backend.server.v2.library.db import backend.server.v2.library.model import backend.server.v2.library.routes app = fastapi.FastAPI() app.include_router(backend.server.v2...
import os from typing import Optional import pytest from docarray import BaseDoc, DocList from docarray.documents import ImageDoc from tests import TOYDATA_DIR @pytest.fixture() def nested_doc_cls(): class MyDoc(BaseDoc): count: Optional[int] text: str class MyDocNested(MyDoc): imag...
import os from typing import Optional import pytest from docarray import BaseDoc, DocList from docarray.documents import ImageDoc from tests import TOYDATA_DIR @pytest.fixture() def nested_doc_cls(): class MyDoc(BaseDoc): count: Optional[int] text: str class MyDocNested(MyDoc): imag...
""" The hub handles the moderation of inter-spoke communication. As the hub and spokes operate in isolated processes, sockets are employed to transmit messages between these processes. Consequently, a Socket class is defined for facilitating communication. """ import json class Socket: """ A class to facilit...
""" The hub handles the moderation of inter-spoke communication. As the hub and spokes operate in isolated processes, sockets are employed to transmit messages between these processes. Consequently, a Socket class is defined for facilitating communication. """ import json class Socket: """ A class to facilit...
# This is different from the TTA of official CenterNet. tta_model = dict( type='DetTTAModel', tta_cfg=dict(nms=dict(type='nms', iou_threshold=0.5), max_per_img=100)) tta_pipeline = [ dict(type='LoadImageFromFile', to_float32=True, backend_args=None), dict( type='TestTimeAug', transform...
# This is different from the TTA of official CenterNet. tta_model = dict( type='DetTTAModel', tta_cfg=dict(nms=dict(type='nms', iou_threshold=0.5), max_per_img=100)) tta_pipeline = [ dict( type='LoadImageFromFile', to_float32=True, file_client_args=dict(backend='disk')), dict( ...
import io from typing import TYPE_CHECKING, Any, Tuple, Type, TypeVar import numpy as np from pydantic import parse_obj_as from pydantic.validators import bytes_validator from docarray.typing.abstract_type import AbstractType from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.audio...
import io from typing import TYPE_CHECKING, Any, Tuple, Type, TypeVar import numpy as np from pydantic import parse_obj_as from pydantic.validators import bytes_validator from docarray.typing.abstract_type import AbstractType from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.audio...
import json import os import pickle import numpy as np import xgboost as xgb kRows = 100 kCols = 10 def generate_data(): X = np.random.randn(kRows, kCols) y = np.random.randn(kRows) return X, y class TestPickling: def run_model_pickling(self, xgb_params) -> str: X, y = generate_data() ...
import json import os import pickle import numpy as np import xgboost as xgb kRows = 100 kCols = 10 def generate_data(): X = np.random.randn(kRows, kCols) y = np.random.randn(kRows) return X, y class TestPickling: def run_model_pickling(self, xgb_params) -> str: X, y = generate_data() ...
# Copyright (c) OpenMMLab. All rights reserved. import torch.nn as nn from mmcv.cnn import ConvModule, is_norm from mmengine.model.utils import caffe2_xavier_init, constant_init, normal_init from torch.nn import BatchNorm2d from mmdet.registry import MODELS class Bottleneck(nn.Module): """Bottleneck block for Di...
# Copyright (c) OpenMMLab. All rights reserved. import torch.nn as nn from mmcv.cnn import (ConvModule, caffe2_xavier_init, constant_init, is_norm, normal_init) from torch.nn import BatchNorm2d from mmdet.registry import MODELS class Bottleneck(nn.Module): """Bottleneck block for DilatedEnc...
import numpy as np import pytest import torch from pydantic import parse_obj_as from docarray import BaseDoc from docarray.documents import PointCloud3D from docarray.utils._internal.misc import is_tf_available from docarray.utils._internal.pydantic import is_pydantic_v2 from tests import TOYDATA_DIR tf_available = i...
import numpy as np import pytest import torch from pydantic import parse_obj_as from docarray import BaseDoc from docarray.documents import PointCloud3D from docarray.utils._internal.misc import is_tf_available from tests import TOYDATA_DIR tf_available = is_tf_available() if tf_available: import tensorflow as tf...
# Licensed to the LF AI & Data foundation under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the "License"); # you may not use this fil...
from docarray.data.torch_dataset import MultiModalDataset __all__ = ['MultiModalDataset']
from __future__ import annotations from typing import TYPE_CHECKING, Literal, Union from pydantic import model_validator from langchain_core.messages import BaseMessage, BaseMessageChunk from langchain_core.outputs.generation import Generation from langchain_core.utils._merge import merge_dicts if TYPE_CHECKING: ...
from __future__ import annotations from typing import Literal, Union from pydantic import model_validator from typing_extensions import Self from langchain_core.messages import BaseMessage, BaseMessageChunk from langchain_core.outputs.generation import Generation from langchain_core.utils._merge import merge_dicts ...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.applications.nasnet import NASNetLarge as NASNetLarge from keras.src.applications.nasnet import NASNetMobile as NASNetMobile from keras.src.applications.nasnet import ( decode_pre...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.applications.nasnet import NASNetLarge from keras.src.applications.nasnet import NASNetMobile from keras.src.applications.nasnet import decode_predictions from keras.src.applications....
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .two_stage import TwoStageDetector @DETECTORS.register_module() class SparseRCNN(TwoStageDetector): r"""Implementation of `Sparse R-CNN: End-to-End Object Detection with Learnable Proposals <https://arxiv.org/abs/2011.12450>`...
from ..builder import DETECTORS from .two_stage import TwoStageDetector @DETECTORS.register_module() class SparseRCNN(TwoStageDetector): r"""Implementation of `Sparse R-CNN: End-to-End Object Detection with Learnable Proposals <https://arxiv.org/abs/2011.12450>`_""" def __init__(self, *args, **kwargs): ...
# Copyright (c) OpenMMLab. All rights reserved. import asyncio from argparse import ArgumentParser from mmdet.apis import (async_inference_detector, inference_detector, init_detector, show_result_pyplot) def parse_args(): parser = ArgumentParser() parser.add_argument('img', help='Imag...
# Copyright (c) OpenMMLab. All rights reserved. import asyncio from argparse import ArgumentParser from mmdet.apis import (async_inference_detector, inference_detector, init_detector, show_result_pyplot) def parse_args(): parser = ArgumentParser() parser.add_argument('img', help='Imag...
import datetime import sys from unittest import TestCase from mmengine import DefaultScope from mmdet.utils import register_all_modules class TestSetupEnv(TestCase): def test_register_all_modules(self): from mmdet.registry import DATASETS # not init default scope sys.modules.pop('mmdet...
import sys from unittest import TestCase from mmengine import DefaultScope from mmdet.utils import register_all_modules class TestSetupEnv(TestCase): def test_register_all_modules(self): from mmdet.registry import DATASETS # not init default scope sys.modules.pop('mmdet.datasets', None...
import json import os from typing import Dict import torch from torch import Tensor, nn class WeightedLayerPooling(nn.Module): """Token embeddings are weighted mean of their different hidden layer representations""" def __init__( self, word_embedding_dimension, num_hidden_layers: int = 12, layer_sta...
import torch from torch import Tensor from torch import nn from typing import Dict import os import json class WeightedLayerPooling(nn.Module): """Token embeddings are weighted mean of their different hidden layer representations""" def __init__( self, word_embedding_dimension, num_hidden_layers: int...
# Copyright (c) OpenMMLab. All rights reserved. from .hub import load_url from .misc import (check_prerequisites, concat_list, deprecated_api_warning, find_latest_checkpoint, has_method, import_modules_from_strings, is_list_of, is_method_overridden, is_seq_of, is...
# Copyright (c) OpenMMLab. All rights reserved. from .hub import load_url from .misc import (check_prerequisites, concat_list, deprecated_api_warning, has_method, import_modules_from_strings, is_list_of, is_method_overridden, is_seq_of, is_str, is_tuple_of, iter_...
def __getattr__(name: str): import warnings warnings.warn( "Torchaudio's I/O functions now support per-call backend dispatch. " "Importing backend implementation directly is no longer guaranteed to work. " "Please use `backend` keyword with load/save/info function, instead of " ...
def __getattr__(name: str): import warnings warnings.warn( "Torchaudio's I/O functions now support par-call bakcend dispatch. " "Importing backend implementation directly is no longer guaranteed to work. " "Please use `backend` keyword with load/save/info function, instead of " ...
import numpy as np from docarray import BaseDocument, DocumentArray, Image, Text from docarray.array.array_stacked import DocumentArrayStacked from docarray.typing import NdArray def test_simple_proto(): class CustomDoc(BaseDocument): text: str tensor: NdArray da = DocumentArray( [Cu...
import numpy as np from docarray import Document, DocumentArray, Image, Text from docarray.array.array_stacked import DocumentArrayStacked from docarray.typing import NdArray def test_simple_proto(): class CustomDoc(Document): text: str tensor: NdArray da = DocumentArray( [CustomDoc(...
"""Tool for the Google Lens""" from typing import Optional from langchain_core.callbacks import CallbackManagerForToolRun from langchain_core.tools import BaseTool from langchain_community.utilities.google_lens import GoogleLensAPIWrapper class GoogleLensQueryRun(BaseTool): """Tool that queries the Google Lens...
"""Tool for the Google Lens""" from typing import Optional from langchain_core.callbacks import CallbackManagerForToolRun from langchain_core.tools import BaseTool from langchain_community.utilities.google_lens import GoogleLensAPIWrapper class GoogleLensQueryRun(BaseTool): # type: ignore[override] """Tool th...
import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder from datasets.utils._hf_hub_fixes import create_repo, delete_repo CI_HUB_USER = "__DUMMY_TRANSFORMERS_USER__" CI_HUB_USER_FULL_NAME = "Dummy User" CI_HUB_USER_T...
import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder from datasets.utils._hf_hub_fixes import create_repo, delete_repo CI_HUB_USER = "__DUMMY_TRANSFORMERS_USER__" CI_HUB_USER_FULL_NAME = "Dummy User" CI_HUB_USER_T...
"""Init file.""" from llama_index.readers.gpt_repo.base import ( GPTRepoReader, get_ignore_list, process_repository, should_ignore, ) __all__ = [ "GPTRepoReader", "get_ignore_list", "process_repository", "should_ignore", ]
"""Init file.""" from llama_index.readers.gpt_repo.base import ( GPTRepoReader, get_ignore_list, process_repository, should_ignore, ) __all__ = [ "GPTRepoReader", "get_ignore_list", "process_repository", "should_ignore", ]
from .PhraseTokenizer import PhraseTokenizer from .WhitespaceTokenizer import WhitespaceTokenizer from .WordTokenizer import ENGLISH_STOP_WORDS, WordTokenizer __all__ = ["WordTokenizer", "WhitespaceTokenizer", "PhraseTokenizer", "ENGLISH_STOP_WORDS"]
from .WordTokenizer import WordTokenizer, ENGLISH_STOP_WORDS from .WhitespaceTokenizer import WhitespaceTokenizer from .PhraseTokenizer import PhraseTokenizer __all__ = ["WordTokenizer", "WhitespaceTokenizer", "PhraseTokenizer", "ENGLISH_STOP_WORDS"]
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.dtype_policies import deserialize from keras.src.dtype_policies import get from keras.src.dtype_policies import serialize from keras.src.dtype_policies.dtype_policy import DTypePolicy...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.dtype_policies import deserialize from keras.src.dtype_policies import get from keras.src.dtype_policies import serialize from keras.src.dtype_policies.dtype_policy import DTypePolicy...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.document_loaders import ( TelegramChatApiLoader, TelegramChatFileLoader, ) from langchain_community.document_loaders.telegram import ( concatenate_rows, ...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.document_loaders import ( TelegramChatApiLoader, TelegramChatFileLoader, ) from langchain_community.document_loaders.telegram import ( concatenate_rows, ...
from docarray import BaseDoc, DocList from docarray.documents import ImageDoc from docarray.typing import NdArray class MyDoc(BaseDoc): embedding: NdArray text: str image: ImageDoc def test_from_to_json(): da = DocList[MyDoc]( [ MyDoc( embedding=[1, 2, 3, 4, 5], t...
from docarray import BaseDoc, DocList from docarray.documents import ImageDoc from docarray.typing import NdArray class MyDoc(BaseDoc): embedding: NdArray text: str image: ImageDoc def test_from_to_json(): da = DocList[MyDoc]( [ MyDoc( embedding=[1, 2, 3, 4, 5], t...
from typing import Union import PIL.Image import torch from torchvision.prototype import features from torchvision.transforms.functional import pil_to_tensor, to_pil_image def erase_image_tensor( image: torch.Tensor, i: int, j: int, h: int, w: int, v: torch.Tensor, inplace: bool = False ) -> torch.Tensor: i...
from typing import Union import PIL.Image import torch from torchvision.prototype import features from torchvision.transforms import functional_tensor as _FT from torchvision.transforms.functional import pil_to_tensor, to_pil_image erase_image_tensor = _FT.erase @torch.jit.unused def erase_image_pil( image: PI...
# In[1]: import pandas as pd # In[2]: data_filename = "data.json" df = pd.read_json(data_filename).T df.tail() # In[3]: all_labels = {lbl for labels in df["labels"] for lbl in labels} all_labels # In[4]: # Add one column per label for label in all_labels: df[label] = df["labels"].apply(lambda labels_list: label ...
# In[1]: # imports and set configuration import pandas as pd from retrieve_prs_data import run exclude_prototype = True data_filename = "10.0_to_11.0-rc2.json" previous_release = "v10.0" current_release = "v11.0-rc2" # In[2]: df = pd.read_json(data_filename).T df.tail() # In[3]: all_labels = {lbl for labels in...
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
# Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
"""Test few shot prompt template.""" import re import pytest from langchain_core.prompts.few_shot_with_templates import FewShotPromptWithTemplates from langchain_core.prompts.prompt import PromptTemplate EXAMPLE_PROMPT = PromptTemplate( input_variables=["question", "answer"], template="{question}: {answer}" ) ...
"""Test few shot prompt template.""" import pytest from langchain_core.prompts.few_shot_with_templates import FewShotPromptWithTemplates from langchain_core.prompts.prompt import PromptTemplate EXAMPLE_PROMPT = PromptTemplate( input_variables=["question", "answer"], template="{question}: {answer}" ) async def ...
from typing import Type from .document import BaseDocument class AnyDocument(BaseDocument): """ AnyDocument is a Document that is not tied to any schema """ def __init__(self, **kwargs): super().__init__() self.__dict__.update(kwargs) @classmethod def _get_nested_document_cl...
from typing import Type from .document import BaseDocument class AnyDocument(BaseDocument): """ AnyDocument is a Document that is not tied to any schema """ def __init__(self, **kwargs): super().__init__() self.__dict__.update(kwargs) @classmethod def _get_nested_document_cl...
# Copyright (c) OpenMMLab. All rights reserved. from .bbox_overlaps import bbox_overlaps from .class_names import (cityscapes_classes, coco_classes, dataset_aliases, get_classes, imagenet_det_classes, imagenet_vid_classes, objects365v1_classes, ...
# Copyright (c) OpenMMLab. All rights reserved. from .bbox_overlaps import bbox_overlaps from .class_names import (cityscapes_classes, coco_classes, dataset_aliases, get_classes, imagenet_det_classes, imagenet_vid_classes, oid_challenge_classes, ...
from keras.src.tree.tree_api import assert_same_paths from keras.src.tree.tree_api import assert_same_structure from keras.src.tree.tree_api import flatten from keras.src.tree.tree_api import flatten_with_path from keras.src.tree.tree_api import is_nested from keras.src.tree.tree_api import lists_to_tuples from keras.s...
from keras.src.tree.tree_api import assert_same_structure from keras.src.tree.tree_api import flatten from keras.src.tree.tree_api import is_nested from keras.src.tree.tree_api import lists_to_tuples from keras.src.tree.tree_api import map_shape_structure from keras.src.tree.tree_api import map_structure from keras.src...
from __future__ import annotations import json from typing import ( Any, Union, ) from langchain_core._api import deprecated from pydantic import PrivateAttr from langchain_anthropic.chat_models import ChatAnthropic SYSTEM_PROMPT_FORMAT = """In this environment you have access to a set of tools you can use ...
import json from typing import ( Any, Union, ) from langchain_core._api import deprecated from pydantic import PrivateAttr from langchain_anthropic.chat_models import ChatAnthropic SYSTEM_PROMPT_FORMAT = """In this environment you have access to a set of tools you can use to answer the user's question. You ...