input
stringlengths
33
5k
output
stringlengths
32
5k
"""base multi modal retriever.""" from abc import abstractmethod from typing import List from llama_index.core.base.base_retriever import BaseRetriever from llama_index.core.image_retriever import BaseImageRetriever from llama_index.core.indices.query.schema import QueryType from llama_index.core.schema import NodeWi...
"""base multi modal retriever.""" from abc import abstractmethod from typing import List from llama_index.core.base.base_retriever import BaseRetriever from llama_index.core.image_retriever import BaseImageRetriever from llama_index.core.indices.query.schema import QueryType from llama_index.core.schema import NodeWit...
# Copyright (c) OpenMMLab. All rights reserved. from .class_names import (cityscapes_classes, coco_classes, dataset_aliases, get_classes, imagenet_det_classes, imagenet_vid_classes, oid_challenge_classes, oid_v6_classes, voc_classes) from .ev...
# Copyright (c) OpenMMLab. All rights reserved. from .class_names import (cityscapes_classes, coco_classes, dataset_aliases, get_classes, imagenet_det_classes, imagenet_vid_classes, voc_classes) from .eval_hooks import DistEvalHook, EvalHook from .mean_ap import avera...
import logging from typing import List, Optional from llama_index.core.schema import Document from llama_index.readers.box import BoxReaderBase from llama_index.readers.box.BoxAPI.box_api import ( box_check_connection, get_box_files_details, get_box_folder_files_details, get_ai_response_from_box_files,...
import logging from typing import List, Optional from llama_index.core.schema import Document from llama_index.readers.box import BoxReaderBase from llama_index.readers.box.BoxAPI.box_api import ( box_check_connection, get_box_files_details, get_box_folder_files_details, get_ai_response_from_box_files,...
from .hnswlib_searcher import HnswlibSearcher
from .hnswlib_searcher import HnswlibSearcher
from datasets import load_dataset from sentence_transformers import SentenceTransformer from sentence_transformers.quantization import quantize_embeddings, semantic_search_faiss # 1. Load the quora corpus with questions dataset = load_dataset("quora", split="train").map( lambda batch: {"text": [text for sample in ...
from sentence_transformers import SentenceTransformer from sentence_transformers.quantization import quantize_embeddings, semantic_search_faiss from datasets import load_dataset # 1. Load the quora corpus with questions dataset = load_dataset("quora", split="train").map( lambda batch: {"text": [text for sample in ...
import strawberry from fastapi import FastAPI from strawberry.fastapi import GraphQLRouter @strawberry.type class User: name: str age: int @strawberry.type class Query: @strawberry.field def user(self) -> User: return User(name="Patrick", age=100) schema = strawberry.Schema(query=Query) ...
import strawberry from fastapi import FastAPI from strawberry.asgi import GraphQL @strawberry.type class User: name: str age: int @strawberry.type class Query: @strawberry.field def user(self) -> User: return User(name="Patrick", age=100) schema = strawberry.Schema(query=Query) graphql_a...
import warnings from sys import platform from typing import Optional import torch import torchaudio dict_format = { torch.uint8: "u8", torch.int16: "s16", torch.int32: "s32", torch.int64: "s64", torch.float32: "flt", torch.float64: "dbl", } @torchaudio._extension.fail_if_no_ffmpeg def play_a...
import warnings from sys import platform from typing import Optional import torch import torchaudio from torchaudio.io import StreamWriter dict_format = { torch.uint8: "u8", torch.int16: "s16", torch.int32: "s32", torch.int64: "s64", torch.float32: "flt", torch.float64: "dbl", } @torchaudio....
import argparse from jina.enums import GatewayProtocolType from jina.helper import parse_host_scheme from jina.logging.predefined import default_logger class NetworkChecker: """Check if a Deployment is running or not.""" def __init__(self, args: 'argparse.Namespace'): """ Create a new :class...
import argparse from jina.enums import GatewayProtocolType from jina.helper import parse_host_scheme from jina.logging.predefined import default_logger class NetworkChecker: """Check if a Deployment is running or not.""" def __init__(self, args: 'argparse.Namespace'): """ Create a new :class...
from langchain_core.tracers.log_stream import ( LogEntry, LogStreamCallbackHandler, RunLog, RunLogPatch, RunState, ) __all__ = ["LogEntry", "LogStreamCallbackHandler", "RunLog", "RunLogPatch", "RunState"]
from langchain_core.tracers.log_stream import ( LogEntry, LogStreamCallbackHandler, RunLog, RunLogPatch, RunState, ) __all__ = ["LogEntry", "RunState", "RunLogPatch", "RunLog", "LogStreamCallbackHandler"]
_base_ = './retinanet_r50-caffe_fpn_ms-1x_coco.py' # training schedule for 2x train_cfg = dict(max_epochs=24) # learning rate policy param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=24, ...
_base_ = './retinanet_r50_caffe_fpn_mstrain_1x_coco.py' # training schedule for 2x train_cfg = dict(max_epochs=24) # learning rate policy param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end=24, ...
from keras.src.backend.jax import core from keras.src.backend.jax import distribution_lib from keras.src.backend.jax import image from keras.src.backend.jax import linalg from keras.src.backend.jax import math from keras.src.backend.jax import nn from keras.src.backend.jax import numpy from keras.src.backend.jax import...
from keras.src.backend.common.name_scope import name_scope from keras.src.backend.jax import core from keras.src.backend.jax import distribution_lib from keras.src.backend.jax import image from keras.src.backend.jax import linalg from keras.src.backend.jax import math from keras.src.backend.jax import nn from keras.src...
""" This file contains deprecated code that can only be used with the old `model.fit`-style Sentence Transformers v2.X training. It exists for backwards compatibility with the `model.old_fit` method, but will be removed in a future version. Nowadays, with Sentence Transformers v3+, it is recommended to use the `Senten...
from __future__ import annotations import gzip from . import InputExample class PairedFilesReader: """Reads in the a Pair Dataset, split in two files""" def __init__(self, filepaths): self.filepaths = filepaths def get_examples(self, max_examples=0): fIns = [] for filepath in s...
"""Anyscale embeddings wrapper.""" from __future__ import annotations from typing import Dict, Optional from langchain_core.utils import convert_to_secret_str, get_from_dict_or_env, pre_init from pydantic import Field, SecretStr from langchain_community.embeddings.openai import OpenAIEmbeddings from langchain_commu...
"""Anyscale embeddings wrapper.""" from __future__ import annotations from typing import Dict, Optional from langchain_core.utils import convert_to_secret_str, get_from_dict_or_env, pre_init from pydantic import Field, SecretStr from langchain_community.embeddings.openai import OpenAIEmbeddings from langchain_commu...
"""Parser for JSON output.""" from __future__ import annotations import json from json import JSONDecodeError from typing import Annotated, Any, Optional, TypeVar, Union import jsonpatch # type: ignore[import-untyped] import pydantic from pydantic import SkipValidation from langchain_core.exceptions import OutputP...
"""Parser for JSON output.""" from __future__ import annotations import json from json import JSONDecodeError from typing import Annotated, Any, Optional, TypeVar, Union import jsonpatch # type: ignore[import] import pydantic from pydantic import SkipValidation from langchain_core.exceptions import OutputParserExc...
from typing import Any, Dict, List, Optional, Sequence, Type, Union import PIL.Image import torch from torchvision import datapoints from torchvision.prototype.datapoints import Label, OneHotLabel from torchvision.transforms.v2 import functional as F, Transform from torchvision.transforms.v2._utils import _setup_fill...
from typing import Any, Dict, List, Optional, Sequence, Type, Union import PIL.Image import torch from torchvision import datapoints from torchvision.prototype.datapoints import Label, OneHotLabel from torchvision.transforms.v2 import functional as F, Transform from torchvision.transforms.v2._utils import _setup_fill...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Optional, Sequence, Union import torch from mmengine.data import BaseDataElement from mmengine.registry import HOOKS from .hook import Hook DATA_BATCH = Optional[Sequence[dict]] @HOOKS.register_module() class EmptyCacheHook(Hook): """Releases a...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Any, Optional, Sequence, Tuple, Union import torch from mmengine.data import BaseDataElement from mmengine.registry import HOOKS from .hook import Hook DATA_BATCH = Optional[Sequence[Tuple[Any, BaseDataElement]]] @HOOKS.register_module() class Empt...
from typing import Any, Optional, Union from huggingface_hub.utils import get_session from .. import config from ..exceptions import DatasetsError from .file_utils import ( get_authentication_headers_for_url, ) from .logging import get_logger logger = get_logger(__name__) class DatasetViewerError(DatasetsErro...
from typing import Any, Dict, List, Optional, Union from huggingface_hub.utils import get_session from .. import config from ..exceptions import DatasetsError from .file_utils import ( get_authentication_headers_for_url, ) from .logging import get_logger logger = get_logger(__name__) class DatasetViewerError(...
import pytest import tensorflow as tf from keras.src import backend from keras.src.backend.tensorflow import random from keras.src.testing import TestCase @pytest.mark.skipif( backend.backend() != "tensorflow", reason="Only applies to TensorFlow random ops.", ) class TFRandomTest(TestCase): def test_cat...
import pytest import tensorflow as tf from keras.src import backend from keras.src.backend.tensorflow import random from keras.src.testing import TestCase @pytest.mark.skipif( backend.backend() != "tensorflow", reason="Only applies to TensorFlow random ops.", ) class TFRandomTest(TestCase): def test_cat...
""" This is a simple application for sentence embeddings: clustering Sentences are mapped to sentence embeddings and then agglomerative clustering with a threshold is applied. """ from sklearn.cluster import AgglomerativeClustering from sentence_transformers import SentenceTransformer embedder = SentenceTransformer...
""" This is a simple application for sentence embeddings: clustering Sentences are mapped to sentence embeddings and then agglomerative clustering with a threshold is applied. """ from sentence_transformers import SentenceTransformer from sklearn.cluster import AgglomerativeClustering import numpy as np embedder = Se...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
# Licensed to the LF AI & Data foundation under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the "License"); # you may not use this fil...
from docarray import DocList from docarray.base_doc.doc import BaseDocWithoutId def test_doc_list(): class A(BaseDocWithoutId): text: str cls_doc_list = DocList[A] assert isinstance(cls_doc_list, type)
"""Intercom reader.""" import json from typing import List from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class IntercomReader(BaseReader): """ Intercom reader. Reads data from a Intercom workspace. Args: personal_access_token (str): Intercom to...
"""Intercom reader.""" import json from typing import List from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class IntercomReader(BaseReader): """Intercom reader. Reads data from a Intercom workspace. Args: personal_access_token (str): Intercom token. ...
import time import http.client import json from typing import List, Optional, Union from llama_index.core.base.base_retriever import BaseRetriever from llama_index.core.callbacks.base import CallbackManager from llama_index.core.schema import NodeWithScore, QueryBundle, TextNode class GalaxiaClient: def __init_...
import time import http.client import json from typing import List, Optional, Union from llama_index.core.base.base_retriever import BaseRetriever from llama_index.core.callbacks.base import CallbackManager from llama_index.core.schema import NodeWithScore, QueryBundle, TextNode class GalaxiaClient: def __init_...
from dataclasses import dataclass, fields import pytest from sklearn.base import ( BaseEstimator, RegressorMixin, TransformerMixin, ) from sklearn.utils import Tags, get_tags from sklearn.utils.estimator_checks import ( check_estimator_tags_renamed, check_valid_tag_types, ) class NoTagsEstimator...
import pytest from sklearn.base import ( BaseEstimator, RegressorMixin, TransformerMixin, ) from sklearn.utils._tags import get_tags class NoTagsEstimator: pass class ClassifierEstimator: # This is to test whether not inheriting from mixins works. _estimator_type = "classifier" class Empt...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Optional, Tuple from mmcv.cnn.bricks import build_plugin_layer from torch import Tensor from mmdet.core.utils.typing import OptConfigType from mmdet.registry import MODELS from .base_roi_extractor import BaseRoIExtractor @MODELS.register_module() cl...
# Copyright (c) OpenMMLab. All rights reserved. from mmcv.cnn.bricks import build_plugin_layer from mmcv.runner import force_fp32 from mmdet.registry import MODELS from .base_roi_extractor import BaseRoIExtractor @MODELS.register_module() class GenericRoIExtractor(BaseRoIExtractor): """Extract RoI features from ...
from torchvision.transforms import InterpolationMode # usort: skip from ._utils import is_simple_tensor # usort: skip from ._meta import ( clamp_bounding_boxes, convert_format_bounding_boxes, get_dimensions_image_tensor, get_dimensions_image_pil, get_dimensions, get_num_frames_video, get...
from torchvision.transforms import InterpolationMode # usort: skip from ._utils import is_simple_tensor # usort: skip from ._meta import ( clamp_bounding_boxes, convert_format_bounding_boxes, get_dimensions_image_tensor, get_dimensions_image_pil, get_dimensions, get_num_frames_video, get...
# Copyright (c) OpenMMLab. All rights reserved. import numpy as np from mmdet.registry import TRANSFORMS @TRANSFORMS.register_module() class InstaBoost: r"""Data augmentation method in `InstaBoost: Boosting Instance Segmentation Via Probability Map Guided Copy-Pasting <https://arxiv.org/abs/1908.07801>`_...
# Copyright (c) OpenMMLab. All rights reserved. import numpy as np from ..builder import PIPELINES @PIPELINES.register_module() class InstaBoost: r"""Data augmentation method in `InstaBoost: Boosting Instance Segmentation Via Probability Map Guided Copy-Pasting <https://arxiv.org/abs/1908.07801>`_. ...
#!/usr/bin/env python3 """Trains a SentencePiece model on transcripts across LRS3 pretrain and trainval. - `[lrs3_path]` is the directory path for the LRS3 cropped face dataset. Example: python train_spm.py --lrs3-path [lrs3_path] """ import io import pathlib from argparse import ArgumentParser, RawTextHelpFormatter...
#!/usr/bin/env python3 """Trains a SentencePiece model on transcripts across LRS3 pretrain and trainval. Example: python train_spm.py --lrs3-path <LRS3-DIRECTORY> """ import io import pathlib from argparse import ArgumentParser, RawTextHelpFormatter import sentencepiece as spm def get_transcript_text(transcript_pa...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from typing import Dict, Iterable, Optional import numpy as np import paddlehub as hub from jina import DocumentArray, Executor, requests from jina_commons.batching import get_docs_batch_generator class TextPa...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from typing import Dict, Optional, Tuple import numpy as np import paddlehub as hub from jina import DocumentArray, Executor, requests from jina_commons.batching import get_docs_batch_generator class TextPaddl...
"""Run smoke tests""" import argparse import torchaudio # noqa: F401 import torchaudio.compliance.kaldi # noqa: F401 import torchaudio.datasets # noqa: F401 import torchaudio.functional # noqa: F401 import torchaudio.models # noqa: F401 import torchaudio.pipelines # noqa: F401 import torchaudio.sox_effects # n...
"""Run smoke tests""" import torchaudio # noqa: F401 import torchaudio.compliance.kaldi # noqa: F401 import torchaudio.datasets # noqa: F401 import torchaudio.functional # noqa: F401 import torchaudio.models # noqa: F401 import torchaudio.pipelines # noqa: F401 import torchaudio.sox_effects # noqa: F401 import ...
"""Utility to lazily import modules.""" from __future__ import annotations import importlib from typing import Any, TYPE_CHECKING class _LazyModule: """Lazily import a module.""" def __init__(self, module_name: str) -> None: self._name = module_name self._module: Any = None def __repr_...
"""Utility to lazily import modules.""" from __future__ import annotations import importlib from typing import Any, TYPE_CHECKING class _LazyModule: """Lazily import a module.""" def __init__(self, module_name: str) -> None: self._name = module_name self._module: Any = None def __repr_...
import pytest from llama_index.core.base.llms.types import ChatMessage from llama_index.core.llms.llm import LLM from llama_index.core.llms.mock import MockLLM from llama_index.core.llms.mock import MockLLMWithNonyieldingChatStream @pytest.fixture() def nonyielding_llm() -> LLM: return MockLLMWithNonyieldingChatS...
import pytest from llama_index.core.base.llms.types import ChatMessage from llama_index.core.llms.llm import LLM from llama_index.core.llms.mock import MockLLM from llama_index.core.llms.mock import MockLLMWithNonyieldingChatStream @pytest.fixture() def nonyielding_llm() -> LLM: return MockLLMWithNonyieldingChatS...
# Copyright (c) OpenMMLab. All rights reserved. from .accuracy import Accuracy, accuracy from .ae_loss import AssociativeEmbeddingLoss from .balanced_l1_loss import BalancedL1Loss, balanced_l1_loss from .cross_entropy_loss import (CrossEntropyLoss, binary_cross_entropy, cross_entropy, m...
# Copyright (c) OpenMMLab. All rights reserved. from .accuracy import Accuracy, accuracy from .ae_loss import AssociativeEmbeddingLoss from .balanced_l1_loss import BalancedL1Loss, balanced_l1_loss from .cross_entropy_loss import (CrossEntropyLoss, binary_cross_entropy, cross_entropy, m...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import subprocess import pytest from jina import Document, DocumentArray, Flow from spacy_text_encoder import SpacyTextEncoder _EMBEDDING_DIM = 96 @pytest.mark.parametrize('request_size', [1, 10, 50, 100]) de...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import subprocess import pytest from jina import Document, DocumentArray, Flow from spacy_text_encoder import SpacyTextEncoder _EMBEDDING_DIM = 96 @pytest.mark.parametrize('request_size', [1, 10, 50, 100]) de...
from __future__ import annotations from sentence_transformers.losses.TripletLoss import TripletDistanceMetric, TripletLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseTripletLoss(TripletLoss): def __init__( self, model: SparseEncoder, distance_metric=TripletDi...
from __future__ import annotations from sentence_transformers.losses.TripletLoss import TripletDistanceMetric, TripletLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseTripletLoss(TripletLoss): def __init__( self, model: SparseEncoder, distance_metric=TripletDi...
from typing import List import torch import torchaudio.prototype.transforms as T from torch.autograd import gradcheck, gradgradcheck from torchaudio_unittest.common_utils import get_spectrogram, get_whitenoise, nested_params, TestBaseMixin class Autograd(TestBaseMixin): def assert_grad( self, tra...
from typing import List import torch import torchaudio.prototype.transforms as T from torch.autograd import gradcheck, gradgradcheck from torchaudio_unittest.common_utils import get_spectrogram, get_whitenoise, nested_params, TestBaseMixin class Autograd(TestBaseMixin): def assert_grad( self, tra...
import importlib import os import re import types from typing import Any, Optional import numpy as np try: import torch # noqa: F401 except ImportError: torch_imported = False else: torch_imported = True try: import tensorflow as tf # type: ignore # noqa: F401 except (ImportError, TypeError): ...
import importlib import os import re import types from typing import Any, Optional import numpy as np try: import torch # noqa: F401 except ImportError: torch_imported = False else: torch_imported = True try: import tensorflow as tf # type: ignore # noqa: F401 except (ImportError, TypeError): ...
import asyncio import os import random import string import tempfile import time import pytest from jina import helper @pytest.fixture(scope='function') def random_workspace_name(): """Generate a random workspace name with digits and letters.""" rand = ''.join(random.choices(string.ascii_uppercase + string....
import asyncio import os import random import string import tempfile import time import pytest from jina import helper @pytest.fixture(scope='function') def random_workspace_name(): """Generate a random workspace name with digits and letters.""" rand = ''.join(random.choices(string.ascii_uppercase + string....
# Copyright (c) OpenMMLab. All rights reserved. import argparse import os.path as osp from mmengine.config import Config, DictAction from mmengine.utils import ProgressBar from mmdet.models.utils import mask2ndarray from mmdet.registry import DATASETS, VISUALIZERS from mmdet.structures.bbox import BaseBoxes from mmde...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import os.path as osp import numpy as np from mmengine.config import Config, DictAction from mmengine.utils import ProgressBar from mmdet.models.utils import mask2ndarray from mmdet.registry import DATASETS, VISUALIZERS from mmdet.structures.bbox import ...
""" This script contains an example how to perform semantic search with Qdrant. You need Qdrant up and running locally: https://qdrant.tech/documentation/quickstart/ Further, you need the Python Qdrant Client installed: https://python-client.qdrant.tech/, e.g.: ``` pip install qdrant-client ``` This script was create...
""" This script contains an example how to perform semantic search with Qdrant. You need Qdrant up and running locally: https://qdrant.tech/documentation/quickstart/ Further, you need the Python Qdrant Client installed: https://python-client.qdrant.tech/, e.g.: ``` pip install qdrant-client ``` This script was create...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import os import subprocess import numpy as np import pytest from jina import Document, DocumentArray, Flow cur_dir = os.path.dirname(os.path.abspath(__file__)) def test_video_torch_encoder(): model_state...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import os import subprocess import numpy as np import pytest from jina import Document, DocumentArray, Flow cur_dir = os.path.dirname(os.path.abspath(__file__)) def test_video_torch_encoder(): model_state...
from docarray.typing.proto_register import _register_proto from docarray.typing.tensor.embedding.embedding_mixin import EmbeddingMixin from docarray.typing.tensor.ndarray import NdArray @_register_proto(proto_type_name='ndarray_embedding') class NdArrayEmbedding(NdArray, EmbeddingMixin): alternative_type = NdArra...
from docarray.typing.tensor.embedding.embedding_mixin import EmbeddingMixin from docarray.typing.tensor.ndarray import NdArray class NdArrayEmbedding(NdArray, EmbeddingMixin): alternative_type = NdArray
import tempfile import os import time import pytest cur_dir = os.path.dirname(os.path.abspath(__file__)) compose_yml = os.path.abspath( os.path.join(cur_dir, 'unit', 'array', 'docker-compose.yml') ) @pytest.fixture(autouse=True) def tmpfile(tmpdir): tmpfile = f'docarray_test_{next(tempfile._get_candidate_na...
import tempfile import os import time import pytest cur_dir = os.path.dirname(os.path.abspath(__file__)) compose_yml = os.path.abspath( os.path.join(cur_dir, 'unit', 'array', 'docker-compose.yml') ) @pytest.fixture(autouse=True) def tmpfile(tmpdir): tmpfile = f'docarray_test_{next(tempfile._get_candidate_na...
from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import ContributorDetails, SchemaField class ReadCsvBlock(Block): class Input(BlockSchema): contents: str = SchemaField( description="The contents of the CSV file to read", placeho...
from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import ContributorDetails, SchemaField class ReadCsvBlock(Block): class Input(BlockSchema): contents: str = SchemaField( description="The contents of the CSV file to read", placeho...
from datetime import datetime, timezone import pytest from prisma.models import CreditTransaction from backend.blocks.llm import AITextGeneratorBlock from backend.data.credit import BetaUserCredit from backend.data.execution import NodeExecutionEntry from backend.data.user import DEFAULT_USER_ID from backend.integrat...
from datetime import datetime import pytest from prisma.models import CreditTransaction from backend.blocks.llm import AITextGeneratorBlock from backend.data.credit import BetaUserCredit from backend.data.execution import NodeExecutionEntry from backend.data.user import DEFAULT_USER_ID from backend.integrations.crede...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
# flake8: noqa # Copyright 2020 The HuggingFace Datasets Authors and the TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
from enum import Enum from typing import Dict, Iterable import torch.nn.functional as F from torch import Tensor, nn from sentence_transformers.SentenceTransformer import SentenceTransformer class TripletDistanceMetric(Enum): """The metric for the triplet loss""" COSINE = lambda x, y: 1 - F.cosine_similari...
from torch import nn, Tensor from typing import Iterable, Dict import torch.nn.functional as F from enum import Enum from ..SentenceTransformer import SentenceTransformer class TripletDistanceMetric(Enum): """The metric for the triplet loss""" COSINE = lambda x, y: 1 - F.cosine_similarity(x, y) EUCLIDEAN...
_base_ = '../rpn/rpn_r50_caffe_fpn_1x_coco.py' model = dict( rpn_head=dict( _delete_=True, type='CascadeRPNHead', num_stages=2, stages=[ dict( type='StageCascadeRPNHead', in_channels=256, feat_channels=256, a...
_base_ = '../rpn/rpn_r50_caffe_fpn_1x_coco.py' model = dict( rpn_head=dict( _delete_=True, type='CascadeRPNHead', num_stages=2, stages=[ dict( type='StageCascadeRPNHead', in_channels=256, feat_channels=256, a...
# Copyright (c) OpenMMLab. All rights reserved. import torch.nn as nn from mmcv.cnn import ConvModule from mmcv.ops import MaskedConv2d from ..builder import HEADS from .guided_anchor_head import FeatureAdaption, GuidedAnchorHead @HEADS.register_module() class GARetinaHead(GuidedAnchorHead): """Guided-Anchor-bas...
# Copyright (c) OpenMMLab. All rights reserved. import torch.nn as nn from mmcv.cnn import ConvModule from mmcv.ops import MaskedConv2d from ..builder import HEADS from .guided_anchor_head import FeatureAdaption, GuidedAnchorHead @HEADS.register_module() class GARetinaHead(GuidedAnchorHead): """Guided-Anchor-bas...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.legacy.saving.serialization import ( deserialize_keras_object as deserialize_keras_object, ) from keras.src.legacy.saving.serialization import ( serialize_keras_object as seri...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.legacy.saving.serialization import deserialize_keras_object from keras.src.legacy.saving.serialization import serialize_keras_object
from dataclasses import dataclass from functools import partial from typing import Callable import torch import torchaudio from torchaudio.prototype.models import conv_tasnet_base @dataclass class SourceSeparationBundle: """torchaudio.prototype.pipelines.SourceSeparationBundle() Dataclass that bundles comp...
from dataclasses import dataclass from functools import partial from typing import Callable import torch import torchaudio from torchaudio.prototype.models import conv_tasnet_base @dataclass class SourceSeparationBundle: """torchaudio.prototype.pipelines.SourceSeparationBundle() Dataclass that bundles comp...
from __future__ import annotations from sentence_transformers.sparse_encoder.data_collator import SparseEncoderDataCollator from sentence_transformers.sparse_encoder.evaluation import ( SparseBinaryClassificationEvaluator, SparseEmbeddingSimilarityEvaluator, SparseInformationRetrievalEvaluator, SparseM...
from __future__ import annotations from sentence_transformers.sparse_encoder.data_collator import SparseEncoderDataCollator from sentence_transformers.sparse_encoder.evaluation import ( SparseEmbeddingSimilarityEvaluator, SparseInformationRetrievalEvaluator, ) from sentence_transformers.sparse_encoder.losses i...
import torch from torch import Tensor from torch import nn from typing import List, Dict import os import json import logging from .tokenizer import WhitespaceTokenizer logger = logging.getLogger(__name__) class BoW(nn.Module): """Implements a Bag-of-Words (BoW) model to derive sentence embeddings. A weigh...
import torch from torch import Tensor from torch import nn from typing import List, Dict import os import json import logging import numpy as np from .tokenizer import WhitespaceTokenizer logger = logging.getLogger(__name__) class BoW(nn.Module): """Implements a Bag-of-Words (BoW) model to derive sentence embed...
import torch import torchaudio.prototype.functional as F from parameterized import parameterized from torch.autograd import gradcheck, gradgradcheck from torchaudio_unittest.common_utils import nested_params, TestBaseMixin class AutogradTestImpl(TestBaseMixin): @nested_params( [F.convolve, F.fftconvolve],...
import torch import torchaudio.prototype.functional as F from parameterized import parameterized from torch.autograd import gradcheck, gradgradcheck from torchaudio_unittest.common_utils import nested_params, TestBaseMixin class AutogradTestImpl(TestBaseMixin): @nested_params( [F.convolve, F.fftconvolve],...
# Copyright (c) OpenMMLab. All rights reserved. from .accuracy import Accuracy, accuracy from .ae_loss import AssociativeEmbeddingLoss from .balanced_l1_loss import BalancedL1Loss, balanced_l1_loss from .cross_entropy_loss import (CrossEntropyLoss, binary_cross_entropy, cross_entropy, m...
# Copyright (c) OpenMMLab. All rights reserved. from .accuracy import Accuracy, accuracy from .ae_loss import AssociativeEmbeddingLoss from .balanced_l1_loss import BalancedL1Loss, balanced_l1_loss from .cross_entropy_loss import (CrossEntropyLoss, binary_cross_entropy, cross_entropy, m...
_base_ = './cascade-rcnn_r50_fpn_1x_coco.py' model = dict( type='CascadeRCNN', backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True),...
_base_ = './cascade_rcnn_r50_fpn_1x_coco.py' model = dict( type='CascadeRCNN', backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True),...
_base_ = '../fast_rcnn/fast_rcnn_r50_fpn_1x_coco.py' model = dict( backbone=dict( type='ResNet', depth=50, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=False), norm_eval=True, style='caffe', in...
_base_ = '../fast_rcnn/fast_rcnn_r50_fpn_1x_coco.py' model = dict( backbone=dict( type='ResNet', depth=50, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=False), norm_eval=True, style='caffe', in...
from torchaudio.models.rnnt import emformer_rnnt_model # https://pytorch.org/audio/master/_modules/torchaudio/models/rnnt.html#emformer_rnnt_base def emformer_rnnt(): return emformer_rnnt_model( input_dim=512, encoding_dim=1024, num_symbols=1024, segment_length=64, right_co...
from torchaudio.models.rnnt import emformer_rnnt_model # https://pytorch.org/audio/master/_modules/torchaudio/models/rnnt.html#emformer_rnnt_base def emformer_rnnt(): return emformer_rnnt_model( input_dim=512, encoding_dim=1024, num_symbols=1024, segment_length=64, right_con...
from collections.abc import Sequence from inspect import signature from typing import Optional, Union from langchain_core.callbacks import Callbacks from langchain_core.documents import ( BaseDocumentCompressor, BaseDocumentTransformer, Document, ) from pydantic import ConfigDict class DocumentCompressor...
from collections.abc import Sequence from inspect import signature from typing import Optional, Union from langchain_core.callbacks import Callbacks from langchain_core.documents import ( BaseDocumentCompressor, BaseDocumentTransformer, Document, ) from pydantic import ConfigDict class DocumentCompressor...
import os import pytest from typing import List from unittest.mock import MagicMock, patch, AsyncMock import uuid from llama_index.core.base.base_selector import ( SelectorResult, SingleSelection, ) from llama_index.core.schema import QueryBundle from llama_index.core.tools import ToolMetadata from llama_index...
import os import pytest from typing import List from unittest.mock import MagicMock, patch, AsyncMock import uuid from llama_index.core.base.base_selector import ( SelectorResult, SingleSelection, ) from llama_index.core.schema import QueryBundle from llama_index.core.tools import ToolMetadata from llama_index...
from ._bounding_box import BoundingBox, BoundingBoxFormat from ._datapoint import FillType, FillTypeJIT, InputType, InputTypeJIT from ._image import Image, ImageType, ImageTypeJIT, TensorImageType, TensorImageTypeJIT from ._label import Label, OneHotLabel from ._mask import Mask from ._video import TensorVideoType, Ten...
from ._bounding_box import BoundingBox, BoundingBoxFormat from ._datapoint import FillType, FillTypeJIT, InputType, InputTypeJIT from ._image import Image, ImageType, ImageTypeJIT, TensorImageType, TensorImageTypeJIT from ._label import Label, OneHotLabel from ._mask import Mask from ._video import TensorVideoType, Ten...
# Copyright (c) OpenMMLab. All rights reserved. from .checkloss_hook import CheckInvalidLossHook from .ema import ExpMomentumEMAHook, LinearMomentumEMAHook from .sync_norm_hook import SyncNormHook from .sync_random_size_hook import SyncRandomSizeHook from .yolox_lrupdater_hook import YOLOXLrUpdaterHook from .yolox_mode...
from .checkloss_hook import CheckInvalidLossHook from .ema import ExpMomentumEMAHook, LinearMomentumEMAHook from .sync_norm_hook import SyncNormHook from .sync_random_size_hook import SyncRandomSizeHook from .yolox_lrupdater_hook import YOLOXLrUpdaterHook from .yolox_mode_switch_hook import YOLOXModeSwitchHook __all__...
""" This file loads sentences from a provided text file. It is expected, that the there is one sentence per line in that text file. TSDAE will be training using these sentences. Checkpoints are stored every 500 steps to the output folder. Usage: python train_tsdae_from_file.py path/to/sentences.txt """ import gzip ...
""" This file loads sentences from a provided text file. It is expected, that the there is one sentence per line in that text file. TSDAE will be training using these sentences. Checkpoints are stored every 500 steps to the output folder. Usage: python train_tsdae_from_file.py path/to/sentences.txt """ from sentenc...
# Copyright (c) OpenMMLab. All rights reserved. import time import unittest from unittest import TestCase import torch from mmengine.logging import MessageHub from mmengine.registry import init_default_scope from parameterized import parameterized from mmdet.registry import MODELS from mmdet.testing import demo_track...
# Copyright (c) OpenMMLab. All rights reserved. import time import unittest from unittest import TestCase import torch from mmengine.logging import MessageHub from mmengine.registry import init_default_scope from parameterized import parameterized from mmdet.registry import MODELS from mmdet.testing import demo_track...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.api import activations from keras.api import applications from keras.api import backend from keras.api import callbacks from keras.api import config from keras.api import constraints from...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.api import _tf_keras from keras.api import activations from keras.api import applications from keras.api import backend from keras.api import callbacks from keras.api import config from k...
from __future__ import annotations from dataclasses import dataclass from sentence_transformers.training_args import SentenceTransformerTrainingArguments @dataclass class SparseEncoderTrainingArguments(SentenceTransformerTrainingArguments): """ SparseEncoderTrainingArguments extends :class:`~SentenceTransfo...
from __future__ import annotations from dataclasses import dataclass from sentence_transformers.training_args import SentenceTransformerTrainingArguments @dataclass class SparseEncoderTrainingArguments(SentenceTransformerTrainingArguments): """ SparseEncoderTrainingArguments extends :class:`~transformers.Tr...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import MODELS from mmdet.utils import ConfigType, OptConfigType, OptMultiConfig from .sparse_rcnn import SparseRCNN @MODELS.register_module() class QueryInst(SparseRCNN): r"""Implementation of `Instances as Queries <http://arxiv.org/abs/2105....
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import MODELS from .sparse_rcnn import SparseRCNN @MODELS.register_module() class QueryInst(SparseRCNN): r"""Implementation of `Instances as Queries <http://arxiv.org/abs/2105.01928>`_""" def __init__(self, backbone, ...
import torch from docarray.typing.tensor.torch_tensor import TorchTensor import copy from docarray import BaseDoc from docarray.typing import TorchEmbedding, TorchTensor def test_set_torch_tensor(): class MyDocument(BaseDoc): tensor: TorchTensor d = MyDocument(tensor=torch.zeros((3, 224, 224))) ...
import torch from docarray import BaseDoc from docarray.typing import TorchEmbedding, TorchTensor def test_set_torch_tensor(): class MyDocument(BaseDoc): tensor: TorchTensor d = MyDocument(tensor=torch.zeros((3, 224, 224))) assert isinstance(d.tensor, TorchTensor) assert isinstance(d.tensor...
# data settings dataset_type = 'CocoCaptionDataset' data_root = 'data/coco/' # Example to use different file client # Method 1: simply set the data root and let the file I/O module # automatically infer from prefix (not support LMDB and Memcache yet) # data_root = 's3://openmmlab/datasets/detection/coco/' # Method ...
# data settings dataset_type = 'COCOCaptionDataset' data_root = 'data/coco/' # Example to use different file client # Method 1: simply set the data root and let the file I/O module # automatically infer from prefix (not support LMDB and Memcache yet) # data_root = 's3://openmmlab/datasets/detection/coco/' # Method ...
"""Pydantic v1 compatibility shim.""" from importlib import metadata from pydantic.v1 import * # noqa: F403 from langchain_core._api.deprecation import warn_deprecated try: _PYDANTIC_MAJOR_VERSION: int = int(metadata.version("pydantic").split(".")[0]) except metadata.PackageNotFoundError: _PYDANTIC_MAJOR_V...
"""Pydantic v1 compatibility shim.""" from importlib import metadata from langchain_core._api.deprecation import warn_deprecated # Create namespaces for pydantic v1 and v2. # This code must stay at the top of the file before other modules may # attempt to import pydantic since it adds pydantic_v1 and pydantic_v2 to ...
from torchaudio._internal.module_utils import dropping_support from ._multi_channel import MVDR, PSD, RTFMVDR, SoudenMVDR from ._transforms import ( AddNoise, AmplitudeToDB, ComputeDeltas, Convolve, Deemphasis, Fade, FFTConvolve, FrequencyMasking, GriffinLim, InverseMelScale, ...
from ._multi_channel import MVDR, PSD, RTFMVDR, SoudenMVDR from ._transforms import ( AddNoise, AmplitudeToDB, ComputeDeltas, Convolve, Deemphasis, Fade, FFTConvolve, FrequencyMasking, GriffinLim, InverseMelScale, InverseSpectrogram, LFCC, Loudness, MelScale, ...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Optional import torch def get_max_cuda_memory(device: Optional[torch.device] = None) -> int: """Returns the maximum GPU memory occupied by tensors in megabytes (MB) for a given device. By default, this returns the peak allocated memory since ...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Optional import torch def get_max_cuda_memory(device: Optional[torch.device] = None) -> int: """Returns the maximum GPU memory occupied by tensors in megabytes (MB) for a given device. By default, this returns the peak allocated memory since ...
_base_ = './reppoints-moment_r50_fpn-gn_head-gn_1x_coco.py' model = dict(bbox_head=dict(transform_method='minmax', use_grid_points=True))
_base_ = './reppoints_moment_r50_fpn_gn-neck+head_1x_coco.py' model = dict(bbox_head=dict(transform_method='minmax', use_grid_points=True))
# Copyright (c) OpenMMLab. All rights reserved. from .dist_utils import (DistOptimizerHook, all_reduce_dict, allreduce_grads, reduce_mean, sync_random_seed) from .misc import (center_of_mass, filter_scores_and_topk, flip_tensor, generate_coordinate, levels_to_images, mask2nda...
# Copyright (c) OpenMMLab. All rights reserved. from .dist_utils import (DistOptimizerHook, all_reduce_dict, allreduce_grads, reduce_mean, sync_random_seed) from .misc import (center_of_mass, filter_scores_and_topk, flip_tensor, generate_coordinate, levels_to_images, mask2nda...
# Copyright (c) OpenMMLab. All rights reserved. from .manager import ManagerMeta, ManagerMixin from .misc import (apply_to, check_prerequisites, concat_list, deprecated_api_warning, deprecated_function, get_object_from_string, has_method, import_modules_from_stri...
# Copyright (c) OpenMMLab. All rights reserved. from .manager import ManagerMeta, ManagerMixin from .misc import (apply_to, check_prerequisites, concat_list, deprecated_api_warning, deprecated_function, has_method, import_modules_from_strings, is_list_of, is_meth...
# In[1]: import pandas as pd # In[2]: # from https://github.com/pytorch/audio/blob/main/.github/process_commit.py primary_labels_mapping = { "BC-breaking": "Backward-incompatible changes", "deprecation": "Deprecations", "bug fix": "Bug Fixes", "new feature": "New Features", "improvement": "Imp...
# In[1]: import pandas as pd # In[2]: # from https://github.com/pytorch/audio/blob/main/.github/process_commit.py primary_labels_mapping = { "BC-breaking": "Backward-incompatible changes", "deprecation": "Deprecations", "bug fix": "Bug Fixes", "new feature": "New Features", "improvement": "Imp...
from __future__ import annotations from collections.abc import Iterable import torch.nn as nn from torch import Tensor from sentence_transformers.losses.CosineSimilarityLoss import CosineSimilarityLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseCosineSimilarityLoss(Cos...
from __future__ import annotations import torch.nn as nn from sentence_transformers.losses.CosineSimilarityLoss import CosineSimilarityLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseCosineSimilarityLoss(CosineSimilarityLoss): def __init__( self, mod...
# Copyright (c) OpenMMLab. All rights reserved. from .anchor import * # noqa: F401, F403 from .bbox import * # noqa: F401, F403 from .evaluation import * # noqa: F401, F403 from .hook import * # noqa: F401, F403 from .mask import * # noqa: F401, F403 from .post_processing import * # noqa: F401, F403 from .utils i...
from .anchor import * # noqa: F401, F403 from .bbox import * # noqa: F401, F403 from .evaluation import * # noqa: F401, F403 from .hook import * # noqa: F401, F403 from .mask import * # noqa: F401, F403 from .post_processing import * # noqa: F401, F403 from .utils import * # noqa: F401, F403
from __future__ import annotations from typing import TYPE_CHECKING, Any, Literal, TypeAlias import numpy as np Device: TypeAlias = Literal["cpu"] if TYPE_CHECKING: # NumPy 1.x on Python 3.10 fails to parse np.dtype[] DType: TypeAlias = np.dtype[ np.bool_ | np.integer[Any] | np.floa...
from __future__ import annotations __all__ = [ "ndarray", "Device", "Dtype", ] import sys from typing import ( Literal, Union, TYPE_CHECKING, ) from numpy import ( ndarray, dtype, int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, ...
# mypy: enable-error-code=unused-ignore from typing_extensions import assert_type, Never from torch import Size class ZeroIndex: def __index__(self) -> int: return 0 tup0: tuple[()] = () tup1: tuple[int] = (1,) tup2: tuple[int, int] = (1, 2) tupN: tuple[int, int, int] = (1, 2, 3) tupX: tuple[Never, .....
from typing_extensions import assert_type from torch import Size s1 = Size([1, 2, 3]) s2 = Size([1, 2, 3]) class ZeroIndex: def __index__(self) -> int: return 0 # __getitem__ assert_type(s1[0], int) assert_type(s1[ZeroIndex()], int) assert_type(s1[:2], Size) # __add__ assert_type(s1 + s2, Size) asser...
"""Google Search tool spec.""" import urllib.parse from typing import Optional import requests from llama_index.core.schema import Document from llama_index.core.tools.tool_spec.base import BaseToolSpec QUERY_URL_TMPL = ( "https://www.googleapis.com/customsearch/v1?key={key}&cx={engine}&q={query}" ) class Goog...
"""Google Search tool spec.""" import urllib.parse from typing import Optional import requests from llama_index.core.schema import Document from llama_index.core.tools.tool_spec.base import BaseToolSpec QUERY_URL_TMPL = ( "https://www.googleapis.com/customsearch/v1?key={key}&cx={engine}&q={query}" ) class Goog...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import os import os.path as osp from mmengine.config import Config, DictAction from mmengine.model import is_model_wrapper from mmengine.registry import RUNNERS from mmengine.runner import Runner from mmengine.runner.checkpoint import load_checkpoint fro...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import os import os.path as osp from mmengine.config import Config, DictAction from mmengine.model import is_model_wrapper from mmengine.registry import RUNNERS from mmengine.runner import Runner from mmengine.runner.checkpoint import load_checkpoint fro...
"""LangChain **Runnable** and the **LangChain Expression Language (LCEL)**. The LangChain Expression Language (LCEL) offers a declarative method to build production-grade programs that harness the power of LLMs. Programs created using LCEL and LangChain Runnables inherently support synchronous, asynchronous, batch, a...
"""LangChain **Runnable** and the **LangChain Expression Language (LCEL)**. The LangChain Expression Language (LCEL) offers a declarative method to build production-grade programs that harness the power of LLMs. Programs created using LCEL and LangChain Runnables inherently support synchronous, asynchronous, batch, a...
from typing import MutableSequence, TYPE_CHECKING, Union, Iterable from docarray import Document if TYPE_CHECKING: from docarray.typing import T class BaseDocumentArray(MutableSequence[Document]): def __init__(self, *args, storage: str = 'memory', **kwargs): super().__init__() self._init_sto...
from typing import MutableSequence, TYPE_CHECKING, Union, Iterable from .. import Document if TYPE_CHECKING: from ..typing import T class BaseDocumentArray(MutableSequence[Document]): def __init__(self, *args, storage: str = 'memory', **kwargs): super().__init__() self._init_storage(*args, *...
# Copyright (c) OpenMMLab. All rights reserved. """MMEngine provides 11 root registries to support using modules across projects. More datails can be found at https://mmengine.readthedocs.io/en/latest/tutorials/registry.html. """ from .registry import Registry # manage all kinds of runners like `EpochBasedRunner` an...
# Copyright (c) OpenMMLab. All rights reserved. """MMEngine provides 11 root registries to support using modules across projects. More datails can be found at https://mmengine.readthedocs.io/en/latest/tutorials/registry.html. """ from .registry import Registry # manage all kinds of runners like `EpochBasedRunner` an...
from langchain_core.tracers.schemas import ( BaseRun, ChainRun, LLMRun, Run, RunTypeEnum, ToolRun, TracerSession, TracerSessionBase, TracerSessionV1, TracerSessionV1Base, TracerSessionV1Create, ) __all__ = [ "BaseRun", "ChainRun", "LLMRun", "Run", "RunTyp...
from langchain_core.tracers.schemas import ( BaseRun, ChainRun, LLMRun, Run, RunTypeEnum, ToolRun, TracerSession, TracerSessionBase, TracerSessionV1, TracerSessionV1Base, TracerSessionV1Create, ) __all__ = [ "RunTypeEnum", "TracerSessionV1Base", "TracerSessionV1C...
import warnings from unittest import mock import numpy as np from conftest import skip_if_backend from keras.src import backend from keras.src import callbacks from keras.src import layers from keras.src import testing from keras.src.models import Sequential from keras.src.utils import numerical_utils try: impor...
import warnings from unittest import mock import numpy as np from keras.src import backend from keras.src import callbacks from keras.src import layers from keras.src import testing from keras.src.models import Sequential from keras.src.utils import numerical_utils try: import requests except ImportError: re...
from __future__ import annotations from collections.abc import Iterable import torch import torch.nn as nn from sentence_transformers.sparse_encoder.losses.CSRReconstructionLoss import CSRReconstructionLoss from sentence_transformers.sparse_encoder.losses.SparseMultipleNegativesRankingLoss import ( SparseMultipl...
from __future__ import annotations from collections.abc import Iterable import torch import torch.nn as nn from sentence_transformers.sparse_encoder.losses.CSRReconstructionLoss import CSRReconstructionLoss from sentence_transformers.sparse_encoder.losses.SparseMultipleNegativesRankingLoss import ( SparseMultipl...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( type='YOLOF', data_preprocessor=dict( type='DetDataPreprocessor', mean=[103.530, 116.280, 123.675], std=[1.0, 1.0, 1.0], bgr_to_rgb=Fals...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( type='YOLOF', backbone=dict( type='ResNet', depth=50, num_stages=4, out_indices=(3, ), frozen_stages=1, norm_cfg=dict(ty...
# Copyright 2025 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writ...
# Copyright 2025 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writ...
import asyncio from itertools import cycle from typing import Any, Optional, Union from uuid import UUID import pytest from pytest_benchmark.fixture import BenchmarkFixture # type: ignore[import-untyped] from typing_extensions import override from langchain_core.callbacks.base import AsyncCallbackHandler from langch...
# ruff: noqa: ARG002 import asyncio from itertools import cycle from typing import Any import pytest from pytest_benchmark.fixture import BenchmarkFixture # type: ignore from langchain_core.callbacks.base import AsyncCallbackHandler from langchain_core.language_models import GenericFakeChatModel from langchain_core....
import os from pathlib import Path from typing import List, Tuple, Union import torchaudio from torch import Tensor from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import extract_archive _RELEASE_CONFIGS = { "release1": { "folder_in_archive":...
import os from pathlib import Path from typing import List, Tuple, Union import torchaudio from torch import Tensor from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import ( extract_archive, ) _RELEASE_CONFIGS = { "release1": { "folder_in_...
""" This script runs the evaluation of an SBERT msmarco model on the MS MARCO dev dataset and reports different performances metrices for cossine similarity & dot-product. Usage: python eval_msmarco.py model_name [max_corpus_size_in_thousands] """ from sentence_transformers import LoggingHandler, SentenceTransformer,...
""" This script runs the evaluation of an SBERT msmarco model on the MS MARCO dev dataset and reports different performances metrices for cossine similarity & dot-product. Usage: python eval_msmarco.py model_name [max_corpus_size_in_thousands] """ from sentence_transformers import LoggingHandler, SentenceTransformer,...
from typing import Iterable, Optional, Type from langchain_core.callbacks import CallbackManagerForToolRun from langchain_core.documents import Document from langchain_core.tools import BaseTool from pydantic import BaseModel, Field from requests.exceptions import HTTPError, ReadTimeout from urllib3.exceptions import ...
from typing import Iterable, Optional, Type from langchain_core.callbacks import CallbackManagerForToolRun from langchain_core.documents import Document from langchain_core.tools import BaseTool from pydantic import BaseModel, Field from requests.exceptions import HTTPError, ReadTimeout from urllib3.exceptions import ...
import matplotlib.pyplot as plt import torch from torchvision.utils import draw_bounding_boxes, draw_segmentation_masks from torchvision import tv_tensors from torchvision.transforms.v2 import functional as F def plot(imgs, row_title=None, **imshow_kwargs): if not isinstance(imgs[0], list): # Make a 2d gr...
import matplotlib.pyplot as plt import torch from torchvision.utils import draw_bounding_boxes, draw_segmentation_masks from torchvision import datapoints from torchvision.transforms.v2 import functional as F def plot(imgs, row_title=None, **imshow_kwargs): if not isinstance(imgs[0], list): # Make a 2d gr...
from typing import Any, cast import pytest from llama_index.core.bridge.pydantic import PrivateAttr from llama_index.core.workflow.context_serializers import JsonSerializer from llama_index.core.workflow.events import Event class _TestEvent(Event): param: str _private_param_1: str = PrivateAttr() _privat...
from typing import Any, cast import pytest from llama_index.core.bridge.pydantic import PrivateAttr from llama_index.core.workflow.context_serializers import JsonSerializer from llama_index.core.workflow.events import Event class _TestEvent(Event): param: str _private_param_1: str = PrivateAttr() _privat...
""" This script contains an example how to perform semantic search with OpenSearch. You need OpenSearch up and running locally: https://docs.opensearch.org/docs/latest/getting-started/quickstart/ Further, you need the Python OpenSearch Client installed: https://docs.opensearch.org/docs/latest/clients/python-low-level...
""" This script contains an example how to perform semantic search with OpenSearch. You need OpenSearch up and running locally: https://docs.opensearch.org/docs/latest/getting-started/quickstart/ Further, you need the Python OpenSearch Client installed: https://docs.opensearch.org/docs/latest/clients/python-low-level...
from __future__ import annotations from collections.abc import Iterable from torch import Tensor from sentence_transformers import util from sentence_transformers.losses.CoSENTLoss import CoSENTLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseCoSENTLoss(CoSENTLoss): ...
from __future__ import annotations from sentence_transformers import util from sentence_transformers.losses.CoSENTLoss import CoSENTLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseCoSENTLoss(CoSENTLoss): def __init__(self, model: SparseEncoder, scale: float = 20.0, s...
from __future__ import annotations from collections.abc import Iterable import torch from torch import Tensor, nn from sentence_transformers import SentenceTransformer class MSELoss(nn.Module): def __init__(self, model: SentenceTransformer) -> None: """ Computes the MSE loss between the compute...
from __future__ import annotations from typing import Iterable import torch from torch import Tensor, nn from sentence_transformers import SentenceTransformer class MSELoss(nn.Module): def __init__(self, model: SentenceTransformer) -> None: """ Computes the MSE loss between the computed sentenc...
from typing import Iterable, Dict from docarray.array.storage.annlite.helper import OffsetMapping from docarray.array.storage.base.getsetdel import BaseGetSetDelMixin from docarray.array.storage.base.helper import Offset2ID from docarray.array.memory import DocumentArrayInMemory from docarray import Document, Document...
from typing import Iterable, Dict from docarray.array.storage.annlite.helper import OffsetMapping from docarray.array.storage.base.getsetdel import BaseGetSetDelMixin from docarray.array.storage.base.helper import Offset2ID from docarray.array.memory import DocumentArrayInMemory from docarray import Document, Document...