input
stringlengths
33
5k
output
stringlengths
32
5k
import numpy as np import orjson import pytest from pydantic.tools import parse_obj_as, schema_json_of from docarray.document.io.json import orjson_dumps from docarray.typing import NdArray def test_proto_tensor(): tensor = parse_obj_as(NdArray, np.zeros((3, 224, 224))) tensor._to_node_protobuf() def tes...
import numpy as np import orjson from pydantic.tools import parse_obj_as, schema_json_of from docarray.document.io.json import orjson_dumps from docarray.typing import NdArray def test_proto_tensor(): tensor = parse_obj_as(NdArray, np.zeros((3, 224, 224))) tensor._to_node_protobuf() def test_from_list():...
""" Make.com API wrapper. Currently cannot load documents. """ from typing import Any, List, Optional import requests from llama_index.core.base.response.schema import Response from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document, NodeWithScore, TextNode class MakeWrap...
"""Make.com API wrapper. Currently cannot load documents. """ from typing import Any, List, Optional import requests from llama_index.core.base.response.schema import Response from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document, NodeWithScore, TextNode class MakeWrapp...
_base_ = '../fast_rcnn/fast-rcnn_r50-caffe_fpn_1x_coco.py' model = dict( roi_head=dict( bbox_head=dict( bbox_coder=dict(target_stds=[0.04, 0.04, 0.08, 0.08]), loss_cls=dict( type='CrossEntropyLoss', use_sigmoid=False, loss_weight=1.5), loss_bbox=dict(type=...
_base_ = '../fast_rcnn/fast-rcnn_r50_fpn_1x_coco.py' model = dict( backbone=dict( type='ResNet', depth=50, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=False), norm_eval=True, style='caffe', in...
import pytest from llama_index.core.llms import ChatMessage from llama_index.core.tools import ToolSelection from llama_index.core.bridge.pydantic import BaseModel, ValidationError from llama_index.core.agent.workflow.workflow_events import ( AgentWorkflowStartEvent, AgentOutput, PydanticConversionWarning,...
from llama_index.core.llms import ChatMessage from llama_index.core.agent.workflow.workflow_events import AgentWorkflowStartEvent from llama_index.core.memory import Memory def test_agent_workflow_start_event(): event = AgentWorkflowStartEvent( user_msg="Hello, world!", chat_history=[ChatMessage(r...
# Copyright (c) OpenMMLab. All rights reserved. from .anchor_free_head import AnchorFreeHead from .anchor_head import AnchorHead from .atss_head import ATSSHead from .atss_vlfusion_head import ATSSVLFusionHead from .autoassign_head import AutoAssignHead from .boxinst_head import BoxInstBboxHead, BoxInstMaskHead from .c...
# Copyright (c) OpenMMLab. All rights reserved. from .anchor_free_head import AnchorFreeHead from .anchor_head import AnchorHead from .atss_head import ATSSHead from .autoassign_head import AutoAssignHead from .boxinst_head import BoxInstBboxHead, BoxInstMaskHead from .cascade_rpn_head import CascadeRPNHead, StageCasca...
# Copyright 2025 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writ...
# Copyright 2024 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writ...
from __future__ import annotations from collections.abc import Iterable import torch import torch.nn as nn import torch.nn.functional as F from sentence_transformers.sparse_encoder import SparseEncoder def normalized_mean_squared_error(reconstruction: torch.Tensor, original_input: torch.Tensor) -> torch.Tensor: ...
from __future__ import annotations from collections.abc import Iterable import torch import torch.nn as nn import torch.nn.functional as F from sentence_transformers.sparse_encoder import SparseEncoder def normalized_mean_squared_error(reconstruction: torch.Tensor, original_input: torch.Tensor) -> torch.Tensor: ...
from typing import Any, Optional, Sequence from llama_index.core.evaluation.base import BaseEvaluator, EvaluationResult from llama_index.core.prompts.mixin import PromptDictType, PromptMixinType from tonic_validate.metrics.answer_consistency_metric import ( AnswerConsistencyMetric, ) from tonic_validate.services....
from typing import Any, Optional, Sequence from llama_index.core.evaluation.base import BaseEvaluator, EvaluationResult from llama_index.core.prompts.mixin import PromptDictType, PromptMixinType from tonic_validate.metrics.answer_consistency_metric import ( AnswerConsistencyMetric, ) from tonic_validate.services....
# Copyright (c) OpenMMLab. All rights reserved. from math import ceil from unittest import TestCase import torch from mmengine import Config from mmengine.data import InstanceData from mmdet import * # noqa from mmdet.models.dense_heads import SSDHead class TestSSDHead(TestCase): def test_ssd_head_loss(self):...
# Copyright (c) OpenMMLab. All rights reserved. from math import ceil from unittest import TestCase import torch from mmengine import Config from mmengine.data import InstanceData from mmdet import * # noqa from mmdet.models.dense_heads import SSDHead class TestSSDHead(TestCase): def test_ssd_head_loss(self):...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.random.random import beta as beta from keras.src.random.random import binomial as binomial from keras.src.random.random import categorical as categorical from keras.src.random.random ...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.random.random import beta from keras.src.random.random import binomial from keras.src.random.random import categorical from keras.src.random.random import dropout from keras.src.rando...
import asyncio import os import random import string import tempfile import time import pytest from jina import helper @pytest.fixture(scope='function') def random_workspace_name(): """Generate a random workspace name with digits and letters.""" rand = ''.join(random.choices(string.ascii_uppercase + string....
import asyncio import os import random import string import tempfile import time import pytest from jina import helper @pytest.fixture(scope='function') def random_workspace_name(): """Generate a random workspace name with digits and letters.""" rand = ''.join(random.choices(string.ascii_uppercase + string....
from typing import Literal from langchain_core.tools import BaseTool from langchain_tests.integration_tests import ToolsIntegrationTests from langchain_tests.unit_tests import ToolsUnitTests class ParrotMultiplyTool(BaseTool): name: str = "ParrotMultiplyTool" description: str = ( "Multiply two numbe...
from typing import Literal from langchain_core.tools import BaseTool from langchain_tests.integration_tests import ToolsIntegrationTests from langchain_tests.unit_tests import ToolsUnitTests class ParrotMultiplyTool(BaseTool): name: str = "ParrotMultiplyTool" description: str = ( "Multiply two numbe...
import os from typing import Any, Optional from llama_index.llms.openai_like import OpenAILike class NetmindLLM(OpenAILike): """ Netmind LLM. Examples: `pip install llama-index-llms-netmind` ```python from llama_index.llms.netmind import NetmindLLM # set api key in env ...
import os from typing import Any, Optional from llama_index.llms.openai_like import OpenAILike class NetmindLLM(OpenAILike): """Netmind LLM. Examples: `pip install llama-index-llms-netmind` ```python from llama_index.llms.netmind import NetmindLLM # set api key in env or in...
"""Oxylabs Web Reader.""" import asyncio from typing import Any, Dict, List, Optional, TYPE_CHECKING from platform import architecture, python_version from importlib.metadata import version from llama_index.core.bridge.pydantic import Field from llama_index.core.readers.base import BasePydanticReader from llama_index...
"""Oxylabs Web Reader.""" import asyncio from typing import Any, List from platform import architecture, python_version from importlib.metadata import version from llama_index.core.readers.base import BasePydanticReader from llama_index.core.schema import Document from markdownify import markdownify from llama_index...
import re from typing import Any, Dict, Union from sentence_transformers import SentenceTransformer class SentenceEvaluator: """ Base class for all evaluators Extend this class and implement __call__ for custom evaluators. """ def __init__(self): self.greater_is_better = True # ...
from sentence_transformers import SentenceTransformer class SentenceEvaluator: """ Base class for all evaluators Extend this class and implement __call__ for custom evaluators. """ def __call__(self, model: SentenceTransformer, output_path: str = None, epoch: int = -1, steps: int = -1) -> float:...
from pathlib import Path from typing import Any, Callable, Optional, Tuple, Union from .folder import default_loader from .utils import download_and_extract_archive from .vision import VisionDataset class SUN397(VisionDataset): """`The SUN397 Data Set <https://vision.princeton.edu/projects/2010/SUN/>`_. Th...
from pathlib import Path from typing import Any, Callable, Optional, Tuple, Union import PIL.Image from .utils import download_and_extract_archive from .vision import VisionDataset class SUN397(VisionDataset): """`The SUN397 Data Set <https://vision.princeton.edu/projects/2010/SUN/>`_. The SUN397 or Scene ...
from PIL import Image from sentence_transformers import SentenceTransformer, models, util ########### image = Image.open("two_dogs_in_snow.jpg") from transformers import CLIPModel, CLIPProcessor model = CLIPModel.from_pretrained("openai/clip-vit-base-patch32") processor = CLIPProcessor.from_pretrained("openai/clip...
from PIL import Image from sentence_transformers import SentenceTransformer, models, util ########### image = Image.open("two_dogs_in_snow.jpg") from transformers import CLIPModel, CLIPProcessor model = CLIPModel.from_pretrained("openai/clip-vit-base-patch32") processor = CLIPProcessor.from_pretrained("openai/clip...
from __future__ import annotations from sentence_transformers.training_args import SentenceTransformerTrainingArguments class CrossEncoderTrainingArguments(SentenceTransformerTrainingArguments): r""" CrossEncoderTrainingArguments extends :class:`~transformers.TrainingArguments` with additional arguments ...
from __future__ import annotations from sentence_transformers.training_args import SentenceTransformerTrainingArguments class CrossEncoderTrainingArguments(SentenceTransformerTrainingArguments): r""" CrossEncoderTrainingArguments extends :class:`~transformers.TrainingArguments` with additional arguments ...
# Copyright (c) OpenMMLab. All rights reserved. from .base_roi_head import BaseRoIHead from .bbox_heads import (BBoxHead, ConvFCBBoxHead, DIIHead, DoubleConvFCBBoxHead, SABLHead, SCNetBBoxHead, Shared2FCBBoxHead, Shared4Conv1FCBBoxHead) from .cascade_roi_head import Cas...
from .base_roi_head import BaseRoIHead from .bbox_heads import (BBoxHead, ConvFCBBoxHead, DIIHead, DoubleConvFCBBoxHead, SABLHead, SCNetBBoxHead, Shared2FCBBoxHead, Shared4Conv1FCBBoxHead) from .cascade_roi_head import CascadeRoIHead from .double_roi_head import DoubleH...
from collections.abc import Sequence from typing import Optional from langchain_core.language_models import BaseLanguageModel from langchain_core.prompts.chat import ChatPromptTemplate from langchain_core.runnables import Runnable, RunnablePassthrough from langchain_core.tools import BaseTool from langchain_core.utils...
from typing import Optional, Sequence from langchain_core.language_models import BaseLanguageModel from langchain_core.prompts.chat import ChatPromptTemplate from langchain_core.runnables import Runnable, RunnablePassthrough from langchain_core.tools import BaseTool from langchain_core.utils.function_calling import co...
"""Test Ollama Chat API wrapper.""" from typing import Any from unittest.mock import patch from langchain_ollama import OllamaLLM MODEL_NAME = "llama3.1" def test_initialization() -> None: """Test integration initialization.""" OllamaLLM(model=MODEL_NAME) def test_model_params() -> None: # Test stand...
"""Test Ollama Chat API wrapper.""" from typing import Any from unittest.mock import patch from langchain_ollama import OllamaLLM MODEL_NAME = "llama3.1" def test_initialization() -> None: """Test integration initialization.""" OllamaLLM(model="llama3") def test_model_params() -> None: # Test standar...
from typing import TypeVar from fastapi import Depends, FastAPI from fastapi.testclient import TestClient from typing_extensions import Annotated app = FastAPI() T = TypeVar("T") Dep = Annotated[T, Depends()] class A: pass class B: pass @app.get("/a") async def a(dep: Dep[A]): return {"cls": dep._...
from typing import TypeVar from fastapi import Depends, FastAPI from fastapi.testclient import TestClient from typing_extensions import Annotated app = FastAPI() T = TypeVar("T") Dep = Annotated[T, Depends()] class A: pass class B: pass @app.get("/a") async def a(dep: Dep[A]): return {"cls": dep._...
# Copyright 2018 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
# Copyright 2018 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
from keras.src import backend from keras.src.utils.module_utils import tensorflow as tf def get_tensor_spec(t, dynamic_batch=False, name=None): """Returns a `TensorSpec` given a single `Tensor` or `TensorSpec`.""" if isinstance(t, tf.TypeSpec): spec = t elif isinstance(t, tf.__internal__.Composite...
from keras.src import backend from keras.src.utils.module_utils import tensorflow as tf def get_tensor_spec(t, dynamic_batch=False, name=None): """Returns a `TensorSpec` given a single `Tensor` or `TensorSpec`.""" if isinstance(t, tf.TypeSpec): spec = t elif isinstance(t, tf.__internal__.Composite...
import numpy as np import pytest from docarray import BaseDoc, DocList from docarray.typing import NdArray @pytest.mark.parametrize('shuffle', [False, True]) @pytest.mark.parametrize('stack', [False, True]) @pytest.mark.parametrize('batch_size,n_batches', [(16, 7), (10, 10)]) def test_batch(shuffle, stack, batch_siz...
import numpy as np import pytest from docarray import BaseDoc, DocList from docarray.typing import NdArray @pytest.mark.parametrize('shuffle', [False, True]) @pytest.mark.parametrize('stack', [False, True]) @pytest.mark.parametrize('batch_size,n_batches', [(16, 7), (10, 10)]) def test_batch(shuffle, stack, batch_siz...
import sys import pytest from hypothesis import given, settings, strategies from xgboost.testing import no_cupy sys.path.append("tests/python") from test_data_iterator import run_data_iterator from test_data_iterator import test_single_batch as cpu_single_batch def test_gpu_single_batch() -> None: cpu_single_b...
import sys import pytest from hypothesis import given, settings, strategies from xgboost.testing import no_cupy sys.path.append("tests/python") from test_data_iterator import run_data_iterator from test_data_iterator import test_single_batch as cpu_single_batch def test_gpu_single_batch() -> None: cpu_single_b...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import MODELS from .two_stage import TwoStageDetector @MODELS.register_module() class MaskRCNN(TwoStageDetector): """Implementation of `Mask R-CNN <https://arxiv.org/abs/1703.06870>`_""" def __init__(self, backbone, ...
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .two_stage import TwoStageDetector @DETECTORS.register_module() class MaskRCNN(TwoStageDetector): """Implementation of `Mask R-CNN <https://arxiv.org/abs/1703.06870>`_""" def __init__(self, backbone, ...
# coding: utf-8 from pathlib import Path import numpy as np preds = [np.loadtxt(str(name)) for name in Path(__file__).absolute().parent.glob("*.pred")] np.testing.assert_allclose(preds[0], preds[1])
# coding: utf-8 from pathlib import Path import numpy as np preds = [np.loadtxt(str(name)) for name in Path(__file__).absolute().parent.glob('*.pred')] np.testing.assert_allclose(preds[0], preds[1])
import warnings from typing import Any, Callable, Dict, List, Optional, Sequence, Union import torch from torch import nn from torchvision import transforms as _transforms from torchvision.prototype.transforms import Transform class Compose(Transform): def __init__(self, transforms: Sequence[Callable]) -> None:...
import warnings from typing import Any, Callable, List, Optional, Sequence, Union import torch from torch import nn from torchvision.prototype.transforms import Transform class Compose(Transform): def __init__(self, transforms: Sequence[Callable]) -> None: super().__init__() if not isinstance(tr...
# Copyright (c) OpenMMLab. All rights reserved. import time from typing import Optional, Sequence, Union from mmengine.data import BaseDataElement from mmengine.registry import HOOKS from .hook import Hook DATA_BATCH = Optional[Sequence[dict]] @HOOKS.register_module() class IterTimerHook(Hook): """A hook that l...
# Copyright (c) OpenMMLab. All rights reserved. import time from typing import Any, Optional, Sequence, Tuple, Union from mmengine.data import BaseDataElement from mmengine.registry import HOOKS from .hook import Hook DATA_BATCH = Optional[Sequence[Tuple[Any, BaseDataElement]]] @HOOKS.register_module() class IterTi...
# Copyright (c) OpenMMLab. All rights reserved. import torch import torch.nn as nn class _BatchNormXd(nn.modules.batchnorm._BatchNorm): """A general BatchNorm layer without input dimension check. Reproduced from @kapily's work: (https://github.com/pytorch/pytorch/issues/41081#issuecomment-783961547) ...
# Copyright (c) OpenMMLab. All rights reserved. import torch import torch.nn as nn class _BatchNormXd(nn.modules.batchnorm._BatchNorm): """A general BatchNorm layer without input dimension check. Reproduced from @kapily's work: (https://github.com/pytorch/pytorch/issues/41081#issuecomment-783961547) ...
PREFIX = """Respond to the human as helpfully and accurately as possible. You have access to the following tools:""" # noqa: E501 FORMAT_INSTRUCTIONS = """Use a json blob to specify a tool by providing an action key (tool name) and an action_input key (tool input). Valid "action" values: "Final Answer" or {tool_names...
# flake8: noqa PREFIX = """Respond to the human as helpfully and accurately as possible. You have access to the following tools:""" FORMAT_INSTRUCTIONS = """Use a json blob to specify a tool by providing an action key (tool name) and an action_input key (tool input). Valid "action" values: "Final Answer" or {tool_name...
# Copyright (c) OpenMMLab. All rights reserved. import torch import torch.nn as nn import torch.nn.functional as F from mmcv.cnn import PLUGIN_LAYERS eps = 1e-6 @PLUGIN_LAYERS.register_module() class DropBlock(nn.Module): """Randomly drop some regions of feature maps. Please refer to the method proposed in...
# Copyright (c) OpenMMLab. All rights reserved. import torch import torch.nn as nn import torch.nn.functional as F from mmcv.cnn import PLUGIN_LAYERS eps = 1e-6 @PLUGIN_LAYERS.register_module() class DropBlock(nn.Module): """Randomly drop some regions of feature maps. Please refer to the method proposed in...
checkpoint_config = dict(interval=1) # yapf:disable log_config = dict( interval=50, hooks=[ dict(type='TextLoggerHook'), # dict(type='TensorboardLoggerHook') ]) # yapf:enable custom_hooks = [dict(type='NumClassCheckHook')] dist_params = dict(backend='nccl') log_level = 'INFO' load_from = No...
checkpoint_config = dict(interval=1) # yapf:disable log_config = dict( interval=50, hooks=[ dict(type='TextLoggerHook'), # dict(type='TensorboardLoggerHook') ]) # yapf:enable custom_hooks = [dict(type='NumClassCheckHook')] dist_params = dict(backend='nccl') log_level = 'INFO' load_from = No...
import logging from typing import Optional, cast from autogpt_libs.auth.models import DEFAULT_USER_ID from autogpt_libs.supabase_integration_credentials_store.types import ( UserIntegrations, UserMetadata, UserMetadataRaw, ) from fastapi import HTTPException from prisma import Json from prisma.models impor...
import logging from typing import Optional, cast from autogpt_libs.supabase_integration_credentials_store.types import ( UserIntegrations, UserMetadata, UserMetadataRaw, ) from fastapi import HTTPException from prisma import Json from prisma.models import User from backend.data.db import prisma from backe...
from typing import Dict, List import numpy as np import pytest from orjson import orjson from docarray import DocList from docarray.base_doc import AnyDoc, BaseDoc from docarray.base_doc.io.json import orjson_dumps_and_decode from docarray.typing import NdArray from docarray.typing.tensor.abstract_tensor import Abstr...
from typing import Dict, List import numpy as np import pytest from orjson import orjson from docarray import DocList from docarray.base_doc import AnyDoc, BaseDoc from docarray.base_doc.io.json import orjson_dumps_and_decode from docarray.typing import NdArray from docarray.typing.tensor.abstract_tensor import Abstr...
from contextlib import contextmanager from functools import partial from unittest.mock import patch import torch from parameterized import parameterized from torchaudio._internal.module_utils import is_module_available from torchaudio_unittest.common_utils import skipIfNoModule, TorchaudioTestCase from .utils import ...
from contextlib import contextmanager from functools import partial from unittest.mock import patch import torch from parameterized import parameterized from torchaudio._internal.module_utils import is_module_available from torchaudio_unittest.common_utils import skipIfNoModule, TorchaudioTestCase from .utils import ...
import numpy as np import torch from docarray import BaseDocument from docarray.document import AnyDocument from docarray.typing import ( AnyUrl, Embedding, ImageUrl, Mesh3DUrl, NdArray, PointCloud3DUrl, TextUrl, TorchTensor, ) def test_proto_all_types(): class Mymmdoc(BaseDocumen...
import numpy as np import torch from docarray import Document from docarray.document import AnyDocument from docarray.typing import ( AnyUrl, Embedding, ImageUrl, Mesh3DUrl, NdArray, PointCloud3DUrl, TextUrl, TorchTensor, ) def test_proto_all_types(): class Mymmdoc(Document): ...
# Copyright (c) OpenMMLab. All rights reserved. from .base_panoptic_fusion_head import \ BasePanopticFusionHead # noqa: F401,F403 from .heuristic_fusion_head import HeuristicFusionHead # noqa: F401,F403 from .maskformer_fusion_head import MaskFormerFusionHead # noqa: F401,F403
# Copyright (c) OpenMMLab. All rights reserved. from .base_panoptic_fusion_head import \ BasePanopticFusionHead # noqa: F401,F403 from .heuristic_fusion_head import HeuristicFusionHead # noqa: F401,F403
# Licensed to the LF AI & Data foundation under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the "License"); # you may not use this fil...
import os import time import uuid import numpy as np import pytest from pydantic import Field from docarray import BaseDoc from docarray.documents import ImageDoc from docarray.typing import NdArray pytestmark = [pytest.mark.slow, pytest.mark.index] cur_dir = os.path.dirname(os.path.abspath(__file__)) compose_yml_v...
from backend.blocks.nvidia._auth import ( NvidiaCredentials, NvidiaCredentialsField, NvidiaCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField from backend.util.request import Requests from backend.util.type import Medi...
from backend.blocks.nvidia._auth import ( NvidiaCredentials, NvidiaCredentialsField, NvidiaCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField from backend.util.request import Requests from backend.util.type import Medi...
# Owner(s): ["module: inductor"] import unittest import torch from torch._inductor import config from torch._inductor.test_case import run_tests, TestCase from torch.testing._internal.common_cuda import TEST_CUDA class MatMulModule(torch.nn.Module): def __init__(self): super().__init__() self.mat...
# Owner(s): ["module: inductor"] import unittest import torch from torch._inductor import config from torch._inductor.test_case import run_tests, TestCase from torch.testing._internal.common_cuda import TEST_CUDA class MatMulModule(torch.nn.Module): def __init__(self): super().__init__() self.mat...
from .rnnt_pipeline import EMFORMER_RNNT_BASE_MUSTC, EMFORMER_RNNT_BASE_TEDLIUM3 from .source_separation_pipeline import CONVTASNET_BASE_LIBRI2MIX __all__ = [ "CONVTASNET_BASE_LIBRI2MIX", "EMFORMER_RNNT_BASE_MUSTC", "EMFORMER_RNNT_BASE_TEDLIUM3", ]
from .rnnt_pipeline import EMFORMER_RNNT_BASE_MUSTC, EMFORMER_RNNT_BASE_TEDLIUM3 __all__ = [ "EMFORMER_RNNT_BASE_MUSTC", "EMFORMER_RNNT_BASE_TEDLIUM3", ]
from typing import Iterable, Dict from docarray.array.storage.annlite.helper import OffsetMapping from docarray.array.storage.base.getsetdel import BaseGetSetDelMixin from docarray.array.storage.base.helper import Offset2ID from docarray.array.memory import DocumentArrayInMemory from docarray import Document, Document...
from typing import Iterable, Dict from docarray.array.storage.annlite.helper import OffsetMapping from docarray.array.storage.base.getsetdel import BaseGetSetDelMixin from docarray.array.storage.base.helper import Offset2ID from docarray.array.memory import DocumentArrayInMemory from docarray import Document, Document...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.core import ConfigType, OptConfigType, OptMultiConfig from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class DETR(SingleStageDetector): r"""Implementation of `DETR: End-to-End Object Detection with ...
# Copyright (c) OpenMMLab. All rights reserved. import warnings import torch from mmdet.registry import MODELS from .single_stage import SingleStageDetector @MODELS.register_module() class DETR(SingleStageDetector): r"""Implementation of `DETR: End-to-End Object Detection with Transformers <https://arxiv.or...
from typing import Any, Dict from backend.data.block import Block from backend.util.request import Requests from ._api import Color, CustomerDetails, OrderItem, Profile class Slant3DBlockBase(Block): """Base block class for Slant3D API interactions""" BASE_URL = "https://www.slant3dapi.com/api" def _g...
from typing import Any, Dict from backend.data.block import Block from backend.util.request import Requests from ._api import Color, CustomerDetails, OrderItem, Profile class Slant3DBlockBase(Block): """Base block class for Slant3D API interactions""" BASE_URL = "https://www.slant3dapi.com/api" def _g...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase import numpy as np from mmengine.registry import init_default_scope from mmdet.registry import TASK_UTILS class TestInterpolateTracklets(TestCase): @classmethod def setUpClass(cls): init_default_scope('mmdet') cls...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase import numpy as np from mmdet.registry import TASK_UTILS from mmdet.utils import register_all_modules class TestInterpolateTracklets(TestCase): @classmethod def setUpClass(cls): register_all_modules() cls.cfg = di...
# coding=utf-8 # Copyright 2025 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2025 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
""" In SecGPT, all messages exchanged among spokes conform to predefined formats, encapsulated within the Message class. """ import json class Message: @staticmethod def function_probe_request(spoke_id, function): """ Create a function probe request message. Args: spoke_i...
""" In SecGPT, all messages exchanged among spokes conform to predefined formats, encapsulated within the Message class. """ import json class Message: @staticmethod def function_probe_request(spoke_id, function): """ Create a function probe request message. Args: spoke_id...
import json import os from typing import Dict from torch import Tensor, nn class Dropout(nn.Module): """Dropout layer. Args: dropout: Sets a dropout value for dense layer. """ def __init__(self, dropout: float = 0.2): super(Dropout, self).__init__() self.dropout = dropout ...
import torch from torch import Tensor from torch import nn from typing import Dict import os import json class Dropout(nn.Module): """Dropout layer. :param dropout: Sets a dropout value for dense layer. """ def __init__(self, dropout: float = 0.2): super(Dropout, self).__init__() self...
_base_ = [ '../_base_/models/cascade_mask_rcnn_r50_fpn.py', '../_base_/datasets/coco_instance.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] checkpoint = 'https://download.openmmlab.com/mmclassification/v0/resnet/resnet50_8xb256-rsb-a1-600e_in1k_20211228-20e21305.pth' # noqa m...
_base_ = [ '../_base_/models/cascade_mask_rcnn_r50_fpn.py', '../_base_/datasets/coco_instance.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] checkpoint = 'https://download.openmmlab.com/mmclassification/v0/resnet/resnet50_8xb256-rsb-a1-600e_in1k_20211228-20e21305.pth' # noqa m...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../common/lsj_100e_coco_instance.py' ] image_size = (1024, 1024) batch_augments = [dict(type='BatchFixedSizePad', size=image_size)] norm_cfg = dict(type='SyncBN', requires_grad=True) # Use MMSyncBN that handles empty tensor in head. It can be changed to # Sy...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../common/lsj_100e_coco_instance.py' ] image_size = (1024, 1024) batch_augments = [dict(type='BatchFixedSizePad', size=image_size)] norm_cfg = dict(type='SyncBN', requires_grad=True) # Use MMSyncBN that handles empty tensor in head. It can be changed to # Sy...
from typing import Optional, Type from langchain_core.callbacks import CallbackManagerForToolRun from pydantic import BaseModel, Field from langchain_community.tools.slack.base import SlackBaseTool class SendMessageSchema(BaseModel): """Input for SendMessageTool.""" message: str = Field( ..., ...
from typing import Optional, Type from langchain_core.callbacks import CallbackManagerForToolRun from pydantic import BaseModel, Field from langchain_community.tools.slack.base import SlackBaseTool class SendMessageSchema(BaseModel): """Input for SendMessageTool.""" message: str = Field( ..., ...
from .conv_emformer import ConvEmformer from .rnnt import conformer_rnnt_base, conformer_rnnt_model __all__ = [ "conformer_rnnt_base", "conformer_rnnt_model", "ConvEmformer", ]
from .rnnt import conformer_rnnt_base, conformer_rnnt_model __all__ = [ "conformer_rnnt_base", "conformer_rnnt_model", ]
from typing import Any from typing_inspect import get_args, is_union_type from docarray.typing.tensor.abstract_tensor import AbstractTensor def is_type_tensor(type_: Any) -> bool: """Return True if type is a type Tensor or an Optional Tensor type.""" return isinstance(type_, type) and issubclass(type_, Abst...
from typing import Any, get_args from typing_inspect import is_union_type from docarray.typing.tensor.abstract_tensor import AbstractTensor def is_type_tensor(type_: Any) -> bool: """Return True if type is a type Tensor or an Optional Tensor type.""" return isinstance(type_, type) and issubclass(type_, Abst...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase from mmengine.registry import MODELS from parameterized import parameterized from mmdet.testing import get_detector_cfg from mmdet.utils import register_all_modules register_all_modules() class TestSemiBase(TestCase): @parameterized...
# Copyright (c) OpenMMLab. All rights reserved. from unittest import TestCase from mmengine.registry import MODELS from parameterized import parameterized from mmdet.testing import get_detector_cfg from mmdet.utils import register_all_modules register_all_modules() class TestSemiBase(TestCase): @parameterized...
from torch.utils.data import IterableDataset import numpy as np from typing import List from ..readers import InputExample import logging logger = logging.getLogger(__name__) class SentenceLabelDataset(IterableDataset): """ This dataset can be used for some specific Triplet Losses like BATCH_HARD_TRIPLET_LOS...
""" """ from torch.utils.data import IterableDataset import numpy as np from typing import List from ..readers import InputExample import logging logger = logging.getLogger(__name__) class SentenceLabelDataset(IterableDataset): """ This dataset can be used for some specific Triplet Losses like BATCH_HARD_TR...
_base_ = './tood_r50_fpn_ms-2x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
_base_ = './tood_r50_fpn_mstrain_2x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
from typing import Dict, Type from llama_index.core.base.embeddings.base import BaseEmbedding from llama_index.core.embeddings.mock_embed_model import MockEmbedding RECOGNIZED_EMBEDDINGS: Dict[str, Type[BaseEmbedding]] = { MockEmbedding.class_name(): MockEmbedding, } # conditionals for llama-cloud support try: ...
from typing import Dict, Type from llama_index.core.base.embeddings.base import BaseEmbedding from llama_index.core.embeddings.mock_embed_model import MockEmbedding RECOGNIZED_EMBEDDINGS: Dict[str, Type[BaseEmbedding]] = { MockEmbedding.class_name(): MockEmbedding, } # conditionals for llama-cloud support try: ...
# Licensed to the LF AI & Data foundation under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the "License"); # you may not use this fil...
import os from pydantic import parse_obj_as from docarray.typing import ImageBytes, ImageTensor, ImageUrl CUR_DIR = os.path.dirname(os.path.abspath(__file__)) PATH_TO_IMAGE_DATA = os.path.join(CUR_DIR, '..', '..', 'toydata', 'image-data') IMAGE_PATHS = { 'png': os.path.join(PATH_TO_IMAGE_DATA, 'so_good.png'), ...
import gzip import logging import os from datetime import datetime from torch.utils.data import DataLoader from sentence_transformers import InputExample, LoggingHandler, SentenceTransformer, evaluation, losses, models, util #### Just some code to print debug information to stdout logging.basicConfig( format="%(...
import gzip import logging import os from datetime import datetime from torch.utils.data import DataLoader from sentence_transformers import InputExample, LoggingHandler, SentenceTransformer, evaluation, losses, models, util #### Just some code to print debug information to stdout logging.basicConfig( format="%(...
import torch from mmdet.models.task_modules import embed_similarity def test_embed_similarity(): """Test embed similarity.""" embeds = torch.rand(2, 3) similarity = embed_similarity(embeds, embeds) assert similarity.shape == (2, 2)
import torch from mmdet.models.task_modules import embed_similarity def test_embed_similarity(): """Test embed similarity.""" embeds = torch.rand(2, 3) similarity = embed_similarity(embeds, embeds) assert similarity.shape == (2, 2) assert torch.allclose(similarity, torch.eye(2))
from ._dsp import ( adsr_envelope, extend_pitch, filter_waveform, frequency_impulse_response, oscillator_bank, sinc_impulse_response, ) from .functional import barkscale_fbanks __all__ = [ "adsr_envelope", "barkscale_fbanks", "extend_pitch", "filter_waveform", "frequency_im...
from ._dsp import ( adsr_envelope, extend_pitch, filter_waveform, frequency_impulse_response, oscillator_bank, sinc_impulse_response, ) from .functional import add_noise, barkscale_fbanks, convolve, deemphasis, fftconvolve, preemphasis, speed __all__ = [ "add_noise", "adsr_envelope", ...
"""Standard LangChain interface tests""" import os from langchain_core.language_models import BaseChatModel from langchain_tests.integration_tests import ChatModelIntegrationTests from langchain_openai import AzureChatOpenAI OPENAI_API_VERSION = os.environ.get("AZURE_OPENAI_API_VERSION", "") OPENAI_API_BASE = os.en...
"""Standard LangChain interface tests""" import os from langchain_core.language_models import BaseChatModel from langchain_tests.integration_tests import ChatModelIntegrationTests from langchain_openai import AzureChatOpenAI OPENAI_API_VERSION = os.environ.get("AZURE_OPENAI_API_VERSION", "") OPENAI_API_BASE = os.en...
from backend.app import run_processes from backend.executor import DatabaseManager, ExecutionScheduler from backend.server.rest_api import AgentServer def main(): """ Run all the processes required for the AutoGPT-server REST API. """ run_processes( DatabaseManager(), ExecutionSchedule...
from backend.app import run_processes from backend.executor import ExecutionScheduler from backend.server.rest_api import AgentServer def main(): """ Run all the processes required for the AutoGPT-server REST API. """ run_processes( ExecutionScheduler(), AgentServer(), ) if __nam...
import random import torch from processing import bits_to_normalized_waveform, normalized_waveform_to_bits from torch.utils.data.dataset import random_split from torchaudio.datasets import LIBRITTS, LJSPEECH from torchaudio.transforms import MuLawEncoding class MapMemoryCache(torch.utils.data.Dataset): r"""Wrap ...
import random import torch from processing import bits_to_normalized_waveform, normalized_waveform_to_bits from torch.utils.data.dataset import random_split from torchaudio.datasets import LJSPEECH, LIBRITTS from torchaudio.transforms import MuLawEncoding class MapMemoryCache(torch.utils.data.Dataset): r"""Wrap ...
import logging from datasets import load_dataset from sentence_transformers import SparseEncoder from sentence_transformers.sparse_encoder.evaluation import SparseBinaryClassificationEvaluator logging.basicConfig(format="%(message)s", level=logging.INFO) # Initialize the SPLADE model model = SparseEncoder("naver/sp...
import logging from datasets import load_dataset from sentence_transformers import SparseEncoder from sentence_transformers.sparse_encoder.evaluation import SparseBinaryClassificationEvaluator logging.basicConfig(format="%(message)s", level=logging.INFO) # Initialize the SPLADE model model = SparseEncoder("naver/sp...
import torch from keras.src.backend import config from keras.src.backend import standardize_dtype from keras.src.backend.common import dtypes from keras.src.backend.torch.core import cast from keras.src.backend.torch.core import convert_to_tensor def cholesky(x): return torch.linalg.cholesky(x) def det(x): ...
import torch from keras.src.backend import config from keras.src.backend import standardize_dtype from keras.src.backend.common import dtypes from keras.src.backend.torch.core import cast from keras.src.backend.torch.core import convert_to_tensor def cholesky(x): return torch.linalg.cholesky(x) def det(x): ...
"""Init file of LlamaIndex.""" __version__ = "0.12.22" import logging from logging import NullHandler from typing import Callable, Optional try: # Force pants to install eval_type_backport on 3.9 import eval_type_backport # noqa # type: ignore except ImportError: pass # response from llama_index.core....
"""Init file of LlamaIndex.""" __version__ = "0.12.21" import logging from logging import NullHandler from typing import Callable, Optional try: # Force pants to install eval_type_backport on 3.9 import eval_type_backport # noqa # type: ignore except ImportError: pass # response from llama_index.core....
import argparse import json import os from datetime import date from pathlib import Path from slack_sdk import WebClient from tabulate import tabulate MAX_LEN_MESSAGE = 2900 # slack endpoint has a limit of 3001 characters parser = argparse.ArgumentParser() parser.add_argument("--slack_channel_name", default="diffu...
import argparse import json import os from datetime import date from pathlib import Path from slack_sdk import WebClient from tabulate import tabulate MAX_LEN_MESSAGE = 2900 # slack endpoint has a limit of 3001 characters parser = argparse.ArgumentParser() parser.add_argument("--slack_channel_name", default="diffu...
from langchain_core.utils.aiter import NoLock, Tee, py_anext __all__ = ["NoLock", "Tee", "py_anext"]
from langchain_core.utils.aiter import NoLock, Tee, py_anext __all__ = ["py_anext", "NoLock", "Tee"]
# Copyright (c) OpenMMLab. All rights reserved. import torch.nn as nn import torch.nn.functional as F from mmdet.registry import MODELS from .utils import weighted_loss @weighted_loss def knowledge_distillation_kl_div_loss(pred, soft_label, ...
# Copyright (c) OpenMMLab. All rights reserved. import mmcv import torch.nn as nn import torch.nn.functional as F from mmdet.registry import MODELS from .utils import weighted_loss @mmcv.jit(derivate=True, coderize=True) @weighted_loss def knowledge_distillation_kl_div_loss(pred, ...
from __future__ import annotations from sentence_transformers.sparse_encoder.data_collator import SparseEncoderDataCollator from sentence_transformers.sparse_encoder.evaluation import ( SparseBinaryClassificationEvaluator, SparseEmbeddingSimilarityEvaluator, SparseInformationRetrievalEvaluator, SparseM...
from __future__ import annotations from sentence_transformers.sparse_encoder.data_collator import SparseEncoderDataCollator from sentence_transformers.sparse_encoder.evaluation import ( SparseBinaryClassificationEvaluator, SparseEmbeddingSimilarityEvaluator, SparseInformationRetrievalEvaluator, SparseM...
from __future__ import annotations from collections.abc import Generator import torch from torch import Tensor, nn from sentence_transformers.cross_encoder import CrossEncoder from sentence_transformers.util import fullname class MultipleNegativesRankingLoss(nn.Module): def __init__( self, mode...
from __future__ import annotations from collections.abc import Generator import torch from torch import Tensor, nn from sentence_transformers.cross_encoder import CrossEncoder class MultipleNegativesRankingLoss(nn.Module): def __init__( self, model: CrossEncoder, num_negatives: int | No...
from xgboost.testing.parse_tree import ( run_split_value_histograms, run_tree_to_df_categorical, ) def test_tree_to_df_categorical() -> None: run_tree_to_df_categorical("hist", "cuda") def test_split_value_histograms() -> None: run_split_value_histograms("hist", "cuda")
import sys sys.path.append("tests/python") from test_parse_tree import TestTreesToDataFrame def test_tree_to_df_categorical(): cputest = TestTreesToDataFrame() cputest.run_tree_to_df_categorical("gpu_hist") def test_split_value_histograms(): cputest = TestTreesToDataFrame() cputest.run_split_value_...
from typing import Dict, Optional, Sequence import torch from jina import DocumentArray, Executor, requests from jina_commons.batching import get_docs_batch_generator from transformers import CLIPModel, CLIPTokenizer class CLIPTextEncoder(Executor): """Encode text into embeddings using the CLIP model.""...
from typing import Dict, Optional, Sequence import torch from jina import DocumentArray, Executor, requests from jina_commons.batching import get_docs_batch_generator from transformers import CLIPModel, CLIPTokenizer class CLIPTextEncoder(Executor): """Encode text into embeddings using a CLIP model. ...
_base_ = './fovea_r50_fpn_4xb4-2x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
_base_ = './fovea_r50_fpn_4x4_2x_coco.py' model = dict( backbone=dict( depth=101, init_cfg=dict(type='Pretrained', checkpoint='torchvision://resnet101')))
# Copyright 2024 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
# Copyright 2024 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
from __future__ import annotations import collections import json import os import string from typing import Iterable from .WordTokenizer import ENGLISH_STOP_WORDS, WordTokenizer class WhitespaceTokenizer(WordTokenizer): """ Simple and fast white-space tokenizer. Splits sentence based on white spaces. P...
from __future__ import annotations import collections import json import os import string from typing import Iterable from .WordTokenizer import ENGLISH_STOP_WORDS, WordTokenizer class WhitespaceTokenizer(WordTokenizer): """ Simple and fast white-space tokenizer. Splits sentence based on white spaces. P...
from io import BytesIO from typing import TYPE_CHECKING, Any, Optional, Tuple, Type, TypeVar import numpy as np from pydantic import parse_obj_as from pydantic.validators import bytes_validator from docarray.typing.abstract_type import AbstractType from docarray.typing.proto_register import _register_proto from docar...
from io import BytesIO from typing import TYPE_CHECKING, Any, Optional, Tuple, Type, TypeVar import numpy as np from pydantic import parse_obj_as from pydantic.validators import bytes_validator from docarray.typing.abstract_type import AbstractType from docarray.typing.proto_register import _register_proto from docar...
from __future__ import annotations import torch.nn as nn from sentence_transformers.losses.CosineSimilarityLoss import CosineSimilarityLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseCosineSimilarityLoss(CosineSimilarityLoss): def __init__( self, mod...
from __future__ import annotations import torch.nn as nn from sentence_transformers.losses.CosineSimilarityLoss import CosineSimilarityLoss from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class SparseCosineSimilarityLoss(CosineSimilarityLoss): def __init__( self, mod...
import types from typing import TYPE_CHECKING from docarray.utils._internal.misc import ( _get_path_from_docarray_root_level, import_library, ) if TYPE_CHECKING: from docarray.index.backends.elastic import ElasticV7DocIndex # noqa: F401 from docarray.index.backends.hnswlib import HnswDocumentIndex #...
from docarray.index.backends.elastic import ElasticV7DocIndex from docarray.index.backends.hnswlib import HnswDocumentIndex __all__ = ['HnswDocumentIndex', 'ElasticV7DocIndex']
import time import unittest from parameterized import parameterized from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig from transformers.testing_utils import require_flash_attn, require_torch_gpu, slow _TEST_PROMPTS = [ "A man is a walking his dog down the street, and a the turn he s...
import time import unittest from parameterized import parameterized from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig from transformers.testing_utils import require_flash_attn, require_torch_gpu, run_slow _TEST_PROMPTS = [ "A man is a walking his dog down the street, and a the turn ...
__version__ = '0.14.9' import os from docarray.document import Document from docarray.array import DocumentArray from docarray.dataclasses import dataclass, field if 'DA_RICH_HANDLER' in os.environ: from rich.traceback import install install()
__version__ = '0.14.8' import os from docarray.document import Document from docarray.array import DocumentArray from docarray.dataclasses import dataclass, field if 'DA_RICH_HANDLER' in os.environ: from rich.traceback import install install()
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import os import numpy as np from jina import Document, Flow, DocumentArray from ...custom_image_torch_encoder import CustomImageTorchEncoder cur_dir = os.path.dirname(os.path.abspath(__file__)) def test_vi...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import os import numpy as np from jina import Document, Flow, DocumentArray try: from custom_image_torch_encoder import CustomImageTorchEncoder except: from jinahub.encoder.custom_image_torch_encoder i...
"""Standard LangChain interface tests""" import os from langchain_core.language_models import BaseChatModel from langchain_tests.integration_tests import ChatModelIntegrationTests from langchain_openai import AzureChatOpenAI OPENAI_API_VERSION = os.environ.get("AZURE_OPENAI_API_VERSION", "") OPENAI_API_BASE = os.en...
"""Standard LangChain interface tests""" import os from langchain_core.language_models import BaseChatModel from langchain_tests.integration_tests import ChatModelIntegrationTests from langchain_openai import AzureChatOpenAI OPENAI_API_VERSION = os.environ.get("AZURE_OPENAI_API_VERSION", "") OPENAI_API_BASE = os.en...
from docarray.typing.embedding import Embedding from docarray.typing.id import ID from docarray.typing.tensor import Tensor from docarray.typing.url import AnyUrl, ImageUrl __all__ = ['Tensor', 'Embedding', 'ImageUrl', 'AnyUrl', 'ID']
from docarray.document.base_node import BaseNode from docarray.typing.ndarray import Embedding, Tensor from docarray.typing.url import ImageUrl __all__ = ['Tensor', 'Embedding', 'BaseNode', 'ImageUrl']
_base_ = [ '../_base_/default_runtime.py', '../_base_/datasets/coco_detection.py' ] data_preprocessor = dict( type='DetDataPreprocessor', mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], bgr_to_rgb=True) # model settings model = dict( type='CornerNet', data_preprocessor=data_pr...
_base_ = [ '../_base_/default_runtime.py', '../_base_/datasets/coco_detection.py' ] data_preprocessor = dict( type='DetDataPreprocessor', mean=[123.675, 116.28, 103.53], std=[58.395, 57.12, 57.375], bgr_to_rgb=True) # model settings model = dict( type='CornerNet', data_preprocessor=data_pr...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.ops.nn import average_pool from keras.src.ops.nn import batch_normalization from keras.src.ops.nn import binary_crossentropy from keras.src.ops.nn import categorical_crossentropy from...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.ops.nn import average_pool from keras.src.ops.nn import batch_normalization from keras.src.ops.nn import binary_crossentropy from keras.src.ops.nn import categorical_crossentropy from...
from langchain_core.tracers import schemas from langchain_core.tracers.schemas import __all__ as schemas_all def test_public_api() -> None: """Test for changes in the public API.""" expected_all = [ "BaseRun", "ChainRun", "LLMRun", "Run", "RunTypeEnum", "ToolRun...
import langchain_core.tracers.schemas as schemas from langchain_core.tracers.schemas import __all__ as schemas_all def test_public_api() -> None: """Test for changes in the public API.""" expected_all = [ "BaseRun", "ChainRun", "LLMRun", "Run", "RunTypeEnum", "T...
import numpy as np from docarray.proto import DocumentProto, NodeProto from docarray.typing import NdArray def test_nested_item_proto(): NodeProto(text='hello') NodeProto(nested=DocumentProto()) def test_nested_optional_item_proto(): NodeProto() def test_ndarray(): original_ndarray = np.zeros((3...
import numpy as np from docarray.proto import DocumentProto, NodeProto from docarray.typing import NdArray def test_nested_item_proto(): NodeProto(text='hello') NodeProto(nested=DocumentProto()) def test_nested_optional_item_proto(): NodeProto() def test_ndarray(): original_ndarray = np.zeros((3...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from typing import Dict, Iterable, List, Union import numpy as np import tensorflow as tf from jina import DocumentArray, Executor, requests from jina.logging.logger import JinaLogger from jina_commons.batching ...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from typing import Dict, Iterable, List, Union import numpy as np import tensorflow as tf from jina import DocumentArray, Executor, requests from jina.logging.logger import JinaLogger from jina_commons.batching ...
import logging import os from argparse import ArgumentParser import sentencepiece as spm from average_checkpoints import ensemble from pytorch_lightning import seed_everything, Trainer from pytorch_lightning.callbacks import LearningRateMonitor, ModelCheckpoint from pytorch_lightning.strategies import DDPStrategy from...
import logging import os from argparse import ArgumentParser import sentencepiece as spm from average_checkpoints import ensemble from pytorch_lightning import seed_everything, Trainer from pytorch_lightning.callbacks import LearningRateMonitor, ModelCheckpoint from pytorch_lightning.strategies import DDPStrategy from...
import sys from os import path from setuptools import find_packages from setuptools import setup if sys.version_info < (3, 7, 0): raise OSError(f'DocArray requires Python >=3.7, but yours is {sys.version}') try: pkg_name = 'docarray' libinfo_py = path.join(pkg_name, '__init__.py') libinfo_content = o...
import sys from os import path from setuptools import find_packages from setuptools import setup if sys.version_info < (3, 7, 0): raise OSError(f'DocArray requires Python >=3.7, but yours is {sys.version}') try: pkg_name = 'docarray' libinfo_py = path.join(pkg_name, '__init__.py') libinfo_content = o...
from langchain_core.tracers.langchain import ( LangChainTracer, get_client, log_error_once, wait_for_all_tracers, ) __all__ = ["LangChainTracer", "get_client", "log_error_once", "wait_for_all_tracers"]
from langchain_core.tracers.langchain import ( LangChainTracer, get_client, log_error_once, wait_for_all_tracers, ) __all__ = ["log_error_once", "wait_for_all_tracers", "get_client", "LangChainTracer"]
_base_ = '../mask_rcnn/mask-rcnn_r50_fpn_1x_coco.py' norm_cfg = dict(type='GN', num_groups=32, requires_grad=True) model = dict( backbone=dict( norm_cfg=norm_cfg, init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://contrib/resnet50_gn')), neck=dict(norm_cfg=norm_cfg), roi_...
_base_ = '../mask_rcnn/mask_rcnn_r50_fpn_1x_coco.py' norm_cfg = dict(type='GN', num_groups=32, requires_grad=True) model = dict( backbone=dict( norm_cfg=norm_cfg, init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://contrib/resnet50_gn')), neck=dict(norm_cfg=norm_cfg), roi_...
from __future__ import annotations from collections.abc import Iterable import torch from torch import Tensor, nn from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class FlopsLoss(nn.Module): def __init__(self, model: SparseEncoder, threshold: float = None) -> None: """ ...
from __future__ import annotations from collections.abc import Iterable import torch from torch import Tensor, nn from sentence_transformers.sparse_encoder.SparseEncoder import SparseEncoder class FlopsLoss(nn.Module): def __init__(self, model: SparseEncoder, threshold: float = None) -> None: """ ...
# Copyright 2019 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
# Copyright 2019 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applica...
""" Given a dataset with parallel sentences, one "english" column and one "non_english" column, this script evaluates a model on the translation task. Given a sentence in the "english" column, the model should find the correct translation in the "non_english" column, based on just the embeddings. It then computes an a...
""" Given a dataset with parallel sentences, one "english" column and one "non_english" column, this script evaluates a model on the translation task. Given a sentence in the "english" column, the model should find the correct translation in the "non_english" column, based on just the embeddings. It then computes an a...
# pylint: disable=invalid-name,unused-import """For compatibility and optional dependencies.""" import importlib.util import logging import sys import types from typing import Any, Sequence, cast import numpy as np from ._typing import _T assert sys.version_info[0] == 3, "Python 2 is no longer supported." def py_s...
# pylint: disable=invalid-name,unused-import """For compatibility and optional dependencies.""" import importlib.util import logging import sys import types from typing import Any, Sequence, cast import numpy as np from ._typing import _T assert sys.version_info[0] == 3, "Python 2 is no longer supported." def py_s...