input
stringlengths
33
5k
output
stringlengths
32
5k
""" This script translates the queries in the MS MARCO dataset to the defined target languages. For machine translation, we use EasyNMT: https://github.com/UKPLab/EasyNMT You can install it via: pip install easynmt Usage: python translate_queries [target_language] """ import logging import os import sys import tarfi...
""" This script translates the queries in the MS MARCO dataset to the defined target languages. For machine translation, we use EasyNMT: https://github.com/UKPLab/EasyNMT You can install it via: pip install easynmt Usage: python translate_queries [target_language] """ import os from sentence_transformers import Logg...
"""Wordpress reader.""" import warnings from typing import List, Optional from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class WordpressReader(BaseReader): """ Wordpress reader. Reads data from a Wordpress workspace. Args: url (str): Base URL o...
"""Wordpress reader.""" import warnings from typing import List, Optional from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class WordpressReader(BaseReader): """ Wordpress reader. Reads data from a Wordpress workspace. Args: url (str): Base URL of...
_base_ = './mask_rcnn_r101_fpn_gn_ws-all_2x_coco.py' # learning policy max_epochs = 24 train_cfg = dict(max_epochs=max_epochs) # learning rate param_scheduler = [ dict( type='LinearLR', start_factor=0.001, by_epoch=False, begin=0, end=500), dict( type='MultiStepLR', begin=0, end...
_base_ = './mask_rcnn_r101_fpn_gn_ws-all_2x_coco.py' # learning policy lr_config = dict(step=[20, 23]) runner = dict(type='EpochBasedRunner', max_epochs=24)
from utils import foo from jina import Executor class DummyHubExecutorAbs(Executor): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) foo()
from helper import foo from jina import Executor class DummyHubExecutorAbs(Executor): def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) foo()
_base_ = './detic_centernet2_r50_fpn_4x_lvis_boxsup.py' dataset_type = ['LVISV1Dataset', 'ImageNetLVISV1Dataset'] image_size_det = (640, 640) image_size_cls = (320, 320) # backend = 'pillow' backend_args = None train_pipeline_det = [ dict(type='LoadImageFromFile', backend_args=backend_args), dict(type='LoadAn...
_base_ = './detic_centernet2_r50_fpn_4x_lvis_boxsup.py' image_size_det = (640, 640) image_size_cls = (320, 320) # backend = 'pillow' backend_args = None train_pipeline_det = [ dict(type='LoadImageFromFile', backend_args=backend_args), dict(type='LoadAnnotations', with_bbox=True, with_mask=True), dict( ...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import subprocess from typing import Dict, Iterable, Optional import spacy from jina import DocumentArray, Executor, requests from jina_commons.batching import get_docs_batch_generator _EXCLUDE_COMPONENTS = [ ...
__copyright__ = "Copyright (c) 2020-2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" import subprocess from typing import Dict, List, Optional import spacy from jina import DocumentArray, Executor, requests from jina_commons.batching import get_docs_batch_generator _EXCLUDE_COMPONENTS = [ '...
"""Simple file node parser.""" from typing import Any, Dict, List, Optional, Sequence, Type from llama_index.core.callbacks.base import CallbackManager from llama_index.core.node_parser.node_utils import build_nodes_from_splits from llama_index.core.node_parser.file.html import HTMLNodeParser from llama_index.core.nod...
"""Simple file node parser.""" from typing import Any, Dict, List, Optional, Sequence, Type from llama_index.core.callbacks.base import CallbackManager from llama_index.core.node_parser.node_utils import build_nodes_from_splits from llama_index.core.node_parser.file.html import HTMLNodeParser from llama_index.core.nod...
from llama_index.core.base.llms.types import ( ChatMessage, ChatResponse, ChatResponseAsyncGen, ChatResponseGen, CompletionResponse, CompletionResponseAsyncGen, CompletionResponseGen, ImageBlock, LLMMetadata, MessageRole, TextBlock, AudioBlock, DocumentBlock, Cach...
from llama_index.core.base.llms.types import ( ChatMessage, ChatResponse, ChatResponseAsyncGen, ChatResponseGen, CompletionResponse, CompletionResponseAsyncGen, CompletionResponseGen, ImageBlock, LLMMetadata, MessageRole, TextBlock, AudioBlock, DocumentBlock, Cach...
from pydantic import BaseModel from inspect import Signature, Parameter from typing import Any, Dict, Optional, List, Callable from llama_index.core.llms import ChatMessage, AudioBlock, TextBlock, MessageRole from llama_index.core.tools import BaseTool def make_function_from_tool_model( model_cls: type[...
from typing import Any, Dict, Optional, List from llama_index.core.llms import ChatMessage, AudioBlock, TextBlock, MessageRole def callback_user_message( messages: Dict[int, List[ChatMessage]], message_id: int, text: Optional[str] = None, audio: Optional[Any] = None, ) -> None: if messag...
from __future__ import annotations import json import logging import re from re import Pattern from typing import Optional, Union from langchain_core.agents import AgentAction, AgentFinish from langchain_core.exceptions import OutputParserException from langchain_core.language_models import BaseLanguageModel from pyd...
from __future__ import annotations import json import logging import re from re import Pattern from typing import Optional, Union from langchain_core.agents import AgentAction, AgentFinish from langchain_core.exceptions import OutputParserException from langchain_core.language_models import BaseLanguageModel from pyd...
import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder CI_HUB_USER = "__DUMMY_TRANSFORMERS_USER__" CI_HUB_USER_FULL_NAME = "Dummy User" CI_HUB_USER_TOKEN = "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" CI_HUB_ENDPOINT = "...
import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder from datasets.utils._hf_hub_fixes import create_repo, delete_repo CI_HUB_USER = "__DUMMY_TRANSFORMERS_USER__" CI_HUB_USER_FULL_NAME = "Dummy User" CI_HUB_USER_T...
import csv import os from pathlib import Path from typing import Tuple, Union import torchaudio from torch import Tensor from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import extract_archive _RELEASE_CONFIGS = { "release1": { "folder_in_arch...
import csv import os from pathlib import Path from typing import Tuple, Union import torchaudio from torch import Tensor from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import extract_archive _RELEASE_CONFIGS = { "release1": { "folder_in_arch...
"""This file only exists to be lazy-imported and avoid V2-related import warnings when just using V1.""" import torch from torchvision import datapoints from torchvision.transforms import v2 class PadIfSmaller(v2.Transform): def __init__(self, size, fill=0): super().__init__() self.size = size ...
"""This file only exists to be lazy-imported and avoid V2-related import warnings when just using V1.""" import torch from torchvision import datapoints from torchvision.transforms import v2 class PadIfSmaller(v2.Transform): def __init__(self, size, fill=0): super().__init__() self.size = size ...
# Copyright 2025 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
"""Question-answering with sources over a vector database.""" import warnings from typing import Any from langchain_core.callbacks import ( AsyncCallbackManagerForChainRun, CallbackManagerForChainRun, ) from langchain_core.documents import Document from langchain_core.vectorstores import VectorStore from pyda...
"""Question-answering with sources over a vector database.""" import warnings from typing import Any from langchain_core.callbacks import ( AsyncCallbackManagerForChainRun, CallbackManagerForChainRun, ) from langchain_core.documents import Document from langchain_core.vectorstores import VectorStore from pyda...
from langchain_exa import ( ExaSearchResults, # type: ignore[import-not-found, import-not-found] ) def test_search_tool() -> None: tool = ExaSearchResults() res = tool.invoke({"query": "best time to visit japan", "num_results": 5}) print(res) # noqa: T201 assert not isinstance(res, str) # str m...
from langchain_exa import ( ExaSearchResults, # type: ignore[import-not-found, import-not-found] ) def test_search_tool() -> None: tool = ExaSearchResults() res = tool.invoke({"query": "best time to visit japan", "num_results": 5}) print(res) # noqa: T201 assert not isinstance(res, str) # str m...
from llama_index.core.base.embeddings.base import BaseEmbedding from llama_index.embeddings.ollama import OllamaEmbedding def test_embedding_class(): emb = OllamaEmbedding( model_name="", client_kwargs={"headers": {"Authorization": "Bearer token"}} ) assert isinstance(emb, BaseEmbedding)
from llama_index.core.base.embeddings.base import BaseEmbedding from llama_index.embeddings.ollama import OllamaEmbedding def test_embedding_class(): emb = OllamaEmbedding(model_name="") assert isinstance(emb, BaseEmbedding)
# Copyright (c) OpenMMLab. All rights reserved. from .gaussian_target import (gather_feat, gaussian_radius, gen_gaussian_target, get_local_maximum, get_topk_from_heatmap, transpose_and_gather_feat) from .image import imrenormalize from .make_divisible import m...
# Copyright (c) OpenMMLab. All rights reserved. from .gaussian_target import (gather_feat, gaussian_radius, gen_gaussian_target, get_local_maximum, get_topk_from_heatmap, transpose_and_gather_feat) from .image import imrenormalize from .make_divisible import m...
import logging import sys import traceback from datasets import Dataset, load_dataset from peft import LoraConfig, TaskType from sentence_transformers import ( SentenceTransformer, SentenceTransformerModelCardData, SentenceTransformerTrainer, SentenceTransformerTrainingArguments, ) from sentence_trans...
import logging import sys import traceback from datasets import Dataset, load_dataset from peft import LoraConfig, TaskType from sentence_transformers import ( SentenceTransformer, SentenceTransformerModelCardData, SentenceTransformerTrainer, SentenceTransformerTrainingArguments, ) from sentence_trans...
from ..utils import is_torch_available if is_torch_available(): from .faster_cache import FasterCacheConfig, apply_faster_cache from .group_offloading import apply_group_offloading from .hooks import HookRegistry, ModelHook from .layerwise_casting import apply_layerwise_casting, apply_layerwise_castin...
from ..utils import is_torch_available if is_torch_available(): from .group_offloading import apply_group_offloading from .hooks import HookRegistry, ModelHook from .layerwise_casting import apply_layerwise_casting, apply_layerwise_casting_hook from .pyramid_attention_broadcast import PyramidAttention...
""" This file runs Masked Language Model. You provide a training file. Each line is interpreted as a sentence / paragraph. Optionally, you can also provide a dev file. The fine-tuned model is stored in the output/model_name folder. Usage: python train_mlm.py model_name data/train_sentences.txt [data/dev_sentences.txt...
""" This file runs Masked Language Model. You provide a training file. Each line is interpreted as a sentence / paragraph. Optionally, you can also provide a dev file. The fine-tuned model is stored in the output/model_name folder. Usage: python train_mlm.py model_name data/train_sentences.txt [data/dev_sentences.txt...
""" This file loads sentences from a provided text file. It is expected, that the there is one sentence per line in that text file. CT will be training using these sentences. Checkpoints are stored every 500 steps to the output folder. Usage: python train_ct_from_file.py path/to/sentences.txt """ import math from s...
""" This file loads sentences from a provided text file. It is expected, that the there is one sentence per line in that text file. CT will be training using these sentences. Checkpoints are stored every 500 steps to the output folder. Usage: python train_ct_from_file.py path/to/sentences.txt """ import math from se...
from typing import List from pydantic import BaseModel from backend.blocks.exa._auth import ( ExaCredentials, ExaCredentialsField, ExaCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField from backend.util.request impor...
from typing import List, Optional from pydantic import BaseModel from backend.blocks.exa._auth import ( ExaCredentials, ExaCredentialsField, ExaCredentialsInput, ) from backend.data.block import Block, BlockCategory, BlockOutput, BlockSchema from backend.data.model import SchemaField from backend.util.req...
_base_ = './htc_x101_32x4d_fpn_16x1_20e_coco.py' model = dict( backbone=dict( type='ResNeXt', groups=64, init_cfg=dict( type='Pretrained', checkpoint='open-mmlab://resnext101_64x4d')))
_base_ = './htc_r50_fpn_1x_coco.py' model = dict( backbone=dict( type='ResNeXt', depth=101, groups=64, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), norm_eval=True, ...
from torchaudio._internal import module_utils as _mod_utils from . import ffmpeg_utils, sox_utils from .download import download_asset if _mod_utils.is_sox_available(): sox_utils.set_verbosity(0) __all__ = [ "download_asset", "sox_utils", "ffmpeg_utils", ]
from torchaudio._internal import module_utils as _mod_utils from . import sox_utils from .download import download_asset if _mod_utils.is_sox_available(): sox_utils.set_verbosity(0) __all__ = [ "download_asset", "sox_utils", ]
import os from pathlib import Path from typing import List, Tuple, Union import torchaudio from torch import Tensor from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import extract_archive _RELEASE_CONFIGS = { "release1": { "folder_in_archive":...
import os from pathlib import Path from typing import List, Tuple, Union import torchaudio from torch import Tensor from torch.hub import download_url_to_file from torch.utils.data import Dataset from torchaudio.datasets.utils import extract_archive _RELEASE_CONFIGS = { "release1": { "folder_in_archive":...
import os class Settings: def __init__(self): self.JWT_SECRET_KEY: str = os.getenv("SUPABASE_JWT_SECRET", "") self.ENABLE_AUTH: bool = os.getenv("ENABLE_AUTH", "false").lower() == "true" self.JWT_ALGORITHM: str = "HS256" @property def is_configured(self) -> bool: return bo...
import os from dotenv import load_dotenv load_dotenv() class Settings: JWT_SECRET_KEY: str = os.getenv("SUPABASE_JWT_SECRET", "") ENABLE_AUTH: bool = os.getenv("ENABLE_AUTH", "false").lower() == "true" JWT_ALGORITHM: str = "HS256" @property def is_configured(self) -> bool: return bool(s...
# Copyright 2025 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# Copyright (c) OpenMMLab. All rights reserved. from .anchor_free_head import AnchorFreeHead from .anchor_head import AnchorHead from .atss_head import ATSSHead from .atss_vlfusion_head import ATSSVLFusionHead from .autoassign_head import AutoAssignHead from .boxinst_head import BoxInstBboxHead, BoxInstMaskHead from .c...
# Copyright (c) OpenMMLab. All rights reserved. from .anchor_free_head import AnchorFreeHead from .anchor_head import AnchorHead from .atss_head import ATSSHead from .atss_vlfusion_head import ATSSVLFusionHead from .autoassign_head import AutoAssignHead from .boxinst_head import BoxInstBboxHead, BoxInstMaskHead from .c...
from setuptools import setup, find_packages with open("README.md", mode="r", encoding="utf-8") as readme_file: readme = readme_file.read() setup( name="sentence-transformers", version="2.6.0.dev0", author="Nils Reimers", author_email="info@nils-reimers.de", description="Multilingual text embe...
from setuptools import setup, find_packages with open("README.md", mode="r", encoding="utf-8") as readme_file: readme = readme_file.read() setup( name="sentence-transformers", version="2.5.0.dev0", author="Nils Reimers", author_email="info@nils-reimers.de", description="Multilingual text embe...
from dataclasses import dataclass, field from typing import Any, Dict, Type import pytest from pydantic import Field from docarray import BaseDoc from docarray.index.abstract import BaseDocIndex from docarray.typing import NdArray pytestmark = pytest.mark.index class SimpleDoc(BaseDoc): tens: NdArray[10] = Fie...
from dataclasses import dataclass, field from typing import Any, Dict, Type import pytest from pydantic import Field from docarray import BaseDoc from docarray.index.abstract import BaseDocIndex from docarray.typing import NdArray pytestmark = pytest.mark.index class SimpleDoc(BaseDoc): tens: NdArray[10] = Fie...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import os from collections import Sequence from pathlib import Path import mmcv import numpy as np from mmcv import Config, DictAction from mmdet.core.utils import mask2ndarray from mmdet.core.visualization import imshow_det_bboxes from mmdet.datasets.bu...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import os from collections import Sequence from pathlib import Path import mmcv from mmcv import Config, DictAction from mmdet.core.utils import mask2ndarray from mmdet.core.visualization import imshow_det_bboxes from mmdet.datasets.builder import build_...
from .vggish_audio_encoder import VggishAudioEncoder
from .vggish_audio_encoder import VggishAudioEncoder
"""Question-answering with sources over a vector database.""" import warnings from typing import Any from langchain_core.callbacks import ( AsyncCallbackManagerForChainRun, CallbackManagerForChainRun, ) from langchain_core.documents import Document from langchain_core.vectorstores import VectorStore from pyda...
"""Question-answering with sources over a vector database.""" import warnings from typing import Any from langchain_core.callbacks import ( AsyncCallbackManagerForChainRun, CallbackManagerForChainRun, ) from langchain_core.documents import Document from langchain_core.vectorstores import VectorStore from pyda...
""" Example of training with Dask on GPU ==================================== """ import cupy as cp import dask_cudf from dask import array as da from dask import dataframe as dd from dask.distributed import Client from dask_cuda import LocalCUDACluster from xgboost import dask as dxgb from xgboost.dask import DaskDMa...
""" Example of training with Dask on GPU ==================================== """ import dask_cudf from dask import array as da from dask import dataframe as dd from dask.distributed import Client from dask_cuda import LocalCUDACluster from xgboost import dask as dxgb from xgboost.dask import DaskDMatrix def using_d...
# Licensed to the LF AI & Data foundation under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the "License"); # you may not use this fil...
import numpy as np import pytest from pydantic import parse_obj_as from docarray.computation.numpy_backend import NumpyCompBackend from docarray.typing import NdArray def test_to_device(): with pytest.raises(NotImplementedError): NumpyCompBackend.to_device(np.random.rand(10, 3), 'meta') @pytest.mark.pa...
# coding=utf-8 # Copyright 2025 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
# coding=utf-8 # Copyright 2024 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
from __future__ import annotations try: from typing import Self except ImportError: from typing_extensions import Self import torch.nn.functional as F from torch import Tensor from sentence_transformers.models.Module import Module class Normalize(Module): """This layer normalizes embeddings to unit len...
from __future__ import annotations import torch.nn.functional as F from torch import Tensor, nn class Normalize(nn.Module): """This layer normalizes embeddings to unit length""" def __init__(self) -> None: super().__init__() def forward(self, features: dict[str, Tensor]) -> dict[str, Tensor]: ...
import numpy as np import pytest from pydantic.tools import parse_obj_as, schema_json_of from docarray.base_doc.io.json import orjson_dumps from docarray.typing import Mesh3DUrl, NdArray from tests import TOYDATA_DIR MESH_FILES = { 'obj': str(TOYDATA_DIR / 'tetrahedron.obj'), 'glb': str(TOYDATA_DIR / 'test.gl...
import numpy as np import pytest from pydantic.tools import parse_obj_as, schema_json_of from docarray.base_document.io.json import orjson_dumps from docarray.typing import Mesh3DUrl, NdArray from tests import TOYDATA_DIR MESH_FILES = { 'obj': str(TOYDATA_DIR / 'tetrahedron.obj'), 'glb': str(TOYDATA_DIR / 'te...
import pytest from langchain_core.utils.env import get_from_dict_or_env def test_get_from_dict_or_env() -> None: assert ( get_from_dict_or_env( { "a": "foo", }, ["a"], "__SOME_KEY_IN_ENV", ) == "foo" ) assert ( ...
import pytest from langchain_core.utils.env import get_from_dict_or_env def test_get_from_dict_or_env() -> None: assert ( get_from_dict_or_env( { "a": "foo", }, ["a"], "__SOME_KEY_IN_ENV", ) == "foo" ) assert ( ...
""" Quantile Regression =================== .. versionadded:: 2.0.0 The script is inspired by this awesome example in sklearn: https://scikit-learn.org/stable/auto_examples/ensemble/plot_gradient_boosting_quantile.html .. note:: The feature is only supported using the Python package. In addition, quantile ...
""" Quantile Regression =================== .. versionadded:: 2.0.0 The script is inspired by this awesome example in sklearn: https://scikit-learn.org/stable/auto_examples/ensemble/plot_gradient_boosting_quantile.html """ import argparse from typing import Dict import numpy as np from sklearn.model_selection i...
from __future__ import annotations from typing import Iterable import torch.nn.functional as F from torch import Tensor, nn from sentence_transformers.SentenceTransformer import SentenceTransformer from .ContrastiveLoss import SiameseDistanceMetric class OnlineContrastiveLoss(nn.Module): def __init__( ...
from __future__ import annotations from typing import Iterable import torch.nn.functional as F from torch import Tensor, nn from sentence_transformers.SentenceTransformer import SentenceTransformer from .ContrastiveLoss import SiameseDistanceMetric class OnlineContrastiveLoss(nn.Module): def __init__( ...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.initializers import deserialize from keras.src.initializers import get from keras.src.initializers import serialize from keras.src.initializers.constant_initializers import Constant f...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.initializers import deserialize from keras.src.initializers import get from keras.src.initializers import serialize from keras.src.initializers.constant_initializers import Constant f...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.saving.file_editor import KerasFileEditor as KerasFileEditor from keras.src.saving.object_registration import ( CustomObjectScope as CustomObjectScope, ) from keras.src.saving.obj...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.saving.file_editor import KerasFileEditor from keras.src.saving.object_registration import CustomObjectScope from keras.src.saving.object_registration import ( CustomObjectScope a...
"""Weaviate Sub-Question Query Engine Pack.""" from typing import Any, Dict, List, Optional from llama_index.core.indices.vector_store import VectorStoreIndex from llama_index.core.llama_pack.base import BaseLlamaPack from llama_index.core.query_engine import SubQuestionQueryEngine from llama_index.core.schema import...
"""Weaviate Sub-Question Query Engine Pack.""" from typing import Any, Dict, List, Optional from llama_index.core.indices.vector_store import VectorStoreIndex from llama_index.core.llama_pack.base import BaseLlamaPack from llama_index.core.query_engine import SubQuestionQueryEngine from llama_index.core.schema impor...
from __future__ import annotations import re import pytest from sentence_transformers import SentenceTransformer from sentence_transformers.evaluation import NanoBEIREvaluator from sentence_transformers.util import is_datasets_available from tests.utils import is_ci if not is_datasets_available(): pytest.skip( ...
from __future__ import annotations import re import pytest from sentence_transformers import SentenceTransformer from sentence_transformers.evaluation import NanoBEIREvaluator from sentence_transformers.util import is_datasets_available from tests.utils import is_ci if not is_datasets_available(): pytest.skip( ...
""" This is a more complex example on performing clustering on large scale dataset. This examples find in a large set of sentences local communities, i.e., groups of sentences that are highly similar. You can freely configure the threshold what is considered as similar. A high threshold will only find extremely simila...
""" This is a more complex example on performing clustering on large scale dataset. This examples find in a large set of sentences local communities, i.e., groups of sentences that are highly similar. You can freely configure the threshold what is considered as similar. A high threshold will only find extremely simila...
"""Standard LangChain interface tests""" from langchain_core.language_models import BaseChatModel from langchain_tests.integration_tests import ( # type: ignore[import-not-found] ChatModelIntegrationTests, # type: ignore[import-not-found] ) from langchain_mistralai import ChatMistralAI class TestMistralStanda...
"""Standard LangChain interface tests""" from typing import Type from langchain_core.language_models import BaseChatModel from langchain_tests.integration_tests import ( # type: ignore[import-not-found] ChatModelIntegrationTests, # type: ignore[import-not-found] ) from langchain_mistralai import ChatMistralAI ...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( type='CenterNet', # use caffe img_norm data_preprocessor=dict( type='DetDataPreprocessor', mean=[103.530, 116.280, 123.675], std=[1.0, 1.0,...
_base_ = [ '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_1x.py', '../_base_/default_runtime.py' ] model = dict( type='CenterNet', # use caffe img_norm data_preprocessor=dict( type='DetDataPreprocessor', mean=[103.530, 116.280, 123.675], std=[1.0, 1.0,...
"""Flat reader.""" from fsspec import AbstractFileSystem from fsspec.implementations.local import LocalFileSystem from pathlib import Path from typing import Any, Dict, List, Optional from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class FlatReader(BaseReader): "...
"""Flat reader.""" from fsspec import AbstractFileSystem from fsspec.implementations.local import LocalFileSystem from pathlib import Path from typing import Any, Dict, List, Optional from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class FlatReader(BaseReader): "...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Optional, Sequence, Union from torch import Tensor from mmdet.registry import TASK_UTILS from mmdet.structures.bbox import (BaseBoxes, HorizontalBoxes, bbox2distance, distance2bbox, get_box_tensor) from .base_bbox_co...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import TASK_UTILS from mmdet.structures.bbox import (HorizontalBoxes, bbox2distance, distance2bbox, get_box_tensor) from .base_bbox_coder import BaseBBoxCoder @TASK_UTILS.register_module() class DistancePointBBoxCod...
import os import sys import pkg_resources from setuptools import setup, find_packages def read_version(fname="whisper/version.py"): exec(compile(open(fname, encoding="utf-8").read(), fname, "exec")) return locals()["__version__"] requirements = [] if sys.platform.startswith("linux"): triton_requirement...
import os import pkg_resources from setuptools import setup, find_packages def read_version(fname="whisper/version.py"): exec(compile(open(fname, encoding="utf-8").read(), fname, "exec")) return locals()["__version__"] setup( name="openai-whisper", py_modules=["whisper"], version=read_version()...
# Copyright (c) OpenMMLab. All rights reserved. from .anchor_free_head import AnchorFreeHead from .anchor_head import AnchorHead from .atss_head import ATSSHead from .autoassign_head import AutoAssignHead from .boxinst_head import BoxInstBboxHead, BoxInstMaskHead from .cascade_rpn_head import CascadeRPNHead, StageCasca...
# Copyright (c) OpenMMLab. All rights reserved. from .anchor_free_head import AnchorFreeHead from .anchor_head import AnchorHead from .atss_head import ATSSHead from .autoassign_head import AutoAssignHead from .boxinst_head import BoxInstBboxHead, BoxInstMaskHead from .cascade_rpn_head import CascadeRPNHead, StageCasca...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Iterator, List, Optional, Sequence, Union from mmengine.registry import EVALUATOR, METRICS from mmengine.structures import BaseDataElement from .metric import BaseMetric @EVALUATOR.register_module() class Evaluator: """Wrapper class to compose mu...
# Copyright (c) OpenMMLab. All rights reserved. from typing import Iterator, List, Optional, Sequence, Union from mmengine.data import BaseDataElement from mmengine.registry import EVALUATOR, METRICS from .metric import BaseMetric @EVALUATOR.register_module() class Evaluator: """Wrapper class to compose multiple...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.chat_message_histories import RedisChatMessageHistory # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handling o...
from typing import TYPE_CHECKING, Any from langchain._api import create_importer if TYPE_CHECKING: from langchain_community.chat_message_histories import RedisChatMessageHistory # Create a way to dynamically look up deprecated imports. # Used to consolidate logic for raising deprecation warnings and # handling o...
from typing import Optional, TypeVar from docarray.base_document import BaseDocument from docarray.documents import Audio from docarray.typing import AnyEmbedding, AnyTensor from docarray.typing.tensor.video.video_tensor import VideoTensor from docarray.typing.url.video_url import VideoUrl T = TypeVar('T', bound='Vid...
from typing import Optional, TypeVar from docarray.base_document import BaseDocument from docarray.documents import Audio from docarray.typing import AnyEmbedding, AnyTensor from docarray.typing.tensor.video.video_tensor import VideoTensor from docarray.typing.url.video_url import VideoUrl T = TypeVar('T', bound='Vid...
_base_ = [ '../_base_/models/retinanet_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] # model settings norm_cfg = dict(type='GN', num_groups=32, requires_grad=True) model = dict( backbone=dict( depth=101, init_c...
_base_ = [ '../_base_/models/retinanet_r50_fpn.py', '../_base_/datasets/coco_detection.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] # model settings norm_cfg = dict(type='GN', num_groups=32, requires_grad=True) model = dict( backbone=dict( depth=101, init_c...
from keras.src.api_export import keras_export @keras_export(["keras.Initializer", "keras.initializers.Initializer"]) class Initializer: """Initializer base class: all Keras initializers inherit from this class. Initializers should implement a `__call__()` method with the following signature: ```pyth...
from keras.src.api_export import keras_export @keras_export(["keras.Initializer", "keras.initializers.Initializer"]) class Initializer: """Initializer base class: all Keras initializers inherit from this class. Initializers should implement a `__call__()` method with the following signature: ```pyth...
# Copyright (c) OpenMMLab. All rights reserved. import argparse import cv2 import torch from mmdet.apis import inference_detector, init_detector def parse_args(): parser = argparse.ArgumentParser(description='MMDetection webcam demo') parser.add_argument('config', help='test config file path') parser.ad...
import argparse import cv2 import torch from mmdet.apis import inference_detector, init_detector def parse_args(): parser = argparse.ArgumentParser(description='MMDetection webcam demo') parser.add_argument('config', help='test config file path') parser.add_argument('checkpoint', help='checkpoint file')...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
""" Top-level module of Jina. The primary function of this module is to import all of the public Jina interfaces into a single place. The interfaces themselves are located in sub-modules, as described below. """ import os as _os import platform as _platform import signal as _signal import sys as _sys import warnings...
from __future__ import annotations import os from . import InputExample class LabelSentenceReader: """Reads in a file that has at least two columns: a label and a sentence. This reader can for example be used with the BatchHardTripletLoss. Maps labels automatically to integers """ def __init__(...
import os from . import InputExample class LabelSentenceReader: """Reads in a file that has at least two columns: a label and a sentence. This reader can for example be used with the BatchHardTripletLoss. Maps labels automatically to integers """ def __init__(self, folder, label_col_idx=0, sente...
import unittest import torch from mmengine.config import Config from mmengine.data import InstanceData from mmengine.testing import assert_allclose from mmdet.core.evaluation import INSTANCE_OFFSET from mmdet.models.seg_heads.panoptic_fusion_heads import HeuristicFusionHead class TestHeuristicFusionHead(unittest.Te...
import unittest import torch from mmengine.config import Config from mmengine.testing import assert_allclose from mmdet.core.evaluation import INSTANCE_OFFSET from mmdet.models.seg_heads.panoptic_fusion_heads import HeuristicFusionHead class TestHeuristicFusionHead(unittest.TestCase): def test_loss(self): ...
# Copyright (c) OpenMMLab. All rights reserved. from .hub import load_url from .misc import (check_prerequisites, concat_list, deprecated_api_warning, has_method, import_modules_from_strings, is_list_of, is_method_overridden, is_seq_of, is_str, is_tuple_of, iter_...
# Copyright (c) OpenMMLab. All rights reserved. from .misc import (check_prerequisites, concat_list, deprecated_api_warning, has_method, import_modules_from_strings, is_list_of, is_method_overridden, is_seq_of, is_str, is_tuple_of, iter_cast, list_cast, mmcv_full...
from dataclasses import dataclass, fields, field from typing import Optional, Tuple, TYPE_CHECKING if TYPE_CHECKING: from docarray.score import NamedScore default_values = dict(value=0.0, op_name='', description='', ref_id='') @dataclass(unsafe_hash=True) class NamedScoreData: _reference_ns: 'NamedScore' = ...
from dataclasses import dataclass, fields, field from typing import Optional, Tuple, TYPE_CHECKING if TYPE_CHECKING: from docarray.score import NamedScore default_values = dict(value=0.0, op_name='', description='', ref_id='') @dataclass(unsafe_hash=True) class NamedScoreData: _reference_ns: 'NamedScore' = ...
from __future__ import annotations import re import pytest from sentence_transformers import SentenceTransformer from sentence_transformers.evaluation import NanoBEIREvaluator from sentence_transformers.util import is_datasets_available if not is_datasets_available(): pytest.skip( reason="Datasets are n...
from __future__ import annotations import re import pytest from sentence_transformers import SentenceTransformer from sentence_transformers.evaluation import NanoBEIREvaluator def test_nanobeir_evaluator(): """Tests that the NanoBERTEvaluator can be loaded and produces expected metrics""" datasets = ["Quor...
import torchaudio _STREAM_READER = [ "StreamReader", ] _STREAM_WRITER = [ "StreamWriter", ] _LAZILY_IMPORTED = _STREAM_READER + _STREAM_WRITER def __getattr__(name: str): if name in _LAZILY_IMPORTED: if not torchaudio._extension._FFMPEG_INITIALIZED: torchaudio._extension._init_ffmp...
import torchaudio _STREAM_READER = [ "StreamReader", "StreamReaderSourceStream", "StreamReaderSourceAudioStream", "StreamReaderSourceVideoStream", "StreamReaderOutputStream", ] _STREAM_WRITER = [ "StreamWriter", ] _LAZILY_IMPORTED = _STREAM_READER + _STREAM_WRITER def __getattr__(name: str...
import pytest from backend.util.service import ( AppService, AppServiceClient, endpoint_to_async, expose, get_service_client, ) TEST_SERVICE_PORT = 8765 class ServiceTest(AppService): def __init__(self): super().__init__() def cleanup(self): pass @classmethod de...
import pytest from backend.util.service import ( AppService, AppServiceClient, endpoint_to_async, expose, get_service_client, ) TEST_SERVICE_PORT = 8765 class ServiceTest(AppService): def __init__(self): super().__init__() def cleanup(self): pass @classmethod de...
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .two_stage import TwoStageDetector @DETECTORS.register_module() class FasterRCNN(TwoStageDetector): """Implementation of `Faster R-CNN <https://arxiv.org/abs/1506.01497>`_""" def __init__(self, backbone, ...
# Copyright (c) OpenMMLab. All rights reserved. from ..builder import DETECTORS from .two_stage import TwoStageDetector @DETECTORS.register_module() class FasterRCNN(TwoStageDetector): """Implementation of `Faster R-CNN <https://arxiv.org/abs/1506.01497>`_""" def __init__(self, backbone, ...
from torchvision.transforms import AutoAugmentPolicy, InterpolationMode # usort: skip from . import functional, utils # usort: skip from ._transform import Transform # usort: skip from ._presets import StereoMatching # usort: skip from ._augment import RandomCutmix, RandomErasing, RandomMixup, SimpleCopyPaste fr...
from torchvision.transforms import AutoAugmentPolicy, InterpolationMode # usort: skip from . import functional, utils # usort: skip from ._transform import Transform # usort: skip from ._presets import StereoMatching # usort: skip from ._augment import RandomCutmix, RandomErasing, RandomMixup, SimpleCopyPaste fr...
""" This script trains sentence transformers with a triplet loss function. As corpus, we use the wikipedia sections dataset that was describd by Dor et al., 2018, Learning Thematic Similarity Metric Using Triplet Networks. """ import logging import traceback from datetime import datetime from datasets import load_da...
""" This script trains sentence transformers with a triplet loss function. As corpus, we use the wikipedia sections dataset that was describd by Dor et al., 2018, Learning Thematic Similarity Metric Using Triplet Networks. """ from sentence_transformers import SentenceTransformer, InputExample, LoggingHandler, losses...
_base_ = [ '../_base_/models/mask-rcnn_r50_fpn.py', '../_base_/datasets/coco_instance.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] model = dict( backbone=dict( frozen_stages=0, norm_cfg=dict(type='SyncBN', requires_grad=True), norm_eval=False, ...
_base_ = [ '../_base_/models/mask_rcnn_r50_fpn.py', '../_base_/datasets/coco_instance.py', '../_base_/schedules/schedule_2x.py', '../_base_/default_runtime.py' ] model = dict( backbone=dict( frozen_stages=0, norm_cfg=dict(type='SyncBN', requires_grad=True), norm_eval=False, ...
from keras.src.backend.common.name_scope import name_scope from keras.src.backend.jax import core from keras.src.backend.jax import distribution_lib from keras.src.backend.jax import image from keras.src.backend.jax import linalg from keras.src.backend.jax import math from keras.src.backend.jax import nn from keras.src...
from keras.src.backend.common.name_scope import name_scope from keras.src.backend.jax import core from keras.src.backend.jax import distribution_lib from keras.src.backend.jax import image from keras.src.backend.jax import linalg from keras.src.backend.jax import math from keras.src.backend.jax import nn from keras.src...
# Copyright 2022 The HuggingFace Datasets Authors and the current dataset script contributor. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2....
# Copyright 2022 The HuggingFace Datasets Authors and the current dataset script contributor. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2....
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.activations import deserialize from keras.src.activations import get from keras.src.activations import serialize from keras.src.activations.activations import celu from keras.src.acti...
"""DO NOT EDIT. This file was autogenerated. Do not edit it by hand, since your modifications would be overwritten. """ from keras.src.activations import deserialize from keras.src.activations import get from keras.src.activations import serialize from keras.src.activations.activations import celu from keras.src.acti...
# Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
""" Example of forwarding evaluation logs to the client =================================================== The example runs on GPU. Two classes are defined to show how to use Dask builtins to forward the logs to the client process. """ import logging import dask import distributed from dask import array as da from...
"""Example of forwarding evaluation logs to the client =================================================== The example runs on GPU. Two classes are defined to show how to use Dask builtins to forward the logs to the client process. """ import logging import dask import distributed from dask import array as da from ...
import os from typing import Any, Callable, Optional, Tuple from PIL import Image from .utils import check_integrity, download_and_extract_archive, download_url from .vision import VisionDataset class SBU(VisionDataset): """`SBU Captioned Photo <http://www.cs.virginia.edu/~vicente/sbucaptions/>`_ Dataset. ...
import os from typing import Any, Callable, Optional, Tuple from PIL import Image from .utils import check_integrity, download_url from .vision import VisionDataset class SBU(VisionDataset): """`SBU Captioned Photo <http://www.cs.virginia.edu/~vicente/sbucaptions/>`_ Dataset. Args: root (string): R...
_base_ = '../cascade_rcnn/cascade_mask_rcnn_r50_fpn_1x_coco.py' model = dict( backbone=dict( _delete_=True, type='HRNet', extra=dict( stage1=dict( num_modules=1, num_branches=1, block='BOTTLENECK', num_blocks=(4, ), ...
_base_ = '../cascade_rcnn/cascade_mask_rcnn_r50_fpn_1x_coco.py' model = dict( backbone=dict( _delete_=True, type='HRNet', extra=dict( stage1=dict( num_modules=1, num_branches=1, block='BOTTLENECK', num_blocks=(4, ), ...
from . import utils from .model import ( hubert_base, hubert_large, hubert_pretrain_base, hubert_pretrain_large, hubert_pretrain_model, hubert_pretrain_xlarge, hubert_xlarge, HuBERTPretrainModel, wav2vec2_base, wav2vec2_large, wav2vec2_large_lv60k, wav2vec2_model, Wav...
from . import utils from .model import ( Wav2Vec2Model, HuBERTPretrainModel, wav2vec2_model, wav2vec2_base, wav2vec2_large, wav2vec2_large_lv60k, hubert_base, hubert_large, hubert_xlarge, hubert_pretrain_model, hubert_pretrain_base, hubert_pretrain_large, hubert_pretr...
from typing import TYPE_CHECKING from ...utils import ( DIFFUSERS_SLOW_IMPORT, OptionalDependencyNotAvailable, _LazyModule, get_objects_from_module, is_opencv_available, is_torch_available, is_transformers_available, ) _dummy_objects = {} _import_structure = {} try: if not (is_trans...
from typing import TYPE_CHECKING from ...utils import ( DIFFUSERS_SLOW_IMPORT, OptionalDependencyNotAvailable, _LazyModule, get_objects_from_module, is_torch_available, is_transformers_available, ) _dummy_objects = {} _import_structure = {} try: if not (is_transformers_available() and i...
_base_ = './mask-rcnn_r50_fpn_gn-ws-all_2x_coco.py' # model settings conv_cfg = dict(type='ConvWS') norm_cfg = dict(type='GN', num_groups=32, requires_grad=True) model = dict( backbone=dict( type='ResNeXt', depth=101, groups=32, base_width=4, num_stages=4, out_indices...
_base_ = './mask_rcnn_r50_fpn_gn_ws-all_2x_coco.py' # model settings conv_cfg = dict(type='ConvWS') norm_cfg = dict(type='GN', num_groups=32, requires_grad=True) model = dict( backbone=dict( type='ResNeXt', depth=101, groups=32, base_width=4, num_stages=4, out_indices...
import os from pathlib import Path import pytest from jina import Executor def test_config(): ranker = Executor.load_config( str(Path(__file__).parents[2] / 'config.yml'), override_with={ 'query_features': ['query'], 'match_features': ['match'], 'relevance_labe...
import os from pathlib import Path import pytest from jina import Executor def test_config(): ranker = Executor.load_config( str(Path(__file__).parents[2] / 'config.yml'), override_with={ 'query_features': ['query'], 'match_features': ['match'], 'relevance_lab...
""" The system trains BERT (or any other transformer model like RoBERTa, DistilBERT etc.) on the SNLI + MultiNLI (AllNLI) dataset with softmax loss function. At every 1000 training steps, the model is evaluated on the STS benchmark dataset Usage: python training_nli.py OR python training_nli.py pretrained_transformer...
""" The system trains BERT (or any other transformer model like RoBERTa, DistilBERT etc.) on the SNLI + MultiNLI (AllNLI) dataset with softmax loss function. At every 1000 training steps, the model is evaluated on the STS benchmark dataset Usage: python training_nli.py OR python training_nli.py pretrained_transformer...
import json import os from typing import Dict from torch import Tensor, nn class Dropout(nn.Module): """Dropout layer. Args: dropout: Sets a dropout value for dense layer. """ def __init__(self, dropout: float = 0.2): super(Dropout, self).__init__() self.dropout = dropout ...
from torch import Tensor from torch import nn from typing import Dict import os import json class Dropout(nn.Module): """Dropout layer. :param dropout: Sets a dropout value for dense layer. """ def __init__(self, dropout: float = 0.2): super(Dropout, self).__init__() self.dropout = d...
import os import time import pytest from docarray import Document from jina import Client, Flow from jina.serve.networking.utils import send_health_check_sync @pytest.fixture def error_log_level(): old_env = os.environ.get('JINA_LOG_LEVEL') os.environ['JINA_LOG_LEVEL'] = 'ERROR' yield os.environ['JI...
import os import time import pytest from docarray import Document from jina import Client, Flow from jina.serve.networking import GrpcConnectionPool @pytest.fixture def error_log_level(): old_env = os.environ.get('JINA_LOG_LEVEL') os.environ['JINA_LOG_LEVEL'] = 'ERROR' yield os.environ['JINA_LOG_LEV...
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from .mongo_handler import MongoHandler from .mongo_storage import MongoDBStorage
__copyright__ = "Copyright (c) 2021 Jina AI Limited. All rights reserved." __license__ = "Apache-2.0" from .mongo_storage import MongoDBStorage from .mongo_handler import MongoHandler
import torch from keras.src.backend.common.stateless_scope import in_stateless_scope from keras.src.ops.operation import Operation class TorchLayer(torch.nn.Module): @property def torch_params(self): if not hasattr(self, "_torch_params"): self._track_variables() return self._torch...
from typing import Iterator from typing import Tuple import torch from keras.src.backend.common.stateless_scope import in_stateless_scope from keras.src.ops.operation import Operation class TorchLayer(torch.nn.Module): @property def torch_params(self): if not hasattr(self, "_torch_params"): ...
_base_ = './cascade-mask-rcnn_r50_fpn_20e_coco.py' model = dict( backbone=dict( type='ResNeXt', depth=101, groups=32, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), style='py...
_base_ = './cascade_mask_rcnn_r50_fpn_20e_coco.py' model = dict( backbone=dict( type='ResNeXt', depth=101, groups=32, base_width=4, num_stages=4, out_indices=(0, 1, 2, 3), frozen_stages=1, norm_cfg=dict(type='BN', requires_grad=True), style='py...
import re import pytest from llama_index.core.workflow.decorators import step from llama_index.core.workflow.errors import WorkflowValidationError from llama_index.core.workflow.events import Event, StartEvent, StopEvent from llama_index.core.workflow.workflow import Workflow def test_decorated_config(workflow): ...
import re import pytest from llama_index.core.workflow.decorators import step from llama_index.core.workflow.errors import WorkflowValidationError from llama_index.core.workflow.events import Event, StartEvent, StopEvent from llama_index.core.workflow.workflow import Workflow def test_decorated_config(workflow): ...
#!/usr/bin/env python import functools as func import glob import os.path as osp import re import numpy as np url_prefix = 'https://github.com/open-mmlab/mmdetection/blob/master/configs' files = sorted(glob.glob('../../configs/*/README.md')) stats = [] titles = [] num_ckpts = 0 for f in files: url = osp.dirnam...
#!/usr/bin/env python import functools as func import glob import os.path as osp import re import numpy as np url_prefix = 'https://github.com/open-mmlab/mmdetection/blob/master/' files = sorted(glob.glob('../configs/*/README.md')) stats = [] titles = [] num_ckpts = 0 for f in files: url = osp.dirname(f.replac...
""" This file evaluates CrossEncoder on the TREC 2019 Deep Learning (DL) Track: https://arxiv.org/abs/2003.07820 TREC 2019 DL is based on the corpus of MS Marco. MS Marco provides a sparse annotation, i.e., usually only a single passage is marked as relevant for a given query. Many other highly relevant passages are n...
""" This file evaluates CrossEncoder on the TREC 2019 Deep Learning (DL) Track: https://arxiv.org/abs/2003.07820 TREC 2019 DL is based on the corpus of MS Marco. MS Marco provides a sparse annotation, i.e., usually only a single passage is marked as relevant for a given query. Many other highly relevant passages are n...
import http.client import json from typing import Optional def list_packages(*, contains: Optional[str] = None): conn = http.client.HTTPSConnection("api.github.com") headers = { "Accept": "application/vnd.github+json", "X-GitHub-Api-Version": "2022-11-28", "User-Agent": "langchain-cli...
import http.client import json from typing import Optional def list_packages(*, contains: Optional[str] = None): conn = http.client.HTTPSConnection("api.github.com") headers = { "Accept": "application/vnd.github+json", "X-GitHub-Api-Version": "2022-11-28", "User-Agent": "langchain-cli...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.models.utils.misc import get_box_tensor from mmdet.registry import TASK_UTILS from mmdet.structures.bbox import HorizontalBoxes from .base_bbox_coder import BaseBBoxCoder @TASK_UTILS.register_module() class PseudoBBoxCoder(BaseBBoxCoder): """Pseudo boundi...
# Copyright (c) OpenMMLab. All rights reserved. from mmdet.registry import TASK_UTILS from .base_bbox_coder import BaseBBoxCoder @TASK_UTILS.register_module() class PseudoBBoxCoder(BaseBBoxCoder): """Pseudo bounding box coder.""" def __init__(self, **kwargs): super(BaseBBoxCoder, self).__init__(**kwa...
from typing import Any, Optional from langchain_core.language_models import BaseLanguageModel from langchain_core.memory import BaseMemory from langchain_core.messages import SystemMessage from langchain_core.prompts.chat import MessagesPlaceholder from langchain_core.tools import BaseTool from langchain.agents.agent...
from typing import Any, Optional from langchain_core.language_models import BaseLanguageModel from langchain_core.memory import BaseMemory from langchain_core.messages import SystemMessage from langchain_core.prompts.chat import MessagesPlaceholder from langchain_core.tools import BaseTool from langchain.agents.agent...
from typing import Union, Dict, Any import google.ai.generativelanguage as glm import google.generativeai as genai from llama_index.core.base.llms.types import ( ChatMessage, ChatResponse, CompletionResponse, ImageBlock, TextBlock, ) from llama_index.core.multi_modal_llms.base import ChatMessage fr...
from typing import Union, Dict, Any import google.ai.generativelanguage as glm import google.generativeai as genai from llama_index.core.base.llms.types import ( ChatMessage, ChatResponse, CompletionResponse, ImageBlock, TextBlock, ) from llama_index.core.multi_modal_llms.base import ChatMessage fr...
import numpy as np import torch from docarray import Document from docarray.document import AnyDocument from docarray.typing import AnyUrl, Embedding, ImageUrl, NdArray, TextUrl, TorchTensor def test_proto_all_types(): class Mymmdoc(Document): tensor: NdArray torch_tensor: TorchTensor emb...
import numpy as np import torch from docarray import Document from docarray.document import AnyDocument from docarray.typing import AnyUrl, Embedding, ImageUrl, Tensor, TextUrl, TorchTensor def test_proto_all_types(): class Mymmdoc(Document): tensor: Tensor torch_tensor: TorchTensor embed...
# Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# Copyright (c) OpenMMLab. All rights reserved. import os.path as osp import unittest from unittest.mock import MagicMock, patch import pytest from mmdet.datasets import DATASETS @patch('mmdet.datasets.CocoDataset.load_annotations', MagicMock()) @patch('mmdet.datasets.CustomDataset.load_annotations', MagicMock()) @...
# Copyright (c) OpenMMLab. All rights reserved. import os import unittest from unittest.mock import MagicMock, patch import pytest from mmdet.datasets import DATASETS @patch('mmdet.datasets.CocoDataset.load_annotations', MagicMock()) @patch('mmdet.datasets.CustomDataset.load_annotations', MagicMock()) @patch('mmdet...
"""Experiment with different models.""" from __future__ import annotations from collections.abc import Sequence from typing import Optional from langchain_core.language_models.llms import BaseLLM from langchain_core.prompts.prompt import PromptTemplate from langchain_core.utils.input import get_color_mapping, print_...
"""Experiment with different models.""" from __future__ import annotations from collections.abc import Sequence from typing import Optional from langchain_core.language_models.llms import BaseLLM from langchain_core.prompts.prompt import PromptTemplate from langchain_core.utils.input import get_color_mapping, print_...
def get_doc_value(): return 'MyExecutorAfterReload'
def get_doc_value(): return 'MyExecutorAfterReload'