method_name
stringlengths
1
78
method_body
stringlengths
3
9.66k
full_code
stringlengths
31
10.7k
docstring
stringlengths
4
4.74k
is_lc_serializable
return False
@classmethod def is_lc_serializable(cls) ->bool: return False
null
_chat
if self.hunyuan_secret_key is None: raise ValueError('Hunyuan secret key is not set.') parameters = {**self._default_params, **kwargs} headers = parameters.pop('headers', {}) timestamp = parameters.pop('timestamp', int(time.time())) expired = parameters.pop('expired', timestamp + 24 * 60 * 60) payload = {'timestamp...
def _chat(self, messages: List[BaseMessage], **kwargs: Any ) ->requests.Response: if self.hunyuan_secret_key is None: raise ValueError('Hunyuan secret key is not set.') parameters = {**self._default_params, **kwargs} headers = parameters.pop('headers', {}) timestamp = parameters.pop('timesta...
null
_run
mailbox = self.account.mailbox() message = mailbox.new_message() message.body = body message.subject = subject message.to.add(to) if cc is not None: message.cc.add(cc) if bcc is not None: message.bcc.add(bcc) message.send() output = 'Message sent: ' + str(message) return output
def _run(self, body: str, to: List[str], subject: str, cc: Optional[List[ str]]=None, bcc: Optional[List[str]]=None, run_manager: Optional[ CallbackManagerForToolRun]=None) ->str: mailbox = self.account.mailbox() message = mailbox.new_message() message.body = body message.subject = subject m...
null
get_aggregate_feedback
"""Return quantiles for the feedback scores. This method calculates and prints the quantiles for the feedback scores across all feedback keys. Returns: A DataFrame containing the quantiles for each feedback key. """ df = self.to_dataframe() to_drop = [col for col in df.colu...
def get_aggregate_feedback(self) ->pd.DataFrame: """Return quantiles for the feedback scores. This method calculates and prints the quantiles for the feedback scores across all feedback keys. Returns: A DataFrame containing the quantiles for each feedback key. """ d...
Return quantiles for the feedback scores. This method calculates and prints the quantiles for the feedback scores across all feedback keys. Returns: A DataFrame containing the quantiles for each feedback key.
_get_anthropic_client
try: import anthropic except ImportError: raise ImportError( 'Could not import anthropic python package. This is needed in order to accurately tokenize the text for anthropic models. Please install it with `pip install anthropic`.' ) return anthropic.Anthropic()
def _get_anthropic_client() ->Any: try: import anthropic except ImportError: raise ImportError( 'Could not import anthropic python package. This is needed in order to accurately tokenize the text for anthropic models. Please install it with `pip install anthropic`.' ) ...
null
embeddings
return self.embedding_function if isinstance(self.embedding_function, Embeddings) else None
@property def embeddings(self) ->Optional[Embeddings]: return self.embedding_function if isinstance(self.embedding_function, Embeddings) else None
null
on_chain_error
"""Run when chain errors. Args: error (Exception or KeyboardInterrupt): The error. """ self.ended = True return self.parent_run_manager.on_chain_error(error, **kwargs)
def on_chain_error(self, error: BaseException, **kwargs: Any) ->None: """Run when chain errors. Args: error (Exception or KeyboardInterrupt): The error. """ self.ended = True return self.parent_run_manager.on_chain_error(error, **kwargs)
Run when chain errors. Args: error (Exception or KeyboardInterrupt): The error.
_import_bigquery
from langchain_community.vectorstores.bigquery_vector_search import BigQueryVectorSearch return BigQueryVectorSearch
def _import_bigquery() ->Any: from langchain_community.vectorstores.bigquery_vector_search import BigQueryVectorSearch return BigQueryVectorSearch
null
_generate
should_stream = stream if stream is not None else self.streaming if should_stream: generation: Optional[GenerationChunk] = None stream_iter = self._stream(prompts[0], stop=stop, run_manager= run_manager, **kwargs) for chunk in stream_iter: if generation is None: generation = chun...
def _generate(self, prompts: List[str], stop: Optional[List[str]]=None, run_manager: Optional[CallbackManagerForLLMRun]=None, stream: Optional[ bool]=None, **kwargs: Any) ->LLMResult: should_stream = stream if stream is not None else self.streaming if should_stream: generation: Optional[Generati...
null
_import_golden_query
from langchain_community.utilities.golden_query import GoldenQueryAPIWrapper return GoldenQueryAPIWrapper
def _import_golden_query() ->Any: from langchain_community.utilities.golden_query import GoldenQueryAPIWrapper return GoldenQueryAPIWrapper
null
test_configurable_fields_prefix_keys
fake_chat = FakeListChatModel(responses=['b']).configurable_fields(responses =ConfigurableFieldMultiOption(id='responses', name='Chat Responses', options={'hello': 'A good morning to you!', 'bye': 'See you later!', 'helpful': 'How can I help you?'}, default=['hello', 'bye']), sleep= ConfigurableField(id...
def test_configurable_fields_prefix_keys() ->None: fake_chat = FakeListChatModel(responses=['b']).configurable_fields( responses=ConfigurableFieldMultiOption(id='responses', name= 'Chat Responses', options={'hello': 'A good morning to you!', 'bye': 'See you later!', 'helpful': 'How can I hel...
null
__init__
self._iterator = iter(iterable) self._buffers: List[Deque[T]] = [deque() for _ in range(n)] self._children = tuple(tee_peer(iterator=self._iterator, buffer=buffer, peers=self._buffers, lock=lock if lock is not None else NoLock()) for buffer in self._buffers)
def __init__(self, iterable: Iterator[T], n: int=2, *, lock: Optional[ ContextManager[Any]]=None): self._iterator = iter(iterable) self._buffers: List[Deque[T]] = [deque() for _ in range(n)] self._children = tuple(tee_peer(iterator=self._iterator, buffer=buffer, peers=self._buffers, lock=lock if...
null
__init__
"""Instantiate AsyncFinalIteratorCallbackHandler. Args: answer_prefix_tokens: Token sequence that prefixes the answer. Default is ["Final", "Answer", ":"] strip_tokens: Ignore white spaces and new lines when comparing answer_prefix_tokens to last tokens? ...
def __init__(self, *, answer_prefix_tokens: Optional[List[str]]=None, strip_tokens: bool=True, stream_prefix: bool=False) ->None: """Instantiate AsyncFinalIteratorCallbackHandler. Args: answer_prefix_tokens: Token sequence that prefixes the answer. Default is ["Final", "Answ...
Instantiate AsyncFinalIteratorCallbackHandler. Args: answer_prefix_tokens: Token sequence that prefixes the answer. Default is ["Final", "Answer", ":"] strip_tokens: Ignore white spaces and new lines when comparing answer_prefix_tokens to last tokens? (to determine if answer has been re...
validate_environment
values['qianfan_ak'] = convert_to_secret_str(get_from_dict_or_env(values, 'qianfan_ak', 'QIANFAN_AK', default='')) values['qianfan_sk'] = convert_to_secret_str(get_from_dict_or_env(values, 'qianfan_sk', 'QIANFAN_SK', default='')) params = {**values.get('init_kwargs', {}), 'model': values['model']} if values['qi...
@root_validator() def validate_environment(cls, values: Dict) ->Dict: values['qianfan_ak'] = convert_to_secret_str(get_from_dict_or_env( values, 'qianfan_ak', 'QIANFAN_AK', default='')) values['qianfan_sk'] = convert_to_secret_str(get_from_dict_or_env( values, 'qianfan_sk', 'QIANFAN_SK', default...
null
on_llm_start
"""Save the prompts in memory when an LLM starts.""" self.prompts.update({str(kwargs['parent_run_id'] or kwargs['run_id']): prompts} )
def on_llm_start(self, serialized: Dict[str, Any], prompts: List[str], ** kwargs: Any) ->None: """Save the prompts in memory when an LLM starts.""" self.prompts.update({str(kwargs['parent_run_id'] or kwargs['run_id']): prompts})
Save the prompts in memory when an LLM starts.
from_existing_index
connection_string = cls.get_connection_string(kwargs) store = cls(connection_string=connection_string, collection_name= collection_name, embedding_function=embedding, pre_delete_collection= pre_delete_collection) return store
@classmethod def from_existing_index(cls: Type[PGEmbedding], embedding: Embeddings, collection_name: str=_LANGCHAIN_DEFAULT_COLLECTION_NAME, pre_delete_collection: bool=False, **kwargs: Any) ->PGEmbedding: connection_string = cls.get_connection_string(kwargs) store = cls(connection_string=connection_str...
null
_format_request_payload
chat_messages = [LlamaContentFormatter._convert_message_to_dict(message) for message in messages] prompt = json.dumps({'input_data': {'input_string': chat_messages, 'parameters': model_kwargs}}) return self.format_request_payload(prompt=prompt, model_kwargs=model_kwargs)
def _format_request_payload(self, messages: List[BaseMessage], model_kwargs: Dict) ->bytes: chat_messages = [LlamaContentFormatter._convert_message_to_dict(message ) for message in messages] prompt = json.dumps({'input_data': {'input_string': chat_messages, 'parameters': model_kwargs}}) ...
null
_make_request_body_for_models
"""Make the request body for generate/retrieve models endpoint""" _model_kwargs = self.model_kwargs or {} _params = {**_model_kwargs, **kwargs} filters = [DALMFilter(**f) for f in _params.get('filters', [])] return dict(model_id=self.model_id, query=prompt, size=_params.get('size', 3), filters=filters, id=self.mod...
def _make_request_body_for_models(self, prompt: str, **kwargs: Mapping[str, Any]) ->Mapping[str, Any]: """Make the request body for generate/retrieve models endpoint""" _model_kwargs = self.model_kwargs or {} _params = {**_model_kwargs, **kwargs} filters = [DALMFilter(**f) for f in _params.get('filt...
Make the request body for generate/retrieve models endpoint
_type
return 'structured_chat_with_retries'
@property def _type(self) ->str: return 'structured_chat_with_retries'
null
_fstring_Constant
assert isinstance(t.value, str) value = t.value.replace('{', '{{').replace('}', '}}') write(value)
def _fstring_Constant(self, t, write): assert isinstance(t.value, str) value = t.value.replace('{', '{{').replace('}', '}}') write(value)
null
on_chain_error
"""Run when chain errors.""" self.step += 1 self.errors += 1
def on_chain_error(self, error: BaseException, **kwargs: Any) ->None: """Run when chain errors.""" self.step += 1 self.errors += 1
Run when chain errors.
search
"""Search for document. If page exists, return the page summary, and a Document object. If page does not exist, return similar entries. """
@abstractmethod def search(self, search: str) ->Union[str, Document]: """Search for document. If page exists, return the page summary, and a Document object. If page does not exist, return similar entries. """
Search for document. If page exists, return the page summary, and a Document object. If page does not exist, return similar entries.
random_string
return str(uuid.uuid4())
def random_string() ->str: return str(uuid.uuid4())
null
test_api_key_masked_when_passed_via_constructor
gpt_router = GPTRouter(gpt_router_api_base='https://example.com', gpt_router_api_key='secret-api-key') print(gpt_router.gpt_router_api_key, end='') captured = capsys.readouterr() assert captured.out == '**********'
def test_api_key_masked_when_passed_via_constructor(capsys: CaptureFixture ) ->None: gpt_router = GPTRouter(gpt_router_api_base='https://example.com', gpt_router_api_key='secret-api-key') print(gpt_router.gpt_router_api_key, end='') captured = capsys.readouterr() assert captured.out == '****...
null
scrape
"""Scrape data from webpage and return it in BeautifulSoup format.""" if parser is None: parser = self.default_parser return self._scrape(self.web_path, parser=parser, bs_kwargs=self.bs_kwargs)
def scrape(self, parser: Union[str, None]=None) ->Any: """Scrape data from webpage and return it in BeautifulSoup format.""" if parser is None: parser = self.default_parser return self._scrape(self.web_path, parser=parser, bs_kwargs=self.bs_kwargs)
Scrape data from webpage and return it in BeautifulSoup format.
_llm_type
"""Return type of llm.""" return 'fake'
@property def _llm_type(self) ->str: """Return type of llm.""" return 'fake'
Return type of llm.
output_keys
return [self.output_key]
@property def output_keys(self) ->List[str]: return [self.output_key]
null
test_api_key_is_secret_string
llm = Nebula(nebula_api_key='secret-api-key') assert isinstance(llm.nebula_api_key, SecretStr) assert llm.nebula_api_key.get_secret_value() == 'secret-api-key'
def test_api_key_is_secret_string() ->None: llm = Nebula(nebula_api_key='secret-api-key') assert isinstance(llm.nebula_api_key, SecretStr) assert llm.nebula_api_key.get_secret_value() == 'secret-api-key'
null
exists
return self.redis_client.exists(f'{self.full_key_prefix}:{key}') == 1
def exists(self, key: str) ->bool: return self.redis_client.exists(f'{self.full_key_prefix}:{key}') == 1
null
_import_office365_messages_search
from langchain_community.tools.office365.messages_search import O365SearchEmails return O365SearchEmails
def _import_office365_messages_search() ->Any: from langchain_community.tools.office365.messages_search import O365SearchEmails return O365SearchEmails
null
_dependable_tweepy_import
try: import tweepy except ImportError: raise ImportError( 'tweepy package not found, please install it with `pip install tweepy`' ) return tweepy
def _dependable_tweepy_import() ->tweepy: try: import tweepy except ImportError: raise ImportError( 'tweepy package not found, please install it with `pip install tweepy`' ) return tweepy
null
tool_func
"""Return the arguments directly.""" return f'{arg1} {arg2} {arg3}'
@tool(args_schema=_MockSchema) def tool_func(arg1: int, arg2: bool, arg3: Optional[dict]=None) ->str: """Return the arguments directly.""" return f'{arg1} {arg2} {arg3}'
Return the arguments directly.
lazy_parse
"""Lazily parse the blob.""" if not self.extract_images: from pdfminer.high_level import extract_text with blob.as_bytes_io() as pdf_file_obj: if self.concatenate_pages: text = extract_text(pdf_file_obj) metadata = {'source': blob.source} yield Document(page_content=t...
def lazy_parse(self, blob: Blob) ->Iterator[Document]: """Lazily parse the blob.""" if not self.extract_images: from pdfminer.high_level import extract_text with blob.as_bytes_io() as pdf_file_obj: if self.concatenate_pages: text = extract_text(pdf_file_obj) ...
Lazily parse the blob.
validate_environment
"""Validate that the python package exists in the environment.""" try: import tokenizers except ImportError: raise ImportError( 'Could not import tokenizers python package. Please install it with `pip install tokenizers`.' ) try: from rwkv.model import RWKV as RWKVMODEL from rwkv.utils i...
@root_validator() def validate_environment(cls, values: Dict) ->Dict: """Validate that the python package exists in the environment.""" try: import tokenizers except ImportError: raise ImportError( 'Could not import tokenizers python package. Please install it with `pip install t...
Validate that the python package exists in the environment.
_import_vertex
from langchain_community.llms.vertexai import VertexAI return VertexAI
def _import_vertex() ->Any: from langchain_community.llms.vertexai import VertexAI return VertexAI
null
_on_llm_start
"""Persist an LLM run.""" if run.parent_run_id is None: run.reference_example_id = self.example_id self._submit(self._persist_run_single, _copy(run))
def _on_llm_start(self, run: Run) ->None: """Persist an LLM run.""" if run.parent_run_id is None: run.reference_example_id = self.example_id self._submit(self._persist_run_single, _copy(run))
Persist an LLM run.
similarity_search
"""Return documents most similar to the query Args: query: String to query the vectorstore with. k: Number of documents to return. Returns: List of documents most similar to the query. """ embedding = self._embedding.embed_query(query) docs = self._connectio...
def similarity_search(self, query: str, k: int=4, **kwargs: Any) ->List[ Document]: """Return documents most similar to the query Args: query: String to query the vectorstore with. k: Number of documents to return. Returns: List of documents most similar to ...
Return documents most similar to the query Args: query: String to query the vectorstore with. k: Number of documents to return. Returns: List of documents most similar to the query.
check_compatibility
"""Check if a vectorstore is compatible with the indexing API.""" methods = ['delete', 'add_documents'] for method in methods: if not hasattr(vector_store, method): return False if getattr(vector_store, 'delete') == VectorStore.delete: return False return True
def check_compatibility(vector_store: VectorStore) ->bool: """Check if a vectorstore is compatible with the indexing API.""" methods = ['delete', 'add_documents'] for method in methods: if not hasattr(vector_store, method): return False if getattr(vector_store, 'delete') == VectorSto...
Check if a vectorstore is compatible with the indexing API.
test_python_text_splitter
splitter = PythonCodeTextSplitter(chunk_size=30, chunk_overlap=0) splits = splitter.split_text(FAKE_PYTHON_TEXT) split_0 = """class Foo: def bar():""" split_1 = 'def foo():' split_2 = 'def testing_func():' split_3 = 'def bar():' expected_splits = [split_0, split_1, split_2, split_3] assert splits == expected_split...
def test_python_text_splitter() ->None: splitter = PythonCodeTextSplitter(chunk_size=30, chunk_overlap=0) splits = splitter.split_text(FAKE_PYTHON_TEXT) split_0 = """class Foo: def bar():""" split_1 = 'def foo():' split_2 = 'def testing_func():' split_3 = 'def bar():' expected_splits = ...
null
_evaluate_string_pairs
"""Evaluate whether output A is preferred to output B. Args: prediction (str): The output string from the first model. prediction_b (str): The output string from the second model. input (str, optional): The input or task string. callbacks (Callbacks, optional): T...
def _evaluate_string_pairs(self, *, prediction: str, prediction_b: str, input: Optional[str]=None, reference: Optional[str]=None, callbacks: Callbacks=None, tags: Optional[List[str]]=None, metadata: Optional[Dict [str, Any]]=None, include_run_info: bool=False, **kwargs: Any) ->dict: """Evaluate whether ...
Evaluate whether output A is preferred to output B. Args: prediction (str): The output string from the first model. prediction_b (str): The output string from the second model. input (str, optional): The input or task string. callbacks (Callbacks, optional): The callbacks to use. reference (str, op...
_keyword
if t.arg is None: self.write('**') else: self.write(t.arg) self.write('=') self.dispatch(t.value)
def _keyword(self, t): if t.arg is None: self.write('**') else: self.write(t.arg) self.write('=') self.dispatch(t.value)
null
loader
nonlocal file_contents assert file_contents is None file_contents = Path(file_path).read_text()
def loader(file_path: str) ->None: nonlocal file_contents assert file_contents is None file_contents = Path(file_path).read_text()
null
test_json_distance_evaluator_evaluate_strings_list_diff_length
prediction = '[{"a": 1, "b": 2}, {"a": 2, "b": 3}]' reference = '[{"a": 1, "b": 2}]' result = json_distance_evaluator._evaluate_strings(prediction=prediction, reference=reference) pytest.approx(len('{"a":2,"b":3}') / len(reference.replace(' ', '')), result['score'])
@pytest.mark.requires('rapidfuzz') def test_json_distance_evaluator_evaluate_strings_list_diff_length( json_distance_evaluator: JsonEditDistanceEvaluator) ->None: prediction = '[{"a": 1, "b": 2}, {"a": 2, "b": 3}]' reference = '[{"a": 1, "b": 2}]' result = json_distance_evaluator._evaluate_strings(predi...
null
_call_after_llm_before_scoring
next_chain_inputs = event.inputs.copy() value = next(iter(event.to_select_from.values())) v = value[event.selected.index ] if event.selected else event.to_select_from.values() next_chain_inputs.update({self.selected_based_on_input_key: str(event. based_on), self.selected_input_key: v}) return next_chain_inputs,...
def _call_after_llm_before_scoring(self, llm_response: str, event: PickBestEvent) ->Tuple[Dict[str, Any], PickBestEvent]: next_chain_inputs = event.inputs.copy() value = next(iter(event.to_select_from.values())) v = value[event.selected.index ] if event.selected else event.to_select_from.values(...
null
get_connection_string
connection_string: str = get_from_dict_or_env(data=kwargs, key= 'connection_string', env_key='PG_CONNECTION_STRING') if not connection_string: raise ValueError( 'Postgres connection string is requiredEither pass it as a parameteror set the PG_CONNECTION_STRING environment variable.' ) return con...
@classmethod def get_connection_string(cls, kwargs: Dict[str, Any]) ->str: connection_string: str = get_from_dict_or_env(data=kwargs, key= 'connection_string', env_key='PG_CONNECTION_STRING') if not connection_string: raise ValueError( 'Postgres connection string is requiredEither pa...
null
get_lc_namespace
"""Get the namespace of the langchain object.""" return ['langchain', 'schema', 'messages']
@classmethod def get_lc_namespace(cls) ->List[str]: """Get the namespace of the langchain object.""" return ['langchain', 'schema', 'messages']
Get the namespace of the langchain object.
test_tiledb_mmr_with_metadatas_and_list_filter
texts = ['foo', 'fou', 'foy', 'foo'] metadatas = [{'page': i} for i in range(len(texts))] docsearch = TileDB.from_texts(texts=texts, metadatas=metadatas, embedding= ConsistentFakeEmbeddings(), index_uri=f'{str(tmp_path)}/flat', index_type='FLAT') query_vec = ConsistentFakeEmbeddings().embed_query(text='foo') ou...
@pytest.mark.requires('tiledb-vector-search') def test_tiledb_mmr_with_metadatas_and_list_filter(tmp_path: Path) ->None: texts = ['foo', 'fou', 'foy', 'foo'] metadatas = [{'page': i} for i in range(len(texts))] docsearch = TileDB.from_texts(texts=texts, metadatas=metadatas, embedding=ConsistentFakeE...
null
test_bittensor_call
"""Test valid call to validator endpoint.""" llm = NIBittensorLLM(system_prompt='Your task is to answer user prompt.') output = llm('Say foo:') assert isinstance(output, str)
def test_bittensor_call() ->None: """Test valid call to validator endpoint.""" llm = NIBittensorLLM(system_prompt='Your task is to answer user prompt.') output = llm('Say foo:') assert isinstance(output, str)
Test valid call to validator endpoint.
validate_environments
"""Validate Arcee environment variables.""" values['arcee_api_key'] = convert_to_secret_str(get_from_dict_or_env(values, 'arcee_api_key', 'ARCEE_API_KEY')) values['arcee_api_url'] = get_from_dict_or_env(values, 'arcee_api_url', 'ARCEE_API_URL') values['arcee_app_url'] = get_from_dict_or_env(values, 'arcee_app_u...
@root_validator(pre=False) def validate_environments(cls, values: Dict) ->Dict: """Validate Arcee environment variables.""" values['arcee_api_key'] = convert_to_secret_str(get_from_dict_or_env( values, 'arcee_api_key', 'ARCEE_API_KEY')) values['arcee_api_url'] = get_from_dict_or_env(values, 'arcee_a...
Validate Arcee environment variables.
_get_embedding_collection_store
global _classes if _classes is not None: return _classes from pgvector.sqlalchemy import Vector class CollectionStore(BaseModel): """Collection store.""" __tablename__ = 'langchain_pg_collection' name = sqlalchemy.Column(sqlalchemy.String) cmetadata = sqlalchemy.Column(JSON) embeddings = relatio...
def _get_embedding_collection_store() ->Any: global _classes if _classes is not None: return _classes from pgvector.sqlalchemy import Vector class CollectionStore(BaseModel): """Collection store.""" __tablename__ = 'langchain_pg_collection' name = sqlalchemy.Column(sqla...
null
test_langchain_together_embedding_query
"""Test cohere embeddings.""" document = 'foo bar' embedding = TogetherEmbeddings(model= 'togethercomputer/m2-bert-80M-8k-retrieval') output = embedding.embed_query(document) assert len(output) > 0
def test_langchain_together_embedding_query() ->None: """Test cohere embeddings.""" document = 'foo bar' embedding = TogetherEmbeddings(model= 'togethercomputer/m2-bert-80M-8k-retrieval') output = embedding.embed_query(document) assert len(output) > 0
Test cohere embeddings.
_make_session
"""Create a context manager for the session, bind to _conn string.""" yield Session(self._bind)
@contextlib.contextmanager def _make_session(self) ->Generator[Session, None, None]: """Create a context manager for the session, bind to _conn string.""" yield Session(self._bind)
Create a context manager for the session, bind to _conn string.
test_ai_message_first
with pytest.raises(ValueError) as info: model.predict_messages([AIMessage(content='ai-msg-1')]) assert info.value.args[0 ] == 'messages list must start with a SystemMessage or UserMessage'
def test_ai_message_first(model: Llama2Chat) ->None: with pytest.raises(ValueError) as info: model.predict_messages([AIMessage(content='ai-msg-1')]) assert info.value.args[0 ] == 'messages list must start with a SystemMessage or UserMessage'
null
process_value
"""Convert a value to a string and add double quotes if it is a string. It required for comparators involving strings. Args: value: The value to convert. Returns: The converted value as a string. """ if isinstance(value, str): return f'"{value}"' else: return str(value)
def process_value(value: Union[int, float, str]) ->str: """Convert a value to a string and add double quotes if it is a string. It required for comparators involving strings. Args: value: The value to convert. Returns: The converted value as a string. """ if isinstance(value, ...
Convert a value to a string and add double quotes if it is a string. It required for comparators involving strings. Args: value: The value to convert. Returns: The converted value as a string.
_on_chain_end
"""Process the Chain Run.""" self._process_end_trace(run)
def _on_chain_end(self, run: 'Run') ->None: """Process the Chain Run.""" self._process_end_trace(run)
Process the Chain Run.
_call
"""Call out to Amazon API Gateway model. Args: prompt: The prompt to pass into the model. stop: Optional list of stop words to use when generating. Returns: The string generated by the model. Example: .. code-block:: python resp...
def _call(self, prompt: str, stop: Optional[List[str]]=None, run_manager: Optional[CallbackManagerForLLMRun]=None, **kwargs: Any) ->str: """Call out to Amazon API Gateway model. Args: prompt: The prompt to pass into the model. stop: Optional list of stop words to use when genera...
Call out to Amazon API Gateway model. Args: prompt: The prompt to pass into the model. stop: Optional list of stop words to use when generating. Returns: The string generated by the model. Example: .. code-block:: python response = se("Tell me a joke.")
requires_reference
""" This evaluator requires a reference. """ return True
@property def requires_reference(self) ->bool: """ This evaluator requires a reference. """ return True
This evaluator requires a reference.
split_text
"""Split text into multiple components."""
@abstractmethod def split_text(self, text: str) ->List[str]: """Split text into multiple components."""
Split text into multiple components.
on_llm_start
if self.__has_valid_config is False: return try: user_id = _get_user_id(metadata) user_props = _get_user_props(metadata) params = kwargs.get('invocation_params', {}) params.update(serialized.get('kwargs', {})) name = params.get('model') or params.get('model_name') or params.get( 'model_i...
def on_llm_start(self, serialized: Dict[str, Any], prompts: List[str], *, run_id: UUID, parent_run_id: Union[UUID, None]=None, tags: Union[List[ str], None]=None, metadata: Union[Dict[str, Any], None]=None, **kwargs: Any ) ->None: if self.__has_valid_config is False: return try: user...
null
_identifying_params
"""Get the identifying parameters.""" return {**{'model_kwargs': self.model_kwargs}}
@property def _identifying_params(self) ->Mapping[str, Any]: """Get the identifying parameters.""" return {**{'model_kwargs': self.model_kwargs}}
Get the identifying parameters.
query
result = self.execute(query, retry=retry) columns = result.keys() d: Dict[str, list] = {} for col_num in range(result.col_size()): col_name = columns[col_num] col_list = result.column_values(col_name) d[col_name] = [x.cast() for x in col_list] return d
def query(self, query: str, retry: int=0) ->Dict[str, Any]: result = self.execute(query, retry=retry) columns = result.keys() d: Dict[str, list] = {} for col_num in range(result.col_size()): col_name = columns[col_num] col_list = result.column_values(col_name) d[col_name] = [x.ca...
null
similarity_search_with_score
"""Run similarity search with Chroma with distance. Args: query (str): Query text to search for. k (int): Number of results to return. Defaults to 4. filter (Optional[Dict[str, str]]): Filter by metadata. Defaults to None. Returns: List[Tuple[Document, f...
def similarity_search_with_score(self, query: str, k: int=DEFAULT_K, filter: Optional[Dict[str, str]]=None, where_document: Optional[Dict[str, str]] =None, **kwargs: Any) ->List[Tuple[Document, float]]: """Run similarity search with Chroma with distance. Args: query (str): Query text to...
Run similarity search with Chroma with distance. Args: query (str): Query text to search for. k (int): Number of results to return. Defaults to 4. filter (Optional[Dict[str, str]]): Filter by metadata. Defaults to None. Returns: List[Tuple[Document, float]]: List of documents most similar to the q...
InputType
for step in self.steps.values(): if step.InputType: return step.InputType return Any
@property def InputType(self) ->Any: for step in self.steps.values(): if step.InputType: return step.InputType return Any
null
lazy_parse
"""Parses a blob lazily. Args: blobs: a Blob to parse This is a long-running operation. A recommended way is to batch documents together and use the `batch_parse()` method. """ yield from self.batch_parse([blob], gcs_output_path=self._gcs_output_path)
def lazy_parse(self, blob: Blob) ->Iterator[Document]: """Parses a blob lazily. Args: blobs: a Blob to parse This is a long-running operation. A recommended way is to batch documents together and use the `batch_parse()` method. """ yield from self.batch_parse([b...
Parses a blob lazily. Args: blobs: a Blob to parse This is a long-running operation. A recommended way is to batch documents together and use the `batch_parse()` method.
foo
"""Docstring Args: bar: str """ assert callbacks is not None return 'foo' + bar
def foo(bar: str, callbacks: Optional[CallbackManagerForToolRun]=None) ->str: """Docstring Args: bar: str """ assert callbacks is not None return 'foo' + bar
Docstring Args: bar: str
lazy_load
"""Lazy load documents""" if self.web_path and self._is_s3_url(self.web_path): blob = Blob(path=self.web_path) else: blob = Blob.from_path(self.file_path) if AmazonTextractPDFLoader._get_number_of_pages(blob) > 1: raise ValueError( f'the file {blob.path} is a multi-page document, ...
def lazy_load(self) ->Iterator[Document]: """Lazy load documents""" if self.web_path and self._is_s3_url(self.web_path): blob = Blob(path=self.web_path) else: blob = Blob.from_path(self.file_path) if AmazonTextractPDFLoader._get_number_of_pages(blob) > 1: raise ValueError...
Lazy load documents
test_compatible_vectorstore_documentation
"""Test which vectorstores are compatible with the indexing API. This serves as a reminder to update the documentation in [1] that specifies which vectorstores are compatible with the indexing API. Ideally if a developer adds a new vectorstore or modifies an existing one in such a way that affects...
def test_compatible_vectorstore_documentation() ->None: """Test which vectorstores are compatible with the indexing API. This serves as a reminder to update the documentation in [1] that specifies which vectorstores are compatible with the indexing API. Ideally if a developer adds a new vectorstor...
Test which vectorstores are compatible with the indexing API. This serves as a reminder to update the documentation in [1] that specifies which vectorstores are compatible with the indexing API. Ideally if a developer adds a new vectorstore or modifies an existing one in such a way that affects its compatibility with...
buffer
return self.chat_memory.messages
@property def buffer(self) ->List[BaseMessage]: return self.chat_memory.messages
null
api_client
return RedditSearchAPIWrapper()
@pytest.fixture def api_client() ->RedditSearchAPIWrapper: return RedditSearchAPIWrapper()
null
test_init_from_client
import chromadb client = chromadb.Client(chromadb.config.Settings()) Chroma(client=client)
def test_init_from_client() ->None: import chromadb client = chromadb.Client(chromadb.config.Settings()) Chroma(client=client)
null
load_chain
"""Unified method for loading a chain from LangChainHub or local fs.""" if (hub_result := try_load_from_hub(path, _load_chain_from_file, 'chains', {'json', 'yaml'}, **kwargs)): return hub_result else: return _load_chain_from_file(path, **kwargs)
def load_chain(path: Union[str, Path], **kwargs: Any) ->Chain: """Unified method for loading a chain from LangChainHub or local fs.""" if (hub_result := try_load_from_hub(path, _load_chain_from_file, 'chains', {'json', 'yaml'}, **kwargs)): return hub_result else: return _load_chain_f...
Unified method for loading a chain from LangChainHub or local fs.
_import_ainetwork_transfer
from langchain_community.tools.ainetwork.transfer import AINTransfer return AINTransfer
def _import_ainetwork_transfer() ->Any: from langchain_community.tools.ainetwork.transfer import AINTransfer return AINTransfer
null
generate
"""Generate LLM result from inputs.""" prompts, stop = self.prep_prompts(input_list, run_manager=run_manager) callbacks = run_manager.get_child() if run_manager else None if isinstance(self.llm, BaseLanguageModel): return self.llm.generate_prompt(prompts, stop, callbacks=callbacks, ** self.llm_kwargs) else:...
def generate(self, input_list: List[Dict[str, Any]], run_manager: Optional[ CallbackManagerForChainRun]=None) ->LLMResult: """Generate LLM result from inputs.""" prompts, stop = self.prep_prompts(input_list, run_manager=run_manager) callbacks = run_manager.get_child() if run_manager else None if isi...
Generate LLM result from inputs.
test_structured_tool_from_function_docstring
"""Test that structured tools can be created from functions.""" def foo(bar: int, baz: str) ->str: """Docstring Args: bar: int baz: str """ raise NotImplementedError() structured_tool = StructuredTool.from_function(foo) assert structured_tool.name == 'foo' assert structur...
def test_structured_tool_from_function_docstring() ->None: """Test that structured tools can be created from functions.""" def foo(bar: int, baz: str) ->str: """Docstring Args: bar: int baz: str """ raise NotImplementedError() structured_tool = Struct...
Test that structured tools can be created from functions.
__init__
"""Initialize Trello loader. Args: client: Trello API client. board_name: The name of the Trello board. include_card_name: Whether to include the name of the card in the document. include_comments: Whether to include the comments on the card in the ...
def __init__(self, client: TrelloClient, board_name: str, *, include_card_name: bool=True, include_comments: bool=True, include_checklist: bool=True, card_filter: Literal['closed', 'open', 'all']='all', extra_metadata: Tuple[str, ...]=('due_date', 'labels', 'list', 'closed')): """Initialize Trello l...
Initialize Trello loader. Args: client: Trello API client. board_name: The name of the Trello board. include_card_name: Whether to include the name of the card in the document. include_comments: Whether to include the comments on the card in the document. include_checklist: Whether to inclu...
_pushText
field = {'textfield': {'text': {'body': text, 'format': 0}}, 'processing_options': {'ml_text': self._config['enable_ml']}} return self._pushField(id, field)
def _pushText(self, id: str, text: str) ->str: field = {'textfield': {'text': {'body': text, 'format': 0}}, 'processing_options': {'ml_text': self._config['enable_ml']}} return self._pushField(id, field)
null
_return
if run_manager: run_manager.on_agent_finish(output, color='green', verbose=self.verbose) final_output = output.return_values if self.return_intermediate_steps: final_output['intermediate_steps'] = intermediate_steps return final_output
def _return(self, output: AgentFinish, intermediate_steps: list, run_manager: Optional[CallbackManagerForChainRun]=None) ->Dict[str, Any]: if run_manager: run_manager.on_agent_finish(output, color='green', verbose=self.verbose ) final_output = output.return_values if self.return_inte...
null
_chain_type
return 'api_chain'
@property def _chain_type(self) ->str: return 'api_chain'
null
test_incremental_delete
"""Test indexing with incremental deletion strategy.""" loader = ToyLoader(documents=[Document(page_content= 'This is a test document.', metadata={'source': '1'}), Document( page_content='This is another document.', metadata={'source': '2'})]) with patch.object(record_manager, 'get_time', return_value=datetime(...
def test_incremental_delete(record_manager: SQLRecordManager, vector_store: InMemoryVectorStore) ->None: """Test indexing with incremental deletion strategy.""" loader = ToyLoader(documents=[Document(page_content= 'This is a test document.', metadata={'source': '1'}), Document( page_content=...
Test indexing with incremental deletion strategy.
test_visit_operation
op = Operation(operator=Operator.AND, arguments=[Comparison(comparator= Comparator.EQ, attribute='foo', value='hello'), Comparison(comparator= Comparator.GTE, attribute='bar', value={'type': 'date', 'date': '2023-09-13'}), Comparison(comparator=Comparator.LTE, attribute='abc', value=1.4)]) expected = {'...
def test_visit_operation() ->None: op = Operation(operator=Operator.AND, arguments=[Comparison(comparator= Comparator.EQ, attribute='foo', value='hello'), Comparison( comparator=Comparator.GTE, attribute='bar', value={'type': 'date', 'date': '2023-09-13'}), Comparison(comparator=Comparator.L...
null
invoke
config = ensure_config(config) return self.get_relevant_documents(input, callbacks=config.get('callbacks'), tags=config.get('tags'), metadata=config.get('metadata'), run_name= config.get('run_name'))
def invoke(self, input: str, config: Optional[RunnableConfig]=None) ->List[ Document]: config = ensure_config(config) return self.get_relevant_documents(input, callbacks=config.get( 'callbacks'), tags=config.get('tags'), metadata=config.get( 'metadata'), run_name=config.get('run_name'))
null
_create_retry_decorator
from grpc import RpcError min_seconds = 1 max_seconds = 60 return retry(reraise=True, stop=stop_after_attempt(llm.max_retries), wait= wait_exponential(multiplier=1, min=min_seconds, max=max_seconds), retry =retry_if_exception_type(RpcError), before_sleep=before_sleep_log( logger, logging.WARNING))
def _create_retry_decorator(llm: ChatYandexGPT) ->Callable[[Any], Any]: from grpc import RpcError min_seconds = 1 max_seconds = 60 return retry(reraise=True, stop=stop_after_attempt(llm.max_retries), wait=wait_exponential(multiplier=1, min=min_seconds, max= max_seconds), retry=retry_if_e...
null
__init__
super().__init__(pydantic_object=LineList)
def __init__(self) ->None: super().__init__(pydantic_object=LineList)
null
test_gpt_router_streaming_callback
"""Test that streaming correctly invokes on_llm_new_token callback.""" callback_handler = FakeCallbackHandler() callback_manager = CallbackManager([callback_handler]) anthropic_claude = GPTRouterModel(name='claude-instant-1.2', provider_name= 'anthropic') chat = GPTRouter(models_priority_list=[anthropic_claude], st...
def test_gpt_router_streaming_callback() ->None: """Test that streaming correctly invokes on_llm_new_token callback.""" callback_handler = FakeCallbackHandler() callback_manager = CallbackManager([callback_handler]) anthropic_claude = GPTRouterModel(name='claude-instant-1.2', provider_name='anth...
Test that streaming correctly invokes on_llm_new_token callback.
parse
try: parsed = json.loads(text, strict=False) except json.JSONDecodeError: preprocessed_text = preprocess_json_input(text) try: parsed = json.loads(preprocessed_text, strict=False) except Exception: return AutoGPTAction(name='ERROR', args={'error': f'Could not parse invalid js...
def parse(self, text: str) ->AutoGPTAction: try: parsed = json.loads(text, strict=False) except json.JSONDecodeError: preprocessed_text = preprocess_json_input(text) try: parsed = json.loads(preprocessed_text, strict=False) except Exception: return AutoGPT...
null
test_indexing_with_no_docs
"""Check edge case when loader returns no new docs.""" loader = ToyLoader(documents=[]) assert index(loader, record_manager, vector_store, cleanup='full') == { 'num_added': 0, 'num_deleted': 0, 'num_skipped': 0, 'num_updated': 0}
def test_indexing_with_no_docs(record_manager: SQLRecordManager, vector_store: VectorStore) ->None: """Check edge case when loader returns no new docs.""" loader = ToyLoader(documents=[]) assert index(loader, record_manager, vector_store, cleanup='full') == { 'num_added': 0, 'num_deleted': 0, 'n...
Check edge case when loader returns no new docs.
_llm_type
return 'wasm-chat'
@property def _llm_type(self) ->str: return 'wasm-chat'
null
__init__
try: import kuzu except ImportError: raise ImportError( 'Could not import Kùzu python package.Please install Kùzu with `pip install kuzu`.' ) self.db = db self.conn = kuzu.Connection(self.db) self.database = database self.refresh_schema()
def __init__(self, db: Any, database: str='kuzu') ->None: try: import kuzu except ImportError: raise ImportError( 'Could not import Kùzu python package.Please install Kùzu with `pip install kuzu`.' ) self.db = db self.conn = kuzu.Connection(self.db) self.datab...
null
__init__
"""Initialize the UpstashRedisStore with HTTP API. Must provide either an Upstash Redis client or a url. Args: client: An Upstash Redis instance url: UPSTASH_REDIS_REST_URL token: UPSTASH_REDIS_REST_TOKEN ttl: time to expire keys in seconds if provided, ...
def __init__(self, *, client: Any=None, url: Optional[str]=None, token: Optional[str]=None, ttl: Optional[int]=None, namespace: Optional[str]=None ) ->None: """Initialize the UpstashRedisStore with HTTP API. Must provide either an Upstash Redis client or a url. Args: client: An...
Initialize the UpstashRedisStore with HTTP API. Must provide either an Upstash Redis client or a url. Args: client: An Upstash Redis instance url: UPSTASH_REDIS_REST_URL token: UPSTASH_REDIS_REST_TOKEN ttl: time to expire keys in seconds if provided, if None keys will never expire namespa...
authorization
if self.beam_client_id: credential_str = self.beam_client_id + ':' + self.beam_client_secret else: credential_str = self.beam_client_secret return base64.b64encode(credential_str.encode()).decode()
@property def authorization(self) ->str: if self.beam_client_id: credential_str = self.beam_client_id + ':' + self.beam_client_secret else: credential_str = self.beam_client_secret return base64.b64encode(credential_str.encode()).decode()
null
test_all_imports
assert set(__all__) == set(EXPECTED_ALL)
def test_all_imports() ->None: assert set(__all__) == set(EXPECTED_ALL)
null
_similarity_search
""" Perform a similarity search. Args: query_emb: Query represented as an embedding Returns: A list of documents most similar to the query """ docs = self._search(query_emb=query_emb, top_k=self.top_k) results = [self._docarray_to_langchain_doc(doc) for doc in d...
def _similarity_search(self, query_emb: np.ndarray) ->List[Document]: """ Perform a similarity search. Args: query_emb: Query represented as an embedding Returns: A list of documents most similar to the query """ docs = self._search(query_emb=query_emb, ...
Perform a similarity search. Args: query_emb: Query represented as an embedding Returns: A list of documents most similar to the query
get_entity_knowledge
"""Get information about an entity.""" import networkx as nx if not self._graph.has_node(entity): return [] results = [] for src, sink in nx.dfs_edges(self._graph, entity, depth_limit=depth): relation = self._graph[src][sink]['relation'] results.append(f'{src} {relation} {sink}') return results
def get_entity_knowledge(self, entity: str, depth: int=1) ->List[str]: """Get information about an entity.""" import networkx as nx if not self._graph.has_node(entity): return [] results = [] for src, sink in nx.dfs_edges(self._graph, entity, depth_limit=depth): relation = self._grap...
Get information about an entity.
test_load_no_result
docs = retriever.get_relevant_documents( 'NORESULTCALL_NORESULTCALL_NORESULTCALL_NORESULTCALL_NORESULTCALL_NORESULTCALL' ) assert not docs
def test_load_no_result(retriever: WikipediaRetriever) ->None: docs = retriever.get_relevant_documents( 'NORESULTCALL_NORESULTCALL_NORESULTCALL_NORESULTCALL_NORESULTCALL_NORESULTCALL' ) assert not docs
null
__getattr__
if name == 'AnalyticDB': return _import_analyticdb() elif name == 'AlibabaCloudOpenSearch': return _import_alibaba_cloud_open_search() elif name == 'AlibabaCloudOpenSearchSettings': return _import_alibaba_cloud_open_search_settings() elif name == 'AzureCosmosDBVectorSearch': return _import_azure_cosmos_...
def __getattr__(name: str) ->Any: if name == 'AnalyticDB': return _import_analyticdb() elif name == 'AlibabaCloudOpenSearch': return _import_alibaba_cloud_open_search() elif name == 'AlibabaCloudOpenSearchSettings': return _import_alibaba_cloud_open_search_settings() elif name ==...
null
similarity_search_with_score
""" Run similarity search with Deep Lake with distance returned. Examples: >>> data = vector_store.similarity_search_with_score( ... query=<your_query>, ... embedding=<your_embedding_function> ... k=<number_of_items_to_return>, ... exec_option=<pr...
def similarity_search_with_score(self, query: str, k: int=4, **kwargs: Any ) ->List[Tuple[Document, float]]: """ Run similarity search with Deep Lake with distance returned. Examples: >>> data = vector_store.similarity_search_with_score( ... query=<your_query>, ... ...
Run similarity search with Deep Lake with distance returned. Examples: >>> data = vector_store.similarity_search_with_score( ... query=<your_query>, ... embedding=<your_embedding_function> ... k=<number_of_items_to_return>, ... exec_option=<preferred_exec_option>, ... ) Args: query (str): Query te...
_run
body = {'this': tool_input} response = requests.post(self.url, data=body) return response.text
def _run(self, tool_input: str, run_manager: Optional[ CallbackManagerForToolRun]=None) ->str: body = {'this': tool_input} response = requests.post(self.url, data=body) return response.text
null
test_action_w_namespace_no_emb
str1 = 'test1' str2 = 'test2' str3 = 'test3' expected = [{'test_namespace': str1}, {'test_namespace': str2}, { 'test_namespace': str3}] assert base.embed([{'test_namespace': str1}, {'test_namespace': str2}, { 'test_namespace': str3}], MockEncoder()) == expected
@pytest.mark.requires('vowpal_wabbit_next') def test_action_w_namespace_no_emb() ->None: str1 = 'test1' str2 = 'test2' str3 = 'test3' expected = [{'test_namespace': str1}, {'test_namespace': str2}, { 'test_namespace': str3}] assert base.embed([{'test_namespace': str1}, {'test_namespace': str...
null
_root_to_dict
"""Converts xml tree to python dictionary.""" result: Dict[str, List[Any]] = {root.tag: []} for child in root: if len(child) == 0: result[root.tag].append({child.tag: child.text}) else: result[root.tag].append(self._root_to_dict(child)) return result
def _root_to_dict(self, root: ET.Element) ->Dict[str, List[Any]]: """Converts xml tree to python dictionary.""" result: Dict[str, List[Any]] = {root.tag: []} for child in root: if len(child) == 0: result[root.tag].append({child.tag: child.text}) else: result[root.tag]...
Converts xml tree to python dictionary.
test_faiss_local_save_load
"""Test end to end serialization.""" texts = ['foo', 'bar', 'baz'] docsearch = FAISS.from_texts(texts, FakeEmbeddings()) temp_timestamp = datetime.datetime.utcnow().strftime('%Y%m%d-%H%M%S') with tempfile.TemporaryDirectory(suffix='_' + temp_timestamp + '/' ) as temp_folder: docsearch.save_local(temp_folder) ...
@pytest.mark.requires('faiss') def test_faiss_local_save_load() ->None: """Test end to end serialization.""" texts = ['foo', 'bar', 'baz'] docsearch = FAISS.from_texts(texts, FakeEmbeddings()) temp_timestamp = datetime.datetime.utcnow().strftime('%Y%m%d-%H%M%S') with tempfile.TemporaryDirectory(suff...
Test end to end serialization.