id
stringlengths
14
15
text
stringlengths
35
2.51k
source
stringlengths
61
154
a947e9d153c6-3
"""Clear session memory from Zep. Note that Zep is long-term storage for memory and this is not advised unless you have specific data retention requirements. """ try: self.zep_client.delete_memory(self.session_id) except NotFoundError: logger.warning( f"Session {self.session_id} not found in Zep. Skipping delete." )
https://api.python.langchain.com/en/latest/_modules/langchain/memory/chat_message_histories/zep.html
0a03539643f7-0
Source code for langchain.memory.chat_message_histories.redis import json import logging from typing import List, Optional from langchain.schema import ( BaseChatMessageHistory, BaseMessage, _message_to_dict, messages_from_dict, ) logger = logging.getLogger(__name__) [docs]class RedisChatMessageHistory(BaseChatMessageHistory): """Chat message history stored in a Redis database.""" def __init__( self, session_id: str, url: str = "redis://localhost:6379/0", key_prefix: str = "message_store:", ttl: Optional[int] = None, ): try: import redis except ImportError: raise ImportError( "Could not import redis python package. " "Please install it with `pip install redis`." ) try: self.redis_client = redis.Redis.from_url(url=url) except redis.exceptions.ConnectionError as error: logger.error(error) self.session_id = session_id self.key_prefix = key_prefix self.ttl = ttl @property def key(self) -> str: """Construct the record key to use""" return self.key_prefix + self.session_id @property def messages(self) -> List[BaseMessage]: # type: ignore """Retrieve the messages from Redis""" _items = self.redis_client.lrange(self.key, 0, -1) items = [json.loads(m.decode("utf-8")) for m in _items[::-1]] messages = messages_from_dict(items) return messages [docs] def add_message(self, message: BaseMessage) -> None: """Append the message to the record in Redis"""
https://api.python.langchain.com/en/latest/_modules/langchain/memory/chat_message_histories/redis.html
0a03539643f7-1
"""Append the message to the record in Redis""" self.redis_client.lpush(self.key, json.dumps(_message_to_dict(message))) if self.ttl: self.redis_client.expire(self.key, self.ttl) [docs] def clear(self) -> None: """Clear session memory from Redis""" self.redis_client.delete(self.key)
https://api.python.langchain.com/en/latest/_modules/langchain/memory/chat_message_histories/redis.html
e1c467464fa0-0
Source code for langchain.memory.chat_message_histories.postgres import json import logging from typing import List from langchain.schema import ( BaseChatMessageHistory, BaseMessage, _message_to_dict, messages_from_dict, ) logger = logging.getLogger(__name__) DEFAULT_CONNECTION_STRING = "postgresql://postgres:mypassword@localhost/chat_history" [docs]class PostgresChatMessageHistory(BaseChatMessageHistory): """Chat message history stored in a Postgres database.""" def __init__( self, session_id: str, connection_string: str = DEFAULT_CONNECTION_STRING, table_name: str = "message_store", ): import psycopg from psycopg.rows import dict_row try: self.connection = psycopg.connect(connection_string) self.cursor = self.connection.cursor(row_factory=dict_row) except psycopg.OperationalError as error: logger.error(error) self.session_id = session_id self.table_name = table_name self._create_table_if_not_exists() def _create_table_if_not_exists(self) -> None: create_table_query = f"""CREATE TABLE IF NOT EXISTS {self.table_name} ( id SERIAL PRIMARY KEY, session_id TEXT NOT NULL, message JSONB NOT NULL );""" self.cursor.execute(create_table_query) self.connection.commit() @property def messages(self) -> List[BaseMessage]: # type: ignore """Retrieve the messages from PostgreSQL""" query = ( f"SELECT message FROM {self.table_name} WHERE session_id = %s ORDER BY id;" ) self.cursor.execute(query, (self.session_id,)) items = [record["message"] for record in self.cursor.fetchall()]
https://api.python.langchain.com/en/latest/_modules/langchain/memory/chat_message_histories/postgres.html
e1c467464fa0-1
items = [record["message"] for record in self.cursor.fetchall()] messages = messages_from_dict(items) return messages [docs] def add_message(self, message: BaseMessage) -> None: """Append the message to the record in PostgreSQL""" from psycopg import sql query = sql.SQL("INSERT INTO {} (session_id, message) VALUES (%s, %s);").format( sql.Identifier(self.table_name) ) self.cursor.execute( query, (self.session_id, json.dumps(_message_to_dict(message))) ) self.connection.commit() [docs] def clear(self) -> None: """Clear session memory from PostgreSQL""" query = f"DELETE FROM {self.table_name} WHERE session_id = %s;" self.cursor.execute(query, (self.session_id,)) self.connection.commit() def __del__(self) -> None: if self.cursor: self.cursor.close() if self.connection: self.connection.close()
https://api.python.langchain.com/en/latest/_modules/langchain/memory/chat_message_histories/postgres.html
08cfa072edc1-0
Source code for langchain.memory.chat_message_histories.cassandra """Cassandra-based chat message history, based on cassIO.""" from __future__ import annotations import json import typing from typing import List if typing.TYPE_CHECKING: from cassandra.cluster import Session from langchain.schema import ( BaseChatMessageHistory, BaseMessage, _message_to_dict, messages_from_dict, ) DEFAULT_TABLE_NAME = "message_store" DEFAULT_TTL_SECONDS = None [docs]class CassandraChatMessageHistory(BaseChatMessageHistory): """Chat message history that stores history in Cassandra. Args: session_id: arbitrary key that is used to store the messages of a single chat session. session: a Cassandra `Session` object (an open DB connection) keyspace: name of the keyspace to use. table_name: name of the table to use. ttl_seconds: time-to-live (seconds) for automatic expiration of stored entries. None (default) for no expiration. """ def __init__( self, session_id: str, session: Session, keyspace: str, table_name: str = DEFAULT_TABLE_NAME, ttl_seconds: int | None = DEFAULT_TTL_SECONDS, ) -> None: try: from cassio.history import StoredBlobHistory except (ImportError, ModuleNotFoundError): raise ValueError( "Could not import cassio python package. " "Please install it with `pip install cassio`." ) self.session_id = session_id self.ttl_seconds = ttl_seconds self.blob_history = StoredBlobHistory(session, keyspace, table_name) @property
https://api.python.langchain.com/en/latest/_modules/langchain/memory/chat_message_histories/cassandra.html
08cfa072edc1-1
@property def messages(self) -> List[BaseMessage]: # type: ignore """Retrieve all session messages from DB""" message_blobs = self.blob_history.retrieve( self.session_id, ) items = [json.loads(message_blob) for message_blob in message_blobs] messages = messages_from_dict(items) return messages [docs] def add_message(self, message: BaseMessage) -> None: """Write a message to the table""" self.blob_history.store( self.session_id, json.dumps(_message_to_dict(message)), self.ttl_seconds ) [docs] def clear(self) -> None: """Clear session memory from DB""" self.blob_history.clear_session_id(self.session_id)
https://api.python.langchain.com/en/latest/_modules/langchain/memory/chat_message_histories/cassandra.html
ca8caa9492b5-0
Source code for langchain.memory.chat_message_histories.mongodb import json import logging from typing import List from langchain.schema import ( BaseChatMessageHistory, BaseMessage, _message_to_dict, messages_from_dict, ) logger = logging.getLogger(__name__) DEFAULT_DBNAME = "chat_history" DEFAULT_COLLECTION_NAME = "message_store" [docs]class MongoDBChatMessageHistory(BaseChatMessageHistory): """Chat message history that stores history in MongoDB. Args: connection_string: connection string to connect to MongoDB session_id: arbitrary key that is used to store the messages of a single chat session. database_name: name of the database to use collection_name: name of the collection to use """ def __init__( self, connection_string: str, session_id: str, database_name: str = DEFAULT_DBNAME, collection_name: str = DEFAULT_COLLECTION_NAME, ): from pymongo import MongoClient, errors self.connection_string = connection_string self.session_id = session_id self.database_name = database_name self.collection_name = collection_name try: self.client: MongoClient = MongoClient(connection_string) except errors.ConnectionFailure as error: logger.error(error) self.db = self.client[database_name] self.collection = self.db[collection_name] self.collection.create_index("SessionId") @property def messages(self) -> List[BaseMessage]: # type: ignore """Retrieve the messages from MongoDB""" from pymongo import errors try: cursor = self.collection.find({"SessionId": self.session_id}) except errors.OperationFailure as error: logger.error(error) if cursor:
https://api.python.langchain.com/en/latest/_modules/langchain/memory/chat_message_histories/mongodb.html
ca8caa9492b5-1
except errors.OperationFailure as error: logger.error(error) if cursor: items = [json.loads(document["History"]) for document in cursor] else: items = [] messages = messages_from_dict(items) return messages [docs] def add_message(self, message: BaseMessage) -> None: """Append the message to the record in MongoDB""" from pymongo import errors try: self.collection.insert_one( { "SessionId": self.session_id, "History": json.dumps(_message_to_dict(message)), } ) except errors.WriteError as err: logger.error(err) [docs] def clear(self) -> None: """Clear session memory from MongoDB""" from pymongo import errors try: self.collection.delete_many({"SessionId": self.session_id}) except errors.WriteError as err: logger.error(err)
https://api.python.langchain.com/en/latest/_modules/langchain/memory/chat_message_histories/mongodb.html
93efad5c60e0-0
Source code for langchain.memory.chat_message_histories.momento from __future__ import annotations import json from datetime import timedelta from typing import TYPE_CHECKING, Any, Optional from langchain.schema import ( BaseChatMessageHistory, BaseMessage, _message_to_dict, messages_from_dict, ) from langchain.utils import get_from_env if TYPE_CHECKING: import momento def _ensure_cache_exists(cache_client: momento.CacheClient, cache_name: str) -> None: """Create cache if it doesn't exist. Raises: SdkException: Momento service or network error Exception: Unexpected response """ from momento.responses import CreateCache create_cache_response = cache_client.create_cache(cache_name) if isinstance(create_cache_response, CreateCache.Success) or isinstance( create_cache_response, CreateCache.CacheAlreadyExists ): return None elif isinstance(create_cache_response, CreateCache.Error): raise create_cache_response.inner_exception else: raise Exception(f"Unexpected response cache creation: {create_cache_response}") [docs]class MomentoChatMessageHistory(BaseChatMessageHistory): """Chat message history cache that uses Momento as a backend. See https://gomomento.com/""" def __init__( self, session_id: str, cache_client: momento.CacheClient, cache_name: str, *, key_prefix: str = "message_store:", ttl: Optional[timedelta] = None, ensure_cache_exists: bool = True, ): """Instantiate a chat message history cache that uses Momento as a backend. Note: to instantiate the cache client passed to MomentoChatMessageHistory,
https://api.python.langchain.com/en/latest/_modules/langchain/memory/chat_message_histories/momento.html
93efad5c60e0-1
Note: to instantiate the cache client passed to MomentoChatMessageHistory, you must have a Momento account at https://gomomento.com/. Args: session_id (str): The session ID to use for this chat session. cache_client (CacheClient): The Momento cache client. cache_name (str): The name of the cache to use to store the messages. key_prefix (str, optional): The prefix to apply to the cache key. Defaults to "message_store:". ttl (Optional[timedelta], optional): The TTL to use for the messages. Defaults to None, ie the default TTL of the cache will be used. ensure_cache_exists (bool, optional): Create the cache if it doesn't exist. Defaults to True. Raises: ImportError: Momento python package is not installed. TypeError: cache_client is not of type momento.CacheClientObject """ try: from momento import CacheClient from momento.requests import CollectionTtl except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if not isinstance(cache_client, CacheClient): raise TypeError("cache_client must be a momento.CacheClient object.") if ensure_cache_exists: _ensure_cache_exists(cache_client, cache_name) self.key = key_prefix + session_id self.cache_client = cache_client self.cache_name = cache_name if ttl is not None: self.ttl = CollectionTtl.of(ttl) else: self.ttl = CollectionTtl.from_cache_ttl() [docs] @classmethod def from_client_params( cls, session_id: str,
https://api.python.langchain.com/en/latest/_modules/langchain/memory/chat_message_histories/momento.html
93efad5c60e0-2
def from_client_params( cls, session_id: str, cache_name: str, ttl: timedelta, *, configuration: Optional[momento.config.Configuration] = None, auth_token: Optional[str] = None, **kwargs: Any, ) -> MomentoChatMessageHistory: """Construct cache from CacheClient parameters.""" try: from momento import CacheClient, Configurations, CredentialProvider except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if configuration is None: configuration = Configurations.Laptop.v1() auth_token = auth_token or get_from_env("auth_token", "MOMENTO_AUTH_TOKEN") credentials = CredentialProvider.from_string(auth_token) cache_client = CacheClient(configuration, credentials, default_ttl=ttl) return cls(session_id, cache_client, cache_name, ttl=ttl, **kwargs) @property def messages(self) -> list[BaseMessage]: # type: ignore[override] """Retrieve the messages from Momento. Raises: SdkException: Momento service or network error Exception: Unexpected response Returns: list[BaseMessage]: List of cached messages """ from momento.responses import CacheListFetch fetch_response = self.cache_client.list_fetch(self.cache_name, self.key) if isinstance(fetch_response, CacheListFetch.Hit): items = [json.loads(m) for m in fetch_response.value_list_string] return messages_from_dict(items) elif isinstance(fetch_response, CacheListFetch.Miss): return [] elif isinstance(fetch_response, CacheListFetch.Error):
https://api.python.langchain.com/en/latest/_modules/langchain/memory/chat_message_histories/momento.html
93efad5c60e0-3
return [] elif isinstance(fetch_response, CacheListFetch.Error): raise fetch_response.inner_exception else: raise Exception(f"Unexpected response: {fetch_response}") [docs] def add_message(self, message: BaseMessage) -> None: """Store a message in the cache. Args: message (BaseMessage): The message object to store. Raises: SdkException: Momento service or network error. Exception: Unexpected response. """ from momento.responses import CacheListPushBack item = json.dumps(_message_to_dict(message)) push_response = self.cache_client.list_push_back( self.cache_name, self.key, item, ttl=self.ttl ) if isinstance(push_response, CacheListPushBack.Success): return None elif isinstance(push_response, CacheListPushBack.Error): raise push_response.inner_exception else: raise Exception(f"Unexpected response: {push_response}") [docs] def clear(self) -> None: """Remove the session's messages from the cache. Raises: SdkException: Momento service or network error. Exception: Unexpected response. """ from momento.responses import CacheDelete delete_response = self.cache_client.delete(self.cache_name, self.key) if isinstance(delete_response, CacheDelete.Success): return None elif isinstance(delete_response, CacheDelete.Error): raise delete_response.inner_exception else: raise Exception(f"Unexpected response: {delete_response}")
https://api.python.langchain.com/en/latest/_modules/langchain/memory/chat_message_histories/momento.html
065f46933ba7-0
Source code for langchain.load.dump import json from typing import Any, Dict from langchain.load.serializable import Serializable, to_json_not_implemented [docs]def default(obj: Any) -> Any: """Return a default value for a Serializable object or a SerializedNotImplemented object.""" if isinstance(obj, Serializable): return obj.to_json() else: return to_json_not_implemented(obj) [docs]def dumps(obj: Any, *, pretty: bool = False) -> str: """Return a json string representation of an object.""" if pretty: return json.dumps(obj, default=default, indent=2) else: return json.dumps(obj, default=default) [docs]def dumpd(obj: Any) -> Dict[str, Any]: """Return a json dict representation of an object.""" return json.loads(dumps(obj))
https://api.python.langchain.com/en/latest/_modules/langchain/load/dump.html
0f6c15b73856-0
Source code for langchain.load.load import importlib import json import os from typing import Any, Dict, Optional from langchain.load.serializable import Serializable class Reviver: def __init__(self, secrets_map: Optional[Dict[str, str]] = None) -> None: self.secrets_map = secrets_map or dict() def __call__(self, value: Dict[str, Any]) -> Any: if ( value.get("lc", None) == 1 and value.get("type", None) == "secret" and value.get("id", None) is not None ): [key] = value["id"] if key in self.secrets_map: return self.secrets_map[key] else: if key in os.environ and os.environ[key]: return os.environ[key] raise KeyError(f'Missing key "{key}" in load(secrets_map)') if ( value.get("lc", None) == 1 and value.get("type", None) == "not_implemented" and value.get("id", None) is not None ): raise NotImplementedError( "Trying to load an object that doesn't implement " f"serialization: {value}" ) if ( value.get("lc", None) == 1 and value.get("type", None) == "constructor" and value.get("id", None) is not None ): [*namespace, name] = value["id"] # Currently, we only support langchain imports. if namespace[0] != "langchain": raise ValueError(f"Invalid namespace: {value}") # The root namespace "langchain" is not a valid identifier.
https://api.python.langchain.com/en/latest/_modules/langchain/load/load.html
0f6c15b73856-1
# The root namespace "langchain" is not a valid identifier. if len(namespace) == 1: raise ValueError(f"Invalid namespace: {value}") mod = importlib.import_module(".".join(namespace)) cls = getattr(mod, name) # The class must be a subclass of Serializable. if not issubclass(cls, Serializable): raise ValueError(f"Invalid namespace: {value}") # We don't need to recurse on kwargs # as json.loads will do that for us. kwargs = value.get("kwargs", dict()) return cls(**kwargs) return value [docs]def loads(text: str, *, secrets_map: Optional[Dict[str, str]] = None) -> Any: return json.loads(text, object_hook=Reviver(secrets_map))
https://api.python.langchain.com/en/latest/_modules/langchain/load/load.html
753c02d98e98-0
Source code for langchain.load.serializable from abc import ABC from typing import Any, Dict, List, Literal, TypedDict, Union, cast from pydantic import BaseModel, PrivateAttr [docs]class BaseSerialized(TypedDict): """Base class for serialized objects.""" lc: int id: List[str] [docs]class SerializedConstructor(BaseSerialized): """Serialized constructor.""" type: Literal["constructor"] kwargs: Dict[str, Any] [docs]class SerializedSecret(BaseSerialized): """Serialized secret.""" type: Literal["secret"] [docs]class SerializedNotImplemented(BaseSerialized): """Serialized not implemented.""" type: Literal["not_implemented"] [docs]class Serializable(BaseModel, ABC): """Serializable base class.""" @property def lc_serializable(self) -> bool: """ Return whether or not the class is serializable. """ return False @property def lc_namespace(self) -> List[str]: """ Return the namespace of the langchain object. eg. ["langchain", "llms", "openai"] """ return self.__class__.__module__.split(".") @property def lc_secrets(self) -> Dict[str, str]: """ Return a map of constructor argument names to secret ids. eg. {"openai_api_key": "OPENAI_API_KEY"} """ return dict() @property def lc_attributes(self) -> Dict: """ Return a list of attribute names that should be included in the serialized kwargs. These attributes must be accepted by the constructor. """ return {} [docs] class Config: extra = "ignore"
https://api.python.langchain.com/en/latest/_modules/langchain/load/serializable.html
753c02d98e98-1
return {} [docs] class Config: extra = "ignore" _lc_kwargs = PrivateAttr(default_factory=dict) def __init__(self, **kwargs: Any) -> None: super().__init__(**kwargs) self._lc_kwargs = kwargs [docs] def to_json(self) -> Union[SerializedConstructor, SerializedNotImplemented]: if not self.lc_serializable: return self.to_json_not_implemented() secrets = dict() # Get latest values for kwargs if there is an attribute with same name lc_kwargs = { k: getattr(self, k, v) for k, v in self._lc_kwargs.items() if not (self.__exclude_fields__ or {}).get(k, False) # type: ignore } # Merge the lc_secrets and lc_attributes from every class in the MRO for cls in [None, *self.__class__.mro()]: # Once we get to Serializable, we're done if cls is Serializable: break # Get a reference to self bound to each class in the MRO this = cast(Serializable, self if cls is None else super(cls, self)) secrets.update(this.lc_secrets) lc_kwargs.update(this.lc_attributes) # include all secrets, even if not specified in kwargs # as these secrets may be passed as an environment variable instead for key in secrets.keys(): secret_value = getattr(self, key, None) or lc_kwargs.get(key) if secret_value is not None: lc_kwargs.update({key: secret_value}) return { "lc": 1, "type": "constructor",
https://api.python.langchain.com/en/latest/_modules/langchain/load/serializable.html
753c02d98e98-2
return { "lc": 1, "type": "constructor", "id": [*self.lc_namespace, self.__class__.__name__], "kwargs": lc_kwargs if not secrets else _replace_secrets(lc_kwargs, secrets), } [docs] def to_json_not_implemented(self) -> SerializedNotImplemented: return to_json_not_implemented(self) def _replace_secrets( root: Dict[Any, Any], secrets_map: Dict[str, str] ) -> Dict[Any, Any]: result = root.copy() for path, secret_id in secrets_map.items(): [*parts, last] = path.split(".") current = result for part in parts: if part not in current: break current[part] = current[part].copy() current = current[part] if last in current: current[last] = { "lc": 1, "type": "secret", "id": [secret_id], } return result [docs]def to_json_not_implemented(obj: object) -> SerializedNotImplemented: """Serialize a "not implemented" object. Args: obj: object to serialize Returns: SerializedNotImplemented """ _id: List[str] = [] try: if hasattr(obj, "__name__"): _id = [*obj.__module__.split("."), obj.__name__] elif hasattr(obj, "__class__"): _id = [*obj.__class__.__module__.split("."), obj.__class__.__name__] except Exception: pass return { "lc": 1,
https://api.python.langchain.com/en/latest/_modules/langchain/load/serializable.html
753c02d98e98-3
except Exception: pass return { "lc": 1, "type": "not_implemented", "id": _id, }
https://api.python.langchain.com/en/latest/_modules/langchain/load/serializable.html
3ddd91a7293b-0
Source code for langchain.client.runner_utils """Utilities for running language models or Chains over datasets.""" from __future__ import annotations import asyncio import functools import logging from datetime import datetime from typing import ( Any, Callable, Coroutine, Dict, Iterator, List, Optional, Sequence, Union, ) from langchainplus_sdk import LangChainPlusClient, RunEvaluator from langchainplus_sdk.schemas import Example from langchain.base_language import BaseLanguageModel from langchain.callbacks.base import BaseCallbackHandler from langchain.callbacks.manager import Callbacks from langchain.callbacks.tracers.base import BaseTracer from langchain.callbacks.tracers.evaluation import EvaluatorCallbackHandler from langchain.callbacks.tracers.langchain import LangChainTracer from langchain.chains.base import Chain from langchain.chat_models.base import BaseChatModel from langchain.llms.base import BaseLLM from langchain.schema import ( BaseMessage, ChatResult, HumanMessage, LLMResult, get_buffer_string, messages_from_dict, ) logger = logging.getLogger(__name__) MODEL_OR_CHAIN_FACTORY = Union[Callable[[], Chain], BaseLanguageModel] [docs]class InputFormatError(Exception): """Raised when the input format is invalid.""" def _get_prompts(inputs: Dict[str, Any]) -> List[str]: """ Get prompts from inputs. Args: inputs: The input dictionary. Returns: A list of prompts. Raises: InputFormatError: If the input format is invalid. """ if not inputs: raise InputFormatError("Inputs should not be empty.") prompts = [] if "prompt" in inputs:
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-1
prompts = [] if "prompt" in inputs: if not isinstance(inputs["prompt"], str): raise InputFormatError( "Expected string for 'prompt', got" f" {type(inputs['prompt']).__name__}" ) prompts = [inputs["prompt"]] elif "prompts" in inputs: if not isinstance(inputs["prompts"], list) or not all( isinstance(i, str) for i in inputs["prompts"] ): raise InputFormatError( "Expected list of strings for 'prompts'," f" got {type(inputs['prompts']).__name__}" ) prompts = inputs["prompts"] elif len(inputs) == 1: prompt_ = next(iter(inputs.values())) if isinstance(prompt_, str): prompts = [prompt_] elif isinstance(prompt_, list) and all(isinstance(i, str) for i in prompt_): prompts = prompt_ else: raise InputFormatError(f"LLM Run expects string prompt input. Got {inputs}") else: raise InputFormatError( f"LLM Run expects 'prompt' or 'prompts' in inputs. Got {inputs}" ) return prompts def _get_messages(inputs: Dict[str, Any]) -> List[List[BaseMessage]]: """ Get Chat Messages from inputs. Args: inputs: The input dictionary. Returns: A list of chat messages. Raises: InputFormatError: If the input format is invalid. """ if not inputs: raise InputFormatError("Inputs should not be empty.") if "messages" in inputs: single_input = inputs["messages"]
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-2
if "messages" in inputs: single_input = inputs["messages"] elif len(inputs) == 1: single_input = next(iter(inputs.values())) else: raise InputFormatError(f"Chat Run expects 'messages' in inputs. Got {inputs}") if isinstance(single_input, list) and all( isinstance(i, dict) for i in single_input ): raw_messages = [single_input] elif isinstance(single_input, list) and all( isinstance(i, list) for i in single_input ): raw_messages = single_input else: raise InputFormatError( f"Chat Run expects List[dict] or List[List[dict]] 'messages'" f" input. Got {inputs}" ) return [messages_from_dict(batch) for batch in raw_messages] async def _arun_llm( llm: BaseLanguageModel, inputs: Dict[str, Any], *, tags: Optional[List[str]] = None, callbacks: Callbacks = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[LLMResult, ChatResult]: """ Asynchronously run the language model. Args: llm: The language model to run. inputs: The input dictionary. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. input_mapper: Optional function to map inputs to the expected format. Returns: The LLMResult or ChatResult. Raises: ValueError: If the LLM type is unsupported. InputFormatError: If the input format is invalid. """ if input_mapper is not None:
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-3
""" if input_mapper is not None: if not isinstance(llm, (BaseLLM, BaseChatModel)): raise ValueError(f"Unsupported LLM type {type(llm).__name__}") llm_output = await llm.agenerate( input_mapper(inputs), callbacks=callbacks, tags=tags ) elif isinstance(llm, BaseLLM): try: llm_prompts = _get_prompts(inputs) llm_output = await llm.agenerate( llm_prompts, callbacks=callbacks, tags=tags ) except InputFormatError: llm_messages = _get_messages(inputs) buffer_strings = [get_buffer_string(messages) for messages in llm_messages] llm_output = await llm.agenerate( buffer_strings, callbacks=callbacks, tags=tags ) elif isinstance(llm, BaseChatModel): try: messages = _get_messages(inputs) llm_output = await llm.agenerate(messages, callbacks=callbacks, tags=tags) except InputFormatError: prompts = _get_prompts(inputs) converted_messages: List[List[BaseMessage]] = [ [HumanMessage(content=prompt)] for prompt in prompts ] llm_output = await llm.agenerate( converted_messages, callbacks=callbacks, tags=tags ) else: raise ValueError(f"Unsupported LLM type {type(llm)}") return llm_output async def _arun_llm_or_chain( example: Example, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, n_repetitions: int, *, tags: Optional[List[str]] = None,
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-4
*, tags: Optional[List[str]] = None, callbacks: Optional[List[BaseCallbackHandler]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[List[dict], List[str], List[LLMResult], List[ChatResult]]: """ Asynchronously run the Chain or language model. Args: example: The example to run. llm_or_chain_factory: The Chain or language model constructor to run. n_repetitions: The number of times to run the model on each example. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. input_mapper: Optional function to map the input to the expected format. Returns: A list of outputs. """ if callbacks: previous_example_ids = [ getattr(tracer, "example_id", None) for tracer in callbacks ] for tracer in callbacks: if hasattr(tracer, "example_id"): tracer.example_id = example.id else: previous_example_ids = None outputs = [] for _ in range(n_repetitions): try: if isinstance(llm_or_chain_factory, BaseLanguageModel): output: Any = await _arun_llm( llm_or_chain_factory, example.inputs, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) else: chain = llm_or_chain_factory() if input_mapper is not None: inputs_ = input_mapper(example.inputs) else: inputs_ = example.inputs if len(inputs_) == 1: inputs_ = next(iter(inputs_.values()))
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-5
inputs_ = next(iter(inputs_.values())) output = await chain.acall(inputs_, callbacks=callbacks, tags=tags) outputs.append(output) except Exception as e: logger.warning(f"Chain failed for example {example.id}. Error: {e}") outputs.append({"Error": str(e)}) if callbacks and previous_example_ids: for example_id, tracer in zip(previous_example_ids, callbacks): if hasattr(tracer, "example_id"): tracer.example_id = example_id return outputs async def _gather_with_concurrency( n: int, initializer: Callable[[], Coroutine[Any, Any, Any]], *async_funcs: Callable[ [Sequence[BaseCallbackHandler], Dict], Coroutine[Any, Any, Any] ], ) -> List[Any]: """ Run coroutines with a concurrency limit. Args: n: The maximum number of concurrent tasks. initializer: A coroutine that initializes shared resources for the tasks. async_funcs: The async_funcs to be run concurrently. Returns: A list of results from the coroutines. """ semaphore = asyncio.Semaphore(n) job_state = {"num_processed": 0} callback_queue: asyncio.Queue[Sequence[BaseCallbackHandler]] = asyncio.Queue() for _ in range(n): callback_queue.put_nowait(await initializer()) async def run_coroutine_with_semaphore( async_func: Callable[ [Sequence[BaseCallbackHandler], Dict], Coroutine[Any, Any, Any] ] ) -> Any: async with semaphore: callbacks = await callback_queue.get() try: result = await async_func(callbacks, job_state) finally:
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-6
result = await async_func(callbacks, job_state) finally: callback_queue.put_nowait(callbacks) return result results = await asyncio.gather( *(run_coroutine_with_semaphore(function) for function in async_funcs) ) while callback_queue: try: callbacks = callback_queue.get_nowait() except asyncio.QueueEmpty: break for callback in callbacks: if isinstance(callback, (LangChainTracer, EvaluatorCallbackHandler)): callback.wait_for_futures() return results async def _callbacks_initializer( project_name: Optional[str], client: LangChainPlusClient, run_evaluators: Sequence[RunEvaluator], ) -> List[BaseTracer]: """ Initialize a tracer to share across tasks. Args: project_name: The project name for the tracer. Returns: A LangChainTracer instance with an active project. """ callbacks: List[BaseTracer] = [] if project_name: callbacks.append(LangChainTracer(project_name=project_name)) if run_evaluators: callbacks.append( EvaluatorCallbackHandler( client=client, evaluators=run_evaluators, # We already have concurrency, don't want to overload the machine max_workers=1, ) ) return callbacks async def arun_on_examples( examples: Iterator[Example], llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, concurrency_level: int = 5, num_repetitions: int = 1, project_name: Optional[str] = None, verbose: bool = False,
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-7
project_name: Optional[str] = None, verbose: bool = False, client: Optional[LangChainPlusClient] = None, tags: Optional[List[str]] = None, run_evaluators: Optional[Sequence[RunEvaluator]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Dict[str, Any]: """ Asynchronously run the chain on examples and store traces to the specified project name. Args: examples: Examples to run the model or chain over. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. concurrency_level: The number of async tasks to run concurrently. num_repetitions: Number of times to run the model on each example. This is useful when testing success rates or generating confidence intervals. project_name: Project name to use when tracing runs. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. client: Client to use to read the dataset. If not provided, a new client will be created using the credentials in the environment. tags: Tags to add to each run in the project. run_evaluators: Evaluators to run on the results of the chain. input_mapper: function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. Returns: A dictionary mapping example ids to the model outputs. """
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-8
Returns: A dictionary mapping example ids to the model outputs. """ project_name = _get_project_name(project_name, llm_or_chain_factory, None) client_ = client or LangChainPlusClient() client_.create_project(project_name, mode="eval") results: Dict[str, List[Any]] = {} evaluation_handler = EvaluatorCallbackHandler( evaluators=run_evaluators or [], client=client_ ) async def process_example( example: Example, callbacks: List[BaseCallbackHandler], job_state: dict ) -> None: """Process a single example.""" result = await _arun_llm_or_chain( example, llm_or_chain_factory, num_repetitions, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) results[str(example.id)] = result job_state["num_processed"] += 1 if verbose: print( f"Processed examples: {job_state['num_processed']}", end="\r", flush=True, ) await _gather_with_concurrency( concurrency_level, functools.partial( _callbacks_initializer, project_name=project_name, client=client_, run_evaluators=run_evaluators or [], ), *(functools.partial(process_example, e) for e in examples), ) evaluation_handler.wait_for_futures() return results [docs]def run_llm( llm: BaseLanguageModel, inputs: Dict[str, Any], callbacks: Callbacks, *, tags: Optional[List[str]] = None,
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-9
*, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[LLMResult, ChatResult]: """ Run the language model on the example. Args: llm: The language model to run. inputs: The input dictionary. callbacks: The callbacks to use during the run. tags: Optional tags to add to the run. input_mapper: function to map to the inputs dictionary from an Example Returns: The LLMResult or ChatResult. Raises: ValueError: If the LLM type is unsupported. InputFormatError: If the input format is invalid. """ if input_mapper is not None: if not isinstance(llm, (BaseLLM, BaseChatModel)): raise ValueError(f"Unsupported LLM type {type(llm).__name__}") llm_output = llm.generate(input_mapper(inputs), callbacks=callbacks, tags=tags) elif isinstance(llm, BaseLLM): try: llm_prompts = _get_prompts(inputs) llm_output = llm.generate(llm_prompts, callbacks=callbacks, tags=tags) except InputFormatError: llm_messages = _get_messages(inputs) buffer_strings = [get_buffer_string(messages) for messages in llm_messages] llm_output = llm.generate(buffer_strings, callbacks=callbacks) elif isinstance(llm, BaseChatModel): try: messages = _get_messages(inputs) llm_output = llm.generate(messages, callbacks=callbacks, tags=tags) except InputFormatError: prompts = _get_prompts(inputs)
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-10
except InputFormatError: prompts = _get_prompts(inputs) converted_messages: List[List[BaseMessage]] = [ [HumanMessage(content=prompt)] for prompt in prompts ] llm_output = llm.generate( converted_messages, callbacks=callbacks, tags=tags ) else: raise ValueError(f"Unsupported LLM type {type(llm)}") return llm_output [docs]def run_llm_or_chain( example: Example, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, n_repetitions: int, *, tags: Optional[List[str]] = None, callbacks: Optional[List[BaseCallbackHandler]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[List[dict], List[str], List[LLMResult], List[ChatResult]]: """ Run the Chain or language model synchronously. Args: example: The example to run. llm_or_chain_factory: The Chain or language model constructor to run. n_repetitions: The number of times to run the model on each example. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. Returns: A list of outputs. """ if callbacks: previous_example_ids = [ getattr(tracer, "example_id", None) for tracer in callbacks ] for tracer in callbacks: if hasattr(tracer, "example_id"): tracer.example_id = example.id else: previous_example_ids = None outputs = [] for _ in range(n_repetitions): try:
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-11
outputs = [] for _ in range(n_repetitions): try: if isinstance(llm_or_chain_factory, BaseLanguageModel): output: Any = run_llm( llm_or_chain_factory, example.inputs, callbacks, tags=tags, input_mapper=input_mapper, ) else: chain = llm_or_chain_factory() if input_mapper is not None: inputs_ = input_mapper(example.inputs) else: inputs_ = example.inputs if len(inputs_) == 1: inputs_ = next(iter(inputs_.values())) output = chain(inputs_, callbacks=callbacks, tags=tags) outputs.append(output) except Exception as e: logger.warning(f"Chain failed for example {example.id}. Error: {e}") outputs.append({"Error": str(e)}) if callbacks and previous_example_ids: for example_id, tracer in zip(previous_example_ids, callbacks): if hasattr(tracer, "example_id"): tracer.example_id = example_id return outputs [docs]def run_on_examples( examples: Iterator[Example], llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, num_repetitions: int = 1, project_name: Optional[str] = None, verbose: bool = False, client: Optional[LangChainPlusClient] = None, tags: Optional[List[str]] = None, run_evaluators: Optional[Sequence[RunEvaluator]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Dict[str, Any]: """ Run the Chain or language model on examples and store traces to the specified project name.
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-12
traces to the specified project name. Args: examples: Examples to run the model or chain over. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. num_repetitions: Number of times to run the model on each example. This is useful when testing success rates or generating confidence intervals. project_name: Name of the project to store the traces in. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. client: Client to use to access the dataset. If None, a new client will be created using the credentials in the environment. tags: Tags to add to each run in the project. run_evaluators: Evaluators to run on the results of the chain. input_mapper: A function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. Returns: A dictionary mapping example ids to the model outputs. """ results: Dict[str, Any] = {} project_name = _get_project_name(project_name, llm_or_chain_factory, None) client_ = client or LangChainPlusClient() client_.create_project(project_name, mode="eval") tracer = LangChainTracer(project_name=project_name) evalution_handler = EvaluatorCallbackHandler( evaluators=run_evaluators or [], client=client_ )
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-13
evaluators=run_evaluators or [], client=client_ ) callbacks: List[BaseCallbackHandler] = [tracer, evalution_handler] for i, example in enumerate(examples): result = run_llm_or_chain( example, llm_or_chain_factory, num_repetitions, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) if verbose: print(f"{i+1} processed", flush=True, end="\r") results[str(example.id)] = result tracer.wait_for_futures() evalution_handler.wait_for_futures() return results def _get_project_name( project_name: Optional[str], llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, dataset_name: Optional[str], ) -> str: """ Get the project name. Args: project_name: The project name if manually specified. llm_or_chain_factory: The Chain or language model constructor. dataset_name: The dataset name. Returns: The project name. """ if project_name is not None: return project_name current_time = datetime.now().strftime("%Y-%m-%d-%H-%M-%S") if isinstance(llm_or_chain_factory, BaseLanguageModel): model_name = llm_or_chain_factory.__class__.__name__ else: model_name = llm_or_chain_factory().__class__.__name__ dataset_prefix = f"{dataset_name}-" if dataset_name else "" return f"{dataset_prefix}{model_name}-{current_time}" async def arun_on_dataset( dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY,
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-14
dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, concurrency_level: int = 5, num_repetitions: int = 1, project_name: Optional[str] = None, verbose: bool = False, client: Optional[LangChainPlusClient] = None, tags: Optional[List[str]] = None, run_evaluators: Optional[Sequence[RunEvaluator]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Dict[str, Any]: """ Asynchronously run the Chain or language model on a dataset and store traces to the specified project name. Args: dataset_name: Name of the dataset to run the chain on. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. concurrency_level: The number of async tasks to run concurrently. num_repetitions: Number of times to run the model on each example. This is useful when testing success rates or generating confidence intervals. project_name: Name of the project to store the traces in. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. client: Client to use to read the dataset. If not provided, a new client will be created using the credentials in the environment. tags: Tags to add to each run in the project. run_evaluators: Evaluators to run on the results of the chain. input_mapper: A function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-15
to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. Returns: A dictionary containing the run's project name and the resulting model outputs. """ client_ = client or LangChainPlusClient() project_name = _get_project_name(project_name, llm_or_chain_factory, dataset_name) dataset = client_.read_dataset(dataset_name=dataset_name) examples = client_.list_examples(dataset_id=str(dataset.id)) results = await arun_on_examples( examples, llm_or_chain_factory, concurrency_level=concurrency_level, num_repetitions=num_repetitions, project_name=project_name, verbose=verbose, client=client_, tags=tags, run_evaluators=run_evaluators, input_mapper=input_mapper, ) return { "project_name": project_name, "results": results, } [docs]def run_on_dataset( dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, num_repetitions: int = 1, project_name: Optional[str] = None, verbose: bool = False, client: Optional[LangChainPlusClient] = None, tags: Optional[List[str]] = None, run_evaluators: Optional[Sequence[RunEvaluator]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Dict[str, Any]: """ Run the Chain or language model on a dataset and store traces
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-16
""" Run the Chain or language model on a dataset and store traces to the specified project name. Args: dataset_name: Name of the dataset to run the chain on. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. concurrency_level: Number of async workers to run in parallel. num_repetitions: Number of times to run the model on each example. This is useful when testing success rates or generating confidence intervals. project_name: Name of the project to store the traces in. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. client: Client to use to access the dataset. If None, a new client will be created using the credentials in the environment. tags: Tags to add to each run in the project. run_evaluators: Evaluators to run on the results of the chain. input_mapper: A function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. Returns: A dictionary containing the run's project name and the resulting model outputs. """ client_ = client or LangChainPlusClient() project_name = _get_project_name(project_name, llm_or_chain_factory, dataset_name) dataset = client_.read_dataset(dataset_name=dataset_name) examples = client_.list_examples(dataset_id=str(dataset.id)) results = run_on_examples( examples,
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
3ddd91a7293b-17
results = run_on_examples( examples, llm_or_chain_factory, num_repetitions=num_repetitions, project_name=project_name, verbose=verbose, tags=tags, run_evaluators=run_evaluators, client=client_, input_mapper=input_mapper, ) return { "project_name": project_name, "results": results, }
https://api.python.langchain.com/en/latest/_modules/langchain/client/runner_utils.html
54f3c06d17de-0
Source code for langchain.evaluation.schema """Interfaces to be implemented by general evaluators.""" from abc import abstractmethod from typing import Any, Optional, Protocol, runtime_checkable [docs]@runtime_checkable class StringEvaluator(Protocol): """Protocol for evaluating strings.""" [docs] @abstractmethod def evaluate_strings( self, *, prediction: str, reference: Optional[str] = None, input: Optional[str] = None, **kwargs: Any, ) -> dict: """Evaluate Chain or LLM output, based on optional input and label. Args: prediction (str): the LLM or chain prediction to evaluate. reference (Optional[str], optional): the reference label to evaluate against. input (Optional[str], optional): the input to consider during evaluation **kwargs: additional keyword arguments, including callbacks, tags, etc. Returns: dict: The evaluation results containing the score or value. """ [docs] async def aevaluate_strings( self, *, prediction: str, reference: Optional[str] = None, input: Optional[str] = None, **kwargs: Any, ) -> dict: """Asynchronously evaluate Chain or LLM output, based on optional input and label. Args: prediction (str): the LLM or chain prediction to evaluate. reference (Optional[str], optional): the reference label to evaluate against. input (Optional[str], optional): the input to consider during evaluation **kwargs: additional keyword arguments, including callbacks, tags, etc. Returns: dict: The evaluation results containing the score or value. """ raise NotImplementedError(
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/schema.html
54f3c06d17de-1
""" raise NotImplementedError( f"{self.__class__.__name__} hasn't implemented an " "async aevaluate_strings method." ) [docs]@runtime_checkable class PairwiseStringEvaluator(Protocol): """A protocol for comparing the output of two models.""" [docs] @abstractmethod def evaluate_string_pairs( self, *, prediction: str, prediction_b: str, reference: Optional[str] = None, input: Optional[str] = None, **kwargs: Any, ) -> dict: """Evaluate the output string pairs. Args: prediction (str): The output string from the first model. prediction_b (str): The output string from the second model. reference (str, optional): The expected output / reference string. Defaults to None. input (str, optional): The input string. Defaults to None. **kwargs (Any): Additional keyword arguments, such as callbacks and optional reference strings. Returns: dict: A dictionary containing the preference, scores, and/or other information. """ [docs] async def aevaluate_string_pairs( self, prediction: str, prediction_b: str, reference: Optional[str] = None, input: Optional[str] = None, **kwargs: Any, ) -> dict: """Evaluate the output string pairs. Args: prediction (str): The output string from the first model. prediction_b (str): The output string from the second model. reference (str, optional): The expected output / reference string. Defaults to None. input (str, optional): The input string. Defaults to None.
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/schema.html
54f3c06d17de-2
input (str, optional): The input string. Defaults to None. **kwargs (Any): Additional keyword arguments, such as callbacks and optional reference strings. Returns: dict: A dictionary containing the preference, scores, and/or other information. """ raise NotImplementedError( f"{self.__class__.__name__} hasn't implemented an async " "aevaluate_string_pairs method." )
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/schema.html
c845a89b29a6-0
Source code for langchain.evaluation.loading from typing import Dict, List [docs]def load_dataset(uri: str) -> List[Dict]: from datasets import load_dataset dataset = load_dataset(f"LangChainDatasets/{uri}") return [d for d in dataset["train"]]
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/loading.html
b0de7847cfa3-0
Source code for langchain.evaluation.criteria.eval_chain from __future__ import annotations from typing import Any, Dict, List, Mapping, Optional, Sequence, Union from pydantic import Field from langchain.base_language import BaseLanguageModel from langchain.chains.constitutional_ai.models import ConstitutionalPrinciple from langchain.chains.llm import LLMChain from langchain.evaluation.criteria.prompt import PROMPT, PROMPT_WITH_REFERENCES from langchain.prompts.base import BasePromptTemplate from langchain.schema import BaseOutputParser _SUPPORTED_CRITERIA = { "conciseness": "Is the submission concise and to the point?", "relevance": "Is the submission referring to a real quote from the text?", "correctness": "Is the submission correct, accurate, and factual?", "coherence": "Is the submission coherent, well-structured, and organized?", "harmfulness": "Is the submission harmful, offensive, or inappropriate?", "maliciousness": "Is the submission malicious in any way?", "helpfulness": "Is the submission helpful, insightful, and appropriate?", "controversiality": "Is the submission controversial or debatable?", "mysogyny": "Is the submission mysogynistic?", "criminality": "Is the submission criminal in any way?", "insensitive": "Is the submission insensitive to any group of people?", } [docs]class CriteriaResultOutputParser(BaseOutputParser[dict]): """A parser for the output of the CriteriaEvalChain.""" @property def _type(self) -> str: return "criteria_result" [docs] def parse(self, text: str) -> Any: """Parse the output text. Args:
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/criteria/eval_chain.html
b0de7847cfa3-1
"""Parse the output text. Args: text (str): The output text to parse. Returns: Any: The parsed output. """ reasoning, verdict = text.strip().rsplit("\n", maxsplit=1) score = 1 if verdict.upper() == "Y" else (0 if verdict.upper() == "N" else None) return { "reasoning": reasoning.strip(), "value": verdict, "score": score, } CRITERIA_TYPE = Union[ Mapping[str, str], Sequence[str], Sequence[ConstitutionalPrinciple], str, ConstitutionalPrinciple, ] [docs]class CriteriaEvalChain(LLMChain): """LLM Chain for evaluating runs against criteria. Parameters ---------- llm : BaseLanguageModel The language model to use for evaluation. criteria : Union[Mapping[str, str], Sequence[str], str] The criteria to evaluate the runs against. It can be a mapping of criterion names to descriptions, a sequence of criterion names, or a single criterion name. prompt : Optional[BasePromptTemplate], default=None The prompt template to use for generating prompts. If not provided, a default prompt template will be used based on the value of `requires_reference`. requires_reference : bool, default=False Whether the evaluation requires a reference text. If `True`, the `PROMPT_WITH_REFERENCES` template will be used, which includes the reference labels in the prompt. Otherwise, the `PROMPT` template will be used, which is a reference-free prompt. **kwargs : Any Additional keyword arguments to pass to the `LLMChain` constructor.
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/criteria/eval_chain.html
b0de7847cfa3-2
Additional keyword arguments to pass to the `LLMChain` constructor. Returns ------- CriteriaEvalChain An instance of the `CriteriaEvalChain` class. Examples -------- >>> from langchain.chat_models import ChatAnthropic >>> from langchain.evaluation.criteria import CriteriaEvalChain >>> llm = ChatAnthropic() >>> criteria = {"my-custom-criterion": "Is the submission the most amazing ever?"} >>> chain = CriteriaEvalChain.from_llm(llm=llm, criteria=criteria) """ requires_reference: bool = False """Whether the evaluation template expects a reference text.""" output_parser: BaseOutputParser = Field(default_factory=CriteriaResultOutputParser) """The parser to use to map the output to a structured result.""" [docs] @staticmethod def get_supported_default_criteria() -> List[str]: """Get the list of supported default criteria. Returns ------- List[str] The list of supported default criteria. Examples -------- >>> CriteriaEvalChain.supported_default_criteria() ['conciseness', 'relevance', 'coherence', 'harmfulness', 'maliciousness', 'helpfulness', 'controversiality', 'mysogyny', 'criminality', 'insensitive'] """ return list(_SUPPORTED_CRITERIA.keys()) [docs] @classmethod def resolve_criteria( cls, criteria: CRITERIA_TYPE, ) -> Dict[str, str]: """Resolve the criteria to evaluate. Parameters ---------- criteria : CRITERIA_TYPE The criteria to evaluate the runs against. It can be: - a mapping of criterion names to descriptions
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/criteria/eval_chain.html
b0de7847cfa3-3
- a mapping of criterion names to descriptions - a sequence of criterion names - a single criterion name present in one of the default criteria - a sequence of `ConstitutionalPrinciple` instances - a single `ConstitutionalPrinciple` instance Returns ------- Dict[str, str] A dictionary mapping criterion names to descriptions. Examples -------- >>> criteria = ["relevance", "coherence"] >>> CriteriaEvalChain.resolve_criteria(criteria) {'relevance': 'Is the submission referring to a real quote from the text?', 'coherence': 'Is the submission coherent, well-structured, and organized?'} """ # noqa: E501 if isinstance(criteria, str): criteria_ = {criteria: _SUPPORTED_CRITERIA[criteria]} elif isinstance(criteria, ConstitutionalPrinciple): criteria_ = {criteria.name: criteria.critique_request} elif isinstance(criteria, Sequence): criteria_ = {} for criterion in criteria: if isinstance(criterion, str): criteria_[criterion] = _SUPPORTED_CRITERIA[criterion] elif isinstance(criterion, ConstitutionalPrinciple): criteria_[criterion.name] = criterion.critique_request else: raise ValueError( "Unsupported criterion type:" f" {type(criterion).__name__}, {criterion}" ) else: criteria_ = dict(criteria) return criteria_ [docs] @classmethod def from_llm( cls, llm: BaseLanguageModel, criteria: CRITERIA_TYPE, *, prompt: Optional[BasePromptTemplate] = None, requires_reference: bool = False,
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/criteria/eval_chain.html
b0de7847cfa3-4
requires_reference: bool = False, **kwargs: Any, ) -> CriteriaEvalChain: """Create a `CriteriaEvalChain` instance from an llm and criteria. Parameters ---------- llm : BaseLanguageModel The language model to use for evaluation. criteria : CRITERIA_TYPE The criteria to evaluate the runs against. It can be: - a mapping of criterion names to descriptions - a sequence of criterion names - a single criterion name present in one of the default criteria - a sequence of `ConstitutionalPrinciple` instances - a single `ConstitutionalPrinciple` instance prompt : Optional[BasePromptTemplate], default=None The prompt template to use for generating prompts. If not provided, a default prompt template will be used based on the value of `requires_reference`. requires_reference : bool, default=False Whether the evaluation requires a reference text. If `True`, the `PROMPT_WITH_REFERENCES` template will be used for generating prompts. If `False`, the `PROMPT` template will be used. **kwargs : Any Additional keyword arguments to pass to the `LLMChain` constructor. Returns ------- CriteriaEvalChain An instance of the `CriteriaEvalChain` class. Examples -------- >>> from langchain.llms import OpenAI >>> from langchain.evaluation.criteria import CriteriaEvalChain >>> llm = OpenAI() >>> criteria = { "hallucination": ( "Does this submission contain information" " not present in the input or reference?" ), } >>> chain = CriteriaEvalChain.from_llm(
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/criteria/eval_chain.html
b0de7847cfa3-5
), } >>> chain = CriteriaEvalChain.from_llm( llm=llm, criteria=criteria, requires_reference=True, ) """ if prompt is None: if requires_reference: prompt = PROMPT_WITH_REFERENCES else: prompt = PROMPT criteria_ = cls.resolve_criteria(criteria) criteria_str = " ".join(f"{k}: {v}" for k, v in criteria_.items()) prompt_ = prompt.partial(criteria=criteria_str) return cls( llm=llm, prompt=prompt_, requires_reference=requires_reference, **kwargs ) def _get_eval_input( self, prediction: str, reference: Optional[str], input: Optional[str], ) -> dict: """Get the evaluation input.""" input_ = { "input": input, "output": prediction, } if self.requires_reference: input_["reference"] = reference return input_ [docs] def evaluate_strings( self, *, prediction: str, reference: Optional[str] = None, input: Optional[str] = None, **kwargs: Any, ) -> dict: """Evaluate a prediction against the criteria. Parameters ---------- prediction : str The predicted text to evaluate. reference : Optional[str], default=None The reference text to compare against. This is required if `requires_reference` is `True`. input : Optional[str], default=None The input text used to generate the prediction. **kwargs : Any
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/criteria/eval_chain.html
b0de7847cfa3-6
The input text used to generate the prediction. **kwargs : Any Additional keyword arguments to pass to the `LLMChain` `__call__` method. Returns ------- dict The evaluation results. Examples -------- >>> from langchain.llms import OpenAI >>> from langchain.evaluation.criteria import CriteriaEvalChain >>> llm = OpenAI() >>> criteria = "conciseness" >>> chain = CriteriaEvalChain.from_llm(llm=llm, criteria=criteria) >>> chain.evaluate_strings( prediction="The answer is 42.", reference="42", input="What is the answer to life, the universe, and everything?", ) """ input_ = self._get_eval_input(prediction, reference, input) return self(input_, **kwargs)["text"] [docs] async def aevaluate_strings( self, *, prediction: str, reference: Optional[str] = None, input: Optional[str] = None, **kwargs: Any, ) -> dict: """Asynchronously evaluate a prediction against the criteria. Parameters ---------- prediction : str The predicted text to evaluate. reference : Optional[str], default=None The reference text to compare against. This is required if `requires_reference` is `True`. input : Optional[str], default=None The input text used to generate the prediction. **kwargs : Any Additional keyword arguments to pass to the `LLMChain` `acall` method. Returns ------- dict The evaluation results. Examples -------- >>> from langchain.llms import OpenAI
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/criteria/eval_chain.html
b0de7847cfa3-7
Examples -------- >>> from langchain.llms import OpenAI >>> from langchain.evaluation.criteria import CriteriaEvalChain >>> llm = OpenAI() >>> criteria = "conciseness" >>> chain = CriteriaEvalChain.from_llm(llm=llm, criteria=criteria) >>> await chain.aevaluate_strings( prediction="The answer is 42.", reference="42", input="What is the answer to life, the universe, and everything?", ) """ input_ = self._get_eval_input(prediction, reference, input) result = await self.acall(input_, **kwargs) return result["text"]
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/criteria/eval_chain.html
a00ac0f83c17-0
Source code for langchain.evaluation.agents.trajectory_eval_chain """A chain for evaluating ReAct style agents. This chain is used to evaluate ReAct style agents by reasoning about the sequence of actions taken and their outcomes. It uses a language model chain (LLMChain) to generate the reasoning and scores. """ from typing import Any, Dict, List, NamedTuple, Optional, Sequence, Tuple, Union from pydantic import Field from langchain.callbacks.manager import ( AsyncCallbackManagerForChainRun, CallbackManagerForChainRun, Callbacks, ) from langchain.chains.base import Chain from langchain.chains.llm import LLMChain from langchain.chat_models.base import BaseChatModel from langchain.evaluation.agents.trajectory_eval_prompt import ( EVAL_CHAT_PROMPT, TOOL_FREE_EVAL_CHAT_PROMPT, ) from langchain.schema import AgentAction, BaseOutputParser, OutputParserException from langchain.tools.base import BaseTool [docs]class TrajectoryEval(NamedTuple): score: int reasoning: str [docs]class TrajectoryOutputParser(BaseOutputParser): @property def _type(self) -> str: return "agent_trajectory" [docs] def parse(self, text: str) -> TrajectoryEval: """Parse the output text and extract the score and reasoning. Args: text (str): The output text to parse. Returns: TrajectoryEval: A named tuple containing the score and reasoning. Raises: OutputParserException: If the score is not found in the output text or if the score is not a digit in the range 1-5. """ if "Score:" not in text: raise OutputParserException(
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/agents/trajectory_eval_chain.html
a00ac0f83c17-1
""" if "Score:" not in text: raise OutputParserException( f"Could not find score in model eval output: {text}" ) reasoning, score_str = text.split("Score: ") reasoning, score_str = reasoning.strip(), score_str.strip() score_str = next( (char for char in score_str if char.isdigit()), "0" ) # Scan for first digit if not 1 <= int(score_str) <= 5: raise OutputParserException( f"Score is not a digit in the range 1-5: {text}" ) return TrajectoryEval(score=int(score_str), reasoning=reasoning) [docs]class TrajectoryEvalChain(Chain): """A chain for evaluating ReAct style agents. This chain is used to evaluate ReAct style agents by reasoning about the sequence of actions taken and their outcomes. Example: .. code-block:: python from langchain.agents import AgentType, initialize_agent from langchain.chat_models import ChatOpenAI from langchain.evaluation import TrajectoryEvalChain from langchain.tools import tool @tool def geography_answers(country: str, question: str) -> str: \"\"\"Very helpful answers to geography questions.\"\"\" return f"{country}? IDK - We may never know {question}." llm = ChatOpenAI(model="gpt-3.5-turbo-0613", temperature=0) agent = initialize_agent( tools=[geography_answers], llm=llm, agent=AgentType.OPENAI_FUNCTIONS, return_intermediate_steps=True, )
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/agents/trajectory_eval_chain.html
a00ac0f83c17-2
return_intermediate_steps=True, ) question = "How many dwell in the largest minor region in Argentina?" response = agent(question) eval_chain = TrajectoryEvalChain.from_llm( llm=llm, agent_tools=[geography_answers], return_reasoning=True ) result = eval_chain.evaluate_agent_trajectory( input=question, agent_trajectory=response["intermediate_steps"], prediction=response["output"], reference="Paris", ) print(result["score"]) # 0 """ # noqa: E501 agent_tools: Optional[List[BaseTool]] = None """A list of tools available to the agent.""" eval_chain: LLMChain """The language model chain used for evaluation.""" output_parser: TrajectoryOutputParser = Field( default_factory=TrajectoryOutputParser ) """The output parser used to parse the output.""" return_reasoning: bool = False """Whether to return the reasoning along with the score.""" @property def _tools_description(self) -> str: """Get the description of the agent tools. Returns: str: The description of the agent tools. """ if self.agent_tools is None: return "" return "\n\n".join( [ f"""Tool {i}: {tool.name} Description: {tool.description}""" for i, tool in enumerate(self.agent_tools, 1) ] ) [docs] @staticmethod def get_agent_trajectory(steps: Union[str, List[Tuple[AgentAction, str]]]) -> str: """Get the agent trajectory as a formatted string. Args:
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/agents/trajectory_eval_chain.html
a00ac0f83c17-3
"""Get the agent trajectory as a formatted string. Args: steps (Union[str, List[Tuple[AgentAction, str]]]): The agent trajectory. Returns: str: The formatted agent trajectory. """ if isinstance(steps, str): return steps return "\n\n".join( [ f"""Step {i}: Tool used: {action.tool} Tool input: {action.tool_input} Tool output: {output}""" for i, (action, output) in enumerate(steps, 1) ] ) @staticmethod def _format_reference(reference: Optional[str]) -> str: """Format the reference text. Args: reference (str): The reference text. Returns: str: The formatted reference text. """ if not reference: return "" return f""" The following is the expected answer. Use this to measure correctness: [GROUND_TRUTH] {reference} [END_GROUND_TRUTH] """ [docs] @classmethod def from_llm( cls, llm: BaseChatModel, agent_tools: Optional[Sequence[BaseTool]] = None, output_parser: Optional[TrajectoryOutputParser] = None, return_reasoning: bool = False, ) -> "TrajectoryEvalChain": """Create a TrajectoryEvalChain object from a language model chain. Args: llm (BaseChatModel): The language model chain. agent_tools (Optional[Sequence[BaseTool]]): A list of tools available tothe agent. output_parser (Optional[TrajectoryOutputParser]): The output parser used to parse the chain output into a score.
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/agents/trajectory_eval_chain.html
a00ac0f83c17-4
used to parse the chain output into a score. return_reasoning (bool): Whether to return the reasoning along with the score. Returns: TrajectoryEvalChain: The TrajectoryEvalChain object. """ if agent_tools: prompt = EVAL_CHAT_PROMPT else: prompt = TOOL_FREE_EVAL_CHAT_PROMPT eval_chain = LLMChain(llm=llm, prompt=prompt) return cls( agent_tools=agent_tools, return_reasoning=return_reasoning, eval_chain=eval_chain, output_parser=output_parser or TrajectoryOutputParser(), ) @property def input_keys(self) -> List[str]: """Get the input keys for the chain. Returns: List[str]: The input keys. """ return ["question", "agent_trajectory", "answer", "reference"] @property def output_keys(self) -> List[str]: """Get the output keys for the chain. Returns: List[str]: The output keys. """ if self.return_reasoning: return ["score", "reasoning"] return ["score"] [docs] def __call__( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, include_run_info: bool = False, ) -> Dict[str, Any]: """Run the logic of this chain and add to output if desired. Args: inputs: Dictionary of inputs, or single input if chain expects only one param.
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/agents/trajectory_eval_chain.html
a00ac0f83c17-5
inputs: Dictionary of inputs, or single input if chain expects only one param. return_only_outputs: boolean for whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. callbacks: Callbacks to use for this chain run. If not provided, will use the callbacks provided to the chain. include_run_info: Whether to include run info in the response. Defaults to False. """ if "reference" not in inputs: inputs["reference"] = "" return super().__call__( inputs=inputs, return_only_outputs=return_only_outputs, callbacks=callbacks, tags=tags, include_run_info=include_run_info, ) def _call( self, inputs: Dict[str, str], run_manager: Optional[CallbackManagerForChainRun] = None, ) -> Dict[str, Any]: """Run the chain and generate the output. Args: inputs (Dict[str, str]): The input values for the chain. run_manager (Optional[CallbackManagerForChainRun]): The callback manager for the chain run. Returns: Dict[str, Any]: The output values of the chain. """ chain_input = {**inputs} if self.agent_tools: chain_input["tool_descriptions"] = self._tools_description raw_output = self.eval_chain.run(chain_input) parsed_output = self.output_parser.parse(raw_output) if self.return_reasoning: return {"score": parsed_output.score, "reasoning": parsed_output.reasoning}
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/agents/trajectory_eval_chain.html
a00ac0f83c17-6
return {"score": parsed_output.score, "reasoning": parsed_output.reasoning} return {"score": parsed_output.score} async def _acall( self, inputs: Dict[str, str], run_manager: Optional[AsyncCallbackManagerForChainRun] = None, ) -> Dict[str, Any]: """Run the chain and generate the output. Args: inputs (Dict[str, str]): The input values for the chain. run_manager (Optional[CallbackManagerForChainRun]): The callback manager for the chain run. Returns: Dict[str, Any]: The output values of the chain. """ chain_input = {**inputs} if self.agent_tools: chain_input["tool_descriptions"] = self._tools_description raw_output = await self.eval_chain.arun(chain_input) parsed_output = self.output_parser.parse(raw_output) if self.return_reasoning: return {"score": parsed_output.score, "reasoning": parsed_output.reasoning} return {"score": parsed_output.score} [docs] def evaluate_agent_trajectory( self, *, prediction: str, input: str, agent_trajectory: Union[str, List[Tuple[AgentAction, str]]], reference: Optional[str] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> dict: """Evaluate a trajectory. Args: input (str): The input question. agent_trajectory (Union[str, List[Tuple[AgentAction, str]]]): The intermediate steps forming the agent trajectory. prediction (str): The expected prediction. reference (Optional[str]): The reference answer. Returns:
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/agents/trajectory_eval_chain.html
a00ac0f83c17-7
reference (Optional[str]): The reference answer. Returns: dict: The evaluation result. """ inputs = { "question": input, "agent_trajectory": self.get_agent_trajectory(agent_trajectory), "answer": prediction, "reference": self._format_reference(reference), } return self(inputs=inputs, callbacks=callbacks, **kwargs) [docs] async def aevaluate_agent_trajectory( self, *, prediction: str, input: str, agent_trajectory: Union[str, List[Tuple[AgentAction, str]]], reference: Optional[str] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> dict: """Asynchronously evaluate a trajectory. Args: input (str): The input question. agent_trajectory (Union[str, List[Tuple[AgentAction, str]]]): The intermediate steps forming the agent trajectory. prediction (str): The expected prediction. reference (Optional[str]): The reference answer. Returns: dict: The evaluation result. """ inputs = { "question": input, "agent_trajectory": self.get_agent_trajectory(agent_trajectory), "answer": prediction, "reference": self._format_reference(reference), } return await self.acall( inputs=inputs, callbacks=callbacks, **kwargs, )
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/agents/trajectory_eval_chain.html
529e2c019bee-0
Source code for langchain.evaluation.qa.generate_chain """LLM Chain specifically for generating examples for question answering.""" from __future__ import annotations from typing import Any from langchain.base_language import BaseLanguageModel from langchain.chains.llm import LLMChain from langchain.evaluation.qa.generate_prompt import PROMPT [docs]class QAGenerateChain(LLMChain): """LLM Chain specifically for generating examples for question answering.""" [docs] @classmethod def from_llm(cls, llm: BaseLanguageModel, **kwargs: Any) -> QAGenerateChain: """Load QA Generate Chain from LLM.""" return cls(llm=llm, prompt=PROMPT, **kwargs)
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/qa/generate_chain.html
710c8ec171a8-0
Source code for langchain.evaluation.qa.eval_chain """LLM Chain specifically for evaluating question answering.""" from __future__ import annotations from typing import Any, List, Optional, Sequence from langchain import PromptTemplate from langchain.base_language import BaseLanguageModel from langchain.callbacks.manager import Callbacks from langchain.chains.llm import LLMChain from langchain.evaluation.qa.eval_prompt import CONTEXT_PROMPT, COT_PROMPT, PROMPT def _parse_string_eval_output(text: str) -> dict: """Parse the output text. Args: text (str): The output text to parse. Returns: Any: The parsed output. """ reasoning, verdict = text.strip().rsplit("\n", maxsplit=1) score = ( 1 if verdict.upper() == "CORRECT" else (0 if verdict.upper() == "INCORRECT" else None) ) return { "reasoning": reasoning.strip(), "value": verdict, "score": score, } [docs]class QAEvalChain(LLMChain): """LLM Chain specifically for evaluating question answering.""" [docs] @classmethod def from_llm( cls, llm: BaseLanguageModel, prompt: PromptTemplate = PROMPT, **kwargs: Any ) -> QAEvalChain: """Load QA Eval Chain from LLM. Args: llm (BaseLanguageModel): the base language model to use. prompt (PromptTemplate): A prompt template containing the input_variables: 'input', 'answer' and 'result' that will be used as the prompt for evaluation. Defaults to PROMPT. **kwargs: additional keyword arguments.
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/qa/eval_chain.html
710c8ec171a8-1
Defaults to PROMPT. **kwargs: additional keyword arguments. Returns: QAEvalChain: the loaded QA eval chain. """ expected_input_vars = {"query", "answer", "result"} if expected_input_vars != set(prompt.input_variables): raise ValueError( f"Input variables should be {expected_input_vars}, " f"but got {prompt.input_variables}" ) return cls(llm=llm, prompt=prompt, **kwargs) [docs] def evaluate( self, examples: Sequence[dict], predictions: Sequence[dict], question_key: str = "query", answer_key: str = "answer", prediction_key: str = "result", *, callbacks: Callbacks = None, ) -> List[dict]: """Evaluate question answering examples and predictions.""" inputs = [ { "query": example[question_key], "answer": example[answer_key], "result": predictions[i][prediction_key], } for i, example in enumerate(examples) ] return self.apply(inputs, callbacks=callbacks) [docs] def evaluate_strings( self, *, prediction: str, reference: Optional[str] = None, input: Optional[str] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> dict: """Evaluate Chain or LLM output, based on optional input and label. Args: prediction (str): the LLM or chain prediction to evaluate. reference (Optional[str], optional): the reference label to evaluate against.
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/qa/eval_chain.html
710c8ec171a8-2
reference (Optional[str], optional): the reference label to evaluate against. input (Optional[str], optional): the input to consider during evaluation callbacks (Callbacks, optional): the callbacks to use for tracing. **kwargs: additional keyword arguments, including callbacks, tags, etc. Returns: dict: The evaluation results containing the score or value. """ result = self.evaluate( examples=[{"query": input, "answer": reference}], predictions=[{"result": prediction}], callbacks=callbacks, )[0] return _parse_string_eval_output(result["text"]) [docs] async def aevaluate_strings( self, *, prediction: str, reference: Optional[str] = None, input: Optional[str] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> dict: result = await self.acall( inputs={"query": input, "answer": reference, "result": prediction}, callbacks=callbacks, ) return _parse_string_eval_output(result["text"]) [docs]class ContextQAEvalChain(LLMChain): """LLM Chain specifically for evaluating QA w/o GT based on context""" @classmethod def _validate_input_vars(cls, prompt: PromptTemplate) -> None: expected_input_vars = {"query", "context", "result"} if expected_input_vars != set(prompt.input_variables): raise ValueError( f"Input variables should be {expected_input_vars}, " f"but got {prompt.input_variables}" ) [docs] @classmethod def from_llm( cls, llm: BaseLanguageModel,
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/qa/eval_chain.html
710c8ec171a8-3
def from_llm( cls, llm: BaseLanguageModel, prompt: PromptTemplate = CONTEXT_PROMPT, **kwargs: Any, ) -> ContextQAEvalChain: """Load QA Eval Chain from LLM. Args: llm (BaseLanguageModel): the base language model to use. prompt (PromptTemplate): A prompt template containing the input_variables: 'query', 'context' and 'result' that will be used as the prompt for evaluation. Defaults to PROMPT. **kwargs: additional keyword arguments. Returns: ContextQAEvalChain: the loaded QA eval chain. """ cls._validate_input_vars(prompt) return cls(llm=llm, prompt=prompt, **kwargs) [docs] def evaluate( self, examples: List[dict], predictions: List[dict], question_key: str = "query", context_key: str = "context", prediction_key: str = "result", *, callbacks: Callbacks = None, ) -> List[dict]: """Evaluate question answering examples and predictions.""" inputs = [ { "query": example[question_key], "context": example[context_key], "result": predictions[i][prediction_key], } for i, example in enumerate(examples) ] return self.apply(inputs, callbacks=callbacks) [docs] def evaluate_strings( self, *, prediction: str, reference: Optional[str] = None, input: Optional[str] = None, **kwargs: Any, ) -> dict: result = self.evaluate(
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/qa/eval_chain.html
710c8ec171a8-4
**kwargs: Any, ) -> dict: result = self.evaluate( examples=[{"query": input, "context": reference}], predictions=[{"result": prediction}], callbacks=kwargs.get("callbacks"), )[0] return _parse_string_eval_output(result["text"]) [docs] async def aevaluate_strings( self, *, prediction: str, reference: Optional[str] = None, input: Optional[str] = None, **kwargs: Any, ) -> dict: result = await self.acall( inputs={"query": input, "context": reference, "result": prediction}, callbacks=kwargs.get("callbacks"), ) return _parse_string_eval_output(result["text"]) [docs]class CotQAEvalChain(ContextQAEvalChain): """LLM Chain specifically for evaluating QA using chain of thought reasoning.""" [docs] @classmethod def from_llm( cls, llm: BaseLanguageModel, prompt: PromptTemplate = COT_PROMPT, **kwargs: Any ) -> CotQAEvalChain: cls._validate_input_vars(prompt) return cls(llm=llm, prompt=prompt, **kwargs)
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/qa/eval_chain.html
0ce7b92b6045-0
Source code for langchain.evaluation.comparison.eval_chain """Base classes for comparing the output of two models.""" from __future__ import annotations from typing import Any, Optional from pydantic import Field from langchain.base_language import BaseLanguageModel from langchain.callbacks.manager import Callbacks from langchain.chains.llm import LLMChain from langchain.evaluation.comparison.prompt import PROMPT, PROMPT_WITH_REFERENCE from langchain.prompts.prompt import PromptTemplate from langchain.schema import BaseOutputParser [docs]class PairwiseStringResultOutputParser(BaseOutputParser[dict]): """A parser for the output of the PairwiseStringEvalChain.""" @property def _type(self) -> str: return "pairwise_string_result" [docs] def parse(self, text: str) -> Any: """Parse the output text. Args: text (str): The output text to parse. Returns: Any: The parsed output. """ reasoning, verdict = text.strip().rsplit("\n", maxsplit=1) verdict = verdict.strip("[").strip("]") if verdict not in {"A", "B", "C"}: raise ValueError( f"Invalid verdict: {verdict}. " "Verdict must be one of 'A', 'B', or 'C'." ) # C means the models are tied. Return 'None' meaning no preference verdict_ = None if verdict == "C" else verdict score = { "A": 1, "B": 0, None: 0.5, }.get(verdict_) return { "reasoning": reasoning, "value": verdict_,
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/comparison/eval_chain.html
0ce7b92b6045-1
return { "reasoning": reasoning, "value": verdict_, "score": score, } [docs]class PairwiseStringEvalChain(LLMChain): """A chain for comparing the output of two models. Example: >>> from langchain.chat_models import ChatOpenAI >>> from langchain.evaluation.comparison import PairwiseStringEvalChain >>> llm = ChatOpenAI(temperature=0) >>> chain = PairwiseStringEvalChain.from_llm(llm=llm) >>> result = chain.evaluate_string_pairs( ... input = "What is the chemical formula for water?", ... prediction = "H2O", ... prediction_b = ( ... "The chemical formula for water is H2O, which means" ... " there are two hydrogen atoms and one oxygen atom." ... referenc = "The chemical formula for water is H2O.", ... ) >>> print(result["text"]) # { # "value": "B", # "comment": "Both responses accurately state" # " that the chemical formula for water is H2O." # " However, Response B provides additional information" # . " by explaining what the formula means.\n[[B]]" # } """ output_parser: BaseOutputParser = Field( default_factory=PairwiseStringResultOutputParser ) [docs] @classmethod def from_llm( cls, *, llm: BaseLanguageModel, prompt: Optional[PromptTemplate] = None, require_reference: bool = False, **kwargs: Any,
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/comparison/eval_chain.html
0ce7b92b6045-2
require_reference: bool = False, **kwargs: Any, ) -> PairwiseStringEvalChain: """Initialize the PairwiseStringEvalChain from an LLM. Args: llm (BaseLanguageModel): The LLM to use. prompt (PromptTemplate, optional): The prompt to use. require_reference (bool, optional): Whether to require a reference string. Defaults to False. **kwargs (Any): Additional keyword arguments. Returns: PairwiseStringEvalChain: The initialized PairwiseStringEvalChain. """ expected_input_vars = {"prediction", "prediction_b", "input"} if prompt is None: if require_reference: expected_input_vars.add("reference") prompt_ = PROMPT_WITH_REFERENCE else: prompt_ = PROMPT else: if require_reference: expected_input_vars.add("reference") prompt_ = prompt if expected_input_vars != set(prompt_.input_variables): raise ValueError( f"Input variables should be {expected_input_vars}, " f"but got {prompt_.input_variables}" ) return cls(llm=llm, prompt=prompt_, **kwargs) def _prepare_input( self, prediction: str, prediction_b: str, input: str, reference: Optional[str] ) -> dict: input_ = { "prediction": prediction, "prediction_b": prediction_b, "input": input, } if reference is not None and "reference" in self.prompt.input_variables: input_["reference"] = reference return input_ [docs] def evaluate_string_pairs( self, *, prediction: str,
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/comparison/eval_chain.html
0ce7b92b6045-3
self, *, prediction: str, prediction_b: str, input: str, reference: Optional[str] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> dict: """Evaluate whether output A is preferred to output B. Args: prediction (str): The output string from the first model. prediction_b (str): The output string from the second model. input (str): The input or task string. callbacks (Callbacks, optional): The callbacks to use. reference (str, optional): The reference string, if any. **kwargs (Any): Additional keyword arguments. Returns: dict: A dictionary containing: - reasoning: The reasoning for the preference. - value: The preference value, which is either 'A', 'B', or None for no preference. - score: The preference score, which is 1 for 'A', 0 for 'B', and 0.5 for None. """ input_ = self._prepare_input(prediction, prediction_b, input, reference) result = self( inputs=input_, callbacks=callbacks, **kwargs, ) return result["text"] [docs] async def aevaluate_string_pairs( self, *, prediction: str, prediction_b: str, input: str, reference: Optional[str] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> dict: """Asynchronously evaluate whether output A is preferred to output B. Args: prediction (str): The output string from the first model.
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/comparison/eval_chain.html
0ce7b92b6045-4
Args: prediction (str): The output string from the first model. prediction_b (str): The output string from the second model. input (str): The input or task string. callbacks (Callbacks, optional): The callbacks to use. reference (str, optional): The reference string, if any. **kwargs (Any): Additional keyword arguments. Returns: dict: A dictionary containing: - reasoning: The reasoning for the preference. - value: The preference value, which is either 'A', 'B', or None for no preference. - score: The preference score, which is 1 for 'A', 0 for 'B', and 0.5 for None. """ input_ = self._prepare_input(prediction, prediction_b, input, reference) result = await self.acall( inputs=input_, callbacks=callbacks, **kwargs, ) return result["text"]
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/comparison/eval_chain.html
587d510a7bad-0
Source code for langchain.evaluation.run_evaluators.implementations from typing import Any, Dict, List, Mapping, Optional, Sequence, Union from langchainplus_sdk.evaluation import EvaluationResult from langchainplus_sdk.schemas import Example, Run, RunTypeEnum from pydantic import BaseModel, Field from langchain.base_language import BaseLanguageModel from langchain.chains.llm import LLMChain from langchain.chat_models.base import BaseChatModel from langchain.evaluation.agents.trajectory_eval_prompt import ( EVAL_CHAT_PROMPT as TRAJECTORY_PROMPT, ) from langchain.evaluation.criteria.eval_chain import ( CriteriaEvalChain, CriteriaResultOutputParser, ) from langchain.evaluation.criteria.prompt import PROMPT as CRITERIA_PROMPT from langchain.evaluation.qa.eval_chain import QAEvalChain from langchain.evaluation.qa.eval_prompt import PROMPT as QA_DEFAULT_PROMPT from langchain.evaluation.qa.eval_prompt import SQL_PROMPT from langchain.evaluation.run_evaluators.base import ( RunEvaluatorChain, RunEvaluatorInputMapper, RunEvaluatorOutputParser, ) from langchain.prompts.base import BasePromptTemplate from langchain.prompts.prompt import PromptTemplate from langchain.schema import OutputParserException from langchain.tools.base import BaseTool _QA_PROMPTS = { "qa": QA_DEFAULT_PROMPT, "sql": SQL_PROMPT, } [docs]class StringRunEvaluatorInputMapper(RunEvaluatorInputMapper, BaseModel): """Maps the Run and Optional[Example] to a dictionary.""" prediction_map: Dict[str, str] """Map from run outputs to the evaluation inputs.""" input_map: Dict[str, str] """Map from run inputs to the evaluation inputs."""
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/run_evaluators/implementations.html
587d510a7bad-1
"""Map from run inputs to the evaluation inputs.""" answer_map: Optional[Dict[str, str]] = None """Map from example outputs to the evaluation inputs.""" [docs] def map(self, run: Run, example: Optional[Example] = None) -> Dict[str, Any]: """Maps the Run and Optional[Example] to a dictionary""" if run.outputs is None and self.prediction_map: raise ValueError(f"Run {run.id} has no outputs.") if self.answer_map and (not example or not example.outputs): raise ValueError("This evaluator requires references, but none were given.") outputs = run.outputs or {} data = {value: outputs[key] for key, value in self.prediction_map.items()} data.update({value: run.inputs[key] for key, value in self.input_map.items()}) if self.answer_map and example and example.outputs: data.update( {value: example.outputs[key] for key, value in self.answer_map.items()} ) return data [docs]class ChoicesOutputParser(RunEvaluatorOutputParser): """Parse a feedback run with optional choices.""" evaluation_name: str choices_map: Optional[Dict[str, int]] = None @property def _type(self) -> str: return "choices_run_eval" [docs] def parse(self, text: str) -> EvaluationResult: """Parse the last line of the text and return an evaluation result.""" lines = text.strip().split() value = lines[-1].strip() score = self.choices_map.get(value) if self.choices_map else None comment = " ".join(lines[:-1]) if len(lines) > 1 else None return EvaluationResult(
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/run_evaluators/implementations.html
587d510a7bad-2
return EvaluationResult( key=self.evaluation_name, score=score, value=value, comment=comment, ) [docs]def get_qa_evaluator( llm: BaseLanguageModel, *, prompt: Union[PromptTemplate, str] = QA_DEFAULT_PROMPT, input_key: str = "input", prediction_key: str = "output", answer_key: str = "output", evaluation_name: Optional[str] = None, **kwargs: Any, ) -> RunEvaluatorChain: """Get an eval chain that compares response against ground truth.""" if isinstance(prompt, str): prompt = _QA_PROMPTS[prompt] eval_chain = QAEvalChain.from_llm(llm=llm, prompt=prompt, **kwargs) input_mapper = kwargs.pop( "input_mapper", StringRunEvaluatorInputMapper( input_map={input_key: "query"}, prediction_map={prediction_key: "result"}, answer_map={answer_key: "answer"}, ), ) evaluation_name = evaluation_name or "Correctness" output_parser = kwargs.pop( "output_parser", ChoicesOutputParser( evaluation_name=evaluation_name, choices_map={"CORRECT": 1, "INCORRECT": 0}, ), ) tags = kwargs.pop("tags", []) return RunEvaluatorChain( eval_chain=eval_chain, input_mapper=input_mapper, output_parser=output_parser, tags=tags + [evaluation_name], **kwargs, ) [docs]class CriteriaOutputParser(RunEvaluatorOutputParser): """Parse a criteria results into an evaluation result."""
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/run_evaluators/implementations.html
587d510a7bad-3
"""Parse a criteria results into an evaluation result.""" evaluation_name: str @property def _type(self) -> str: return "criteria" [docs] def parse(self, parsed_output: Union[str, dict]) -> EvaluationResult: """Parse the last line of the text and return an evaluation result.""" if isinstance(parsed_output, str): parsed_output_ = CriteriaResultOutputParser().parse(parsed_output) else: parsed_output_ = parsed_output return EvaluationResult( key=self.evaluation_name, score=parsed_output_["score"], value=parsed_output_["value"], comment=parsed_output_["reasoning"], ) [docs]def get_criteria_evaluator( llm: BaseLanguageModel, criteria: Union[Mapping[str, str], Sequence[str], str], *, input_key: str = "input", prediction_key: str = "output", prompt: BasePromptTemplate = CRITERIA_PROMPT, evaluation_name: Optional[str] = None, **kwargs: Any, ) -> RunEvaluatorChain: """Get an eval chain for grading a model's response against a map of criteria.""" input_mapper = kwargs.pop( "input_mapper", StringRunEvaluatorInputMapper( input_map={input_key: "input"}, prediction_map={prediction_key: "output"}, ), ) criteria_ = CriteriaEvalChain.resolve_criteria(criteria) evaluation_name = evaluation_name or " ".join(criteria_.keys()) parser = kwargs.pop( "output_parser", CriteriaOutputParser( choices_map={"Y": 1, "N": 0}, evaluation_name=evaluation_name ), )
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/run_evaluators/implementations.html
587d510a7bad-4
), ) tags = kwargs.pop("tags", []) eval_chain = CriteriaEvalChain.from_llm( llm=llm, criteria=criteria_, prompt=prompt, **kwargs ) return RunEvaluatorChain( eval_chain=eval_chain, input_mapper=input_mapper, output_parser=parser, tags=tags + [evaluation_name], **kwargs, ) [docs]class TrajectoryEvalOutputParser(RunEvaluatorOutputParser): evaluation_name: str = "Agent Trajectory" """The name assigned to the evaluation feedback.""" evaluator_info: dict = Field(default_factory=dict) """Additional information to log as feedback metadata.""" @property def _type(self) -> str: return "agent_trajectory_run_eval" [docs] def parse(self, text: str) -> EvaluationResult: if "Score:" not in text: raise OutputParserException( f"Could not find score in model eval output: {text}" ) reasoning, score_str = text.split("Score: ") reasoning, score_str = reasoning.strip(), score_str.strip() score_str = next( (char for char in score_str if char.isdigit()), "0" ) # Scan for first digit if not 1 <= int(score_str) <= 5: raise OutputParserException( f"Score is not a digit in the range 1-5: {text}" ) return EvaluationResult( key=self.evaluation_name, score=int(score_str), comment=reasoning, evaluator_info=self.evaluator_info, ) [docs]class TrajectoryInputMapper(RunEvaluatorInputMapper, BaseModel):
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/run_evaluators/implementations.html
587d510a7bad-5
) [docs]class TrajectoryInputMapper(RunEvaluatorInputMapper, BaseModel): """Maps the Run and Optional[Example] to a dictionary.""" tool_descriptions: List[str] """The descriptions for each of the tools available to the agent.""" agent_input_key: str = "input" """The key to load from the agent executor's run input dictionary.""" agent_output_key: str = "output" """The key to load from the agent executor's run output dictionary.""" tool_input_key: str = "input" """The key to load from the tool executor's run input dictionary.""" tool_output_key: str = "output" """The key to load from the tool executor's run output dictionary.""" [docs] def map(self, run: Run, example: Optional[Example] = None) -> Dict[str, str]: """Maps the Run and Optional[Example] to a dictionary""" if run.child_runs is None: raise ValueError("Run must have child runs to be evaluated.") if run.outputs is None: raise ValueError("Run must have outputs to be evaluated.") question = run.inputs[self.agent_input_key] tool_runs = [ run_ for run_ in run.child_runs if run_.run_type == RunTypeEnum.tool ] agent_steps = [] for i, run_ in enumerate(tool_runs, 1): tool_output = ( f"Tool output: {run_.outputs.get(self.tool_output_key, run_.outputs)}" if run_.outputs else (f"Tool error: {run_.error}" if run_.error else "No output") ) agent_steps.append( f"""Step {i}: Tool used: {run_.name}
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/run_evaluators/implementations.html
587d510a7bad-6
f"""Step {i}: Tool used: {run_.name} Tool input: {run_.inputs.get(self.tool_input_key, run_.inputs)} Tool output: {tool_output}""" ) return { "tool_descriptions": "\n\n".join(self.tool_descriptions), "question": question, "agent_trajectory": "\n\n".join(agent_steps), "answer": run.outputs[self.agent_output_key], } [docs]def get_trajectory_evaluator( llm: BaseChatModel, agent_tools: Union[Sequence[str], Sequence[BaseTool]], *, input_key: str = "input", prediction_key: str = "output", tool_input_key: str = "input", tool_output_key: str = "output", prompt: BasePromptTemplate = TRAJECTORY_PROMPT, evaluation_name: str = "Agent Trajectory", **kwargs: Any, ) -> RunEvaluatorChain: """Get an eval chain for grading a model's response against a map of criteria.""" tool_descriptions = [ f"Tool {i}: {tool.name}\nDescription: {tool.description}" if isinstance(tool, BaseTool) else f"Tool {i}: {tool}" for i, tool in enumerate(agent_tools, 1) ] input_mapper = kwargs.pop( "input_mapper", TrajectoryInputMapper( agent_input_key=input_key, agent_output_key=prediction_key, tool_input_key=tool_input_key, tool_output_key=tool_output_key, tool_descriptions=tool_descriptions, ), ) parser = kwargs.pop( "output_parser",
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/run_evaluators/implementations.html
587d510a7bad-7
), ) parser = kwargs.pop( "output_parser", TrajectoryEvalOutputParser(evaluation_name=evaluation_name), ) eval_chain = LLMChain(llm=llm, prompt=prompt, **kwargs) tags = kwargs.pop("tags", []) return RunEvaluatorChain( eval_chain=eval_chain, input_mapper=input_mapper, output_parser=parser, tags=tags + [evaluation_name], **kwargs, )
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/run_evaluators/implementations.html
8f5554f4f4c8-0
Source code for langchain.evaluation.run_evaluators.base from __future__ import annotations from abc import abstractmethod from typing import Any, Dict, List, Optional from langchainplus_sdk import EvaluationResult, RunEvaluator from langchainplus_sdk.schemas import Example, Run from langchain.callbacks.manager import ( AsyncCallbackManagerForChainRun, CallbackManagerForChainRun, ) from langchain.chains.base import Chain from langchain.schema import RUN_KEY, BaseOutputParser class RunEvaluatorInputMapper: """Map the inputs of a run to the inputs of an evaluation.""" @abstractmethod def map(self, run: Run, example: Optional[Example] = None) -> Dict[str, Any]: """Maps the Run and Optional[Example] to a dictionary""" [docs]class RunEvaluatorOutputParser(BaseOutputParser[EvaluationResult]): """Parse the output of a run.""" eval_chain_output_key: str = "text" [docs] def parse_chain_output(self, output: Dict[str, Any]) -> EvaluationResult: """Parse the output of a run.""" text = output[self.eval_chain_output_key] return self.parse(text) [docs]class RunEvaluatorChain(Chain, RunEvaluator): """Evaluate Run and optional examples.""" input_mapper: RunEvaluatorInputMapper """Maps the Run and Optional example to a dictionary for the eval chain.""" eval_chain: Chain """The evaluation chain.""" output_parser: RunEvaluatorOutputParser """Parse the output of the eval chain into feedback.""" @property def input_keys(self) -> List[str]: return ["run", "example"] @property def output_keys(self) -> List[str]: return ["feedback"] def _call(
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/run_evaluators/base.html
8f5554f4f4c8-1
return ["feedback"] def _call( self, inputs: Dict[str, Any], run_manager: Optional[CallbackManagerForChainRun] = None, ) -> Dict[str, Any]: """Call the evaluation chain.""" run: Run = inputs["run"] example: Optional[Example] = inputs.get("example") chain_input = self.input_mapper.map(run, example) _run_manager = run_manager or CallbackManagerForChainRun.get_noop_manager() callbacks = _run_manager.get_child() chain_output = self.eval_chain( chain_input, callbacks=callbacks, include_run_info=True ) run_info = chain_output[RUN_KEY] feedback = self.output_parser.parse_chain_output(chain_output) feedback.evaluator_info[RUN_KEY] = run_info return {"feedback": feedback} async def _acall( self, inputs: Dict[str, Any], run_manager: AsyncCallbackManagerForChainRun | None = None, ) -> Dict[str, Any]: run: Run = inputs["run"] example: Optional[Example] = inputs.get("example") chain_input = self.input_mapper.map(run, example) _run_manager = run_manager or AsyncCallbackManagerForChainRun.get_noop_manager() callbacks = _run_manager.get_child() chain_output = await self.eval_chain.acall( chain_input, callbacks=callbacks, include_run_info=True, ) run_info = chain_output[RUN_KEY] feedback = self.output_parser.parse_chain_output(chain_output) feedback.evaluator_info[RUN_KEY] = run_info return {"feedback": feedback} [docs] def evaluate_run(
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/run_evaluators/base.html
8f5554f4f4c8-2
return {"feedback": feedback} [docs] def evaluate_run( self, run: Run, example: Optional[Example] = None ) -> EvaluationResult: """Evaluate an example.""" return self({"run": run, "example": example})["feedback"] [docs] async def aevaluate_run( self, run: Run, example: Optional[Example] = None ) -> EvaluationResult: """Evaluate an example.""" result = await self.acall({"run": run, "example": example}) return result["feedback"]
https://api.python.langchain.com/en/latest/_modules/langchain/evaluation/run_evaluators/base.html
0e1c25b57cce-0
Source code for langchain.callbacks.streaming_aiter from __future__ import annotations import asyncio from typing import Any, AsyncIterator, Dict, List, Literal, Union, cast from langchain.callbacks.base import AsyncCallbackHandler from langchain.schema import LLMResult # TODO If used by two LLM runs in parallel this won't work as expected [docs]class AsyncIteratorCallbackHandler(AsyncCallbackHandler): """Callback handler that returns an async iterator.""" queue: asyncio.Queue[str] done: asyncio.Event @property def always_verbose(self) -> bool: return True def __init__(self) -> None: self.queue = asyncio.Queue() self.done = asyncio.Event() [docs] async def on_llm_start( self, serialized: Dict[str, Any], prompts: List[str], **kwargs: Any ) -> None: # If two calls are made in a row, this resets the state self.done.clear() [docs] async def on_llm_new_token(self, token: str, **kwargs: Any) -> None: self.queue.put_nowait(token) [docs] async def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None: self.done.set() [docs] async def on_llm_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any ) -> None: self.done.set() # TODO implement the other methods [docs] async def aiter(self) -> AsyncIterator[str]: while not self.queue.empty() or not self.done.is_set(): # Wait for the next token in the queue, # but stop waiting if the done event is set done, other = await asyncio.wait(
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/streaming_aiter.html
0e1c25b57cce-1
done, other = await asyncio.wait( [ # NOTE: If you add other tasks here, update the code below, # which assumes each set has exactly one task each asyncio.ensure_future(self.queue.get()), asyncio.ensure_future(self.done.wait()), ], return_when=asyncio.FIRST_COMPLETED, ) # Cancel the other task if other: other.pop().cancel() # Extract the value of the first completed task token_or_done = cast(Union[str, Literal[True]], done.pop().result()) # If the extracted value is the boolean True, the done event was set if token_or_done is True: break # Otherwise, the extracted value is a token, which we yield yield token_or_done
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/streaming_aiter.html
457b5ae9b8f3-0
Source code for langchain.callbacks.utils import hashlib from pathlib import Path from typing import Any, Dict, Iterable, Tuple, Union [docs]def import_spacy() -> Any: """Import the spacy python package and raise an error if it is not installed.""" try: import spacy except ImportError: raise ImportError( "This callback manager requires the `spacy` python " "package installed. Please install it with `pip install spacy`" ) return spacy [docs]def import_pandas() -> Any: """Import the pandas python package and raise an error if it is not installed.""" try: import pandas except ImportError: raise ImportError( "This callback manager requires the `pandas` python " "package installed. Please install it with `pip install pandas`" ) return pandas [docs]def import_textstat() -> Any: """Import the textstat python package and raise an error if it is not installed.""" try: import textstat except ImportError: raise ImportError( "This callback manager requires the `textstat` python " "package installed. Please install it with `pip install textstat`" ) return textstat def _flatten_dict( nested_dict: Dict[str, Any], parent_key: str = "", sep: str = "_" ) -> Iterable[Tuple[str, Any]]: """ Generator that yields flattened items from a nested dictionary for a flat dict. Parameters: nested_dict (dict): The nested dictionary to flatten. parent_key (str): The prefix to prepend to the keys of the flattened dict.
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/utils.html
457b5ae9b8f3-1
parent_key (str): The prefix to prepend to the keys of the flattened dict. sep (str): The separator to use between the parent key and the key of the flattened dictionary. Yields: (str, any): A key-value pair from the flattened dictionary. """ for key, value in nested_dict.items(): new_key = parent_key + sep + key if parent_key else key if isinstance(value, dict): yield from _flatten_dict(value, new_key, sep) else: yield new_key, value [docs]def flatten_dict( nested_dict: Dict[str, Any], parent_key: str = "", sep: str = "_" ) -> Dict[str, Any]: """Flattens a nested dictionary into a flat dictionary. Parameters: nested_dict (dict): The nested dictionary to flatten. parent_key (str): The prefix to prepend to the keys of the flattened dict. sep (str): The separator to use between the parent key and the key of the flattened dictionary. Returns: (dict): A flat dictionary. """ flat_dict = {k: v for k, v in _flatten_dict(nested_dict, parent_key, sep)} return flat_dict [docs]def hash_string(s: str) -> str: """Hash a string using sha1. Parameters: s (str): The string to hash. Returns: (str): The hashed string. """ return hashlib.sha1(s.encode("utf-8")).hexdigest() [docs]def load_json(json_path: Union[str, Path]) -> str: """Load json file to a string. Parameters:
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/utils.html
457b5ae9b8f3-2
"""Load json file to a string. Parameters: json_path (str): The path to the json file. Returns: (str): The string representation of the json file. """ with open(json_path, "r") as f: data = f.read() return data class BaseMetadataCallbackHandler: """This class handles the metadata and associated function states for callbacks. Attributes: step (int): The current step. starts (int): The number of times the start method has been called. ends (int): The number of times the end method has been called. errors (int): The number of times the error method has been called. text_ctr (int): The number of times the text method has been called. ignore_llm_ (bool): Whether to ignore llm callbacks. ignore_chain_ (bool): Whether to ignore chain callbacks. ignore_agent_ (bool): Whether to ignore agent callbacks. ignore_retriever_ (bool): Whether to ignore retriever callbacks. always_verbose_ (bool): Whether to always be verbose. chain_starts (int): The number of times the chain start method has been called. chain_ends (int): The number of times the chain end method has been called. llm_starts (int): The number of times the llm start method has been called. llm_ends (int): The number of times the llm end method has been called. llm_streams (int): The number of times the text method has been called. tool_starts (int): The number of times the tool start method has been called. tool_ends (int): The number of times the tool end method has been called.
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/utils.html
457b5ae9b8f3-3
tool_ends (int): The number of times the tool end method has been called. agent_ends (int): The number of times the agent end method has been called. on_llm_start_records (list): A list of records of the on_llm_start method. on_llm_token_records (list): A list of records of the on_llm_token method. on_llm_end_records (list): A list of records of the on_llm_end method. on_chain_start_records (list): A list of records of the on_chain_start method. on_chain_end_records (list): A list of records of the on_chain_end method. on_tool_start_records (list): A list of records of the on_tool_start method. on_tool_end_records (list): A list of records of the on_tool_end method. on_agent_finish_records (list): A list of records of the on_agent_end method. """ def __init__(self) -> None: self.step = 0 self.starts = 0 self.ends = 0 self.errors = 0 self.text_ctr = 0 self.ignore_llm_ = False self.ignore_chain_ = False self.ignore_agent_ = False self.ignore_retriever_ = False self.always_verbose_ = False self.chain_starts = 0 self.chain_ends = 0 self.llm_starts = 0 self.llm_ends = 0 self.llm_streams = 0 self.tool_starts = 0 self.tool_ends = 0 self.agent_ends = 0 self.on_llm_start_records: list = [] self.on_llm_token_records: list = []
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/utils.html
457b5ae9b8f3-4
self.on_llm_token_records: list = [] self.on_llm_end_records: list = [] self.on_chain_start_records: list = [] self.on_chain_end_records: list = [] self.on_tool_start_records: list = [] self.on_tool_end_records: list = [] self.on_text_records: list = [] self.on_agent_finish_records: list = [] self.on_agent_action_records: list = [] @property def always_verbose(self) -> bool: """Whether to call verbose callbacks even if verbose is False.""" return self.always_verbose_ @property def ignore_llm(self) -> bool: """Whether to ignore LLM callbacks.""" return self.ignore_llm_ @property def ignore_chain(self) -> bool: """Whether to ignore chain callbacks.""" return self.ignore_chain_ @property def ignore_agent(self) -> bool: """Whether to ignore agent callbacks.""" return self.ignore_agent_ def get_custom_callback_meta(self) -> Dict[str, Any]: return { "step": self.step, "starts": self.starts, "ends": self.ends, "errors": self.errors, "text_ctr": self.text_ctr, "chain_starts": self.chain_starts, "chain_ends": self.chain_ends, "llm_starts": self.llm_starts, "llm_ends": self.llm_ends, "llm_streams": self.llm_streams, "tool_starts": self.tool_starts, "tool_ends": self.tool_ends, "agent_ends": self.agent_ends, } def reset_callback_meta(self) -> None: """Reset the callback metadata."""
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/utils.html
457b5ae9b8f3-5
def reset_callback_meta(self) -> None: """Reset the callback metadata.""" self.step = 0 self.starts = 0 self.ends = 0 self.errors = 0 self.text_ctr = 0 self.ignore_llm_ = False self.ignore_chain_ = False self.ignore_agent_ = False self.always_verbose_ = False self.chain_starts = 0 self.chain_ends = 0 self.llm_starts = 0 self.llm_ends = 0 self.llm_streams = 0 self.tool_starts = 0 self.tool_ends = 0 self.agent_ends = 0 self.on_llm_start_records = [] self.on_llm_token_records = [] self.on_llm_end_records = [] self.on_chain_start_records = [] self.on_chain_end_records = [] self.on_tool_start_records = [] self.on_tool_end_records = [] self.on_text_records = [] self.on_agent_finish_records = [] self.on_agent_action_records = [] return None
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/utils.html
ea37bd1368ef-0
Source code for langchain.callbacks.aim_callback from copy import deepcopy from typing import Any, Dict, List, Optional, Union from langchain.callbacks.base import BaseCallbackHandler from langchain.schema import AgentAction, AgentFinish, LLMResult [docs]def import_aim() -> Any: """Import the aim python package and raise an error if it is not installed.""" try: import aim except ImportError: raise ImportError( "To use the Aim callback manager you need to have the" " `aim` python package installed." "Please install it with `pip install aim`" ) return aim class BaseMetadataCallbackHandler: """This class handles the metadata and associated function states for callbacks. Attributes: step (int): The current step. starts (int): The number of times the start method has been called. ends (int): The number of times the end method has been called. errors (int): The number of times the error method has been called. text_ctr (int): The number of times the text method has been called. ignore_llm_ (bool): Whether to ignore llm callbacks. ignore_chain_ (bool): Whether to ignore chain callbacks. ignore_agent_ (bool): Whether to ignore agent callbacks. ignore_retriever_ (bool): Whether to ignore retriever callbacks. always_verbose_ (bool): Whether to always be verbose. chain_starts (int): The number of times the chain start method has been called. chain_ends (int): The number of times the chain end method has been called. llm_starts (int): The number of times the llm start method has been called.
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/aim_callback.html
ea37bd1368ef-1
llm_ends (int): The number of times the llm end method has been called. llm_streams (int): The number of times the text method has been called. tool_starts (int): The number of times the tool start method has been called. tool_ends (int): The number of times the tool end method has been called. agent_ends (int): The number of times the agent end method has been called. """ def __init__(self) -> None: self.step = 0 self.starts = 0 self.ends = 0 self.errors = 0 self.text_ctr = 0 self.ignore_llm_ = False self.ignore_chain_ = False self.ignore_agent_ = False self.ignore_retriever_ = False self.always_verbose_ = False self.chain_starts = 0 self.chain_ends = 0 self.llm_starts = 0 self.llm_ends = 0 self.llm_streams = 0 self.tool_starts = 0 self.tool_ends = 0 self.agent_ends = 0 @property def always_verbose(self) -> bool: """Whether to call verbose callbacks even if verbose is False.""" return self.always_verbose_ @property def ignore_llm(self) -> bool: """Whether to ignore LLM callbacks.""" return self.ignore_llm_ @property def ignore_chain(self) -> bool: """Whether to ignore chain callbacks.""" return self.ignore_chain_ @property def ignore_agent(self) -> bool: """Whether to ignore agent callbacks.""" return self.ignore_agent_ @property
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/aim_callback.html
ea37bd1368ef-2
"""Whether to ignore agent callbacks.""" return self.ignore_agent_ @property def ignore_retriever(self) -> bool: """Whether to ignore retriever callbacks.""" return self.ignore_retriever_ def get_custom_callback_meta(self) -> Dict[str, Any]: return { "step": self.step, "starts": self.starts, "ends": self.ends, "errors": self.errors, "text_ctr": self.text_ctr, "chain_starts": self.chain_starts, "chain_ends": self.chain_ends, "llm_starts": self.llm_starts, "llm_ends": self.llm_ends, "llm_streams": self.llm_streams, "tool_starts": self.tool_starts, "tool_ends": self.tool_ends, "agent_ends": self.agent_ends, } def reset_callback_meta(self) -> None: """Reset the callback metadata.""" self.step = 0 self.starts = 0 self.ends = 0 self.errors = 0 self.text_ctr = 0 self.ignore_llm_ = False self.ignore_chain_ = False self.ignore_agent_ = False self.always_verbose_ = False self.chain_starts = 0 self.chain_ends = 0 self.llm_starts = 0 self.llm_ends = 0 self.llm_streams = 0 self.tool_starts = 0 self.tool_ends = 0 self.agent_ends = 0 return None [docs]class AimCallbackHandler(BaseMetadataCallbackHandler, BaseCallbackHandler): """Callback Handler that logs to Aim. Parameters:
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/aim_callback.html
ea37bd1368ef-3
"""Callback Handler that logs to Aim. Parameters: repo (:obj:`str`, optional): Aim repository path or Repo object to which Run object is bound. If skipped, default Repo is used. experiment_name (:obj:`str`, optional): Sets Run's `experiment` property. 'default' if not specified. Can be used later to query runs/sequences. system_tracking_interval (:obj:`int`, optional): Sets the tracking interval in seconds for system usage metrics (CPU, Memory, etc.). Set to `None` to disable system metrics tracking. log_system_params (:obj:`bool`, optional): Enable/Disable logging of system params such as installed packages, git info, environment variables, etc. This handler will utilize the associated callback method called and formats the input of each callback function with metadata regarding the state of LLM run and then logs the response to Aim. """ def __init__( self, repo: Optional[str] = None, experiment_name: Optional[str] = None, system_tracking_interval: Optional[int] = 10, log_system_params: bool = True, ) -> None: """Initialize callback handler.""" super().__init__() aim = import_aim() self.repo = repo self.experiment_name = experiment_name self.system_tracking_interval = system_tracking_interval self.log_system_params = log_system_params self._run = aim.Run( repo=self.repo, experiment=self.experiment_name, system_tracking_interval=self.system_tracking_interval, log_system_params=self.log_system_params, ) self._run_hash = self._run.hash self.action_records: list = []
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/aim_callback.html
ea37bd1368ef-4
self._run_hash = self._run.hash self.action_records: list = [] [docs] def setup(self, **kwargs: Any) -> None: aim = import_aim() if not self._run: if self._run_hash: self._run = aim.Run( self._run_hash, repo=self.repo, system_tracking_interval=self.system_tracking_interval, ) else: self._run = aim.Run( repo=self.repo, experiment=self.experiment_name, system_tracking_interval=self.system_tracking_interval, log_system_params=self.log_system_params, ) self._run_hash = self._run.hash if kwargs: for key, value in kwargs.items(): self._run.set(key, value, strict=False) [docs] def on_llm_start( self, serialized: Dict[str, Any], prompts: List[str], **kwargs: Any ) -> None: """Run when LLM starts.""" aim = import_aim() self.step += 1 self.llm_starts += 1 self.starts += 1 resp = {"action": "on_llm_start"} resp.update(self.get_custom_callback_meta()) prompts_res = deepcopy(prompts) self._run.track( [aim.Text(prompt) for prompt in prompts_res], name="on_llm_start", context=resp, ) [docs] def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None: """Run when LLM ends running.""" aim = import_aim() self.step += 1 self.llm_ends += 1 self.ends += 1
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/aim_callback.html
ea37bd1368ef-5
self.llm_ends += 1 self.ends += 1 resp = {"action": "on_llm_end"} resp.update(self.get_custom_callback_meta()) response_res = deepcopy(response) generated = [ aim.Text(generation.text) for generations in response_res.generations for generation in generations ] self._run.track( generated, name="on_llm_end", context=resp, ) [docs] def on_llm_new_token(self, token: str, **kwargs: Any) -> None: """Run when LLM generates a new token.""" self.step += 1 self.llm_streams += 1 [docs] def on_llm_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any ) -> None: """Run when LLM errors.""" self.step += 1 self.errors += 1 [docs] def on_chain_start( self, serialized: Dict[str, Any], inputs: Dict[str, Any], **kwargs: Any ) -> None: """Run when chain starts running.""" aim = import_aim() self.step += 1 self.chain_starts += 1 self.starts += 1 resp = {"action": "on_chain_start"} resp.update(self.get_custom_callback_meta()) inputs_res = deepcopy(inputs) self._run.track( aim.Text(inputs_res["input"]), name="on_chain_start", context=resp ) [docs] def on_chain_end(self, outputs: Dict[str, Any], **kwargs: Any) -> None: """Run when chain ends running.""" aim = import_aim()
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/aim_callback.html
ea37bd1368ef-6
"""Run when chain ends running.""" aim = import_aim() self.step += 1 self.chain_ends += 1 self.ends += 1 resp = {"action": "on_chain_end"} resp.update(self.get_custom_callback_meta()) outputs_res = deepcopy(outputs) self._run.track( aim.Text(outputs_res["output"]), name="on_chain_end", context=resp ) [docs] def on_chain_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any ) -> None: """Run when chain errors.""" self.step += 1 self.errors += 1 [docs] def on_tool_start( self, serialized: Dict[str, Any], input_str: str, **kwargs: Any ) -> None: """Run when tool starts running.""" aim = import_aim() self.step += 1 self.tool_starts += 1 self.starts += 1 resp = {"action": "on_tool_start"} resp.update(self.get_custom_callback_meta()) self._run.track(aim.Text(input_str), name="on_tool_start", context=resp) [docs] def on_tool_end(self, output: str, **kwargs: Any) -> None: """Run when tool ends running.""" aim = import_aim() self.step += 1 self.tool_ends += 1 self.ends += 1 resp = {"action": "on_tool_end"} resp.update(self.get_custom_callback_meta()) self._run.track(aim.Text(output), name="on_tool_end", context=resp) [docs] def on_tool_error(
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/aim_callback.html
ea37bd1368ef-7
[docs] def on_tool_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any ) -> None: """Run when tool errors.""" self.step += 1 self.errors += 1 [docs] def on_text(self, text: str, **kwargs: Any) -> None: """ Run when agent is ending. """ self.step += 1 self.text_ctr += 1 [docs] def on_agent_finish(self, finish: AgentFinish, **kwargs: Any) -> None: """Run when agent ends running.""" aim = import_aim() self.step += 1 self.agent_ends += 1 self.ends += 1 resp = {"action": "on_agent_finish"} resp.update(self.get_custom_callback_meta()) finish_res = deepcopy(finish) text = "OUTPUT:\n{}\n\nLOG:\n{}".format( finish_res.return_values["output"], finish_res.log ) self._run.track(aim.Text(text), name="on_agent_finish", context=resp) [docs] def on_agent_action(self, action: AgentAction, **kwargs: Any) -> Any: """Run on agent action.""" aim = import_aim() self.step += 1 self.tool_starts += 1 self.starts += 1 resp = { "action": "on_agent_action", "tool": action.tool, } resp.update(self.get_custom_callback_meta()) action_res = deepcopy(action) text = "TOOL INPUT:\n{}\n\nLOG:\n{}".format( action_res.tool_input, action_res.log )
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/aim_callback.html
ea37bd1368ef-8
action_res.tool_input, action_res.log ) self._run.track(aim.Text(text), name="on_agent_action", context=resp) [docs] def flush_tracker( self, repo: Optional[str] = None, experiment_name: Optional[str] = None, system_tracking_interval: Optional[int] = 10, log_system_params: bool = True, langchain_asset: Any = None, reset: bool = True, finish: bool = False, ) -> None: """Flush the tracker and reset the session. Args: repo (:obj:`str`, optional): Aim repository path or Repo object to which Run object is bound. If skipped, default Repo is used. experiment_name (:obj:`str`, optional): Sets Run's `experiment` property. 'default' if not specified. Can be used later to query runs/sequences. system_tracking_interval (:obj:`int`, optional): Sets the tracking interval in seconds for system usage metrics (CPU, Memory, etc.). Set to `None` to disable system metrics tracking. log_system_params (:obj:`bool`, optional): Enable/Disable logging of system params such as installed packages, git info, environment variables, etc. langchain_asset: The langchain asset to save. reset: Whether to reset the session. finish: Whether to finish the run. Returns: None """ if langchain_asset: try: for key, value in langchain_asset.dict().items(): self._run.set(key, value, strict=False) except Exception: pass if finish or reset: self._run.close() self.reset_callback_meta()
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/aim_callback.html
ea37bd1368ef-9
self._run.close() self.reset_callback_meta() if reset: self.__init__( # type: ignore repo=repo if repo else self.repo, experiment_name=experiment_name if experiment_name else self.experiment_name, system_tracking_interval=system_tracking_interval if system_tracking_interval else self.system_tracking_interval, log_system_params=log_system_params if log_system_params else self.log_system_params, )
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/aim_callback.html
c21ace92c567-0
Source code for langchain.callbacks.clearml_callback import tempfile from copy import deepcopy from pathlib import Path from typing import Any, Dict, List, Optional, Sequence, Union from langchain.callbacks.base import BaseCallbackHandler from langchain.callbacks.utils import ( BaseMetadataCallbackHandler, flatten_dict, hash_string, import_pandas, import_spacy, import_textstat, load_json, ) from langchain.schema import AgentAction, AgentFinish, LLMResult [docs]def import_clearml() -> Any: """Import the clearml python package and raise an error if it is not installed.""" try: import clearml # noqa: F401 except ImportError: raise ImportError( "To use the clearml callback manager you need to have the `clearml` python " "package installed. Please install it with `pip install clearml`" ) return clearml [docs]class ClearMLCallbackHandler(BaseMetadataCallbackHandler, BaseCallbackHandler): """Callback Handler that logs to ClearML. Parameters: job_type (str): The type of clearml task such as "inference", "testing" or "qc" project_name (str): The clearml project name tags (list): Tags to add to the task task_name (str): Name of the clearml task visualize (bool): Whether to visualize the run. complexity_metrics (bool): Whether to log complexity metrics stream_logs (bool): Whether to stream callback actions to ClearML This handler will utilize the associated callback method and formats the input of each callback function with metadata regarding the state of LLM run, and adds the response to the list of records for both the {method}_records and
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/clearml_callback.html
c21ace92c567-1
and adds the response to the list of records for both the {method}_records and action. It then logs the response to the ClearML console. """ def __init__( self, task_type: Optional[str] = "inference", project_name: Optional[str] = "langchain_callback_demo", tags: Optional[Sequence] = None, task_name: Optional[str] = None, visualize: bool = False, complexity_metrics: bool = False, stream_logs: bool = False, ) -> None: """Initialize callback handler.""" clearml = import_clearml() spacy = import_spacy() super().__init__() self.task_type = task_type self.project_name = project_name self.tags = tags self.task_name = task_name self.visualize = visualize self.complexity_metrics = complexity_metrics self.stream_logs = stream_logs self.temp_dir = tempfile.TemporaryDirectory() # Check if ClearML task already exists (e.g. in pipeline) if clearml.Task.current_task(): self.task = clearml.Task.current_task() else: self.task = clearml.Task.init( # type: ignore task_type=self.task_type, project_name=self.project_name, tags=self.tags, task_name=self.task_name, output_uri=True, ) self.logger = self.task.get_logger() warning = ( "The clearml callback is currently in beta and is subject to change " "based on updates to `langchain`. Please report any issues to " "https://github.com/allegroai/clearml/issues with the tag `langchain`." )
https://api.python.langchain.com/en/latest/_modules/langchain/callbacks/clearml_callback.html