id stringlengths 14 16 | text stringlengths 29 2.73k | source stringlengths 49 117 |
|---|---|---|
288269394a4e-1 | session_id = str(uuid4()) # This is a unique identifier for the user
# Load your OpenAI key from a .env file
from dotenv import load_dotenv
load_dotenv()
True
Initialize the Zep Chat Message History Class and initialize the Agent#
ddg = DuckDuckGoSearchRun()
tools = [ddg]
# Set up Zep Chat History
zep_chat_history = Z... | https://python.langchain.com/en/latest/modules/memory/examples/zep_memory.html |
288269394a4e-2 | "The most well-known adaptation of Octavia Butler's work is the FX series"
" Kindred, based on her novel of the same name."
),
},
{"role": "human", "content": "Who were her contemporaries?"},
{
"role": "ai",
"content": (
"Octavia Butler's contemporaries includ... | https://python.langchain.com/en/latest/modules/memory/examples/zep_memory.html |
288269394a4e-3 | ),
},
]
for msg in test_history:
zep_chat_history.append(
HumanMessage(content=msg["content"])
if msg["role"] == "human"
else AIMessage(content=msg["content"])
)
Run the agent#
Doing so will automatically add the input and response to the Zep memory.
agent_chain.run(
input="WWhat... | https://python.langchain.com/en/latest/modules/memory/examples/zep_memory.html |
288269394a4e-4 | Ursula K. Le Guin, Samuel R. Delany, and Joanna Russ.
{'role': 'human', 'content': 'What awards did she win?', 'uuid': '9fa75c3c-edae-41e3-b9bc-9fcf16b523c9', 'created_at': '2023-05-25T15:09:41.91662Z', 'token_count': 8}
{'role': 'ai', 'content': 'Octavia Butler won the Hugo Award, the Nebula Award, and the MacArthur F... | https://python.langchain.com/en/latest/modules/memory/examples/zep_memory.html |
288269394a4e-5 | {'role': 'human', 'content': "Write a short synopsis of Butler's book, Parable of the Sower. What is it about?", 'uuid': '5678d056-7f05-4e70-b8e5-f85efa56db01', 'created_at': '2023-05-25T15:09:41.938974Z', 'token_count': 23}
{'role': 'ai', 'content': 'Parable of the Sower is a science fiction novel by Octavia Butler, p... | https://python.langchain.com/en/latest/modules/memory/examples/zep_memory.html |
288269394a4e-6 | {'role': 'ai', 'content': 'Parable of the Sower is a prescient novel that speaks to the challenges facing contemporary society, such as climate change, economic inequality, and the rise of authoritarianism. It is a cautionary tale that warns of the dangers of ignoring these issues and the importance of taking action to... | https://python.langchain.com/en/latest/modules/memory/examples/zep_memory.html |
288269394a4e-7 | {'uuid': '52cfe3e8-b800-4dd8-a7dd-8e9e4764dfc8', 'created_at': '2023-05-25T15:09:41.913856Z', 'role': 'ai', 'content': "Octavia Butler's contemporaries included Ursula K. Le Guin, Samuel R. Delany, and Joanna Russ.", 'token_count': 27} 0.852352466457884
{'uuid': 'd40da612-0867-4a43-92ec-778b86490a39', 'created_at': '20... | https://python.langchain.com/en/latest/modules/memory/examples/zep_memory.html |
288269394a4e-8 | {'uuid': '862107de-8f6f-43c0-91fa-4441f01b2b3a', 'created_at': '2023-05-25T15:09:41.898149Z', 'role': 'human', 'content': 'Which books of hers were made into movies?', 'token_count': 11} 0.7954322970428519
{'uuid': '97164506-90fe-4c71-9539-69ebcd1d90a2', 'created_at': '2023-05-25T15:09:41.90887Z', 'role': 'human', 'con... | https://python.langchain.com/en/latest/modules/memory/examples/zep_memory.html |
288269394a4e-9 | previous
Redis Chat Message History
next
Indexes
Contents
REACT Agent Chat Message History Example
Initialize the Zep Chat Message History Class and initialize the Agent
Add some history data
Run the agent
Inspect the Zep memory
Vector search over the Zep memory
By Harrison Chase
© Copyright 2023, Harris... | https://python.langchain.com/en/latest/modules/memory/examples/zep_memory.html |
537b0345419d-0 | .ipynb
.pdf
Momento
Momento#
This notebook goes over how to use Momento Cache to store chat message history using the MomentoChatMessageHistory class. See the Momento docs for more detail on how to get set up with Momento.
Note that, by default we will create a cache if one with the given name doesn’t already exist.
Yo... | https://python.langchain.com/en/latest/modules/memory/examples/momento_chat_message_history.html |
c6ea23b40e2c-0 | .ipynb
.pdf
Dynamodb Chat Message History
Contents
DynamoDBChatMessageHistory
Agent with DynamoDB Memory
Dynamodb Chat Message History#
This notebook goes over how to use Dynamodb to store chat message history.
First make sure you have correctly configured the AWS CLI. Then make sure you have installed boto3.
Next, c... | https://python.langchain.com/en/latest/modules/memory/examples/dynamodb_chat_message_history.html |
c6ea23b40e2c-1 | from langchain.agents import initialize_agent
from langchain.agents import AgentType
from langchain.utilities import PythonREPL
from getpass import getpass
message_history = DynamoDBChatMessageHistory(table_name="SessionTable", session_id="1")
memory = ConversationBufferMemory(memory_key="chat_history", chat_memory=mes... | https://python.langchain.com/en/latest/modules/memory/examples/dynamodb_chat_message_history.html |
c6ea23b40e2c-2 | }
Observation: X Corp. (2023–present)Twitter, Inc. (2006–2023)
Thought:{
"action": "Final Answer",
"action_input": "X Corp. (2023–present)Twitter, Inc. (2006–2023)"
}
> Finished chain.
'X Corp. (2023–present)Twitter, Inc. (2006–2023)'
agent_chain.run(input="My name is Bob.")
> Entering new AgentExecutor chain..... | https://python.langchain.com/en/latest/modules/memory/examples/dynamodb_chat_message_history.html |
ea1defa2cd25-0 | .ipynb
.pdf
Cassandra Chat Message History
Cassandra Chat Message History#
This notebook goes over how to use Cassandra to store chat message history.
Cassandra is a distributed database that is well suited for storing large amounts of data.
It is a good choice for storing chat message history because it is easy to sca... | https://python.langchain.com/en/latest/modules/memory/examples/cassandra_chat_message_history.html |
15838d9dbe1e-0 | .ipynb
.pdf
Entity Memory with SQLite storage
Entity Memory with SQLite storage#
In this walkthrough we’ll create a simple conversation chain which uses ConversationEntityMemory backed by a SqliteEntityStore.
from langchain.chains import ConversationChain
from langchain.llms import OpenAI
from langchain.memory import C... | https://python.langchain.com/en/latest/modules/memory/examples/entity_memory_with_sqlite.html |
15838d9dbe1e-1 | You are constantly learning and improving, and your capabilities are constantly evolving. You are able to process and understand large amounts of text, and can use this knowledge to provide accurate and informative responses to a wide range of questions. You have access to some personalized information provided by the ... | https://python.langchain.com/en/latest/modules/memory/examples/entity_memory_with_sqlite.html |
4d76a36d814a-0 | .ipynb
.pdf
How to create a custom Memory class
How to create a custom Memory class#
Although there are a few predefined types of memory in LangChain, it is highly possible you will want to add your own type of memory that is optimal for your application. This notebook covers how to do that.
For this notebook, we will ... | https://python.langchain.com/en/latest/modules/memory/examples/custom_memory.html |
4d76a36d814a-1 | return [self.memory_key]
def load_memory_variables(self, inputs: Dict[str, Any]) -> Dict[str, str]:
"""Load the memory variables, in this case the entity key."""
# Get the input text and run through spacy
doc = nlp(inputs[list(inputs.keys())[0]])
# Extract known information about ent... | https://python.langchain.com/en/latest/modules/memory/examples/custom_memory.html |
4d76a36d814a-2 | Conversation:
Human: {input}
AI:"""
prompt = PromptTemplate(
input_variables=["entities", "input"], template=template
)
And now we put it all together!
llm = OpenAI(temperature=0)
conversation = ConversationChain(llm=llm, prompt=prompt, verbose=True, memory=SpacyEntityMemory())
In the first example, with no prior k... | https://python.langchain.com/en/latest/modules/memory/examples/custom_memory.html |
4d76a36d814a-3 | Conversation:
Human: What do you think Harrison's favorite subject in college was?
AI:
> Finished ConversationChain chain.
' From what I know about Harrison, I believe his favorite subject in college was machine learning. He has expressed a strong interest in the subject and has mentioned it often.'
Again, please note ... | https://python.langchain.com/en/latest/modules/memory/examples/custom_memory.html |
6e97adfe5289-0 | .ipynb
.pdf
Postgres Chat Message History
Postgres Chat Message History#
This notebook goes over how to use Postgres to store chat message history.
from langchain.memory import PostgresChatMessageHistory
history = PostgresChatMessageHistory(connection_string="postgresql://postgres:mypassword@localhost/chat_history", se... | https://python.langchain.com/en/latest/modules/memory/examples/postgres_chat_message_history.html |
c78b5ef673ef-0 | .ipynb
.pdf
Redis Chat Message History
Redis Chat Message History#
This notebook goes over how to use Redis to store chat message history.
from langchain.memory import RedisChatMessageHistory
history = RedisChatMessageHistory("foo")
history.add_user_message("hi!")
history.add_ai_message("whats up?")
history.messages
[A... | https://python.langchain.com/en/latest/modules/memory/examples/redis_chat_message_history.html |
9916889ff898-0 | .ipynb
.pdf
How to add Memory to an Agent
How to add Memory to an Agent#
This notebook goes over adding memory to an Agent. Before going through this notebook, please walkthrough the following notebooks, as this will build on top of both of them:
Adding memory to an LLM Chain
Custom Agents
In order to add a memory to a... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory.html |
9916889ff898-1 | )
memory = ConversationBufferMemory(memory_key="chat_history")
We can now construct the LLMChain, with the Memory object, and then create the agent.
llm_chain = LLMChain(llm=OpenAI(temperature=0), prompt=prompt)
agent = ZeroShotAgent(llm_chain=llm_chain, tools=tools, verbose=True)
agent_chain = AgentExecutor.from_agent... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory.html |
9916889ff898-2 | Action: Search
Action Input: Population of Canada
Observation: The current population of Canada is 38,566,192 as of Saturday, December 31, 2022, based on Worldometer elaboration of the latest United Nations data. · Canada ... Additional information related to Canadian population trends can be found on Statistics Canada... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory.html |
9916889ff898-3 | > Finished AgentExecutor chain.
'The current population of Canada is 38,566,192 as of Saturday, December 31, 2022, based on Worldometer elaboration of the latest United Nations data.'
To test the memory of this agent, we can ask a followup question that relies on information in the previous exchange to be answered corr... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory.html |
9916889ff898-4 | Action: Search
Action Input: National Anthem of Canada
Observation: Jun 7, 2010 ... https://twitter.com/CanadaImmigrantCanadian National Anthem O Canada in HQ - complete with lyrics, captions, vocals & music.LYRICS:O Canada! Nov 23, 2022 ... After 100 years of tradition, O Canada was proclaimed Canada's national anthem... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory.html |
9916889ff898-5 | Thought: I now know the final answer.
Final Answer: The national anthem of Canada is called "O Canada".
> Finished AgentExecutor chain.
'The national anthem of Canada is called "O Canada".'
We can see that the agent remembered that the previous question was about Canada, and properly asked Google Search what the name o... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory.html |
9916889ff898-6 | Action: Search
Action Input: Population of Canada
Observation: The current population of Canada is 38,566,192 as of Saturday, December 31, 2022, based on Worldometer elaboration of the latest United Nations data. · Canada ... Additional information related to Canadian population trends can be found on Statistics Canada... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory.html |
9916889ff898-7 | > Finished AgentExecutor chain.
'The current population of Canada is 38,566,192 as of Saturday, December 31, 2022, based on Worldometer elaboration of the latest United Nations data.'
agent_without_memory.run("what is their national anthem called?")
> Entering new AgentExecutor chain...
Thought: I should look up the an... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory.html |
9916889ff898-8 | Action: Search
Action Input: national anthem of [country]
Observation: Most nation states have an anthem, defined as "a song, as of praise, devotion, or patriotism"; most anthems are either marches or hymns in style. List of all countries around the world with its national anthem. ... Title and lyrics in the language o... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory.html |
9916889ff898-9 | Thought: I now know the final answer
Final Answer: The national anthem of [country] is [name of anthem].
> Finished AgentExecutor chain.
'The national anthem of [country] is [name of anthem].'
previous
How to add memory to a Multi-Input Chain
next
Adding Message Memory backed by a database to an Agent
By Harrison Chase... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory.html |
c347bbb365a2-0 | .ipynb
.pdf
Motörhead Memory
Contents
Setup
Motörhead Memory#
Motörhead is a memory server implemented in Rust. It automatically handles incremental summarization in the background and allows for stateless applications.
Setup#
See instructions at Motörhead for running the server locally.
from langchain.memory.motorhe... | https://python.langchain.com/en/latest/modules/memory/examples/motorhead_memory.html |
c347bbb365a2-1 | Human: whats my name?
AI:
> Finished chain.
' You said your name is Bob. Is that correct?'
llm_chain.run("whats for dinner?")
> Entering new LLMChain chain...
Prompt after formatting:
You are a chatbot having a conversation with a human.
Human: hi im bob
AI: Hi Bob, nice to meet you! How are you doing today?
Human: wh... | https://python.langchain.com/en/latest/modules/memory/examples/motorhead_memory.html |
27b0de54cd79-0 | .ipynb
.pdf
Mongodb Chat Message History
Mongodb Chat Message History#
This notebook goes over how to use Mongodb to store chat message history.
MongoDB is a source-available cross-platform document-oriented database program. Classified as a NoSQL database program, MongoDB uses JSON-like documents with optional schemas... | https://python.langchain.com/en/latest/modules/memory/examples/mongodb_chat_message_history.html |
1256a78a49b0-0 | .ipynb
.pdf
Adding Message Memory backed by a database to an Agent
Adding Message Memory backed by a database to an Agent#
This notebook goes over adding memory to an Agent where the memory uses an external message store. Before going through this notebook, please walkthrough the following notebooks, as this will build... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory_in_db.html |
1256a78a49b0-1 | {chat_history}
Question: {input}
{agent_scratchpad}"""
prompt = ZeroShotAgent.create_prompt(
tools,
prefix=prefix,
suffix=suffix,
input_variables=["input", "chat_history", "agent_scratchpad"]
)
Now we can create the ChatMessageHistory backed by the database.
message_history = RedisChatMessageHistory(... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory_in_db.html |
1256a78a49b0-2 | Action: Search
Action Input: Population of Canada
Observation: The current population of Canada is 38,566,192 as of Saturday, December 31, 2022, based on Worldometer elaboration of the latest United Nations data. · Canada ... Additional information related to Canadian population trends can be found on Statistics Canada... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory_in_db.html |
1256a78a49b0-3 | > Finished AgentExecutor chain.
'The current population of Canada is 38,566,192 as of Saturday, December 31, 2022, based on Worldometer elaboration of the latest United Nations data.'
To test the memory of this agent, we can ask a followup question that relies on information in the previous exchange to be answered corr... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory_in_db.html |
1256a78a49b0-4 | Action: Search
Action Input: National Anthem of Canada
Observation: Jun 7, 2010 ... https://twitter.com/CanadaImmigrantCanadian National Anthem O Canada in HQ - complete with lyrics, captions, vocals & music.LYRICS:O Canada! Nov 23, 2022 ... After 100 years of tradition, O Canada was proclaimed Canada's national anthem... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory_in_db.html |
1256a78a49b0-5 | Thought: I now know the final answer.
Final Answer: The national anthem of Canada is called "O Canada".
> Finished AgentExecutor chain.
'The national anthem of Canada is called "O Canada".'
We can see that the agent remembered that the previous question was about Canada, and properly asked Google Search what the name o... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory_in_db.html |
1256a78a49b0-6 | Action: Search
Action Input: Population of Canada
Observation: The current population of Canada is 38,566,192 as of Saturday, December 31, 2022, based on Worldometer elaboration of the latest United Nations data. · Canada ... Additional information related to Canadian population trends can be found on Statistics Canada... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory_in_db.html |
1256a78a49b0-7 | > Finished AgentExecutor chain.
'The current population of Canada is 38,566,192 as of Saturday, December 31, 2022, based on Worldometer elaboration of the latest United Nations data.'
agent_without_memory.run("what is their national anthem called?")
> Entering new AgentExecutor chain...
Thought: I should look up the an... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory_in_db.html |
1256a78a49b0-8 | Action: Search
Action Input: national anthem of [country]
Observation: Most nation states have an anthem, defined as "a song, as of praise, devotion, or patriotism"; most anthems are either marches or hymns in style. List of all countries around the world with its national anthem. ... Title and lyrics in the language o... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory_in_db.html |
1256a78a49b0-9 | Thought: I now know the final answer
Final Answer: The national anthem of [country] is [name of anthem].
> Finished AgentExecutor chain.
'The national anthem of [country] is [name of anthem].'
previous
How to add Memory to an Agent
next
Cassandra Chat Message History
By Harrison Chase
© Copyright 2023, Harri... | https://python.langchain.com/en/latest/modules/memory/examples/agent_with_memory_in_db.html |
d036019b4137-0 | .ipynb
.pdf
VectorStore-Backed Memory
Contents
Initialize your VectorStore
Create your the VectorStoreRetrieverMemory
Using in a chain
VectorStore-Backed Memory#
VectorStoreRetrieverMemory stores memories in a VectorDB and queries the top-K most “salient” docs every time it is called.
This differs from most of the ot... | https://python.langchain.com/en/latest/modules/memory/types/vectorstore_retriever_memory.html |
d036019b4137-1 | memory = VectorStoreRetrieverMemory(retriever=retriever)
# When added to an agent, the memory object can save pertinent information from conversations or used tools
memory.save_context({"input": "My favorite food is pizza"}, {"output": "thats good to know"})
memory.save_context({"input": "My favorite sport is soccer"},... | https://python.langchain.com/en/latest/modules/memory/types/vectorstore_retriever_memory.html |
d036019b4137-2 | memory=memory,
verbose=True
)
conversation_with_summary.predict(input="Hi, my name is Perry, what's up?")
> Entering new ConversationChain chain...
Prompt after formatting:
The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from its context.... | https://python.langchain.com/en/latest/modules/memory/types/vectorstore_retriever_memory.html |
d036019b4137-3 | > Entering new ConversationChain chain...
Prompt after formatting:
The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from its context. If the AI does not know the answer to a question, it truthfully says it does not know.
Relevant pieces of pre... | https://python.langchain.com/en/latest/modules/memory/types/vectorstore_retriever_memory.html |
d036019b4137-4 | By Harrison Chase
© Copyright 2023, Harrison Chase.
Last updated on Jun 02, 2023. | https://python.langchain.com/en/latest/modules/memory/types/vectorstore_retriever_memory.html |
4aaeac0bd8f1-0 | .ipynb
.pdf
ConversationBufferWindowMemory
Contents
Using in a chain
ConversationBufferWindowMemory#
ConversationBufferWindowMemory keeps a list of the interactions of the conversation over time. It only uses the last K interactions. This can be useful for keeping a sliding window of the most recent interactions, so ... | https://python.langchain.com/en/latest/modules/memory/types/buffer_window.html |
4aaeac0bd8f1-1 | memory=ConversationBufferWindowMemory(k=2),
verbose=True
)
conversation_with_summary.predict(input="Hi, what's up?")
> Entering new ConversationChain chain...
Prompt after formatting:
The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from ... | https://python.langchain.com/en/latest/modules/memory/types/buffer_window.html |
4aaeac0bd8f1-2 | Current conversation:
Human: Hi, what's up?
AI: Hi there! I'm doing great. I'm currently helping a customer with a technical issue. How about you?
Human: What's their issues?
AI: The customer is having trouble connecting to their Wi-Fi network. I'm helping them troubleshoot the issue and get them connected.
Human: Is... | https://python.langchain.com/en/latest/modules/memory/types/buffer_window.html |
91c89f183176-0 | .ipynb
.pdf
ConversationBufferMemory
Contents
Using in a chain
ConversationBufferMemory#
This notebook shows how to use ConversationBufferMemory. This memory allows for storing of messages and then extracts the messages in a variable.
We can first extract it as a string.
from langchain.memory import ConversationBuffe... | https://python.langchain.com/en/latest/modules/memory/types/buffer.html |
91c89f183176-1 | Current conversation:
Human: Hi there!
AI:
> Finished chain.
" Hi there! It's nice to meet you. How can I help you today?"
conversation.predict(input="I'm doing well! Just having a conversation with an AI.")
> Entering new ConversationChain chain...
Prompt after formatting:
The following is a friendly conversation betw... | https://python.langchain.com/en/latest/modules/memory/types/buffer.html |
91c89f183176-2 | Human: Tell me about yourself.
AI:
> Finished chain.
" Sure! I'm an AI created to help people with their everyday tasks. I'm programmed to understand natural language and provide helpful information. I'm also constantly learning and updating my knowledge base so I can provide more accurate and helpful answers."
And tha... | https://python.langchain.com/en/latest/modules/memory/types/buffer.html |
8ab0d904b90b-0 | .ipynb
.pdf
Entity Memory
Contents
Using in a chain
Inspecting the memory store
Entity Memory#
This notebook shows how to work with a memory module that remembers things about specific entities. It extracts information on entities (using LLMs) and builds up its knowledge about that entity over time (also using LLMs).... | https://python.langchain.com/en/latest/modules/memory/types/entity_summary_memory.html |
8ab0d904b90b-1 | 'entities': {'Sam': 'Sam is working on a hackathon project with Deven.'}}
Using in a chain#
Let’s now use it in a chain!
from langchain.chains import ConversationChain
from langchain.memory import ConversationEntityMemory
from langchain.memory.prompt import ENTITY_MEMORY_CONVERSATION_TEMPLATE
from pydantic import BaseM... | https://python.langchain.com/en/latest/modules/memory/types/entity_summary_memory.html |
8ab0d904b90b-2 | Overall, you are a powerful tool that can help with a wide range of tasks and provide valuable insights and information on a wide range of topics. Whether the human needs help with a specific question or just wants to have a conversation about a particular topic, you are here to assist.
Context:
{'Deven': 'Deven is wor... | https://python.langchain.com/en/latest/modules/memory/types/entity_summary_memory.html |
8ab0d904b90b-3 | You are constantly learning and improving, and your capabilities are constantly evolving. You are able to process and understand large amounts of text, and can use this knowledge to provide accurate and informative responses to a wide range of questions. You have access to some personalized information provided by the ... | https://python.langchain.com/en/latest/modules/memory/types/entity_summary_memory.html |
8ab0d904b90b-4 | You are an assistant to a human, powered by a large language model trained by OpenAI.
You are designed to be able to assist with a wide range of tasks, from answering simple questions to providing in-depth explanations and discussions on a wide range of topics. As a language model, you are able to generate human-like t... | https://python.langchain.com/en/latest/modules/memory/types/entity_summary_memory.html |
8ab0d904b90b-5 | AI: That sounds like a great project! What kind of project are they working on?
Human: They are trying to add more complex memory structures to Langchain
AI: That sounds like an interesting project! What kind of memory structures are they trying to add?
Last line:
Human: They are adding in a key-value store for entit... | https://python.langchain.com/en/latest/modules/memory/types/entity_summary_memory.html |
8ab0d904b90b-6 | Context:
{'Deven': 'Deven is working on a hackathon project with Sam, which they are entering into a hackathon. They are trying to add more complex memory structures to Langchain, including a key-value store for entities mentioned so far in the conversation.', 'Sam': 'Sam is working on a hackathon project with Deven, t... | https://python.langchain.com/en/latest/modules/memory/types/entity_summary_memory.html |
8ab0d904b90b-7 | 'Deven': 'Deven is working on a hackathon project with Sam, which they are '
'entering into a hackathon. They are trying to add more complex '
'memory structures to Langchain, including a key-value store for '
'entities mentioned so far in the conversation, and seem to be '
'work... | https://python.langchain.com/en/latest/modules/memory/types/entity_summary_memory.html |
8ab0d904b90b-8 | You are constantly learning and improving, and your capabilities are constantly evolving. You are able to process and understand large amounts of text, and can use this knowledge to provide accurate and informative responses to a wide range of questions. You have access to some personalized information provided by the ... | https://python.langchain.com/en/latest/modules/memory/types/entity_summary_memory.html |
8ab0d904b90b-9 | Human: Sam is the founder of a company called Daimon.
AI:
That's impressive! It sounds like Sam is a very successful entrepreneur. What kind of company is Daimon?
Last line:
Human: Sam is the founder of a company called Daimon.
You:
> Finished chain.
" That's impressive! It sounds like Sam is a very successful entrepr... | https://python.langchain.com/en/latest/modules/memory/types/entity_summary_memory.html |
8ab0d904b90b-10 | 'for entities mentioned so far in the conversation. They seem to have '
'a great idea for how the key-value store can help, and Sam is also '
'the founder of a successful company called Daimon.'}
conversation.predict(input="What do you know about Sam?")
> Entering new ConversationChain chain...
Prompt a... | https://python.langchain.com/en/latest/modules/memory/types/entity_summary_memory.html |
8ab0d904b90b-11 | Context:
{'Deven': 'Deven is working on a hackathon project with Sam, which they are entering into a hackathon. They are trying to add more complex memory structures to Langchain, including a key-value store for entities mentioned so far in the conversation, and seem to be working hard on this project with a great idea... | https://python.langchain.com/en/latest/modules/memory/types/entity_summary_memory.html |
8ab0d904b90b-12 | Last line:
Human: What do you know about Sam?
You:
> Finished chain.
' Sam is the founder of a successful company called Daimon. He is also working on a hackathon project with Deven to add more complex memory structures to Langchain. They seem to have a great idea for how the key-value store can help.'
previous
Convers... | https://python.langchain.com/en/latest/modules/memory/types/entity_summary_memory.html |
406662fccf3d-0 | .ipynb
.pdf
ConversationTokenBufferMemory
Contents
Using in a chain
ConversationTokenBufferMemory#
ConversationTokenBufferMemory keeps a buffer of recent interactions in memory, and uses token length rather than number of interactions to determine when to flush interactions.
Let’s first walk through how to use the ut... | https://python.langchain.com/en/latest/modules/memory/types/token_buffer.html |
406662fccf3d-1 | > Entering new ConversationChain chain...
Prompt after formatting:
The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from its context. If the AI does not know the answer to a question, it truthfully says it does not know.
Current conversation:
... | https://python.langchain.com/en/latest/modules/memory/types/token_buffer.html |
406662fccf3d-2 | AI: Sounds like a productive day! What kind of documentation are you writing?
Human: For LangChain! Have you heard of it?
AI:
> Finished chain.
" Yes, I have heard of LangChain! It is a decentralized language-learning platform that connects native speakers and learners in real time. Is that the documentation you're wr... | https://python.langchain.com/en/latest/modules/memory/types/token_buffer.html |
1a34a9441f2d-0 | .ipynb
.pdf
Conversation Knowledge Graph Memory
Contents
Using in a chain
Conversation Knowledge Graph Memory#
This type of memory uses a knowledge graph to recreate memory.
Let’s first walk through how to use the utilities
from langchain.memory import ConversationKGMemory
from langchain.llms import OpenAI
llm = Open... | https://python.langchain.com/en/latest/modules/memory/types/kg.html |
1a34a9441f2d-1 | llm = OpenAI(temperature=0)
from langchain.prompts.prompt import PromptTemplate
from langchain.chains import ConversationChain
template = """The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from its context.
If the AI does not know the answer... | https://python.langchain.com/en/latest/modules/memory/types/kg.html |
1a34a9441f2d-2 | > Entering new ConversationChain chain...
Prompt after formatting:
The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from its context.
If the AI does not know the answer to a question, it truthfully says it does not know. The AI ONLY uses info... | https://python.langchain.com/en/latest/modules/memory/types/kg.html |
b8f53041aea3-0 | .ipynb
.pdf
ConversationSummaryBufferMemory
Contents
Using in a chain
ConversationSummaryBufferMemory#
ConversationSummaryBufferMemory combines the last two ideas. It keeps a buffer of recent interactions in memory, but rather than just completely flushing old interactions it compiles them into a summary and uses bot... | https://python.langchain.com/en/latest/modules/memory/types/summary_buffer.html |
b8f53041aea3-1 | from langchain.chains import ConversationChain
conversation_with_summary = ConversationChain(
llm=llm,
# We set a very low max_token_limit for the purposes of testing.
memory=ConversationSummaryBufferMemory(llm=OpenAI(), max_token_limit=40),
verbose=True
)
conversation_with_summary.predict(input="Hi, w... | https://python.langchain.com/en/latest/modules/memory/types/summary_buffer.html |
b8f53041aea3-2 | > Entering new ConversationChain chain...
Prompt after formatting:
The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from its context. If the AI does not know the answer to a question, it truthfully says it does not know.
Current conversation:
... | https://python.langchain.com/en/latest/modules/memory/types/summary_buffer.html |
b8f53041aea3-3 | Human: Haha nope, although a lot of people confuse it for that
AI:
> Finished chain.
' Oh, okay. What is LangChain?'
previous
ConversationSummaryMemory
next
ConversationTokenBufferMemory
Contents
Using in a chain
By Harrison Chase
© Copyright 2023, Harrison Chase.
Last updated on Jun 02, 2023. | https://python.langchain.com/en/latest/modules/memory/types/summary_buffer.html |
f084ccdd4c84-0 | .ipynb
.pdf
ConversationSummaryMemory
Contents
Initializing with messages
Using in a chain
ConversationSummaryMemory#
Now let’s take a look at using a slightly more complex type of memory - ConversationSummaryMemory. This type of memory creates a summary of the conversation over time. This can be useful for condensin... | https://python.langchain.com/en/latest/modules/memory/types/summary.html |
f084ccdd4c84-1 | history.add_user_message("hi")
history.add_ai_message("hi there!")
memory = ConversationSummaryMemory.from_messages(llm=OpenAI(temperature=0), chat_memory=history, return_messages=True)
memory.buffer
'\nThe human greets the AI, to which the AI responds with a friendly greeting.'
Using in a chain#
Let’s walk through an ... | https://python.langchain.com/en/latest/modules/memory/types/summary.html |
f084ccdd4c84-2 | Human: Tell me more about it!
AI:
> Finished chain.
" Sure! The customer is having trouble with their computer not connecting to the internet. I'm helping them troubleshoot the issue and figure out what the problem is. So far, we've tried resetting the router and checking the network settings, but the issue still persi... | https://python.langchain.com/en/latest/modules/memory/types/summary.html |
80a7eebac8a4-0 | .rst
.pdf
Example Selectors
Example Selectors#
Note
Conceptual Guide
If you have a large number of examples, you may need to select which ones to include in the prompt. The ExampleSelector is the class responsible for doing so.
The base interface is defined as below:
class BaseExampleSelector(ABC):
"""Interface for... | https://python.langchain.com/en/latest/modules/prompts/example_selectors.html |
72977666f022-0 | .rst
.pdf
Prompt Templates
Prompt Templates#
Note
Conceptual Guide
Language models take text as input - that text is commonly referred to as a prompt.
Typically this is not simply a hardcoded string but rather a combination of a template, some examples, and user input.
LangChain provides several classes and functions t... | https://python.langchain.com/en/latest/modules/prompts/prompt_templates.html |
f13f4b6335f5-0 | .rst
.pdf
Output Parsers
Output Parsers#
Note
Conceptual Guide
Language models output text. But many times you may want to get more structured information than just text back. This is where output parsers come in.
Output parsers are classes that help structure language model responses. There are two main methods an out... | https://python.langchain.com/en/latest/modules/prompts/output_parsers.html |
6aafda63ed5c-0 | .ipynb
.pdf
Chat Prompt Template
Contents
Format output
Different types of MessagePromptTemplate
Chat Prompt Template#
Chat Models takes a list of chat messages as input - this list commonly referred to as a prompt.
These chat messages differ from raw string (which you would pass into a LLM model) in that every messa... | https://python.langchain.com/en/latest/modules/prompts/chat_prompt_template.html |
6aafda63ed5c-1 | input_variables=["input_language", "output_language"],
)
system_message_prompt_2 = SystemMessagePromptTemplate(prompt=prompt)
assert system_message_prompt == system_message_prompt_2
After that, you can build a ChatPromptTemplate from one or more MessagePromptTemplates. You can use ChatPromptTemplate’s format_prompt – t... | https://python.langchain.com/en/latest/modules/prompts/chat_prompt_template.html |
6aafda63ed5c-2 | [SystemMessage(content='You are a helpful assistant that translates English to French.', additional_kwargs={}),
HumanMessage(content='I love programming.', additional_kwargs={})]
Different types of MessagePromptTemplate#
LangChain provides different types of MessagePromptTemplate. The most commonly used are AIMessageP... | https://python.langchain.com/en/latest/modules/prompts/chat_prompt_template.html |
6aafda63ed5c-3 | 3. Practice, practice, practice: The best way to learn programming is through hands-on experience\
""")
chat_prompt.format_prompt(conversation=[human_message, ai_message], word_count="10").to_messages()
[HumanMessage(content='What is the best way to learn programming?', additional_kwargs={}),
AIMessage(content='1. Cho... | https://python.langchain.com/en/latest/modules/prompts/chat_prompt_template.html |
48c2549ef01a-0 | .ipynb
.pdf
Getting Started
Contents
PromptTemplates
to_string
to_messages
Getting Started#
This section contains everything related to prompts. A prompt is the value passed into the Language Model. This value can either be a string (for LLMs) or a list of messages (for Chat Models).
The data types of these prompts a... | https://python.langchain.com/en/latest/modules/prompts/getting_started.html |
48c2549ef01a-1 | string_prompt_value.to_string()
'tell me a joke about soccer'
chat_prompt_value.to_string()
'Human: tell me a joke about soccer'
to_messages#
This is what is called when passing to ChatModel (which expects a list of messages)
string_prompt_value.to_messages()
[HumanMessage(content='tell me a joke about soccer', additio... | https://python.langchain.com/en/latest/modules/prompts/getting_started.html |
a3ad87c8ba96-0 | .rst
.pdf
How-To Guides
How-To Guides#
If you’re new to the library, you may want to start with the Quickstart.
The user guide here shows more advanced workflows and how to use the library in different ways.
Connecting to a Feature Store
How to create a custom prompt template
How to create a prompt template that uses f... | https://python.langchain.com/en/latest/modules/prompts/prompt_templates/how_to_guides.html |
d868fbc8b065-0 | .md
.pdf
Getting Started
Contents
What is a prompt template?
Create a prompt template
Template formats
Validate template
Serialize prompt template
Pass few shot examples to a prompt template
Select examples for a prompt template
Getting Started#
In this tutorial, we will learn about:
what a prompt template is, and wh... | https://python.langchain.com/en/latest/modules/prompts/prompt_templates/getting_started.html |
d868fbc8b065-1 | no_input_prompt.format()
# -> "Tell me a joke."
# An example prompt with one input variable
one_input_prompt = PromptTemplate(input_variables=["adjective"], template="Tell me a {adjective} joke.")
one_input_prompt.format(adjective="funny")
# -> "Tell me a funny joke."
# An example prompt with multiple input variables
m... | https://python.langchain.com/en/latest/modules/prompts/prompt_templates/getting_started.html |
d868fbc8b065-2 | # -> Tell me a funny joke about chickens.
Currently, PromptTemplate only supports jinja2 and f-string templating format. If there is any other templating format that you would like to use, feel free to open an issue in the Github page.
Validate template#
By default, PromptTemplate will validate the template string by c... | https://python.langchain.com/en/latest/modules/prompts/prompt_templates/getting_started.html |
d868fbc8b065-3 | To generate a prompt with few shot examples, you can use the FewShotPromptTemplate. This class takes in a PromptTemplate and a list of few shot examples. It then formats the prompt template with the few shot examples.
In this example, we’ll create a prompt to generate word antonyms.
from langchain import PromptTemplate... | https://python.langchain.com/en/latest/modules/prompts/prompt_templates/getting_started.html |
d868fbc8b065-4 | input_variables=["input"],
# The example_separator is the string we will use to join the prefix, examples, and suffix together with.
example_separator="\n",
)
# We can now generate a prompt using the `format` method.
print(few_shot_prompt.format(input="big"))
# -> Give the antonym of every input
# ->
# -> Word... | https://python.langchain.com/en/latest/modules/prompts/prompt_templates/getting_started.html |
d868fbc8b065-5 | {"word": "windy", "antonym": "calm"},
]
# We'll use the `LengthBasedExampleSelector` to select the examples.
example_selector = LengthBasedExampleSelector(
# These are the examples is has available to choose from.
examples=examples,
# This is the PromptTemplate being used to format the examples.
exampl... | https://python.langchain.com/en/latest/modules/prompts/prompt_templates/getting_started.html |
d868fbc8b065-6 | # -> Antonym: lethargic
# ->
# -> Word: sunny
# -> Antonym: gloomy
# ->
# -> Word: windy
# -> Antonym: calm
# ->
# -> Word: big
# -> Antonym:
In contrast, if we provide a very long input, the LengthBasedExampleSelector will select fewer examples to include in the prompt.
long_string = "big and huge and massive and larg... | https://python.langchain.com/en/latest/modules/prompts/prompt_templates/getting_started.html |
7cd9c88f0159-0 | .ipynb
.pdf
Connecting to a Feature Store
Contents
Feast
Load Feast Store
Prompts
Use in a chain
Tecton
Prerequisites
Define and Load Features
Prompts
Use in a chain
Featureform
Initialize Featureform
Prompts
Use in a chain
Connecting to a Feature Store#
Feature stores are a concept from traditional machine learning ... | https://python.langchain.com/en/latest/modules/prompts/prompt_templates/examples/connecting_to_a_feature_store.html |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.